mirror of
https://github.com/paboyle/Grid.git
synced 2024-11-09 23:45:36 +00:00
Deprecate UVM
This commit is contained in:
parent
11be10d2c0
commit
066544281f
@ -168,6 +168,7 @@ public:
|
|||||||
template<class vobj>
|
template<class vobj>
|
||||||
void FFT_dim(Lattice<vobj> &result,const Lattice<vobj> &source,int dim, int sign){
|
void FFT_dim(Lattice<vobj> &result,const Lattice<vobj> &source,int dim, int sign){
|
||||||
#ifndef HAVE_FFTW
|
#ifndef HAVE_FFTW
|
||||||
|
std::cerr << "FFTW is not compiled but is called"<<std::endl;
|
||||||
assert(0);
|
assert(0);
|
||||||
#else
|
#else
|
||||||
conformable(result.Grid(),vgrid);
|
conformable(result.Grid(),vgrid);
|
||||||
@ -190,7 +191,8 @@ public:
|
|||||||
|
|
||||||
Lattice<sobj> pgbuf(&pencil_g);
|
Lattice<sobj> pgbuf(&pencil_g);
|
||||||
autoView(pgbuf_v , pgbuf, CpuWrite);
|
autoView(pgbuf_v , pgbuf, CpuWrite);
|
||||||
|
std::cout << "CPU view" << std::endl;
|
||||||
|
|
||||||
typedef typename FFTW<scalar>::FFTW_scalar FFTW_scalar;
|
typedef typename FFTW<scalar>::FFTW_scalar FFTW_scalar;
|
||||||
typedef typename FFTW<scalar>::FFTW_plan FFTW_plan;
|
typedef typename FFTW<scalar>::FFTW_plan FFTW_plan;
|
||||||
|
|
||||||
@ -213,6 +215,7 @@ public:
|
|||||||
else if ( sign == forward ) div = 1.0;
|
else if ( sign == forward ) div = 1.0;
|
||||||
else assert(0);
|
else assert(0);
|
||||||
|
|
||||||
|
std::cout << "Making FFTW plan" << std::endl;
|
||||||
FFTW_plan p;
|
FFTW_plan p;
|
||||||
{
|
{
|
||||||
FFTW_scalar *in = (FFTW_scalar *)&pgbuf_v[0];
|
FFTW_scalar *in = (FFTW_scalar *)&pgbuf_v[0];
|
||||||
@ -226,6 +229,7 @@ public:
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Barrel shift and collect global pencil
|
// Barrel shift and collect global pencil
|
||||||
|
std::cout << "Making pencil" << std::endl;
|
||||||
Coordinate lcoor(Nd), gcoor(Nd);
|
Coordinate lcoor(Nd), gcoor(Nd);
|
||||||
result = source;
|
result = source;
|
||||||
int pc = processor_coor[dim];
|
int pc = processor_coor[dim];
|
||||||
@ -247,6 +251,7 @@ public:
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
std::cout << "Looping orthog" << std::endl;
|
||||||
// Loop over orthog coords
|
// Loop over orthog coords
|
||||||
int NN=pencil_g.lSites();
|
int NN=pencil_g.lSites();
|
||||||
GridStopWatch timer;
|
GridStopWatch timer;
|
||||||
@ -269,6 +274,7 @@ public:
|
|||||||
usec += timer.useconds();
|
usec += timer.useconds();
|
||||||
flops+= flops_call*NN;
|
flops+= flops_call*NN;
|
||||||
|
|
||||||
|
std::cout << "Writing back results " << std::endl;
|
||||||
// writing out result
|
// writing out result
|
||||||
{
|
{
|
||||||
autoView(pgbuf_v,pgbuf,CpuRead);
|
autoView(pgbuf_v,pgbuf,CpuRead);
|
||||||
@ -285,6 +291,7 @@ public:
|
|||||||
}
|
}
|
||||||
result = result*div;
|
result = result*div;
|
||||||
|
|
||||||
|
std::cout << "Destroying plan " << std::endl;
|
||||||
// destroying plan
|
// destroying plan
|
||||||
FFTW<scalar>::fftw_destroy_plan(p);
|
FFTW<scalar>::fftw_destroy_plan(p);
|
||||||
#endif
|
#endif
|
||||||
|
@ -102,11 +102,11 @@ public:
|
|||||||
assert(mass.size()==nshift);
|
assert(mass.size()==nshift);
|
||||||
assert(mresidual.size()==nshift);
|
assert(mresidual.size()==nshift);
|
||||||
|
|
||||||
// dynamic sized arrays on stack; 2d is a pain with vector
|
// remove dynamic sized arrays on stack; 2d is a pain with vector
|
||||||
RealD bs[nshift];
|
std::vector<RealD> bs(nshift);
|
||||||
RealD rsq[nshift];
|
std::vector<RealD> rsq(nshift);
|
||||||
RealD z[nshift][2];
|
std::vector<std::array<RealD,2> > z(nshift);
|
||||||
int converged[nshift];
|
std::vector<int> converged(nshift);
|
||||||
|
|
||||||
const int primary =0;
|
const int primary =0;
|
||||||
|
|
||||||
|
@ -123,11 +123,11 @@ public:
|
|||||||
assert(mresidual.size()==nshift);
|
assert(mresidual.size()==nshift);
|
||||||
|
|
||||||
// dynamic sized arrays on stack; 2d is a pain with vector
|
// dynamic sized arrays on stack; 2d is a pain with vector
|
||||||
RealD bs[nshift];
|
std::vector<RealD> bs(nshift);
|
||||||
RealD rsq[nshift];
|
std::vector<RealD> rsq(nshift);
|
||||||
RealD rsqf[nshift];
|
std::vector<RealD> rsqf(nshift);
|
||||||
RealD z[nshift][2];
|
std::vector<std::array<RealD,2> > z(nshift);
|
||||||
int converged[nshift];
|
std::vector<int> converged(nshift);
|
||||||
|
|
||||||
const int primary =0;
|
const int primary =0;
|
||||||
|
|
||||||
|
@ -156,11 +156,11 @@ public:
|
|||||||
assert(mresidual.size()==nshift);
|
assert(mresidual.size()==nshift);
|
||||||
|
|
||||||
// dynamic sized arrays on stack; 2d is a pain with vector
|
// dynamic sized arrays on stack; 2d is a pain with vector
|
||||||
RealD bs[nshift];
|
std::vector<RealD> bs(nshift);
|
||||||
RealD rsq[nshift];
|
std::vector<RealD> rsq(nshift);
|
||||||
RealD rsqf[nshift];
|
std::vector<RealD> rsqf(nshift);
|
||||||
RealD z[nshift][2];
|
std::vector<std::array<RealD,2> > z(nshift);
|
||||||
int converged[nshift];
|
std::vector<int> converged(nshift);
|
||||||
|
|
||||||
const int primary =0;
|
const int primary =0;
|
||||||
|
|
||||||
|
@ -99,7 +99,7 @@ public:
|
|||||||
CoarseMatrix AselfInvEven;
|
CoarseMatrix AselfInvEven;
|
||||||
CoarseMatrix AselfInvOdd;
|
CoarseMatrix AselfInvOdd;
|
||||||
|
|
||||||
Vector<RealD> dag_factor;
|
deviceVector<RealD> dag_factor;
|
||||||
|
|
||||||
///////////////////////
|
///////////////////////
|
||||||
// Interface
|
// Interface
|
||||||
@ -124,9 +124,13 @@ public:
|
|||||||
int npoint = geom.npoint;
|
int npoint = geom.npoint;
|
||||||
typedef LatticeView<Cobj> Aview;
|
typedef LatticeView<Cobj> Aview;
|
||||||
|
|
||||||
Vector<Aview> AcceleratorViewContainer;
|
deviceVector<Aview> AcceleratorViewContainer(geom.npoint);
|
||||||
|
hostVector<Aview> hAcceleratorViewContainer(geom.npoint);
|
||||||
|
|
||||||
for(int p=0;p<geom.npoint;p++) AcceleratorViewContainer.push_back(A[p].View(AcceleratorRead));
|
for(int p=0;p<geom.npoint;p++) {
|
||||||
|
hAcceleratorViewContainer[p] = A[p].View(AcceleratorRead);
|
||||||
|
acceleratorPut(AcceleratorViewContainer[p],hAcceleratorViewContainer[p]);
|
||||||
|
}
|
||||||
Aview *Aview_p = & AcceleratorViewContainer[0];
|
Aview *Aview_p = & AcceleratorViewContainer[0];
|
||||||
|
|
||||||
const int Nsimd = CComplex::Nsimd();
|
const int Nsimd = CComplex::Nsimd();
|
||||||
@ -161,7 +165,7 @@ public:
|
|||||||
coalescedWrite(out_v[ss](b),res);
|
coalescedWrite(out_v[ss](b),res);
|
||||||
});
|
});
|
||||||
|
|
||||||
for(int p=0;p<geom.npoint;p++) AcceleratorViewContainer[p].ViewClose();
|
for(int p=0;p<geom.npoint;p++) hAcceleratorViewContainer[p].ViewClose();
|
||||||
};
|
};
|
||||||
|
|
||||||
void Mdag (const CoarseVector &in, CoarseVector &out)
|
void Mdag (const CoarseVector &in, CoarseVector &out)
|
||||||
@ -190,9 +194,14 @@ public:
|
|||||||
int npoint = geom.npoint;
|
int npoint = geom.npoint;
|
||||||
typedef LatticeView<Cobj> Aview;
|
typedef LatticeView<Cobj> Aview;
|
||||||
|
|
||||||
Vector<Aview> AcceleratorViewContainer;
|
|
||||||
|
|
||||||
for(int p=0;p<geom.npoint;p++) AcceleratorViewContainer.push_back(A[p].View(AcceleratorRead));
|
deviceVector<Aview> AcceleratorViewContainer(geom.npoint);
|
||||||
|
hostVector<Aview> hAcceleratorViewContainer(geom.npoint);
|
||||||
|
|
||||||
|
for(int p=0;p<geom.npoint;p++) {
|
||||||
|
hAcceleratorViewContainer[p] = A[p].View(AcceleratorRead);
|
||||||
|
acceleratorPut(AcceleratorViewContainer[p],hAcceleratorViewContainer[p]);
|
||||||
|
}
|
||||||
Aview *Aview_p = & AcceleratorViewContainer[0];
|
Aview *Aview_p = & AcceleratorViewContainer[0];
|
||||||
|
|
||||||
const int Nsimd = CComplex::Nsimd();
|
const int Nsimd = CComplex::Nsimd();
|
||||||
@ -201,10 +210,10 @@ public:
|
|||||||
|
|
||||||
int osites=Grid()->oSites();
|
int osites=Grid()->oSites();
|
||||||
|
|
||||||
Vector<int> points(geom.npoint, 0);
|
deviceVector<int> points(geom.npoint);
|
||||||
for(int p=0; p<geom.npoint; p++)
|
for(int p=0; p<geom.npoint; p++) {
|
||||||
points[p] = geom.points_dagger[p];
|
acceleratorPut(points[p],geom.points_dagger[p]);
|
||||||
|
}
|
||||||
auto points_p = &points[0];
|
auto points_p = &points[0];
|
||||||
|
|
||||||
RealD* dag_factor_p = &dag_factor[0];
|
RealD* dag_factor_p = &dag_factor[0];
|
||||||
@ -236,7 +245,7 @@ public:
|
|||||||
coalescedWrite(out_v[ss](b),res);
|
coalescedWrite(out_v[ss](b),res);
|
||||||
});
|
});
|
||||||
|
|
||||||
for(int p=0;p<geom.npoint;p++) AcceleratorViewContainer[p].ViewClose();
|
for(int p=0;p<geom.npoint;p++) hAcceleratorViewContainer[p].ViewClose();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MdirComms(const CoarseVector &in)
|
void MdirComms(const CoarseVector &in)
|
||||||
@ -251,8 +260,14 @@ public:
|
|||||||
out.Checkerboard() = in.Checkerboard();
|
out.Checkerboard() = in.Checkerboard();
|
||||||
|
|
||||||
typedef LatticeView<Cobj> Aview;
|
typedef LatticeView<Cobj> Aview;
|
||||||
Vector<Aview> AcceleratorViewContainer;
|
|
||||||
for(int p=0;p<geom.npoint;p++) AcceleratorViewContainer.push_back(A[p].View(AcceleratorRead));
|
deviceVector<Aview> AcceleratorViewContainer(geom.npoint);
|
||||||
|
hostVector<Aview> hAcceleratorViewContainer(geom.npoint);
|
||||||
|
|
||||||
|
for(int p=0;p<geom.npoint;p++) {
|
||||||
|
hAcceleratorViewContainer[p] = A[p].View(AcceleratorRead);
|
||||||
|
acceleratorPut(AcceleratorViewContainer[p],hAcceleratorViewContainer[p]);
|
||||||
|
}
|
||||||
Aview *Aview_p = & AcceleratorViewContainer[0];
|
Aview *Aview_p = & AcceleratorViewContainer[0];
|
||||||
|
|
||||||
autoView( out_v , out, AcceleratorWrite);
|
autoView( out_v , out, AcceleratorWrite);
|
||||||
@ -285,7 +300,7 @@ public:
|
|||||||
}
|
}
|
||||||
coalescedWrite(out_v[ss](b),res);
|
coalescedWrite(out_v[ss](b),res);
|
||||||
});
|
});
|
||||||
for(int p=0;p<geom.npoint;p++) AcceleratorViewContainer[p].ViewClose();
|
for(int p=0;p<geom.npoint;p++) hAcceleratorViewContainer[p].ViewClose();
|
||||||
}
|
}
|
||||||
void MdirAll(const CoarseVector &in,std::vector<CoarseVector> &out)
|
void MdirAll(const CoarseVector &in,std::vector<CoarseVector> &out)
|
||||||
{
|
{
|
||||||
@ -469,14 +484,20 @@ public:
|
|||||||
|
|
||||||
// determine in what order we need the points
|
// determine in what order we need the points
|
||||||
int npoint = geom.npoint-1;
|
int npoint = geom.npoint-1;
|
||||||
Vector<int> points(npoint, 0);
|
deviceVector<int> points(npoint);
|
||||||
for(int p=0; p<npoint; p++)
|
for(int p=0; p<npoint; p++) {
|
||||||
points[p] = (dag && !hermitian) ? geom.points_dagger[p] : p;
|
int val = (dag && !hermitian) ? geom.points_dagger[p] : p;
|
||||||
|
acceleratorPut(points[p], val);
|
||||||
|
}
|
||||||
auto points_p = &points[0];
|
auto points_p = &points[0];
|
||||||
|
|
||||||
Vector<Aview> AcceleratorViewContainer;
|
deviceVector<Aview> AcceleratorViewContainer(geom.npoint);
|
||||||
for(int p=0;p<npoint;p++) AcceleratorViewContainer.push_back(a[p].View(AcceleratorRead));
|
hostVector<Aview> hAcceleratorViewContainer(geom.npoint);
|
||||||
|
|
||||||
|
for(int p=0;p<geom.npoint;p++) {
|
||||||
|
hAcceleratorViewContainer[p] = a[p].View(AcceleratorRead);
|
||||||
|
acceleratorPut(AcceleratorViewContainer[p],hAcceleratorViewContainer[p]);
|
||||||
|
}
|
||||||
Aview *Aview_p = & AcceleratorViewContainer[0];
|
Aview *Aview_p = & AcceleratorViewContainer[0];
|
||||||
|
|
||||||
const int Nsimd = CComplex::Nsimd();
|
const int Nsimd = CComplex::Nsimd();
|
||||||
@ -539,7 +560,7 @@ public:
|
|||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
for(int p=0;p<npoint;p++) AcceleratorViewContainer[p].ViewClose();
|
for(int p=0;p<npoint;p++) hAcceleratorViewContainer[p].ViewClose();
|
||||||
}
|
}
|
||||||
|
|
||||||
CoarsenedMatrix(GridCartesian &CoarseGrid, int hermitian_=0) :
|
CoarsenedMatrix(GridCartesian &CoarseGrid, int hermitian_=0) :
|
||||||
@ -590,11 +611,13 @@ public:
|
|||||||
}
|
}
|
||||||
|
|
||||||
// GPU readable prefactor
|
// GPU readable prefactor
|
||||||
|
std::vector<RealD> h_dag_factor(nbasis*nbasis);
|
||||||
thread_for(i, nbasis*nbasis, {
|
thread_for(i, nbasis*nbasis, {
|
||||||
int j = i/nbasis;
|
int j = i/nbasis;
|
||||||
int k = i%nbasis;
|
int k = i%nbasis;
|
||||||
dag_factor[i] = dag_factor_eigen(j, k);
|
h_dag_factor[i] = dag_factor_eigen(j, k);
|
||||||
});
|
});
|
||||||
|
acceleratorCopyToDevice(&h_dag_factor[0],&dag_factor[0],dag_factor.size()*sizeof(RealD));
|
||||||
}
|
}
|
||||||
|
|
||||||
void CoarsenOperator(GridBase *FineGrid,LinearOperatorBase<Lattice<Fobj> > &linop,
|
void CoarsenOperator(GridBase *FineGrid,LinearOperatorBase<Lattice<Fobj> > &linop,
|
||||||
|
@ -174,21 +174,11 @@ template<typename _Tp> inline bool operator!=(const devAllocator<_Tp>&, const d
|
|||||||
////////////////////////////////////////////////////////////////////////////////
|
////////////////////////////////////////////////////////////////////////////////
|
||||||
// Template typedefs
|
// Template typedefs
|
||||||
////////////////////////////////////////////////////////////////////////////////
|
////////////////////////////////////////////////////////////////////////////////
|
||||||
#ifdef ACCELERATOR_CSHIFT
|
template<class T> using hostVector = std::vector<T,alignedAllocator<T> >; // Needs autoview
|
||||||
// Cshift on device
|
template<class T> using Vector = std::vector<T,uvmAllocator<T> >; //
|
||||||
template<class T> using cshiftAllocator = devAllocator<T>;
|
template<class T> using uvmVector = std::vector<T,uvmAllocator<T> >; // auto migrating page
|
||||||
#else
|
template<class T> using deviceVector = std::vector<T,devAllocator<T> >; // device vector
|
||||||
// Cshift on host
|
|
||||||
template<class T> using cshiftAllocator = std::allocator<T>;
|
|
||||||
#endif
|
|
||||||
|
|
||||||
template<class T> using Vector = std::vector<T,uvmAllocator<T> >;
|
|
||||||
template<class T> using stencilVector = std::vector<T,alignedAllocator<T> >;
|
|
||||||
template<class T> using commVector = std::vector<T,devAllocator<T> >;
|
|
||||||
template<class T> using deviceVector = std::vector<T,devAllocator<T> >;
|
|
||||||
template<class T> using cshiftVector = std::vector<T,cshiftAllocator<T> >;
|
|
||||||
|
|
||||||
/*
|
|
||||||
template<class T> class vecView
|
template<class T> class vecView
|
||||||
{
|
{
|
||||||
protected:
|
protected:
|
||||||
@ -197,8 +187,9 @@ template<class T> class vecView
|
|||||||
ViewMode mode;
|
ViewMode mode;
|
||||||
void * cpu_ptr;
|
void * cpu_ptr;
|
||||||
public:
|
public:
|
||||||
|
// Rvalue accessor
|
||||||
accelerator_inline T & operator[](size_t i) const { return this->data[i]; };
|
accelerator_inline T & operator[](size_t i) const { return this->data[i]; };
|
||||||
vecView(std::vector<T> &refer_to_me,ViewMode _mode)
|
vecView(Vector<T> &refer_to_me,ViewMode _mode)
|
||||||
{
|
{
|
||||||
cpu_ptr = &refer_to_me[0];
|
cpu_ptr = &refer_to_me[0];
|
||||||
size = refer_to_me.size();
|
size = refer_to_me.size();
|
||||||
@ -214,26 +205,15 @@ template<class T> class vecView
|
|||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
template<class T> vecView<T> VectorView(std::vector<T> &vec,ViewMode _mode)
|
template<class T> vecView<T> VectorView(Vector<T> &vec,ViewMode _mode)
|
||||||
{
|
{
|
||||||
vecView<T> ret(vec,_mode); // does the open
|
vecView<T> ret(vec,_mode); // does the open
|
||||||
return ret; // must be closed
|
return ret; // must be closed
|
||||||
}
|
}
|
||||||
|
|
||||||
// Little autoscope assister
|
|
||||||
template<class View>
|
|
||||||
class VectorViewCloser
|
|
||||||
{
|
|
||||||
View v; // Take a copy of view and call view close when I go out of scope automatically
|
|
||||||
public:
|
|
||||||
VectorViewCloser(View &_v) : v(_v) {};
|
|
||||||
~VectorViewCloser() { auto ptr = v.cpu_ptr; v.ViewClose(); MemoryManager::NotifyDeletion(ptr);}
|
|
||||||
};
|
|
||||||
|
|
||||||
#define autoVecView(v_v,v,mode) \
|
#define autoVecView(v_v,v,mode) \
|
||||||
auto v_v = VectorView(v,mode); \
|
auto v_v = VectorView(v,mode); \
|
||||||
ViewCloser<decltype(v_v)> _autoView##v_v(v_v);
|
ViewCloser<decltype(v_v)> _autoView##v_v(v_v);
|
||||||
*/
|
|
||||||
|
|
||||||
NAMESPACE_END(Grid);
|
NAMESPACE_END(Grid);
|
||||||
|
|
||||||
|
@ -15,10 +15,10 @@ void check_huge_pages(void *Buf,uint64_t BYTES)
|
|||||||
uint64_t virt_pfn = (uint64_t)Buf / page_size;
|
uint64_t virt_pfn = (uint64_t)Buf / page_size;
|
||||||
off_t offset = sizeof(uint64_t) * virt_pfn;
|
off_t offset = sizeof(uint64_t) * virt_pfn;
|
||||||
uint64_t npages = (BYTES + page_size-1) / page_size;
|
uint64_t npages = (BYTES + page_size-1) / page_size;
|
||||||
uint64_t pagedata[npages];
|
std::vector<uint64_t> pagedata(npages);
|
||||||
uint64_t ret = lseek(fd, offset, SEEK_SET);
|
uint64_t ret = lseek(fd, offset, SEEK_SET);
|
||||||
assert(ret == offset);
|
assert(ret == offset);
|
||||||
ret = ::read(fd, pagedata, sizeof(uint64_t)*npages);
|
ret = ::read(fd, &pagedata[0], sizeof(uint64_t)*npages);
|
||||||
assert(ret == sizeof(uint64_t) * npages);
|
assert(ret == sizeof(uint64_t) * npages);
|
||||||
int nhugepages = npages / 512;
|
int nhugepages = npages / 512;
|
||||||
int n4ktotal, nnothuge;
|
int n4ktotal, nnothuge;
|
||||||
|
@ -51,7 +51,6 @@ Author: Peter Boyle <paboyle@ph.ed.ac.uk>
|
|||||||
#endif
|
#endif
|
||||||
|
|
||||||
NAMESPACE_BEGIN(Grid);
|
NAMESPACE_BEGIN(Grid);
|
||||||
|
|
||||||
template<class Expression,typename std::enable_if<is_lattice_expr<Expression>::value,void>::type * = nullptr>
|
template<class Expression,typename std::enable_if<is_lattice_expr<Expression>::value,void>::type * = nullptr>
|
||||||
auto Cshift(const Expression &expr,int dim,int shift) -> decltype(closure(expr))
|
auto Cshift(const Expression &expr,int dim,int shift) -> decltype(closure(expr))
|
||||||
{
|
{
|
||||||
|
@ -30,12 +30,11 @@ Author: Peter Boyle <paboyle@ph.ed.ac.uk>
|
|||||||
NAMESPACE_BEGIN(Grid);
|
NAMESPACE_BEGIN(Grid);
|
||||||
|
|
||||||
extern std::vector<std::pair<int,int> > Cshift_table;
|
extern std::vector<std::pair<int,int> > Cshift_table;
|
||||||
extern commVector<std::pair<int,int> > Cshift_table_device;
|
extern deviceVector<std::pair<int,int> > Cshift_table_device;
|
||||||
|
|
||||||
inline std::pair<int,int> *MapCshiftTable(void)
|
inline std::pair<int,int> *MapCshiftTable(void)
|
||||||
{
|
{
|
||||||
// GPU version
|
// GPU version
|
||||||
#ifdef ACCELERATOR_CSHIFT
|
|
||||||
uint64_t sz=Cshift_table.size();
|
uint64_t sz=Cshift_table.size();
|
||||||
if (Cshift_table_device.size()!=sz ) {
|
if (Cshift_table_device.size()!=sz ) {
|
||||||
Cshift_table_device.resize(sz);
|
Cshift_table_device.resize(sz);
|
||||||
@ -45,16 +44,13 @@ inline std::pair<int,int> *MapCshiftTable(void)
|
|||||||
sizeof(Cshift_table[0])*sz);
|
sizeof(Cshift_table[0])*sz);
|
||||||
|
|
||||||
return &Cshift_table_device[0];
|
return &Cshift_table_device[0];
|
||||||
#else
|
|
||||||
return &Cshift_table[0];
|
|
||||||
#endif
|
|
||||||
// CPU version use identify map
|
// CPU version use identify map
|
||||||
}
|
}
|
||||||
///////////////////////////////////////////////////////////////////
|
///////////////////////////////////////////////////////////////////
|
||||||
// Gather for when there is no need to SIMD split
|
// Gather for when there is no need to SIMD split
|
||||||
///////////////////////////////////////////////////////////////////
|
///////////////////////////////////////////////////////////////////
|
||||||
template<class vobj> void
|
template<class vobj> void
|
||||||
Gather_plane_simple (const Lattice<vobj> &rhs,cshiftVector<vobj> &buffer,int dimension,int plane,int cbmask, int off=0)
|
Gather_plane_simple (const Lattice<vobj> &rhs,deviceVector<vobj> &buffer,int dimension,int plane,int cbmask, int off=0)
|
||||||
{
|
{
|
||||||
int rd = rhs.Grid()->_rdimensions[dimension];
|
int rd = rhs.Grid()->_rdimensions[dimension];
|
||||||
|
|
||||||
@ -94,17 +90,10 @@ Gather_plane_simple (const Lattice<vobj> &rhs,cshiftVector<vobj> &buffer,int dim
|
|||||||
{
|
{
|
||||||
auto buffer_p = & buffer[0];
|
auto buffer_p = & buffer[0];
|
||||||
auto table = MapCshiftTable();
|
auto table = MapCshiftTable();
|
||||||
#ifdef ACCELERATOR_CSHIFT
|
|
||||||
autoView(rhs_v , rhs, AcceleratorRead);
|
autoView(rhs_v , rhs, AcceleratorRead);
|
||||||
accelerator_for(i,ent,vobj::Nsimd(),{
|
accelerator_for(i,ent,vobj::Nsimd(),{
|
||||||
coalescedWrite(buffer_p[table[i].first],coalescedRead(rhs_v[table[i].second]));
|
coalescedWrite(buffer_p[table[i].first],coalescedRead(rhs_v[table[i].second]));
|
||||||
});
|
});
|
||||||
#else
|
|
||||||
autoView(rhs_v , rhs, CpuRead);
|
|
||||||
thread_for(i,ent,{
|
|
||||||
buffer_p[table[i].first]=rhs_v[table[i].second];
|
|
||||||
});
|
|
||||||
#endif
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -129,7 +118,6 @@ Gather_plane_extract(const Lattice<vobj> &rhs,
|
|||||||
int n1=rhs.Grid()->_slice_stride[dimension];
|
int n1=rhs.Grid()->_slice_stride[dimension];
|
||||||
|
|
||||||
if ( cbmask ==0x3){
|
if ( cbmask ==0x3){
|
||||||
#ifdef ACCELERATOR_CSHIFT
|
|
||||||
autoView(rhs_v , rhs, AcceleratorRead);
|
autoView(rhs_v , rhs, AcceleratorRead);
|
||||||
accelerator_for(nn,e1*e2,1,{
|
accelerator_for(nn,e1*e2,1,{
|
||||||
int n = nn%e1;
|
int n = nn%e1;
|
||||||
@ -140,21 +128,10 @@ Gather_plane_extract(const Lattice<vobj> &rhs,
|
|||||||
vobj temp =rhs_v[so+o+b];
|
vobj temp =rhs_v[so+o+b];
|
||||||
extract<vobj>(temp,pointers,offset);
|
extract<vobj>(temp,pointers,offset);
|
||||||
});
|
});
|
||||||
#else
|
|
||||||
autoView(rhs_v , rhs, CpuRead);
|
|
||||||
thread_for2d(n,e1,b,e2,{
|
|
||||||
int o = n*n1;
|
|
||||||
int offset = b+n*e2;
|
|
||||||
|
|
||||||
vobj temp =rhs_v[so+o+b];
|
|
||||||
extract<vobj>(temp,pointers,offset);
|
|
||||||
});
|
|
||||||
#endif
|
|
||||||
} else {
|
} else {
|
||||||
Coordinate rdim=rhs.Grid()->_rdimensions;
|
Coordinate rdim=rhs.Grid()->_rdimensions;
|
||||||
Coordinate cdm =rhs.Grid()->_checker_dim_mask;
|
Coordinate cdm =rhs.Grid()->_checker_dim_mask;
|
||||||
std::cout << " Dense packed buffer WARNING " <<std::endl; // Does this get called twice once for each cb?
|
std::cout << " Dense packed buffer WARNING " <<std::endl; // Does this get called twice once for each cb?
|
||||||
#ifdef ACCELERATOR_CSHIFT
|
|
||||||
autoView(rhs_v , rhs, AcceleratorRead);
|
autoView(rhs_v , rhs, AcceleratorRead);
|
||||||
accelerator_for(nn,e1*e2,1,{
|
accelerator_for(nn,e1*e2,1,{
|
||||||
int n = nn%e1;
|
int n = nn%e1;
|
||||||
@ -175,33 +152,13 @@ Gather_plane_extract(const Lattice<vobj> &rhs,
|
|||||||
extract<vobj>(temp,pointers,offset);
|
extract<vobj>(temp,pointers,offset);
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
#else
|
|
||||||
autoView(rhs_v , rhs, CpuRead);
|
|
||||||
thread_for2d(n,e1,b,e2,{
|
|
||||||
|
|
||||||
Coordinate coor;
|
|
||||||
|
|
||||||
int o=n*n1;
|
|
||||||
int oindex = o+b;
|
|
||||||
|
|
||||||
int cb = RedBlackCheckerBoardFromOindex(oindex, rdim, cdm);
|
|
||||||
|
|
||||||
int ocb=1<<cb;
|
|
||||||
int offset = b+n*e2;
|
|
||||||
|
|
||||||
if ( ocb & cbmask ) {
|
|
||||||
vobj temp =rhs_v[so+o+b];
|
|
||||||
extract<vobj>(temp,pointers,offset);
|
|
||||||
}
|
|
||||||
});
|
|
||||||
#endif
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
//////////////////////////////////////////////////////
|
//////////////////////////////////////////////////////
|
||||||
// Scatter for when there is no need to SIMD split
|
// Scatter for when there is no need to SIMD split
|
||||||
//////////////////////////////////////////////////////
|
//////////////////////////////////////////////////////
|
||||||
template<class vobj> void Scatter_plane_simple (Lattice<vobj> &rhs,cshiftVector<vobj> &buffer, int dimension,int plane,int cbmask)
|
template<class vobj> void Scatter_plane_simple (Lattice<vobj> &rhs,deviceVector<vobj> &buffer, int dimension,int plane,int cbmask)
|
||||||
{
|
{
|
||||||
int rd = rhs.Grid()->_rdimensions[dimension];
|
int rd = rhs.Grid()->_rdimensions[dimension];
|
||||||
|
|
||||||
@ -245,17 +202,10 @@ template<class vobj> void Scatter_plane_simple (Lattice<vobj> &rhs,cshiftVector<
|
|||||||
{
|
{
|
||||||
auto buffer_p = & buffer[0];
|
auto buffer_p = & buffer[0];
|
||||||
auto table = MapCshiftTable();
|
auto table = MapCshiftTable();
|
||||||
#ifdef ACCELERATOR_CSHIFT
|
|
||||||
autoView( rhs_v, rhs, AcceleratorWrite);
|
autoView( rhs_v, rhs, AcceleratorWrite);
|
||||||
accelerator_for(i,ent,vobj::Nsimd(),{
|
accelerator_for(i,ent,vobj::Nsimd(),{
|
||||||
coalescedWrite(rhs_v[table[i].first],coalescedRead(buffer_p[table[i].second]));
|
coalescedWrite(rhs_v[table[i].first],coalescedRead(buffer_p[table[i].second]));
|
||||||
});
|
});
|
||||||
#else
|
|
||||||
autoView( rhs_v, rhs, CpuWrite);
|
|
||||||
thread_for(i,ent,{
|
|
||||||
rhs_v[table[i].first]=buffer_p[table[i].second];
|
|
||||||
});
|
|
||||||
#endif
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -278,7 +228,6 @@ template<class vobj> void Scatter_plane_merge(Lattice<vobj> &rhs,ExtractPointerA
|
|||||||
if(cbmask ==0x3 ) {
|
if(cbmask ==0x3 ) {
|
||||||
int _slice_stride = rhs.Grid()->_slice_stride[dimension];
|
int _slice_stride = rhs.Grid()->_slice_stride[dimension];
|
||||||
int _slice_block = rhs.Grid()->_slice_block[dimension];
|
int _slice_block = rhs.Grid()->_slice_block[dimension];
|
||||||
#ifdef ACCELERATOR_CSHIFT
|
|
||||||
autoView( rhs_v , rhs, AcceleratorWrite);
|
autoView( rhs_v , rhs, AcceleratorWrite);
|
||||||
accelerator_for(nn,e1*e2,1,{
|
accelerator_for(nn,e1*e2,1,{
|
||||||
int n = nn%e1;
|
int n = nn%e1;
|
||||||
@ -287,14 +236,6 @@ template<class vobj> void Scatter_plane_merge(Lattice<vobj> &rhs,ExtractPointerA
|
|||||||
int offset = b+n*_slice_block;
|
int offset = b+n*_slice_block;
|
||||||
merge(rhs_v[so+o+b],pointers,offset);
|
merge(rhs_v[so+o+b],pointers,offset);
|
||||||
});
|
});
|
||||||
#else
|
|
||||||
autoView( rhs_v , rhs, CpuWrite);
|
|
||||||
thread_for2d(n,e1,b,e2,{
|
|
||||||
int o = n*_slice_stride;
|
|
||||||
int offset = b+n*_slice_block;
|
|
||||||
merge(rhs_v[so+o+b],pointers,offset);
|
|
||||||
});
|
|
||||||
#endif
|
|
||||||
} else {
|
} else {
|
||||||
|
|
||||||
// Case of SIMD split AND checker dim cannot currently be hit, except in
|
// Case of SIMD split AND checker dim cannot currently be hit, except in
|
||||||
@ -360,19 +301,11 @@ template<class vobj> void Copy_plane(Lattice<vobj>& lhs,const Lattice<vobj> &rhs
|
|||||||
|
|
||||||
{
|
{
|
||||||
auto table = MapCshiftTable();
|
auto table = MapCshiftTable();
|
||||||
#ifdef ACCELERATOR_CSHIFT
|
|
||||||
autoView(rhs_v , rhs, AcceleratorRead);
|
autoView(rhs_v , rhs, AcceleratorRead);
|
||||||
autoView(lhs_v , lhs, AcceleratorWrite);
|
autoView(lhs_v , lhs, AcceleratorWrite);
|
||||||
accelerator_for(i,ent,vobj::Nsimd(),{
|
accelerator_for(i,ent,vobj::Nsimd(),{
|
||||||
coalescedWrite(lhs_v[table[i].first],coalescedRead(rhs_v[table[i].second]));
|
coalescedWrite(lhs_v[table[i].first],coalescedRead(rhs_v[table[i].second]));
|
||||||
});
|
});
|
||||||
#else
|
|
||||||
autoView(rhs_v , rhs, CpuRead);
|
|
||||||
autoView(lhs_v , lhs, CpuWrite);
|
|
||||||
thread_for(i,ent,{
|
|
||||||
lhs_v[table[i].first]=rhs_v[table[i].second];
|
|
||||||
});
|
|
||||||
#endif
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -412,19 +345,11 @@ template<class vobj> void Copy_plane_permute(Lattice<vobj>& lhs,const Lattice<vo
|
|||||||
|
|
||||||
{
|
{
|
||||||
auto table = MapCshiftTable();
|
auto table = MapCshiftTable();
|
||||||
#ifdef ACCELERATOR_CSHIFT
|
|
||||||
autoView( rhs_v, rhs, AcceleratorRead);
|
autoView( rhs_v, rhs, AcceleratorRead);
|
||||||
autoView( lhs_v, lhs, AcceleratorWrite);
|
autoView( lhs_v, lhs, AcceleratorWrite);
|
||||||
accelerator_for(i,ent,1,{
|
accelerator_for(i,ent,1,{
|
||||||
permute(lhs_v[table[i].first],rhs_v[table[i].second],permute_type);
|
permute(lhs_v[table[i].first],rhs_v[table[i].second],permute_type);
|
||||||
});
|
});
|
||||||
#else
|
|
||||||
autoView( rhs_v, rhs, CpuRead);
|
|
||||||
autoView( lhs_v, lhs, CpuWrite);
|
|
||||||
thread_for(i,ent,{
|
|
||||||
permute(lhs_v[table[i].first],rhs_v[table[i].second],permute_type);
|
|
||||||
});
|
|
||||||
#endif
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -55,13 +55,13 @@ template<class vobj> Lattice<vobj> Cshift(const Lattice<vobj> &rhs,int dimension
|
|||||||
RealD t1,t0;
|
RealD t1,t0;
|
||||||
t0=usecond();
|
t0=usecond();
|
||||||
if ( !comm_dim ) {
|
if ( !comm_dim ) {
|
||||||
//std::cout << "CSHIFT: Cshift_local" <<std::endl;
|
std::cout << "CSHIFT: Cshift_local" <<std::endl;
|
||||||
Cshift_local(ret,rhs,dimension,shift); // Handles checkerboarding
|
Cshift_local(ret,rhs,dimension,shift); // Handles checkerboarding
|
||||||
} else if ( splice_dim ) {
|
} else if ( splice_dim ) {
|
||||||
//std::cout << "CSHIFT: Cshift_comms_simd call - splice_dim = " << splice_dim << " shift " << shift << " dimension = " << dimension << std::endl;
|
std::cout << "CSHIFT: Cshift_comms_simd call - splice_dim = " << splice_dim << " shift " << shift << " dimension = " << dimension << std::endl;
|
||||||
Cshift_comms_simd(ret,rhs,dimension,shift);
|
Cshift_comms_simd(ret,rhs,dimension,shift);
|
||||||
} else {
|
} else {
|
||||||
//std::cout << "CSHIFT: Cshift_comms" <<std::endl;
|
std::cout << "CSHIFT: Cshift_comms" <<std::endl;
|
||||||
Cshift_comms(ret,rhs,dimension,shift);
|
Cshift_comms(ret,rhs,dimension,shift);
|
||||||
}
|
}
|
||||||
t1=usecond();
|
t1=usecond();
|
||||||
@ -76,12 +76,12 @@ template<class vobj> void Cshift_comms(Lattice<vobj>& ret,const Lattice<vobj> &r
|
|||||||
sshift[0] = rhs.Grid()->CheckerBoardShiftForCB(rhs.Checkerboard(),dimension,shift,Even);
|
sshift[0] = rhs.Grid()->CheckerBoardShiftForCB(rhs.Checkerboard(),dimension,shift,Even);
|
||||||
sshift[1] = rhs.Grid()->CheckerBoardShiftForCB(rhs.Checkerboard(),dimension,shift,Odd);
|
sshift[1] = rhs.Grid()->CheckerBoardShiftForCB(rhs.Checkerboard(),dimension,shift,Odd);
|
||||||
|
|
||||||
// std::cout << "Cshift_comms dim "<<dimension<<"cb "<<rhs.Checkerboard()<<"shift "<<shift<<" sshift " << sshift[0]<<" "<<sshift[1]<<std::endl;
|
std::cout << "Cshift_comms dim "<<dimension<<"cb "<<rhs.Checkerboard()<<"shift "<<shift<<" sshift " << sshift[0]<<" "<<sshift[1]<<std::endl;
|
||||||
if ( sshift[0] == sshift[1] ) {
|
if ( sshift[0] == sshift[1] ) {
|
||||||
// std::cout << "Single pass Cshift_comms" <<std::endl;
|
std::cout << "Single pass Cshift_comms" <<std::endl;
|
||||||
Cshift_comms(ret,rhs,dimension,shift,0x3);
|
Cshift_comms(ret,rhs,dimension,shift,0x3);
|
||||||
} else {
|
} else {
|
||||||
// std::cout << "Two pass Cshift_comms" <<std::endl;
|
std::cout << "Two pass Cshift_comms" <<std::endl;
|
||||||
Cshift_comms(ret,rhs,dimension,shift,0x1);// if checkerboard is unfavourable take two passes
|
Cshift_comms(ret,rhs,dimension,shift,0x1);// if checkerboard is unfavourable take two passes
|
||||||
Cshift_comms(ret,rhs,dimension,shift,0x2);// both with block stride loop iteration
|
Cshift_comms(ret,rhs,dimension,shift,0x2);// both with block stride loop iteration
|
||||||
}
|
}
|
||||||
@ -94,18 +94,16 @@ template<class vobj> void Cshift_comms_simd(Lattice<vobj>& ret,const Lattice<vob
|
|||||||
sshift[0] = rhs.Grid()->CheckerBoardShiftForCB(rhs.Checkerboard(),dimension,shift,Even);
|
sshift[0] = rhs.Grid()->CheckerBoardShiftForCB(rhs.Checkerboard(),dimension,shift,Even);
|
||||||
sshift[1] = rhs.Grid()->CheckerBoardShiftForCB(rhs.Checkerboard(),dimension,shift,Odd);
|
sshift[1] = rhs.Grid()->CheckerBoardShiftForCB(rhs.Checkerboard(),dimension,shift,Odd);
|
||||||
|
|
||||||
//std::cout << "Cshift_comms_simd dim "<<dimension<<"cb "<<rhs.checkerboard<<"shift "<<shift<<" sshift " << sshift[0]<<" "<<sshift[1]<<std::endl;
|
std::cout << "Cshift_comms_simd dim "<<dimension<<"cb "<<rhs.Checkerboard()<<"shift "<<shift<<" sshift " << sshift[0]<<" "<<sshift[1]<<std::endl;
|
||||||
if ( sshift[0] == sshift[1] ) {
|
if ( sshift[0] == sshift[1] ) {
|
||||||
//std::cout << "Single pass Cshift_comms" <<std::endl;
|
std::cout << "Single pass Cshift_comms" <<std::endl;
|
||||||
Cshift_comms_simd(ret,rhs,dimension,shift,0x3);
|
Cshift_comms_simd(ret,rhs,dimension,shift,0x3);
|
||||||
} else {
|
} else {
|
||||||
//std::cout << "Two pass Cshift_comms" <<std::endl;
|
std::cout << "Two pass Cshift_comms" <<std::endl;
|
||||||
Cshift_comms_simd(ret,rhs,dimension,shift,0x1);// if checkerboard is unfavourable take two passes
|
Cshift_comms_simd(ret,rhs,dimension,shift,0x1);// if checkerboard is unfavourable take two passes
|
||||||
Cshift_comms_simd(ret,rhs,dimension,shift,0x2);// both with block stride loop iteration
|
Cshift_comms_simd(ret,rhs,dimension,shift,0x2);// both with block stride loop iteration
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
#define ACCELERATOR_CSHIFT_NO_COPY
|
|
||||||
#ifdef ACCELERATOR_CSHIFT_NO_COPY
|
|
||||||
template<class vobj> void Cshift_comms(Lattice<vobj> &ret,const Lattice<vobj> &rhs,int dimension,int shift,int cbmask)
|
template<class vobj> void Cshift_comms(Lattice<vobj> &ret,const Lattice<vobj> &rhs,int dimension,int shift,int cbmask)
|
||||||
{
|
{
|
||||||
typedef typename vobj::vector_type vector_type;
|
typedef typename vobj::vector_type vector_type;
|
||||||
@ -125,8 +123,8 @@ template<class vobj> void Cshift_comms(Lattice<vobj> &ret,const Lattice<vobj> &r
|
|||||||
assert(shift<fd);
|
assert(shift<fd);
|
||||||
|
|
||||||
int buffer_size = rhs.Grid()->_slice_nblock[dimension]*rhs.Grid()->_slice_block[dimension];
|
int buffer_size = rhs.Grid()->_slice_nblock[dimension]*rhs.Grid()->_slice_block[dimension];
|
||||||
static cshiftVector<vobj> send_buf; send_buf.resize(buffer_size);
|
static deviceVector<vobj> send_buf; send_buf.resize(buffer_size);
|
||||||
static cshiftVector<vobj> recv_buf; recv_buf.resize(buffer_size);
|
static deviceVector<vobj> recv_buf; recv_buf.resize(buffer_size);
|
||||||
|
|
||||||
int cb= (cbmask==0x2)? Odd : Even;
|
int cb= (cbmask==0x2)? Odd : Even;
|
||||||
int sshift= rhs.Grid()->CheckerBoardShiftForCB(rhs.Checkerboard(),dimension,shift,cb);
|
int sshift= rhs.Grid()->CheckerBoardShiftForCB(rhs.Checkerboard(),dimension,shift,cb);
|
||||||
@ -161,7 +159,7 @@ template<class vobj> void Cshift_comms(Lattice<vobj> &ret,const Lattice<vobj> &r
|
|||||||
grid->ShiftedRanks(dimension,comm_proc,xmit_to_rank,recv_from_rank);
|
grid->ShiftedRanks(dimension,comm_proc,xmit_to_rank,recv_from_rank);
|
||||||
|
|
||||||
tcomms-=usecond();
|
tcomms-=usecond();
|
||||||
// grid->Barrier();
|
grid->Barrier();
|
||||||
|
|
||||||
grid->SendToRecvFrom((void *)&send_buf[0],
|
grid->SendToRecvFrom((void *)&send_buf[0],
|
||||||
xmit_to_rank,
|
xmit_to_rank,
|
||||||
@ -169,7 +167,7 @@ template<class vobj> void Cshift_comms(Lattice<vobj> &ret,const Lattice<vobj> &r
|
|||||||
recv_from_rank,
|
recv_from_rank,
|
||||||
bytes);
|
bytes);
|
||||||
xbytes+=bytes;
|
xbytes+=bytes;
|
||||||
// grid->Barrier();
|
grid->Barrier();
|
||||||
tcomms+=usecond();
|
tcomms+=usecond();
|
||||||
|
|
||||||
tscatter-=usecond();
|
tscatter-=usecond();
|
||||||
@ -177,13 +175,11 @@ template<class vobj> void Cshift_comms(Lattice<vobj> &ret,const Lattice<vobj> &r
|
|||||||
tscatter+=usecond();
|
tscatter+=usecond();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
/*
|
|
||||||
std::cout << GridLogPerformance << " Cshift copy "<<tcopy/1e3<<" ms"<<std::endl;
|
std::cout << GridLogPerformance << " Cshift copy "<<tcopy/1e3<<" ms"<<std::endl;
|
||||||
std::cout << GridLogPerformance << " Cshift gather "<<tgather/1e3<<" ms"<<std::endl;
|
std::cout << GridLogPerformance << " Cshift gather "<<tgather/1e3<<" ms"<<std::endl;
|
||||||
std::cout << GridLogPerformance << " Cshift scatter "<<tscatter/1e3<<" ms"<<std::endl;
|
std::cout << GridLogPerformance << " Cshift scatter "<<tscatter/1e3<<" ms"<<std::endl;
|
||||||
std::cout << GridLogPerformance << " Cshift comm "<<tcomms/1e3<<" ms"<<std::endl;
|
std::cout << GridLogPerformance << " Cshift comm "<<tcomms/1e3<<" ms"<<std::endl;
|
||||||
std::cout << GridLogPerformance << " Cshift BW "<<(2.0*xbytes)/tcomms<<" MB/s "<<2*xbytes<< " Bytes "<<std::endl;
|
std::cout << GridLogPerformance << " Cshift BW "<<(2.0*xbytes)/tcomms<<" MB/s "<<2*xbytes<< " Bytes "<<std::endl;
|
||||||
*/
|
|
||||||
}
|
}
|
||||||
|
|
||||||
template<class vobj> void Cshift_comms_simd(Lattice<vobj> &ret,const Lattice<vobj> &rhs,int dimension,int shift,int cbmask)
|
template<class vobj> void Cshift_comms_simd(Lattice<vobj> &ret,const Lattice<vobj> &rhs,int dimension,int shift,int cbmask)
|
||||||
@ -201,9 +197,9 @@ template<class vobj> void Cshift_comms_simd(Lattice<vobj> &ret,const Lattice<vo
|
|||||||
int simd_layout = grid->_simd_layout[dimension];
|
int simd_layout = grid->_simd_layout[dimension];
|
||||||
int comm_dim = grid->_processors[dimension] >1 ;
|
int comm_dim = grid->_processors[dimension] >1 ;
|
||||||
|
|
||||||
//std::cout << "Cshift_comms_simd dim "<< dimension << " fd "<<fd<<" rd "<<rd
|
std::cout << "Cshift_comms_simd dim "<< dimension << " fd "<<fd<<" rd "<<rd
|
||||||
// << " ld "<<ld<<" pd " << pd<<" simd_layout "<<simd_layout
|
<< " ld "<<ld<<" pd " << pd<<" simd_layout "<<simd_layout
|
||||||
// << " comm_dim " << comm_dim << " cbmask " << cbmask <<std::endl;
|
<< " comm_dim " << comm_dim << " cbmask " << cbmask <<std::endl;
|
||||||
|
|
||||||
assert(comm_dim==1);
|
assert(comm_dim==1);
|
||||||
assert(simd_layout==2);
|
assert(simd_layout==2);
|
||||||
@ -224,8 +220,8 @@ template<class vobj> void Cshift_comms_simd(Lattice<vobj> &ret,const Lattice<vo
|
|||||||
int buffer_size = grid->_slice_nblock[dimension]*grid->_slice_block[dimension];
|
int buffer_size = grid->_slice_nblock[dimension]*grid->_slice_block[dimension];
|
||||||
// int words = sizeof(vobj)/sizeof(vector_type);
|
// int words = sizeof(vobj)/sizeof(vector_type);
|
||||||
|
|
||||||
static std::vector<cshiftVector<scalar_object> > send_buf_extract; send_buf_extract.resize(Nsimd);
|
static std::vector<deviceVector<scalar_object> > send_buf_extract; send_buf_extract.resize(Nsimd);
|
||||||
static std::vector<cshiftVector<scalar_object> > recv_buf_extract; recv_buf_extract.resize(Nsimd);
|
static std::vector<deviceVector<scalar_object> > recv_buf_extract; recv_buf_extract.resize(Nsimd);
|
||||||
scalar_object * recv_buf_extract_mpi;
|
scalar_object * recv_buf_extract_mpi;
|
||||||
scalar_object * send_buf_extract_mpi;
|
scalar_object * send_buf_extract_mpi;
|
||||||
|
|
||||||
@ -281,7 +277,7 @@ template<class vobj> void Cshift_comms_simd(Lattice<vobj> &ret,const Lattice<vo
|
|||||||
grid->ShiftedRanks(dimension,nbr_proc,xmit_to_rank,recv_from_rank);
|
grid->ShiftedRanks(dimension,nbr_proc,xmit_to_rank,recv_from_rank);
|
||||||
|
|
||||||
tcomms-=usecond();
|
tcomms-=usecond();
|
||||||
// grid->Barrier();
|
grid->Barrier();
|
||||||
|
|
||||||
send_buf_extract_mpi = &send_buf_extract[nbr_lane][0];
|
send_buf_extract_mpi = &send_buf_extract[nbr_lane][0];
|
||||||
recv_buf_extract_mpi = &recv_buf_extract[i][0];
|
recv_buf_extract_mpi = &recv_buf_extract[i][0];
|
||||||
@ -292,7 +288,7 @@ template<class vobj> void Cshift_comms_simd(Lattice<vobj> &ret,const Lattice<vo
|
|||||||
bytes);
|
bytes);
|
||||||
|
|
||||||
xbytes+=bytes;
|
xbytes+=bytes;
|
||||||
// grid->Barrier();
|
grid->Barrier();
|
||||||
tcomms+=usecond();
|
tcomms+=usecond();
|
||||||
|
|
||||||
rpointers[i] = &recv_buf_extract[i][0];
|
rpointers[i] = &recv_buf_extract[i][0];
|
||||||
@ -305,242 +301,12 @@ template<class vobj> void Cshift_comms_simd(Lattice<vobj> &ret,const Lattice<vo
|
|||||||
Scatter_plane_merge(ret,rpointers,dimension,x,cbmask);
|
Scatter_plane_merge(ret,rpointers,dimension,x,cbmask);
|
||||||
tscatter+=usecond();
|
tscatter+=usecond();
|
||||||
}
|
}
|
||||||
/*
|
|
||||||
std::cout << GridLogPerformance << " Cshift (s) copy "<<tcopy/1e3<<" ms"<<std::endl;
|
std::cout << GridLogPerformance << " Cshift (s) copy "<<tcopy/1e3<<" ms"<<std::endl;
|
||||||
std::cout << GridLogPerformance << " Cshift (s) gather "<<tgather/1e3<<" ms"<<std::endl;
|
std::cout << GridLogPerformance << " Cshift (s) gather "<<tgather/1e3<<" ms"<<std::endl;
|
||||||
std::cout << GridLogPerformance << " Cshift (s) scatter "<<tscatter/1e3<<" ms"<<std::endl;
|
std::cout << GridLogPerformance << " Cshift (s) scatter "<<tscatter/1e3<<" ms"<<std::endl;
|
||||||
std::cout << GridLogPerformance << " Cshift (s) comm "<<tcomms/1e3<<" ms"<<std::endl;
|
std::cout << GridLogPerformance << " Cshift (s) comm "<<tcomms/1e3<<" ms"<<std::endl;
|
||||||
std::cout << GridLogPerformance << " Cshift BW "<<(2.0*xbytes)/tcomms<<" MB/s "<<2*xbytes<< " Bytes "<<std::endl;
|
std::cout << GridLogPerformance << " Cshift BW "<<(2.0*xbytes)/tcomms<<" MB/s "<<2*xbytes<< " Bytes "<<std::endl;
|
||||||
*/
|
|
||||||
}
|
}
|
||||||
#else
|
|
||||||
template<class vobj> void Cshift_comms(Lattice<vobj> &ret,const Lattice<vobj> &rhs,int dimension,int shift,int cbmask)
|
|
||||||
{
|
|
||||||
typedef typename vobj::vector_type vector_type;
|
|
||||||
typedef typename vobj::scalar_type scalar_type;
|
|
||||||
|
|
||||||
GridBase *grid=rhs.Grid();
|
|
||||||
Lattice<vobj> temp(rhs.Grid());
|
|
||||||
|
|
||||||
int fd = rhs.Grid()->_fdimensions[dimension];
|
|
||||||
int rd = rhs.Grid()->_rdimensions[dimension];
|
|
||||||
int pd = rhs.Grid()->_processors[dimension];
|
|
||||||
int simd_layout = rhs.Grid()->_simd_layout[dimension];
|
|
||||||
int comm_dim = rhs.Grid()->_processors[dimension] >1 ;
|
|
||||||
assert(simd_layout==1);
|
|
||||||
assert(comm_dim==1);
|
|
||||||
assert(shift>=0);
|
|
||||||
assert(shift<fd);
|
|
||||||
RealD tcopy=0.0;
|
|
||||||
RealD tgather=0.0;
|
|
||||||
RealD tscatter=0.0;
|
|
||||||
RealD tcomms=0.0;
|
|
||||||
uint64_t xbytes=0;
|
|
||||||
|
|
||||||
int buffer_size = rhs.Grid()->_slice_nblock[dimension]*rhs.Grid()->_slice_block[dimension];
|
|
||||||
static cshiftVector<vobj> send_buf_v; send_buf_v.resize(buffer_size);
|
|
||||||
static cshiftVector<vobj> recv_buf_v; recv_buf_v.resize(buffer_size);
|
|
||||||
vobj *send_buf;
|
|
||||||
vobj *recv_buf;
|
|
||||||
{
|
|
||||||
grid->ShmBufferFreeAll();
|
|
||||||
size_t bytes = buffer_size*sizeof(vobj);
|
|
||||||
send_buf=(vobj *)grid->ShmBufferMalloc(bytes);
|
|
||||||
recv_buf=(vobj *)grid->ShmBufferMalloc(bytes);
|
|
||||||
}
|
|
||||||
|
|
||||||
int cb= (cbmask==0x2)? Odd : Even;
|
|
||||||
int sshift= rhs.Grid()->CheckerBoardShiftForCB(rhs.Checkerboard(),dimension,shift,cb);
|
|
||||||
|
|
||||||
for(int x=0;x<rd;x++){
|
|
||||||
|
|
||||||
int sx = (x+sshift)%rd;
|
|
||||||
int comm_proc = ((x+sshift)/rd)%pd;
|
|
||||||
|
|
||||||
if (comm_proc==0) {
|
|
||||||
|
|
||||||
tcopy-=usecond();
|
|
||||||
Copy_plane(ret,rhs,dimension,x,sx,cbmask);
|
|
||||||
tcopy+=usecond();
|
|
||||||
|
|
||||||
} else {
|
|
||||||
|
|
||||||
int words = buffer_size;
|
|
||||||
if (cbmask != 0x3) words=words>>1;
|
|
||||||
|
|
||||||
int bytes = words * sizeof(vobj);
|
|
||||||
|
|
||||||
tgather-=usecond();
|
|
||||||
Gather_plane_simple (rhs,send_buf_v,dimension,sx,cbmask);
|
|
||||||
tgather+=usecond();
|
|
||||||
|
|
||||||
// int rank = grid->_processor;
|
|
||||||
int recv_from_rank;
|
|
||||||
int xmit_to_rank;
|
|
||||||
grid->ShiftedRanks(dimension,comm_proc,xmit_to_rank,recv_from_rank);
|
|
||||||
|
|
||||||
|
|
||||||
tcomms-=usecond();
|
|
||||||
// grid->Barrier();
|
|
||||||
|
|
||||||
acceleratorCopyDeviceToDevice((void *)&send_buf_v[0],(void *)&send_buf[0],bytes);
|
|
||||||
grid->SendToRecvFrom((void *)&send_buf[0],
|
|
||||||
xmit_to_rank,
|
|
||||||
(void *)&recv_buf[0],
|
|
||||||
recv_from_rank,
|
|
||||||
bytes);
|
|
||||||
xbytes+=bytes;
|
|
||||||
acceleratorCopyDeviceToDevice((void *)&recv_buf[0],(void *)&recv_buf_v[0],bytes);
|
|
||||||
|
|
||||||
// grid->Barrier();
|
|
||||||
tcomms+=usecond();
|
|
||||||
|
|
||||||
tscatter-=usecond();
|
|
||||||
Scatter_plane_simple (ret,recv_buf_v,dimension,x,cbmask);
|
|
||||||
tscatter+=usecond();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
/*
|
|
||||||
std::cout << GridLogPerformance << " Cshift copy "<<tcopy/1e3<<" ms"<<std::endl;
|
|
||||||
std::cout << GridLogPerformance << " Cshift gather "<<tgather/1e3<<" ms"<<std::endl;
|
|
||||||
std::cout << GridLogPerformance << " Cshift scatter "<<tscatter/1e3<<" ms"<<std::endl;
|
|
||||||
std::cout << GridLogPerformance << " Cshift comm "<<tcomms/1e3<<" ms"<<std::endl;
|
|
||||||
std::cout << GridLogPerformance << " Cshift BW "<<(2.0*xbytes)/tcomms<<" MB/s "<<2*xbytes<< " Bytes "<<std::endl;
|
|
||||||
*/
|
|
||||||
}
|
|
||||||
|
|
||||||
template<class vobj> void Cshift_comms_simd(Lattice<vobj> &ret,const Lattice<vobj> &rhs,int dimension,int shift,int cbmask)
|
|
||||||
{
|
|
||||||
GridBase *grid=rhs.Grid();
|
|
||||||
const int Nsimd = grid->Nsimd();
|
|
||||||
typedef typename vobj::vector_type vector_type;
|
|
||||||
typedef typename vobj::scalar_object scalar_object;
|
|
||||||
typedef typename vobj::scalar_type scalar_type;
|
|
||||||
|
|
||||||
int fd = grid->_fdimensions[dimension];
|
|
||||||
int rd = grid->_rdimensions[dimension];
|
|
||||||
int ld = grid->_ldimensions[dimension];
|
|
||||||
int pd = grid->_processors[dimension];
|
|
||||||
int simd_layout = grid->_simd_layout[dimension];
|
|
||||||
int comm_dim = grid->_processors[dimension] >1 ;
|
|
||||||
|
|
||||||
//std::cout << "Cshift_comms_simd dim "<< dimension << " fd "<<fd<<" rd "<<rd
|
|
||||||
// << " ld "<<ld<<" pd " << pd<<" simd_layout "<<simd_layout
|
|
||||||
// << " comm_dim " << comm_dim << " cbmask " << cbmask <<std::endl;
|
|
||||||
|
|
||||||
assert(comm_dim==1);
|
|
||||||
assert(simd_layout==2);
|
|
||||||
assert(shift>=0);
|
|
||||||
assert(shift<fd);
|
|
||||||
RealD tcopy=0.0;
|
|
||||||
RealD tgather=0.0;
|
|
||||||
RealD tscatter=0.0;
|
|
||||||
RealD tcomms=0.0;
|
|
||||||
uint64_t xbytes=0;
|
|
||||||
|
|
||||||
int permute_type=grid->PermuteType(dimension);
|
|
||||||
|
|
||||||
///////////////////////////////////////////////
|
|
||||||
// Simd direction uses an extract/merge pair
|
|
||||||
///////////////////////////////////////////////
|
|
||||||
int buffer_size = grid->_slice_nblock[dimension]*grid->_slice_block[dimension];
|
|
||||||
// int words = sizeof(vobj)/sizeof(vector_type);
|
|
||||||
|
|
||||||
static std::vector<cshiftVector<scalar_object> > send_buf_extract; send_buf_extract.resize(Nsimd);
|
|
||||||
static std::vector<cshiftVector<scalar_object> > recv_buf_extract; recv_buf_extract.resize(Nsimd);
|
|
||||||
scalar_object * recv_buf_extract_mpi;
|
|
||||||
scalar_object * send_buf_extract_mpi;
|
|
||||||
{
|
|
||||||
size_t bytes = sizeof(scalar_object)*buffer_size;
|
|
||||||
grid->ShmBufferFreeAll();
|
|
||||||
send_buf_extract_mpi = (scalar_object *)grid->ShmBufferMalloc(bytes);
|
|
||||||
recv_buf_extract_mpi = (scalar_object *)grid->ShmBufferMalloc(bytes);
|
|
||||||
}
|
|
||||||
for(int s=0;s<Nsimd;s++){
|
|
||||||
send_buf_extract[s].resize(buffer_size);
|
|
||||||
recv_buf_extract[s].resize(buffer_size);
|
|
||||||
}
|
|
||||||
|
|
||||||
int bytes = buffer_size*sizeof(scalar_object);
|
|
||||||
|
|
||||||
ExtractPointerArray<scalar_object> pointers(Nsimd); //
|
|
||||||
ExtractPointerArray<scalar_object> rpointers(Nsimd); // received pointers
|
|
||||||
|
|
||||||
///////////////////////////////////////////
|
|
||||||
// Work out what to send where
|
|
||||||
///////////////////////////////////////////
|
|
||||||
int cb = (cbmask==0x2)? Odd : Even;
|
|
||||||
int sshift= grid->CheckerBoardShiftForCB(rhs.Checkerboard(),dimension,shift,cb);
|
|
||||||
|
|
||||||
// loop over outer coord planes orthog to dim
|
|
||||||
for(int x=0;x<rd;x++){
|
|
||||||
|
|
||||||
// FIXME call local permute copy if none are offnode.
|
|
||||||
for(int i=0;i<Nsimd;i++){
|
|
||||||
pointers[i] = &send_buf_extract[i][0];
|
|
||||||
}
|
|
||||||
tgather-=usecond();
|
|
||||||
int sx = (x+sshift)%rd;
|
|
||||||
Gather_plane_extract(rhs,pointers,dimension,sx,cbmask);
|
|
||||||
tgather+=usecond();
|
|
||||||
|
|
||||||
for(int i=0;i<Nsimd;i++){
|
|
||||||
|
|
||||||
int inner_bit = (Nsimd>>(permute_type+1));
|
|
||||||
int ic= (i&inner_bit)? 1:0;
|
|
||||||
|
|
||||||
int my_coor = rd*ic + x;
|
|
||||||
int nbr_coor = my_coor+sshift;
|
|
||||||
int nbr_proc = ((nbr_coor)/ld) % pd;// relative shift in processors
|
|
||||||
|
|
||||||
int nbr_ic = (nbr_coor%ld)/rd; // inner coord of peer
|
|
||||||
int nbr_ox = (nbr_coor%rd); // outer coord of peer
|
|
||||||
int nbr_lane = (i&(~inner_bit));
|
|
||||||
|
|
||||||
int recv_from_rank;
|
|
||||||
int xmit_to_rank;
|
|
||||||
|
|
||||||
if (nbr_ic) nbr_lane|=inner_bit;
|
|
||||||
|
|
||||||
assert (sx == nbr_ox);
|
|
||||||
|
|
||||||
if(nbr_proc){
|
|
||||||
grid->ShiftedRanks(dimension,nbr_proc,xmit_to_rank,recv_from_rank);
|
|
||||||
|
|
||||||
tcomms-=usecond();
|
|
||||||
// grid->Barrier();
|
|
||||||
|
|
||||||
acceleratorCopyDeviceToDevice((void *)&send_buf_extract[nbr_lane][0],(void *)send_buf_extract_mpi,bytes);
|
|
||||||
grid->SendToRecvFrom((void *)send_buf_extract_mpi,
|
|
||||||
xmit_to_rank,
|
|
||||||
(void *)recv_buf_extract_mpi,
|
|
||||||
recv_from_rank,
|
|
||||||
bytes);
|
|
||||||
acceleratorCopyDeviceToDevice((void *)recv_buf_extract_mpi,(void *)&recv_buf_extract[i][0],bytes);
|
|
||||||
xbytes+=bytes;
|
|
||||||
|
|
||||||
// grid->Barrier();
|
|
||||||
tcomms+=usecond();
|
|
||||||
rpointers[i] = &recv_buf_extract[i][0];
|
|
||||||
} else {
|
|
||||||
rpointers[i] = &send_buf_extract[nbr_lane][0];
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
|
||||||
tscatter-=usecond();
|
|
||||||
Scatter_plane_merge(ret,rpointers,dimension,x,cbmask);
|
|
||||||
tscatter+=usecond();
|
|
||||||
|
|
||||||
}
|
|
||||||
/*
|
|
||||||
std::cout << GridLogPerformance << " Cshift (s) copy "<<tcopy/1e3<<" ms"<<std::endl;
|
|
||||||
std::cout << GridLogPerformance << " Cshift (s) gather "<<tgather/1e3<<" ms"<<std::endl;
|
|
||||||
std::cout << GridLogPerformance << " Cshift (s) scatter "<<tscatter/1e3<<" ms"<<std::endl;
|
|
||||||
std::cout << GridLogPerformance << " Cshift (s) comm "<<tcomms/1e3<<" ms"<<std::endl;
|
|
||||||
std::cout << GridLogPerformance << " Cshift BW "<<(2.0*xbytes)/tcomms<<" MB/s"<<std::endl;
|
|
||||||
*/
|
|
||||||
}
|
|
||||||
#endif
|
|
||||||
NAMESPACE_END(Grid);
|
NAMESPACE_END(Grid);
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
@ -1,5 +1,5 @@
|
|||||||
#include <Grid/GridCore.h>
|
#include <Grid/GridCore.h>
|
||||||
NAMESPACE_BEGIN(Grid);
|
NAMESPACE_BEGIN(Grid);
|
||||||
std::vector<std::pair<int,int> > Cshift_table;
|
std::vector<std::pair<int,int> > Cshift_table;
|
||||||
commVector<std::pair<int,int> > Cshift_table_device;
|
deviceVector<std::pair<int,int> > Cshift_table_device;
|
||||||
NAMESPACE_END(Grid);
|
NAMESPACE_END(Grid);
|
||||||
|
@ -53,36 +53,19 @@ void basisRotate(VField &basis,Matrix& Qt,int j0, int j1, int k0,int k1,int Nm)
|
|||||||
typedef decltype(basis[0]) Field;
|
typedef decltype(basis[0]) Field;
|
||||||
typedef decltype(basis[0].View(AcceleratorRead)) View;
|
typedef decltype(basis[0].View(AcceleratorRead)) View;
|
||||||
|
|
||||||
Vector<View> basis_v; basis_v.reserve(basis.size());
|
hostVector<View> h_basis_v(basis.size());
|
||||||
typedef typename std::remove_reference<decltype(basis_v[0][0])>::type vobj;
|
deviceVector<View> d_basis_v(basis.size());
|
||||||
|
typedef typename std::remove_reference<decltype(h_basis_v[0][0])>::type vobj;
|
||||||
typedef typename std::remove_reference<decltype(Qt(0,0))>::type Coeff_t;
|
typedef typename std::remove_reference<decltype(Qt(0,0))>::type Coeff_t;
|
||||||
|
|
||||||
GridBase* grid = basis[0].Grid();
|
GridBase* grid = basis[0].Grid();
|
||||||
|
|
||||||
for(int k=0;k<basis.size();k++){
|
for(int k=0;k<basis.size();k++){
|
||||||
basis_v.push_back(basis[k].View(AcceleratorWrite));
|
h_basis_v[k] = basis[k].View(AcceleratorWrite);
|
||||||
|
acceleratorPut(d_basis_v[k],h_basis_v[k]);
|
||||||
}
|
}
|
||||||
|
|
||||||
#if ( !(defined(GRID_CUDA) || defined(GRID_HIP) || defined(GRID_SYCL)) )
|
View *basis_vp = &d_basis_v[0];
|
||||||
int max_threads = thread_max();
|
|
||||||
Vector < vobj > Bt(Nm * max_threads);
|
|
||||||
thread_region
|
|
||||||
{
|
|
||||||
vobj* B = &Bt[Nm * thread_num()];
|
|
||||||
thread_for_in_region(ss, grid->oSites(),{
|
|
||||||
for(int j=j0; j<j1; ++j) B[j]=0.;
|
|
||||||
|
|
||||||
for(int j=j0; j<j1; ++j){
|
|
||||||
for(int k=k0; k<k1; ++k){
|
|
||||||
B[j] +=Qt(j,k) * basis_v[k][ss];
|
|
||||||
}
|
|
||||||
}
|
|
||||||
for(int j=j0; j<j1; ++j){
|
|
||||||
basis_v[j][ss] = B[j];
|
|
||||||
}
|
|
||||||
});
|
|
||||||
}
|
|
||||||
#else
|
|
||||||
View *basis_vp = &basis_v[0];
|
|
||||||
|
|
||||||
int nrot = j1-j0;
|
int nrot = j1-j0;
|
||||||
if (!nrot) // edge case not handled gracefully by Cuda
|
if (!nrot) // edge case not handled gracefully by Cuda
|
||||||
@ -91,17 +74,19 @@ void basisRotate(VField &basis,Matrix& Qt,int j0, int j1, int k0,int k1,int Nm)
|
|||||||
uint64_t oSites =grid->oSites();
|
uint64_t oSites =grid->oSites();
|
||||||
uint64_t siteBlock=(grid->oSites()+nrot-1)/nrot; // Maximum 1 additional vector overhead
|
uint64_t siteBlock=(grid->oSites()+nrot-1)/nrot; // Maximum 1 additional vector overhead
|
||||||
|
|
||||||
Vector <vobj> Bt(siteBlock * nrot);
|
deviceVector <vobj> Bt(siteBlock * nrot);
|
||||||
auto Bp=&Bt[0];
|
auto Bp=&Bt[0];
|
||||||
|
|
||||||
// GPU readable copy of matrix
|
// GPU readable copy of matrix
|
||||||
Vector<Coeff_t> Qt_jv(Nm*Nm);
|
hostVector<Coeff_t> h_Qt_jv(Nm*Nm);
|
||||||
|
deviceVector<Coeff_t> Qt_jv(Nm*Nm);
|
||||||
Coeff_t *Qt_p = & Qt_jv[0];
|
Coeff_t *Qt_p = & Qt_jv[0];
|
||||||
thread_for(i,Nm*Nm,{
|
thread_for(i,Nm*Nm,{
|
||||||
int j = i/Nm;
|
int j = i/Nm;
|
||||||
int k = i%Nm;
|
int k = i%Nm;
|
||||||
Qt_p[i]=Qt(j,k);
|
h_Qt_jv[i]=Qt(j,k);
|
||||||
});
|
});
|
||||||
|
acceleratorCopyToDevice(&h_Qt_jv[0],Qt_p,Nm*Nm*sizeof(Coeff_t));
|
||||||
|
|
||||||
// Block the loop to keep storage footprint down
|
// Block the loop to keep storage footprint down
|
||||||
for(uint64_t s=0;s<oSites;s+=siteBlock){
|
for(uint64_t s=0;s<oSites;s+=siteBlock){
|
||||||
@ -137,9 +122,8 @@ void basisRotate(VField &basis,Matrix& Qt,int j0, int j1, int k0,int k1,int Nm)
|
|||||||
coalescedWrite(basis_vp[jj][sss],coalescedRead(Bp[ss*nrot+j]));
|
coalescedWrite(basis_vp[jj][sss],coalescedRead(Bp[ss*nrot+j]));
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
#endif
|
|
||||||
|
|
||||||
for(int k=0;k<basis.size();k++) basis_v[k].ViewClose();
|
for(int k=0;k<basis.size();k++) h_basis_v[k].ViewClose();
|
||||||
}
|
}
|
||||||
|
|
||||||
// Extract a single rotated vector
|
// Extract a single rotated vector
|
||||||
@ -152,16 +136,19 @@ void basisRotateJ(Field &result,std::vector<Field> &basis,Eigen::MatrixXd& Qt,in
|
|||||||
|
|
||||||
result.Checkerboard() = basis[0].Checkerboard();
|
result.Checkerboard() = basis[0].Checkerboard();
|
||||||
|
|
||||||
Vector<View> basis_v; basis_v.reserve(basis.size());
|
hostVector<View> h_basis_v(basis.size());
|
||||||
|
deviceVector<View> d_basis_v(basis.size());
|
||||||
for(int k=0;k<basis.size();k++){
|
for(int k=0;k<basis.size();k++){
|
||||||
basis_v.push_back(basis[k].View(AcceleratorRead));
|
h_basis_v[k]=basis[k].View(AcceleratorRead);
|
||||||
|
acceleratorPut(d_basis_v[k],h_basis_v[k]);
|
||||||
}
|
}
|
||||||
vobj zz=Zero();
|
|
||||||
Vector<double> Qt_jv(Nm);
|
|
||||||
double * Qt_j = & Qt_jv[0];
|
|
||||||
for(int k=0;k<Nm;++k) Qt_j[k]=Qt(j,k);
|
|
||||||
|
|
||||||
auto basis_vp=& basis_v[0];
|
vobj zz=Zero();
|
||||||
|
deviceVector<double> Qt_jv(Nm);
|
||||||
|
double * Qt_j = & Qt_jv[0];
|
||||||
|
for(int k=0;k<Nm;++k) acceleratorPut(Qt_j[k],Qt(j,k));
|
||||||
|
|
||||||
|
auto basis_vp=& d_basis_v[0];
|
||||||
autoView(result_v,result,AcceleratorWrite);
|
autoView(result_v,result,AcceleratorWrite);
|
||||||
accelerator_for(ss, grid->oSites(),vobj::Nsimd(),{
|
accelerator_for(ss, grid->oSites(),vobj::Nsimd(),{
|
||||||
vobj zzz=Zero();
|
vobj zzz=Zero();
|
||||||
@ -171,7 +158,7 @@ void basisRotateJ(Field &result,std::vector<Field> &basis,Eigen::MatrixXd& Qt,in
|
|||||||
}
|
}
|
||||||
coalescedWrite(result_v[ss], B);
|
coalescedWrite(result_v[ss], B);
|
||||||
});
|
});
|
||||||
for(int k=0;k<basis.size();k++) basis_v[k].ViewClose();
|
for(int k=0;k<basis.size();k++) h_basis_v[k].ViewClose();
|
||||||
}
|
}
|
||||||
|
|
||||||
template<class Field>
|
template<class Field>
|
||||||
|
@ -46,7 +46,7 @@ inline typename vobj::scalar_object sum_cpu(const vobj *arg, Integer osites)
|
|||||||
// const int Nsimd = vobj::Nsimd();
|
// const int Nsimd = vobj::Nsimd();
|
||||||
const int nthread = GridThread::GetThreads();
|
const int nthread = GridThread::GetThreads();
|
||||||
|
|
||||||
Vector<sobj> sumarray(nthread);
|
std::vector<sobj> sumarray(nthread);
|
||||||
for(int i=0;i<nthread;i++){
|
for(int i=0;i<nthread;i++){
|
||||||
sumarray[i]=Zero();
|
sumarray[i]=Zero();
|
||||||
}
|
}
|
||||||
@ -75,7 +75,7 @@ inline typename vobj::scalar_objectD sumD_cpu(const vobj *arg, Integer osites)
|
|||||||
|
|
||||||
const int nthread = GridThread::GetThreads();
|
const int nthread = GridThread::GetThreads();
|
||||||
|
|
||||||
Vector<sobj> sumarray(nthread);
|
std::vector<sobj> sumarray(nthread);
|
||||||
for(int i=0;i<nthread;i++){
|
for(int i=0;i<nthread;i++){
|
||||||
sumarray[i]=Zero();
|
sumarray[i]=Zero();
|
||||||
}
|
}
|
||||||
@ -343,18 +343,6 @@ axpby_norm_fast(Lattice<vobj> &z,sobj a,sobj b,const Lattice<vobj> &x,const Latt
|
|||||||
autoView( x_v, x, AcceleratorRead);
|
autoView( x_v, x, AcceleratorRead);
|
||||||
autoView( y_v, y, AcceleratorRead);
|
autoView( y_v, y, AcceleratorRead);
|
||||||
autoView( z_v, z, AcceleratorWrite);
|
autoView( z_v, z, AcceleratorWrite);
|
||||||
#if 0
|
|
||||||
typedef decltype(innerProductD(x_v[0],y_v[0])) inner_t;
|
|
||||||
Vector<inner_t> inner_tmp(sites);
|
|
||||||
auto inner_tmp_v = &inner_tmp[0];
|
|
||||||
|
|
||||||
accelerator_for( ss, sites, nsimd,{
|
|
||||||
auto tmp = a*x_v(ss)+b*y_v(ss);
|
|
||||||
coalescedWrite(inner_tmp_v[ss],innerProductD(tmp,tmp));
|
|
||||||
coalescedWrite(z_v[ss],tmp);
|
|
||||||
});
|
|
||||||
nrm = real(TensorRemove(sum(inner_tmp_v,sites)));
|
|
||||||
#else
|
|
||||||
typedef decltype(innerProduct(x_v[0],y_v[0])) inner_t;
|
typedef decltype(innerProduct(x_v[0],y_v[0])) inner_t;
|
||||||
deviceVector<inner_t> inner_tmp;
|
deviceVector<inner_t> inner_tmp;
|
||||||
inner_tmp.resize(sites);
|
inner_tmp.resize(sites);
|
||||||
@ -366,7 +354,6 @@ axpby_norm_fast(Lattice<vobj> &z,sobj a,sobj b,const Lattice<vobj> &x,const Latt
|
|||||||
coalescedWrite(z_v[ss],tmp);
|
coalescedWrite(z_v[ss],tmp);
|
||||||
});
|
});
|
||||||
nrm = real(TensorRemove(sumD(inner_tmp_v,sites)));
|
nrm = real(TensorRemove(sumD(inner_tmp_v,sites)));
|
||||||
#endif
|
|
||||||
grid->GlobalSum(nrm);
|
grid->GlobalSum(nrm);
|
||||||
return nrm;
|
return nrm;
|
||||||
}
|
}
|
||||||
@ -377,7 +364,7 @@ innerProductNorm(ComplexD& ip, RealD &nrm, const Lattice<vobj> &left,const Latti
|
|||||||
conformable(left,right);
|
conformable(left,right);
|
||||||
|
|
||||||
typedef typename vobj::vector_typeD vector_type;
|
typedef typename vobj::vector_typeD vector_type;
|
||||||
Vector<ComplexD> tmp(2);
|
std::vector<ComplexD> tmp(2);
|
||||||
|
|
||||||
GridBase *grid = left.Grid();
|
GridBase *grid = left.Grid();
|
||||||
|
|
||||||
@ -387,8 +374,8 @@ innerProductNorm(ComplexD& ip, RealD &nrm, const Lattice<vobj> &left,const Latti
|
|||||||
// GPU
|
// GPU
|
||||||
typedef decltype(innerProductD(vobj(),vobj())) inner_t;
|
typedef decltype(innerProductD(vobj(),vobj())) inner_t;
|
||||||
typedef decltype(innerProductD(vobj(),vobj())) norm_t;
|
typedef decltype(innerProductD(vobj(),vobj())) norm_t;
|
||||||
Vector<inner_t> inner_tmp(sites);
|
deviceVector<inner_t> inner_tmp(sites);
|
||||||
Vector<norm_t> norm_tmp(sites);
|
deviceVector<norm_t> norm_tmp(sites);
|
||||||
auto inner_tmp_v = &inner_tmp[0];
|
auto inner_tmp_v = &inner_tmp[0];
|
||||||
auto norm_tmp_v = &norm_tmp[0];
|
auto norm_tmp_v = &norm_tmp[0];
|
||||||
{
|
{
|
||||||
@ -438,7 +425,9 @@ inline auto sum(const LatticeTrinaryExpression<Op,T1,T2,T3> & expr)
|
|||||||
// sliceSum, sliceInnerProduct, sliceAxpy, sliceNorm etc...
|
// sliceSum, sliceInnerProduct, sliceAxpy, sliceNorm etc...
|
||||||
//////////////////////////////////////////////////////////////////////////////////////////////////////////////
|
//////////////////////////////////////////////////////////////////////////////////////////////////////////////
|
||||||
|
|
||||||
template<class vobj> inline void sliceSum(const Lattice<vobj> &Data,std::vector<typename vobj::scalar_object> &result,int orthogdim)
|
template<class vobj> inline void sliceSum(const Lattice<vobj> &Data,
|
||||||
|
std::vector<typename vobj::scalar_object> &result,
|
||||||
|
int orthogdim)
|
||||||
{
|
{
|
||||||
///////////////////////////////////////////////////////
|
///////////////////////////////////////////////////////
|
||||||
// FIXME precision promoted summation
|
// FIXME precision promoted summation
|
||||||
@ -460,8 +449,8 @@ template<class vobj> inline void sliceSum(const Lattice<vobj> &Data,std::vector<
|
|||||||
int ld=grid->_ldimensions[orthogdim];
|
int ld=grid->_ldimensions[orthogdim];
|
||||||
int rd=grid->_rdimensions[orthogdim];
|
int rd=grid->_rdimensions[orthogdim];
|
||||||
|
|
||||||
Vector<vobj> lvSum(rd); // will locally sum vectors first
|
std::vector<vobj> lvSum(rd); // will locally sum vectors first
|
||||||
Vector<sobj> lsSum(ld,Zero()); // sum across these down to scalars
|
std::vector<sobj> lsSum(ld,Zero()); // sum across these down to scalars
|
||||||
ExtractBuffer<sobj> extracted(Nsimd); // splitting the SIMD
|
ExtractBuffer<sobj> extracted(Nsimd); // splitting the SIMD
|
||||||
|
|
||||||
result.resize(fd); // And then global sum to return the same vector to every node
|
result.resize(fd); // And then global sum to return the same vector to every node
|
||||||
@ -552,8 +541,8 @@ static void sliceInnerProductVector( std::vector<ComplexD> & result, const Latti
|
|||||||
int ld=grid->_ldimensions[orthogdim];
|
int ld=grid->_ldimensions[orthogdim];
|
||||||
int rd=grid->_rdimensions[orthogdim];
|
int rd=grid->_rdimensions[orthogdim];
|
||||||
|
|
||||||
Vector<vector_type> lvSum(rd); // will locally sum vectors first
|
std::vector<vector_type> lvSum(rd); // will locally sum vectors first
|
||||||
Vector<scalar_type > lsSum(ld,scalar_type(0.0)); // sum across these down to scalars
|
std::vector<scalar_type > lsSum(ld,scalar_type(0.0)); // sum across these down to scalars
|
||||||
ExtractBuffer<iScalar<scalar_type> > extracted(Nsimd); // splitting the SIMD
|
ExtractBuffer<iScalar<scalar_type> > extracted(Nsimd); // splitting the SIMD
|
||||||
|
|
||||||
result.resize(fd); // And then global sum to return the same vector to every node for IO to file
|
result.resize(fd); // And then global sum to return the same vector to every node for IO to file
|
||||||
|
@ -214,22 +214,12 @@ inline typename vobj::scalar_objectD sumD_gpu_small(const vobj *lat, Integer osi
|
|||||||
// Move out of UVM
|
// Move out of UVM
|
||||||
// Turns out I had messed up the synchronise after move to compute stream
|
// Turns out I had messed up the synchronise after move to compute stream
|
||||||
// as running this on the default stream fools the synchronise
|
// as running this on the default stream fools the synchronise
|
||||||
#undef UVM_BLOCK_BUFFER
|
deviceVector<sobj> buffer(numBlocks);
|
||||||
#ifndef UVM_BLOCK_BUFFER
|
|
||||||
commVector<sobj> buffer(numBlocks);
|
|
||||||
sobj *buffer_v = &buffer[0];
|
sobj *buffer_v = &buffer[0];
|
||||||
sobj result;
|
sobj result;
|
||||||
reduceKernel<<< numBlocks, numThreads, smemSize, computeStream >>>(lat, buffer_v, size);
|
reduceKernel<<< numBlocks, numThreads, smemSize, computeStream >>>(lat, buffer_v, size);
|
||||||
accelerator_barrier();
|
accelerator_barrier();
|
||||||
acceleratorCopyFromDevice(buffer_v,&result,sizeof(result));
|
acceleratorCopyFromDevice(buffer_v,&result,sizeof(result));
|
||||||
#else
|
|
||||||
Vector<sobj> buffer(numBlocks);
|
|
||||||
sobj *buffer_v = &buffer[0];
|
|
||||||
sobj result;
|
|
||||||
reduceKernel<<< numBlocks, numThreads, smemSize, computeStream >>>(lat, buffer_v, size);
|
|
||||||
accelerator_barrier();
|
|
||||||
result = *buffer_v;
|
|
||||||
#endif
|
|
||||||
return result;
|
return result;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -244,7 +234,7 @@ inline typename vobj::scalar_objectD sumD_gpu_large(const vobj *lat, Integer osi
|
|||||||
|
|
||||||
const int words = sizeof(vobj)/sizeof(vector);
|
const int words = sizeof(vobj)/sizeof(vector);
|
||||||
|
|
||||||
Vector<vector> buffer(osites);
|
deviceVector<vector> buffer(osites);
|
||||||
vector *dat = (vector *)lat;
|
vector *dat = (vector *)lat;
|
||||||
vector *buf = &buffer[0];
|
vector *buf = &buffer[0];
|
||||||
iScalar<vector> *tbuf =(iScalar<vector> *) &buffer[0];
|
iScalar<vector> *tbuf =(iScalar<vector> *) &buffer[0];
|
||||||
|
@ -10,7 +10,7 @@ inline typename vobj::scalar_objectD sumD_gpu_tensor(const vobj *lat, Integer os
|
|||||||
{
|
{
|
||||||
typedef typename vobj::scalar_object sobj;
|
typedef typename vobj::scalar_object sobj;
|
||||||
typedef typename vobj::scalar_objectD sobjD;
|
typedef typename vobj::scalar_objectD sobjD;
|
||||||
#if 1
|
|
||||||
sobj identity; zeroit(identity);
|
sobj identity; zeroit(identity);
|
||||||
sobj ret; zeroit(ret);
|
sobj ret; zeroit(ret);
|
||||||
Integer nsimd= vobj::Nsimd();
|
Integer nsimd= vobj::Nsimd();
|
||||||
@ -28,32 +28,6 @@ inline typename vobj::scalar_objectD sumD_gpu_tensor(const vobj *lat, Integer os
|
|||||||
}
|
}
|
||||||
sobjD dret; convertType(dret,ret);
|
sobjD dret; convertType(dret,ret);
|
||||||
return dret;
|
return dret;
|
||||||
#else
|
|
||||||
static Vector<sobj> mysum;
|
|
||||||
mysum.resize(1);
|
|
||||||
sobj *mysum_p = & mysum[0];
|
|
||||||
sobj identity; zeroit(identity);
|
|
||||||
acceleratorPut(mysum[0],identity);
|
|
||||||
sobj ret ;
|
|
||||||
|
|
||||||
Integer nsimd= vobj::Nsimd();
|
|
||||||
|
|
||||||
const cl::sycl::property_list PropList ({ cl::sycl::property::reduction::initialize_to_identity() });
|
|
||||||
theGridAccelerator->submit([&](cl::sycl::handler &cgh) {
|
|
||||||
auto Reduction = cl::sycl::reduction(mysum_p,identity,std::plus<>(),PropList);
|
|
||||||
cgh.parallel_for(cl::sycl::range<1>{osites},
|
|
||||||
Reduction,
|
|
||||||
[=] (cl::sycl::id<1> item, auto &sum) {
|
|
||||||
auto osite = item[0];
|
|
||||||
sum +=Reduce(lat[osite]);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
theGridAccelerator->wait();
|
|
||||||
ret = mysum[0];
|
|
||||||
// free(mysum,*theGridAccelerator);
|
|
||||||
sobjD dret; convertType(dret,ret);
|
|
||||||
return dret;
|
|
||||||
#endif
|
|
||||||
}
|
}
|
||||||
|
|
||||||
template <class vobj>
|
template <class vobj>
|
||||||
@ -97,7 +71,6 @@ inline typename vobj::scalar_object sum_gpu_large(const vobj *lat, Integer osite
|
|||||||
|
|
||||||
template<class Word> Word svm_xor(Word *vec,uint64_t L)
|
template<class Word> Word svm_xor(Word *vec,uint64_t L)
|
||||||
{
|
{
|
||||||
#if 1
|
|
||||||
Word identity; identity=0;
|
Word identity; identity=0;
|
||||||
Word ret = 0;
|
Word ret = 0;
|
||||||
{
|
{
|
||||||
@ -113,60 +86,7 @@ template<class Word> Word svm_xor(Word *vec,uint64_t L)
|
|||||||
}
|
}
|
||||||
theGridAccelerator->wait();
|
theGridAccelerator->wait();
|
||||||
return ret;
|
return ret;
|
||||||
#else
|
|
||||||
static Vector<Word> d_sum;
|
|
||||||
d_sum.resize(1);
|
|
||||||
Word *d_sum_p=&d_sum[0];
|
|
||||||
Word identity; identity=0;
|
|
||||||
acceleratorPut(d_sum[0],identity);
|
|
||||||
const cl::sycl::property_list PropList ({ cl::sycl::property::reduction::initialize_to_identity() });
|
|
||||||
theGridAccelerator->submit([&](cl::sycl::handler &cgh) {
|
|
||||||
auto Reduction = cl::sycl::reduction(d_sum_p,identity,std::bit_xor<>(),PropList);
|
|
||||||
cgh.parallel_for(cl::sycl::range<1>{L},
|
|
||||||
Reduction,
|
|
||||||
[=] (cl::sycl::id<1> index, auto &sum) {
|
|
||||||
sum^=vec[index];
|
|
||||||
});
|
|
||||||
});
|
|
||||||
theGridAccelerator->wait();
|
|
||||||
Word ret = acceleratorGet(d_sum[0]);
|
|
||||||
// free(d_sum,*theGridAccelerator);
|
|
||||||
return ret;
|
|
||||||
#endif
|
|
||||||
}
|
}
|
||||||
|
|
||||||
NAMESPACE_END(Grid);
|
NAMESPACE_END(Grid);
|
||||||
|
|
||||||
/*
|
|
||||||
|
|
||||||
template <class vobj>
|
|
||||||
inline typename vobj::scalar_objectD sumD_gpu_repack(const vobj *lat, Integer osites)
|
|
||||||
{
|
|
||||||
typedef typename vobj::vector_type vector;
|
|
||||||
typedef typename vobj::scalar_type scalar;
|
|
||||||
|
|
||||||
typedef typename vobj::scalar_typeD scalarD;
|
|
||||||
typedef typename vobj::scalar_objectD sobjD;
|
|
||||||
|
|
||||||
sobjD ret;
|
|
||||||
scalarD *ret_p = (scalarD *)&ret;
|
|
||||||
|
|
||||||
const int nsimd = vobj::Nsimd();
|
|
||||||
const int words = sizeof(vobj)/sizeof(vector);
|
|
||||||
|
|
||||||
Vector<scalar> buffer(osites*nsimd);
|
|
||||||
scalar *buf = &buffer[0];
|
|
||||||
vector *dat = (vector *)lat;
|
|
||||||
|
|
||||||
for(int w=0;w<words;w++) {
|
|
||||||
|
|
||||||
accelerator_for(ss,osites,nsimd,{
|
|
||||||
int lane = acceleratorSIMTlane(nsimd);
|
|
||||||
buf[ss*nsimd+lane] = dat[ss*words+w].getlane(lane);
|
|
||||||
});
|
|
||||||
//Precision change at this point is to late to gain precision
|
|
||||||
ret_p[w] = svm_reduce(buf,nsimd*osites);
|
|
||||||
}
|
|
||||||
return ret;
|
|
||||||
}
|
|
||||||
*/
|
|
||||||
|
@ -21,9 +21,18 @@ NAMESPACE_BEGIN(Grid);
|
|||||||
|
|
||||||
|
|
||||||
#if defined(GRID_CUDA) || defined(GRID_HIP)
|
#if defined(GRID_CUDA) || defined(GRID_HIP)
|
||||||
template<class vobj> inline void sliceSumReduction_cub_small(const vobj *Data, Vector<vobj> &lvSum, const int rd, const int e1, const int e2, const int stride, const int ostride, const int Nsimd) {
|
template<class vobj>
|
||||||
|
inline void sliceSumReduction_cub_small(const vobj *Data,
|
||||||
|
std::vector<vobj> &lvSum,
|
||||||
|
const int rd,
|
||||||
|
const int e1,
|
||||||
|
const int e2,
|
||||||
|
const int stride,
|
||||||
|
const int ostride,
|
||||||
|
const int Nsimd)
|
||||||
|
{
|
||||||
size_t subvol_size = e1*e2;
|
size_t subvol_size = e1*e2;
|
||||||
commVector<vobj> reduction_buffer(rd*subvol_size);
|
deviceVector<vobj> reduction_buffer(rd*subvol_size);
|
||||||
auto rb_p = &reduction_buffer[0];
|
auto rb_p = &reduction_buffer[0];
|
||||||
vobj zero_init;
|
vobj zero_init;
|
||||||
zeroit(zero_init);
|
zeroit(zero_init);
|
||||||
@ -94,7 +103,15 @@ template<class vobj> inline void sliceSumReduction_cub_small(const vobj *Data, V
|
|||||||
|
|
||||||
|
|
||||||
#if defined(GRID_SYCL)
|
#if defined(GRID_SYCL)
|
||||||
template<class vobj> inline void sliceSumReduction_sycl_small(const vobj *Data, Vector <vobj> &lvSum, const int &rd, const int &e1, const int &e2, const int &stride, const int &ostride, const int &Nsimd)
|
template<class vobj>
|
||||||
|
inline void sliceSumReduction_sycl_small(const vobj *Data,
|
||||||
|
std::vector <vobj> &lvSum,
|
||||||
|
const int &rd,
|
||||||
|
const int &e1,
|
||||||
|
const int &e2,
|
||||||
|
const int &stride,
|
||||||
|
const int &ostride,
|
||||||
|
const int &Nsimd)
|
||||||
{
|
{
|
||||||
size_t subvol_size = e1*e2;
|
size_t subvol_size = e1*e2;
|
||||||
|
|
||||||
@ -105,7 +122,7 @@ template<class vobj> inline void sliceSumReduction_sycl_small(const vobj *Data,
|
|||||||
mysum[r] = vobj_zero;
|
mysum[r] = vobj_zero;
|
||||||
}
|
}
|
||||||
|
|
||||||
commVector<vobj> reduction_buffer(rd*subvol_size);
|
deviceVector<vobj> reduction_buffer(rd*subvol_size);
|
||||||
|
|
||||||
auto rb_p = &reduction_buffer[0];
|
auto rb_p = &reduction_buffer[0];
|
||||||
|
|
||||||
@ -144,14 +161,23 @@ template<class vobj> inline void sliceSumReduction_sycl_small(const vobj *Data,
|
|||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
template<class vobj> inline void sliceSumReduction_large(const vobj *Data, Vector<vobj> &lvSum, const int rd, const int e1, const int e2, const int stride, const int ostride, const int Nsimd) {
|
template<class vobj>
|
||||||
|
inline void sliceSumReduction_large(const vobj *Data,
|
||||||
|
std::vector<vobj> &lvSum,
|
||||||
|
const int rd,
|
||||||
|
const int e1,
|
||||||
|
const int e2,
|
||||||
|
const int stride,
|
||||||
|
const int ostride,
|
||||||
|
const int Nsimd)
|
||||||
|
{
|
||||||
typedef typename vobj::vector_type vector;
|
typedef typename vobj::vector_type vector;
|
||||||
const int words = sizeof(vobj)/sizeof(vector);
|
const int words = sizeof(vobj)/sizeof(vector);
|
||||||
const int osites = rd*e1*e2;
|
const int osites = rd*e1*e2;
|
||||||
commVector<vector>buffer(osites);
|
deviceVector<vector>buffer(osites);
|
||||||
vector *dat = (vector *)Data;
|
vector *dat = (vector *)Data;
|
||||||
vector *buf = &buffer[0];
|
vector *buf = &buffer[0];
|
||||||
Vector<vector> lvSum_small(rd);
|
std::vector<vector> lvSum_small(rd);
|
||||||
vector *lvSum_ptr = (vector *)&lvSum[0];
|
vector *lvSum_ptr = (vector *)&lvSum[0];
|
||||||
|
|
||||||
for (int w = 0; w < words; w++) {
|
for (int w = 0; w < words; w++) {
|
||||||
@ -168,13 +194,18 @@ template<class vobj> inline void sliceSumReduction_large(const vobj *Data, Vecto
|
|||||||
for (int r = 0; r < rd; r++) {
|
for (int r = 0; r < rd; r++) {
|
||||||
lvSum_ptr[w+words*r]=lvSum_small[r];
|
lvSum_ptr[w+words*r]=lvSum_small[r];
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
template<class vobj> inline void sliceSumReduction_gpu(const Lattice<vobj> &Data, Vector<vobj> &lvSum, const int rd, const int e1, const int e2, const int stride, const int ostride, const int Nsimd)
|
template<class vobj>
|
||||||
|
inline void sliceSumReduction_gpu(const Lattice<vobj> &Data,
|
||||||
|
std::vector<vobj> &lvSum,
|
||||||
|
const int rd,
|
||||||
|
const int e1,
|
||||||
|
const int e2,
|
||||||
|
const int stride,
|
||||||
|
const int ostride,
|
||||||
|
const int Nsimd)
|
||||||
{
|
{
|
||||||
autoView(Data_v, Data, AcceleratorRead); //reduction libraries cannot deal with large vobjs so we split into small/large case.
|
autoView(Data_v, Data, AcceleratorRead); //reduction libraries cannot deal with large vobjs so we split into small/large case.
|
||||||
if constexpr (sizeof(vobj) <= 256) {
|
if constexpr (sizeof(vobj) <= 256) {
|
||||||
@ -192,7 +223,15 @@ template<class vobj> inline void sliceSumReduction_gpu(const Lattice<vobj> &Data
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
template<class vobj> inline void sliceSumReduction_cpu(const Lattice<vobj> &Data, Vector<vobj> &lvSum, const int &rd, const int &e1, const int &e2, const int &stride, const int &ostride, const int &Nsimd)
|
template<class vobj>
|
||||||
|
inline void sliceSumReduction_cpu(const Lattice<vobj> &Data,
|
||||||
|
std::vector<vobj> &lvSum,
|
||||||
|
const int &rd,
|
||||||
|
const int &e1,
|
||||||
|
const int &e2,
|
||||||
|
const int &stride,
|
||||||
|
const int &ostride,
|
||||||
|
const int &Nsimd)
|
||||||
{
|
{
|
||||||
// sum over reduced dimension planes, breaking out orthog dir
|
// sum over reduced dimension planes, breaking out orthog dir
|
||||||
// Parallel over orthog direction
|
// Parallel over orthog direction
|
||||||
@ -208,16 +247,20 @@ template<class vobj> inline void sliceSumReduction_cpu(const Lattice<vobj> &Data
|
|||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
template<class vobj> inline void sliceSumReduction(const Lattice<vobj> &Data, Vector<vobj> &lvSum, const int &rd, const int &e1, const int &e2, const int &stride, const int &ostride, const int &Nsimd)
|
template<class vobj> inline void sliceSumReduction(const Lattice<vobj> &Data,
|
||||||
|
std::vector<vobj> &lvSum,
|
||||||
|
const int &rd,
|
||||||
|
const int &e1,
|
||||||
|
const int &e2,
|
||||||
|
const int &stride,
|
||||||
|
const int &ostride,
|
||||||
|
const int &Nsimd)
|
||||||
{
|
{
|
||||||
#if defined(GRID_CUDA) || defined(GRID_HIP) || defined(GRID_SYCL)
|
#if defined(GRID_CUDA) || defined(GRID_HIP) || defined(GRID_SYCL)
|
||||||
|
|
||||||
sliceSumReduction_gpu(Data, lvSum, rd, e1, e2, stride, ostride, Nsimd);
|
sliceSumReduction_gpu(Data, lvSum, rd, e1, e2, stride, ostride, Nsimd);
|
||||||
|
#else
|
||||||
#else
|
|
||||||
sliceSumReduction_cpu(Data, lvSum, rd, e1, e2, stride, ostride, Nsimd);
|
sliceSumReduction_cpu(Data, lvSum, rd, e1, e2, stride, ostride, Nsimd);
|
||||||
|
#endif
|
||||||
#endif
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
@ -54,7 +54,7 @@ struct CshiftImplGauge: public CshiftImplBase<typename Gimpl::GaugeLinkField::ve
|
|||||||
*
|
*
|
||||||
*/
|
*/
|
||||||
|
|
||||||
template<class vobj> inline void ScatterSlice(const cshiftVector<vobj> &buf,
|
template<class vobj> inline void ScatterSlice(const deviceVector<vobj> &buf,
|
||||||
Lattice<vobj> &lat,
|
Lattice<vobj> &lat,
|
||||||
int x,
|
int x,
|
||||||
int dim,
|
int dim,
|
||||||
@ -140,7 +140,7 @@ template<class vobj> inline void ScatterSlice(const cshiftVector<vobj> &buf,
|
|||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
template<class vobj> inline void GatherSlice(cshiftVector<vobj> &buf,
|
template<class vobj> inline void GatherSlice(deviceVector<vobj> &buf,
|
||||||
const Lattice<vobj> &lat,
|
const Lattice<vobj> &lat,
|
||||||
int x,
|
int x,
|
||||||
int dim,
|
int dim,
|
||||||
@ -462,8 +462,8 @@ public:
|
|||||||
int rNsimd = Nsimd / simd[dimension];
|
int rNsimd = Nsimd / simd[dimension];
|
||||||
assert( buffer_size == from.Grid()->_slice_nblock[dimension]*from.Grid()->_slice_block[dimension] / simd[dimension]);
|
assert( buffer_size == from.Grid()->_slice_nblock[dimension]*from.Grid()->_slice_block[dimension] / simd[dimension]);
|
||||||
|
|
||||||
static cshiftVector<vobj> send_buf;
|
static deviceVector<vobj> send_buf;
|
||||||
static cshiftVector<vobj> recv_buf;
|
static deviceVector<vobj> recv_buf;
|
||||||
send_buf.resize(buffer_size*2*depth);
|
send_buf.resize(buffer_size*2*depth);
|
||||||
recv_buf.resize(buffer_size*2*depth);
|
recv_buf.resize(buffer_size*2*depth);
|
||||||
|
|
||||||
|
@ -90,16 +90,16 @@ public:
|
|||||||
void M5D(const FermionField &psi,
|
void M5D(const FermionField &psi,
|
||||||
const FermionField &phi,
|
const FermionField &phi,
|
||||||
FermionField &chi,
|
FermionField &chi,
|
||||||
Vector<Coeff_t> &lower,
|
std::vector<Coeff_t> &lower,
|
||||||
Vector<Coeff_t> &diag,
|
std::vector<Coeff_t> &diag,
|
||||||
Vector<Coeff_t> &upper);
|
std::vector<Coeff_t> &upper);
|
||||||
|
|
||||||
void M5Ddag(const FermionField &psi,
|
void M5Ddag(const FermionField &psi,
|
||||||
const FermionField &phi,
|
const FermionField &phi,
|
||||||
FermionField &chi,
|
FermionField &chi,
|
||||||
Vector<Coeff_t> &lower,
|
std::vector<Coeff_t> &lower,
|
||||||
Vector<Coeff_t> &diag,
|
std::vector<Coeff_t> &diag,
|
||||||
Vector<Coeff_t> &upper);
|
std::vector<Coeff_t> &upper);
|
||||||
|
|
||||||
virtual void Instantiatable(void)=0;
|
virtual void Instantiatable(void)=0;
|
||||||
|
|
||||||
@ -119,35 +119,35 @@ public:
|
|||||||
RealD mass_plus, mass_minus;
|
RealD mass_plus, mass_minus;
|
||||||
|
|
||||||
// Save arguments to SetCoefficientsInternal
|
// Save arguments to SetCoefficientsInternal
|
||||||
Vector<Coeff_t> _gamma;
|
std::vector<Coeff_t> _gamma;
|
||||||
RealD _zolo_hi;
|
RealD _zolo_hi;
|
||||||
RealD _b;
|
RealD _b;
|
||||||
RealD _c;
|
RealD _c;
|
||||||
|
|
||||||
// Cayley form Moebius (tanh and zolotarev)
|
// Cayley form Moebius (tanh and zolotarev)
|
||||||
Vector<Coeff_t> omega;
|
std::vector<Coeff_t> omega;
|
||||||
Vector<Coeff_t> bs; // S dependent coeffs
|
std::vector<Coeff_t> bs; // S dependent coeffs
|
||||||
Vector<Coeff_t> cs;
|
std::vector<Coeff_t> cs;
|
||||||
Vector<Coeff_t> as;
|
std::vector<Coeff_t> as;
|
||||||
// For preconditioning Cayley form
|
// For preconditioning Cayley form
|
||||||
Vector<Coeff_t> bee;
|
std::vector<Coeff_t> bee;
|
||||||
Vector<Coeff_t> cee;
|
std::vector<Coeff_t> cee;
|
||||||
Vector<Coeff_t> aee;
|
std::vector<Coeff_t> aee;
|
||||||
Vector<Coeff_t> beo;
|
std::vector<Coeff_t> beo;
|
||||||
Vector<Coeff_t> ceo;
|
std::vector<Coeff_t> ceo;
|
||||||
Vector<Coeff_t> aeo;
|
std::vector<Coeff_t> aeo;
|
||||||
// LDU factorisation of the eeoo matrix
|
// LDU factorisation of the eeoo matrix
|
||||||
Vector<Coeff_t> lee;
|
std::vector<Coeff_t> lee;
|
||||||
Vector<Coeff_t> leem;
|
std::vector<Coeff_t> leem;
|
||||||
Vector<Coeff_t> uee;
|
std::vector<Coeff_t> uee;
|
||||||
Vector<Coeff_t> ueem;
|
std::vector<Coeff_t> ueem;
|
||||||
Vector<Coeff_t> dee;
|
std::vector<Coeff_t> dee;
|
||||||
|
|
||||||
// Matrices of 5d ee inverse params
|
// Matrices of 5d ee inverse params
|
||||||
Vector<iSinglet<Simd> > MatpInv;
|
// std::vector<iSinglet<Simd> > MatpInv;
|
||||||
Vector<iSinglet<Simd> > MatmInv;
|
// std::vector<iSinglet<Simd> > MatmInv;
|
||||||
Vector<iSinglet<Simd> > MatpInvDag;
|
// std::vector<iSinglet<Simd> > MatpInvDag;
|
||||||
Vector<iSinglet<Simd> > MatmInvDag;
|
// std::vector<iSinglet<Simd> > MatmInvDag;
|
||||||
|
|
||||||
///////////////////////////////////////////////////////////////
|
///////////////////////////////////////////////////////////////
|
||||||
// Conserved current utilities
|
// Conserved current utilities
|
||||||
@ -187,7 +187,7 @@ public:
|
|||||||
protected:
|
protected:
|
||||||
virtual void SetCoefficientsZolotarev(RealD zolohi,Approx::zolotarev_data *zdata,RealD b,RealD c);
|
virtual void SetCoefficientsZolotarev(RealD zolohi,Approx::zolotarev_data *zdata,RealD b,RealD c);
|
||||||
virtual void SetCoefficientsTanh(Approx::zolotarev_data *zdata,RealD b,RealD c);
|
virtual void SetCoefficientsTanh(Approx::zolotarev_data *zdata,RealD b,RealD c);
|
||||||
virtual void SetCoefficientsInternal(RealD zolo_hi,Vector<Coeff_t> & gamma,RealD b,RealD c);
|
virtual void SetCoefficientsInternal(RealD zolo_hi,std::vector<Coeff_t> & gamma,RealD b,RealD c);
|
||||||
};
|
};
|
||||||
|
|
||||||
NAMESPACE_END(Grid);
|
NAMESPACE_END(Grid);
|
||||||
|
@ -90,12 +90,12 @@ protected:
|
|||||||
RealD mass;
|
RealD mass;
|
||||||
RealD R;
|
RealD R;
|
||||||
RealD ZoloHiInv;
|
RealD ZoloHiInv;
|
||||||
Vector<double> Beta;
|
std::vector<double> Beta;
|
||||||
Vector<double> cc;;
|
std::vector<double> cc;;
|
||||||
Vector<double> cc_d;;
|
std::vector<double> cc_d;;
|
||||||
Vector<double> sqrt_cc;
|
std::vector<double> sqrt_cc;
|
||||||
Vector<double> See;
|
std::vector<double> See;
|
||||||
Vector<double> Aee;
|
std::vector<double> Aee;
|
||||||
|
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -69,10 +69,10 @@ public:
|
|||||||
// Instantiate different versions depending on Impl
|
// Instantiate different versions depending on Impl
|
||||||
/////////////////////////////////////////////////////
|
/////////////////////////////////////////////////////
|
||||||
void M5D(const FermionField& psi, const FermionField& phi, FermionField& chi,
|
void M5D(const FermionField& psi, const FermionField& phi, FermionField& chi,
|
||||||
Vector<Coeff_t>& lower, Vector<Coeff_t>& diag, Vector<Coeff_t>& upper);
|
std::vector<Coeff_t>& lower, std::vector<Coeff_t>& diag, std::vector<Coeff_t>& upper);
|
||||||
|
|
||||||
void M5Ddag(const FermionField& psi, const FermionField& phi, FermionField& chi,
|
void M5Ddag(const FermionField& psi, const FermionField& phi, FermionField& chi,
|
||||||
Vector<Coeff_t>& lower, Vector<Coeff_t>& diag, Vector<Coeff_t>& upper);
|
std::vector<Coeff_t>& lower, std::vector<Coeff_t>& diag, std::vector<Coeff_t>& upper);
|
||||||
|
|
||||||
virtual void RefreshShiftCoefficients(RealD new_shift);
|
virtual void RefreshShiftCoefficients(RealD new_shift);
|
||||||
|
|
||||||
@ -83,7 +83,7 @@ public:
|
|||||||
RealD _M5, const ImplParams& p=ImplParams());
|
RealD _M5, const ImplParams& p=ImplParams());
|
||||||
|
|
||||||
protected:
|
protected:
|
||||||
void SetCoefficientsInternal(RealD zolo_hi, Vector<Coeff_t>& gamma, RealD b, RealD c);
|
void SetCoefficientsInternal(RealD zolo_hi, std::vector<Coeff_t>& gamma, RealD b, RealD c);
|
||||||
};
|
};
|
||||||
|
|
||||||
NAMESPACE_END(Grid);
|
NAMESPACE_END(Grid);
|
||||||
|
@ -102,11 +102,11 @@ public:
|
|||||||
GaugeField &mat,
|
GaugeField &mat,
|
||||||
const FermionField &A, const FermionField &B, int dag);
|
const FermionField &A, const FermionField &B, int dag);
|
||||||
|
|
||||||
void DhopInternal(StencilImpl &st, LebesgueOrder &lo, DoubledGaugeField &U,DoubledGaugeField &UUU,
|
void DhopInternal(StencilImpl &st, DoubledGaugeField &U,DoubledGaugeField &UUU,
|
||||||
const FermionField &in, FermionField &out, int dag);
|
const FermionField &in, FermionField &out, int dag);
|
||||||
void DhopInternalSerialComms(StencilImpl &st, LebesgueOrder &lo, DoubledGaugeField &U,DoubledGaugeField &UUU,
|
void DhopInternalSerialComms(StencilImpl &st, DoubledGaugeField &U,DoubledGaugeField &UUU,
|
||||||
const FermionField &in, FermionField &out, int dag);
|
const FermionField &in, FermionField &out, int dag);
|
||||||
void DhopInternalOverlappedComms(StencilImpl &st, LebesgueOrder &lo, DoubledGaugeField &U,DoubledGaugeField &UUU,
|
void DhopInternalOverlappedComms(StencilImpl &st, DoubledGaugeField &U,DoubledGaugeField &UUU,
|
||||||
const FermionField &in, FermionField &out, int dag);
|
const FermionField &in, FermionField &out, int dag);
|
||||||
|
|
||||||
//////////////////////////////////////////////////////////////////////////
|
//////////////////////////////////////////////////////////////////////////
|
||||||
@ -164,8 +164,6 @@ public:
|
|||||||
DoubledGaugeField UUUmuEven;
|
DoubledGaugeField UUUmuEven;
|
||||||
DoubledGaugeField UUUmuOdd;
|
DoubledGaugeField UUUmuOdd;
|
||||||
|
|
||||||
LebesgueOrder Lebesgue;
|
|
||||||
LebesgueOrder LebesgueEvenOdd;
|
|
||||||
|
|
||||||
///////////////////////////////////////////////////////////////
|
///////////////////////////////////////////////////////////////
|
||||||
// Conserved current utilities
|
// Conserved current utilities
|
||||||
|
@ -100,7 +100,6 @@ public:
|
|||||||
int dag);
|
int dag);
|
||||||
|
|
||||||
void DhopInternal(StencilImpl & st,
|
void DhopInternal(StencilImpl & st,
|
||||||
LebesgueOrder &lo,
|
|
||||||
DoubledGaugeField &U,
|
DoubledGaugeField &U,
|
||||||
DoubledGaugeField &UUU,
|
DoubledGaugeField &UUU,
|
||||||
const FermionField &in,
|
const FermionField &in,
|
||||||
@ -108,7 +107,6 @@ public:
|
|||||||
int dag);
|
int dag);
|
||||||
|
|
||||||
void DhopInternalOverlappedComms(StencilImpl & st,
|
void DhopInternalOverlappedComms(StencilImpl & st,
|
||||||
LebesgueOrder &lo,
|
|
||||||
DoubledGaugeField &U,
|
DoubledGaugeField &U,
|
||||||
DoubledGaugeField &UUU,
|
DoubledGaugeField &UUU,
|
||||||
const FermionField &in,
|
const FermionField &in,
|
||||||
@ -116,7 +114,6 @@ public:
|
|||||||
int dag);
|
int dag);
|
||||||
|
|
||||||
void DhopInternalSerialComms(StencilImpl & st,
|
void DhopInternalSerialComms(StencilImpl & st,
|
||||||
LebesgueOrder &lo,
|
|
||||||
DoubledGaugeField &U,
|
DoubledGaugeField &U,
|
||||||
DoubledGaugeField &UUU,
|
DoubledGaugeField &UUU,
|
||||||
const FermionField &in,
|
const FermionField &in,
|
||||||
@ -192,8 +189,6 @@ public:
|
|||||||
DoubledGaugeField UUUmuEven;
|
DoubledGaugeField UUUmuEven;
|
||||||
DoubledGaugeField UUUmuOdd;
|
DoubledGaugeField UUUmuOdd;
|
||||||
|
|
||||||
LebesgueOrder Lebesgue;
|
|
||||||
LebesgueOrder LebesgueEvenOdd;
|
|
||||||
|
|
||||||
// Comms buffer
|
// Comms buffer
|
||||||
// std::vector<SiteHalfSpinor,alignedAllocator<SiteHalfSpinor> > comm_buf;
|
// std::vector<SiteHalfSpinor,alignedAllocator<SiteHalfSpinor> > comm_buf;
|
||||||
|
@ -42,11 +42,11 @@ public:
|
|||||||
|
|
||||||
public:
|
public:
|
||||||
// Shift operator coefficients for red-black preconditioned Mobius EOFA
|
// Shift operator coefficients for red-black preconditioned Mobius EOFA
|
||||||
Vector<Coeff_t> Mooee_shift;
|
std::vector<Coeff_t> Mooee_shift;
|
||||||
Vector<Coeff_t> MooeeInv_shift_lc;
|
std::vector<Coeff_t> MooeeInv_shift_lc;
|
||||||
Vector<Coeff_t> MooeeInv_shift_norm;
|
std::vector<Coeff_t> MooeeInv_shift_norm;
|
||||||
Vector<Coeff_t> MooeeInvDag_shift_lc;
|
std::vector<Coeff_t> MooeeInvDag_shift_lc;
|
||||||
Vector<Coeff_t> MooeeInvDag_shift_norm;
|
std::vector<Coeff_t> MooeeInvDag_shift_norm;
|
||||||
|
|
||||||
virtual void Instantiatable(void) {};
|
virtual void Instantiatable(void) {};
|
||||||
|
|
||||||
@ -74,18 +74,18 @@ public:
|
|||||||
// Instantiate different versions depending on Impl
|
// Instantiate different versions depending on Impl
|
||||||
/////////////////////////////////////////////////////
|
/////////////////////////////////////////////////////
|
||||||
void M5D(const FermionField& psi, const FermionField& phi, FermionField& chi,
|
void M5D(const FermionField& psi, const FermionField& phi, FermionField& chi,
|
||||||
Vector<Coeff_t>& lower, Vector<Coeff_t>& diag, Vector<Coeff_t>& upper);
|
std::vector<Coeff_t>& lower, std::vector<Coeff_t>& diag, std::vector<Coeff_t>& upper);
|
||||||
|
|
||||||
void M5D_shift(const FermionField& psi, const FermionField& phi, FermionField& chi,
|
void M5D_shift(const FermionField& psi, const FermionField& phi, FermionField& chi,
|
||||||
Vector<Coeff_t>& lower, Vector<Coeff_t>& diag, Vector<Coeff_t>& upper,
|
std::vector<Coeff_t>& lower, std::vector<Coeff_t>& diag, std::vector<Coeff_t>& upper,
|
||||||
Vector<Coeff_t>& shift_coeffs);
|
std::vector<Coeff_t>& shift_coeffs);
|
||||||
|
|
||||||
void M5Ddag(const FermionField& psi, const FermionField& phi, FermionField& chi,
|
void M5Ddag(const FermionField& psi, const FermionField& phi, FermionField& chi,
|
||||||
Vector<Coeff_t>& lower, Vector<Coeff_t>& diag, Vector<Coeff_t>& upper);
|
std::vector<Coeff_t>& lower, std::vector<Coeff_t>& diag, std::vector<Coeff_t>& upper);
|
||||||
|
|
||||||
void M5Ddag_shift(const FermionField& psi, const FermionField& phi, FermionField& chi,
|
void M5Ddag_shift(const FermionField& psi, const FermionField& phi, FermionField& chi,
|
||||||
Vector<Coeff_t>& lower, Vector<Coeff_t>& diag, Vector<Coeff_t>& upper,
|
std::vector<Coeff_t>& lower, std::vector<Coeff_t>& diag, std::vector<Coeff_t>& upper,
|
||||||
Vector<Coeff_t>& shift_coeffs);
|
std::vector<Coeff_t>& shift_coeffs);
|
||||||
|
|
||||||
virtual void RefreshShiftCoefficients(RealD new_shift);
|
virtual void RefreshShiftCoefficients(RealD new_shift);
|
||||||
|
|
||||||
|
@ -102,11 +102,11 @@ public:
|
|||||||
GaugeField &mat,
|
GaugeField &mat,
|
||||||
const FermionField &A, const FermionField &B, int dag);
|
const FermionField &A, const FermionField &B, int dag);
|
||||||
|
|
||||||
void DhopInternal(StencilImpl &st, LebesgueOrder &lo, DoubledGaugeField &U,
|
void DhopInternal(StencilImpl &st, DoubledGaugeField &U,
|
||||||
const FermionField &in, FermionField &out, int dag);
|
const FermionField &in, FermionField &out, int dag);
|
||||||
void DhopInternalSerialComms(StencilImpl &st, LebesgueOrder &lo, DoubledGaugeField &U,
|
void DhopInternalSerialComms(StencilImpl &st, DoubledGaugeField &U,
|
||||||
const FermionField &in, FermionField &out, int dag);
|
const FermionField &in, FermionField &out, int dag);
|
||||||
void DhopInternalOverlappedComms(StencilImpl &st, LebesgueOrder &lo, DoubledGaugeField &U,
|
void DhopInternalOverlappedComms(StencilImpl &st, DoubledGaugeField &U,
|
||||||
const FermionField &in, FermionField &out, int dag);
|
const FermionField &in, FermionField &out, int dag);
|
||||||
|
|
||||||
//////////////////////////////////////////////////////////////////////////
|
//////////////////////////////////////////////////////////////////////////
|
||||||
@ -152,9 +152,6 @@ public:
|
|||||||
DoubledGaugeField UmuEven;
|
DoubledGaugeField UmuEven;
|
||||||
DoubledGaugeField UmuOdd;
|
DoubledGaugeField UmuOdd;
|
||||||
|
|
||||||
LebesgueOrder Lebesgue;
|
|
||||||
LebesgueOrder LebesgueEvenOdd;
|
|
||||||
|
|
||||||
///////////////////////////////////////////////////////////////
|
///////////////////////////////////////////////////////////////
|
||||||
// Conserved current utilities
|
// Conserved current utilities
|
||||||
///////////////////////////////////////////////////////////////
|
///////////////////////////////////////////////////////////////
|
||||||
|
@ -94,8 +94,8 @@ protected:
|
|||||||
RealD R;
|
RealD R;
|
||||||
RealD amax;
|
RealD amax;
|
||||||
RealD scale;
|
RealD scale;
|
||||||
Vector<double> p;
|
std::vector<double> p;
|
||||||
Vector<double> q;
|
std::vector<double> q;
|
||||||
|
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -35,7 +35,7 @@ template<class Matrix, class Field>
|
|||||||
class KappaSimilarityTransform {
|
class KappaSimilarityTransform {
|
||||||
public:
|
public:
|
||||||
INHERIT_IMPL_TYPES(Matrix);
|
INHERIT_IMPL_TYPES(Matrix);
|
||||||
Vector<Coeff_t> kappa, kappaDag, kappaInv, kappaInvDag;
|
std::vector<Coeff_t> kappa, kappaDag, kappaInv, kappaInvDag;
|
||||||
|
|
||||||
KappaSimilarityTransform (Matrix &zmob) {
|
KappaSimilarityTransform (Matrix &zmob) {
|
||||||
for (int i=0;i<(int)zmob.bs.size();i++) {
|
for (int i=0;i<(int)zmob.bs.size();i++) {
|
||||||
|
@ -49,10 +49,10 @@ template<class Impl> class StaggeredKernels : public FermionOperator<Impl> , pub
|
|||||||
|
|
||||||
public:
|
public:
|
||||||
|
|
||||||
void DhopImproved(StencilImpl &st, LebesgueOrder &lo,
|
void DhopImproved(StencilImpl &st,
|
||||||
DoubledGaugeField &U, DoubledGaugeField &UUU,
|
DoubledGaugeField &U, DoubledGaugeField &UUU,
|
||||||
const FermionField &in, FermionField &out, int dag, int interior,int exterior);
|
const FermionField &in, FermionField &out, int dag, int interior,int exterior);
|
||||||
void DhopNaive(StencilImpl &st, LebesgueOrder &lo,
|
void DhopNaive(StencilImpl &st,
|
||||||
DoubledGaugeField &U,
|
DoubledGaugeField &U,
|
||||||
const FermionField &in, FermionField &out, int dag, int interior,int exterior);
|
const FermionField &in, FermionField &out, int dag, int interior,int exterior);
|
||||||
|
|
||||||
|
@ -47,7 +47,7 @@ public:
|
|||||||
static int PartialCompressionFactor(GridBase *grid) { return 1;}
|
static int PartialCompressionFactor(GridBase *grid) { return 1;}
|
||||||
#endif
|
#endif
|
||||||
template<class vobj,class cobj,class compressor>
|
template<class vobj,class cobj,class compressor>
|
||||||
static void Gather_plane_simple (commVector<std::pair<int,int> >& table,
|
static void Gather_plane_simple (deviceVector<std::pair<int,int> >& table,
|
||||||
const Lattice<vobj> &rhs,
|
const Lattice<vobj> &rhs,
|
||||||
cobj *buffer,
|
cobj *buffer,
|
||||||
compressor &compress,
|
compressor &compress,
|
||||||
@ -109,7 +109,7 @@ public:
|
|||||||
// Reorder the fifth dim to be s=Ls-1 , s=0, s=1,...,Ls-2.
|
// Reorder the fifth dim to be s=Ls-1 , s=0, s=1,...,Ls-2.
|
||||||
////////////////////////////////////////////////////////////////////////////////////////////
|
////////////////////////////////////////////////////////////////////////////////////////////
|
||||||
template<class vobj,class cobj,class compressor>
|
template<class vobj,class cobj,class compressor>
|
||||||
static void Gather_plane_exchange(commVector<std::pair<int,int> >& table,const Lattice<vobj> &rhs,
|
static void Gather_plane_exchange(deviceVector<std::pair<int,int> >& table,const Lattice<vobj> &rhs,
|
||||||
std::vector<cobj *> pointers,int dimension,int plane,int cbmask,
|
std::vector<cobj *> pointers,int dimension,int plane,int cbmask,
|
||||||
compressor &compress,int type,int partial)
|
compressor &compress,int type,int partial)
|
||||||
{
|
{
|
||||||
@ -197,7 +197,7 @@ public:
|
|||||||
#endif
|
#endif
|
||||||
|
|
||||||
template<class vobj,class cobj,class compressor>
|
template<class vobj,class cobj,class compressor>
|
||||||
static void Gather_plane_simple (commVector<std::pair<int,int> >& table,
|
static void Gather_plane_simple (deviceVector<std::pair<int,int> >& table,
|
||||||
const Lattice<vobj> &rhs,
|
const Lattice<vobj> &rhs,
|
||||||
cobj *buffer,
|
cobj *buffer,
|
||||||
compressor &compress,
|
compressor &compress,
|
||||||
@ -208,7 +208,7 @@ public:
|
|||||||
else FaceGatherSimple::Gather_plane_simple(table,rhs,buffer,compress,off,so,partial);
|
else FaceGatherSimple::Gather_plane_simple(table,rhs,buffer,compress,off,so,partial);
|
||||||
}
|
}
|
||||||
template<class vobj,class cobj,class compressor>
|
template<class vobj,class cobj,class compressor>
|
||||||
static void Gather_plane_exchange(commVector<std::pair<int,int> >& table,const Lattice<vobj> &rhs,
|
static void Gather_plane_exchange(deviceVector<std::pair<int,int> >& table,const Lattice<vobj> &rhs,
|
||||||
std::vector<cobj *> pointers,int dimension,int plane,int cbmask,
|
std::vector<cobj *> pointers,int dimension,int plane,int cbmask,
|
||||||
compressor &compress,int type,int partial)
|
compressor &compress,int type,int partial)
|
||||||
{
|
{
|
||||||
@ -402,7 +402,6 @@ public:
|
|||||||
|
|
||||||
typedef CartesianStencil<vobj,cobj,Parameters> Base;
|
typedef CartesianStencil<vobj,cobj,Parameters> Base;
|
||||||
typedef typename Base::View_type View_type;
|
typedef typename Base::View_type View_type;
|
||||||
typedef typename Base::StencilVector StencilVector;
|
|
||||||
|
|
||||||
// Vector<int> surface_list;
|
// Vector<int> surface_list;
|
||||||
WilsonStencil(GridBase *grid,
|
WilsonStencil(GridBase *grid,
|
||||||
|
@ -126,14 +126,17 @@ public:
|
|||||||
void DerivInternal(StencilImpl &st, DoubledGaugeField &U, GaugeField &mat,
|
void DerivInternal(StencilImpl &st, DoubledGaugeField &U, GaugeField &mat,
|
||||||
const FermionField &A, const FermionField &B, int dag);
|
const FermionField &A, const FermionField &B, int dag);
|
||||||
|
|
||||||
void DhopInternal(StencilImpl &st, LebesgueOrder &lo, DoubledGaugeField &U,
|
void DhopInternal(StencilImpl &st,
|
||||||
|
DoubledGaugeField &U,
|
||||||
const FermionField &in, FermionField &out, int dag);
|
const FermionField &in, FermionField &out, int dag);
|
||||||
|
|
||||||
void DhopInternalSerial(StencilImpl &st, LebesgueOrder &lo, DoubledGaugeField &U,
|
void DhopInternalSerial(StencilImpl &st,
|
||||||
const FermionField &in, FermionField &out, int dag);
|
DoubledGaugeField &U,
|
||||||
|
const FermionField &in, FermionField &out, int dag);
|
||||||
|
|
||||||
void DhopInternalOverlappedComms(StencilImpl &st, LebesgueOrder &lo, DoubledGaugeField &U,
|
void DhopInternalOverlappedComms(StencilImpl &st,
|
||||||
const FermionField &in, FermionField &out, int dag);
|
DoubledGaugeField &U,
|
||||||
|
const FermionField &in, FermionField &out, int dag);
|
||||||
|
|
||||||
// Constructor
|
// Constructor
|
||||||
WilsonFermion(GaugeField &_Umu, GridCartesian &Fgrid,
|
WilsonFermion(GaugeField &_Umu, GridCartesian &Fgrid,
|
||||||
@ -168,9 +171,6 @@ public:
|
|||||||
DoubledGaugeField UmuEven;
|
DoubledGaugeField UmuEven;
|
||||||
DoubledGaugeField UmuOdd;
|
DoubledGaugeField UmuOdd;
|
||||||
|
|
||||||
LebesgueOrder Lebesgue;
|
|
||||||
LebesgueOrder LebesgueEvenOdd;
|
|
||||||
|
|
||||||
WilsonAnisotropyCoefficients anisotropyCoeff;
|
WilsonAnisotropyCoefficients anisotropyCoeff;
|
||||||
|
|
||||||
///////////////////////////////////////////////////////////////
|
///////////////////////////////////////////////////////////////
|
||||||
|
@ -135,21 +135,18 @@ public:
|
|||||||
int dag);
|
int dag);
|
||||||
|
|
||||||
void DhopInternal(StencilImpl & st,
|
void DhopInternal(StencilImpl & st,
|
||||||
LebesgueOrder &lo,
|
|
||||||
DoubledGaugeField &U,
|
DoubledGaugeField &U,
|
||||||
const FermionField &in,
|
const FermionField &in,
|
||||||
FermionField &out,
|
FermionField &out,
|
||||||
int dag);
|
int dag);
|
||||||
|
|
||||||
void DhopInternalOverlappedComms(StencilImpl & st,
|
void DhopInternalOverlappedComms(StencilImpl & st,
|
||||||
LebesgueOrder &lo,
|
|
||||||
DoubledGaugeField &U,
|
DoubledGaugeField &U,
|
||||||
const FermionField &in,
|
const FermionField &in,
|
||||||
FermionField &out,
|
FermionField &out,
|
||||||
int dag);
|
int dag);
|
||||||
|
|
||||||
void DhopInternalSerialComms(StencilImpl & st,
|
void DhopInternalSerialComms(StencilImpl & st,
|
||||||
LebesgueOrder &lo,
|
|
||||||
DoubledGaugeField &U,
|
DoubledGaugeField &U,
|
||||||
const FermionField &in,
|
const FermionField &in,
|
||||||
FermionField &out,
|
FermionField &out,
|
||||||
@ -203,9 +200,6 @@ public:
|
|||||||
DoubledGaugeField UmuEven;
|
DoubledGaugeField UmuEven;
|
||||||
DoubledGaugeField UmuOdd;
|
DoubledGaugeField UmuOdd;
|
||||||
|
|
||||||
LebesgueOrder Lebesgue;
|
|
||||||
LebesgueOrder LebesgueEvenOdd;
|
|
||||||
|
|
||||||
// Comms buffer
|
// Comms buffer
|
||||||
// std::vector<SiteHalfSpinor,alignedAllocator<SiteHalfSpinor> > comm_buf;
|
// std::vector<SiteHalfSpinor,alignedAllocator<SiteHalfSpinor> > comm_buf;
|
||||||
|
|
||||||
|
@ -58,7 +58,7 @@ public:
|
|||||||
{
|
{
|
||||||
// RealD eps = 1.0;
|
// RealD eps = 1.0;
|
||||||
std::cout<<GridLogMessage << "ZMobiusFermion (b="<<b<<",c="<<c<<") with Ls= "<<this->Ls<<" gamma passed in"<<std::endl;
|
std::cout<<GridLogMessage << "ZMobiusFermion (b="<<b<<",c="<<c<<") with Ls= "<<this->Ls<<" gamma passed in"<<std::endl;
|
||||||
Vector<Coeff_t> zgamma(this->Ls);
|
std::vector<Coeff_t> zgamma(this->Ls);
|
||||||
for(int s=0;s<this->Ls;s++){
|
for(int s=0;s<this->Ls;s++){
|
||||||
zgamma[s] = gamma[s];
|
zgamma[s] = gamma[s];
|
||||||
}
|
}
|
||||||
|
@ -1,3 +1,5 @@
|
|||||||
|
#if 0
|
||||||
|
|
||||||
/*************************************************************************************
|
/*************************************************************************************
|
||||||
|
|
||||||
Grid physics library, www.github.com/paboyle/Grid
|
Grid physics library, www.github.com/paboyle/Grid
|
||||||
@ -818,3 +820,5 @@ CayleyFermion5D<Impl>::MooeeInternal(const FermionField &psi, FermionField &chi,
|
|||||||
}
|
}
|
||||||
|
|
||||||
NAMESPACE_END(Grid);
|
NAMESPACE_END(Grid);
|
||||||
|
|
||||||
|
#endif
|
@ -1,3 +1,4 @@
|
|||||||
|
#if 0
|
||||||
/*************************************************************************************
|
/*************************************************************************************
|
||||||
|
|
||||||
Grid physics library, www.github.com/paboyle/Grid
|
Grid physics library, www.github.com/paboyle/Grid
|
||||||
@ -241,3 +242,4 @@ void LebesgueOrder::ZGraph(void)
|
|||||||
}
|
}
|
||||||
NAMESPACE_END(Grid);
|
NAMESPACE_END(Grid);
|
||||||
|
|
||||||
|
#endif
|
@ -72,7 +72,7 @@ public:
|
|||||||
void ThreadInterleave(void);
|
void ThreadInterleave(void);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
Vector<IndexInteger> _LebesgueReorder;
|
deviceVector<IndexInteger> _LebesgueReorder;
|
||||||
|
|
||||||
};
|
};
|
||||||
|
|
@ -156,18 +156,18 @@ template<class Impl>
|
|||||||
void CayleyFermion5D<Impl>::M5D (const FermionField &psi, FermionField &chi)
|
void CayleyFermion5D<Impl>::M5D (const FermionField &psi, FermionField &chi)
|
||||||
{
|
{
|
||||||
int Ls=this->Ls;
|
int Ls=this->Ls;
|
||||||
Vector<Coeff_t> diag (Ls,1.0);
|
std::vector<Coeff_t> diag (Ls,1.0);
|
||||||
Vector<Coeff_t> upper(Ls,-1.0); upper[Ls-1]=mass_minus;
|
std::vector<Coeff_t> upper(Ls,-1.0); upper[Ls-1]=mass_minus;
|
||||||
Vector<Coeff_t> lower(Ls,-1.0); lower[0] =mass_plus;
|
std::vector<Coeff_t> lower(Ls,-1.0); lower[0] =mass_plus;
|
||||||
M5D(psi,chi,chi,lower,diag,upper);
|
M5D(psi,chi,chi,lower,diag,upper);
|
||||||
}
|
}
|
||||||
template<class Impl>
|
template<class Impl>
|
||||||
void CayleyFermion5D<Impl>::Meooe5D (const FermionField &psi, FermionField &Din)
|
void CayleyFermion5D<Impl>::Meooe5D (const FermionField &psi, FermionField &Din)
|
||||||
{
|
{
|
||||||
int Ls=this->Ls;
|
int Ls=this->Ls;
|
||||||
Vector<Coeff_t> diag = bs;
|
std::vector<Coeff_t> diag = bs;
|
||||||
Vector<Coeff_t> upper= cs;
|
std::vector<Coeff_t> upper= cs;
|
||||||
Vector<Coeff_t> lower= cs;
|
std::vector<Coeff_t> lower= cs;
|
||||||
upper[Ls-1]=-mass_minus*upper[Ls-1];
|
upper[Ls-1]=-mass_minus*upper[Ls-1];
|
||||||
lower[0] =-mass_plus*lower[0];
|
lower[0] =-mass_plus*lower[0];
|
||||||
M5D(psi,psi,Din,lower,diag,upper);
|
M5D(psi,psi,Din,lower,diag,upper);
|
||||||
@ -176,9 +176,9 @@ void CayleyFermion5D<Impl>::Meooe5D (const FermionField &psi, FermionField &D
|
|||||||
template<class Impl> void CayleyFermion5D<Impl>::Meo5D (const FermionField &psi, FermionField &chi)
|
template<class Impl> void CayleyFermion5D<Impl>::Meo5D (const FermionField &psi, FermionField &chi)
|
||||||
{
|
{
|
||||||
int Ls=this->Ls;
|
int Ls=this->Ls;
|
||||||
Vector<Coeff_t> diag = beo;
|
std::vector<Coeff_t> diag = beo;
|
||||||
Vector<Coeff_t> upper(Ls);
|
std::vector<Coeff_t> upper(Ls);
|
||||||
Vector<Coeff_t> lower(Ls);
|
std::vector<Coeff_t> lower(Ls);
|
||||||
for(int i=0;i<Ls;i++) {
|
for(int i=0;i<Ls;i++) {
|
||||||
upper[i]=-ceo[i];
|
upper[i]=-ceo[i];
|
||||||
lower[i]=-ceo[i];
|
lower[i]=-ceo[i];
|
||||||
@ -191,9 +191,9 @@ template<class Impl>
|
|||||||
void CayleyFermion5D<Impl>::Mooee (const FermionField &psi, FermionField &chi)
|
void CayleyFermion5D<Impl>::Mooee (const FermionField &psi, FermionField &chi)
|
||||||
{
|
{
|
||||||
int Ls=this->Ls;
|
int Ls=this->Ls;
|
||||||
Vector<Coeff_t> diag = bee;
|
std::vector<Coeff_t> diag = bee;
|
||||||
Vector<Coeff_t> upper(Ls);
|
std::vector<Coeff_t> upper(Ls);
|
||||||
Vector<Coeff_t> lower(Ls);
|
std::vector<Coeff_t> lower(Ls);
|
||||||
for(int i=0;i<Ls;i++) {
|
for(int i=0;i<Ls;i++) {
|
||||||
upper[i]=-cee[i];
|
upper[i]=-cee[i];
|
||||||
lower[i]=-cee[i];
|
lower[i]=-cee[i];
|
||||||
@ -206,9 +206,9 @@ template<class Impl>
|
|||||||
void CayleyFermion5D<Impl>::MooeeDag (const FermionField &psi, FermionField &chi)
|
void CayleyFermion5D<Impl>::MooeeDag (const FermionField &psi, FermionField &chi)
|
||||||
{
|
{
|
||||||
int Ls=this->Ls;
|
int Ls=this->Ls;
|
||||||
Vector<Coeff_t> diag = bee;
|
std::vector<Coeff_t> diag = bee;
|
||||||
Vector<Coeff_t> upper(Ls);
|
std::vector<Coeff_t> upper(Ls);
|
||||||
Vector<Coeff_t> lower(Ls);
|
std::vector<Coeff_t> lower(Ls);
|
||||||
|
|
||||||
for (int s=0;s<Ls;s++){
|
for (int s=0;s<Ls;s++){
|
||||||
// Assemble the 5d matrix
|
// Assemble the 5d matrix
|
||||||
@ -236,9 +236,9 @@ template<class Impl>
|
|||||||
void CayleyFermion5D<Impl>::M5Ddag (const FermionField &psi, FermionField &chi)
|
void CayleyFermion5D<Impl>::M5Ddag (const FermionField &psi, FermionField &chi)
|
||||||
{
|
{
|
||||||
int Ls=this->Ls;
|
int Ls=this->Ls;
|
||||||
Vector<Coeff_t> diag(Ls,1.0);
|
std::vector<Coeff_t> diag(Ls,1.0);
|
||||||
Vector<Coeff_t> upper(Ls,-1.0);
|
std::vector<Coeff_t> upper(Ls,-1.0);
|
||||||
Vector<Coeff_t> lower(Ls,-1.0);
|
std::vector<Coeff_t> lower(Ls,-1.0);
|
||||||
upper[Ls-1]=-mass_plus*upper[Ls-1];
|
upper[Ls-1]=-mass_plus*upper[Ls-1];
|
||||||
lower[0] =-mass_minus*lower[0];
|
lower[0] =-mass_minus*lower[0];
|
||||||
M5Ddag(psi,chi,chi,lower,diag,upper);
|
M5Ddag(psi,chi,chi,lower,diag,upper);
|
||||||
@ -248,9 +248,9 @@ template<class Impl>
|
|||||||
void CayleyFermion5D<Impl>::MeooeDag5D (const FermionField &psi, FermionField &Din)
|
void CayleyFermion5D<Impl>::MeooeDag5D (const FermionField &psi, FermionField &Din)
|
||||||
{
|
{
|
||||||
int Ls=this->Ls;
|
int Ls=this->Ls;
|
||||||
Vector<Coeff_t> diag =bs;
|
std::vector<Coeff_t> diag =bs;
|
||||||
Vector<Coeff_t> upper=cs;
|
std::vector<Coeff_t> upper=cs;
|
||||||
Vector<Coeff_t> lower=cs;
|
std::vector<Coeff_t> lower=cs;
|
||||||
|
|
||||||
for (int s=0;s<Ls;s++){
|
for (int s=0;s<Ls;s++){
|
||||||
if ( s== 0 ) {
|
if ( s== 0 ) {
|
||||||
@ -394,7 +394,7 @@ void CayleyFermion5D<Impl>::MeoDeriv(GaugeField &mat,const FermionField &U,const
|
|||||||
template<class Impl>
|
template<class Impl>
|
||||||
void CayleyFermion5D<Impl>::SetCoefficientsTanh(Approx::zolotarev_data *zdata,RealD b,RealD c)
|
void CayleyFermion5D<Impl>::SetCoefficientsTanh(Approx::zolotarev_data *zdata,RealD b,RealD c)
|
||||||
{
|
{
|
||||||
Vector<Coeff_t> gamma(this->Ls);
|
std::vector<Coeff_t> gamma(this->Ls);
|
||||||
for(int s=0;s<this->Ls;s++) gamma[s] = zdata->gamma[s];
|
for(int s=0;s<this->Ls;s++) gamma[s] = zdata->gamma[s];
|
||||||
SetCoefficientsInternal(1.0,gamma,b,c);
|
SetCoefficientsInternal(1.0,gamma,b,c);
|
||||||
}
|
}
|
||||||
@ -402,13 +402,13 @@ void CayleyFermion5D<Impl>::SetCoefficientsTanh(Approx::zolotarev_data *zdata,Re
|
|||||||
template<class Impl>
|
template<class Impl>
|
||||||
void CayleyFermion5D<Impl>::SetCoefficientsZolotarev(RealD zolo_hi,Approx::zolotarev_data *zdata,RealD b,RealD c)
|
void CayleyFermion5D<Impl>::SetCoefficientsZolotarev(RealD zolo_hi,Approx::zolotarev_data *zdata,RealD b,RealD c)
|
||||||
{
|
{
|
||||||
Vector<Coeff_t> gamma(this->Ls);
|
std::vector<Coeff_t> gamma(this->Ls);
|
||||||
for(int s=0;s<this->Ls;s++) gamma[s] = zdata->gamma[s];
|
for(int s=0;s<this->Ls;s++) gamma[s] = zdata->gamma[s];
|
||||||
SetCoefficientsInternal(zolo_hi,gamma,b,c);
|
SetCoefficientsInternal(zolo_hi,gamma,b,c);
|
||||||
}
|
}
|
||||||
//Zolo
|
//Zolo
|
||||||
template<class Impl>
|
template<class Impl>
|
||||||
void CayleyFermion5D<Impl>::SetCoefficientsInternal(RealD zolo_hi,Vector<Coeff_t> & gamma,RealD b,RealD c)
|
void CayleyFermion5D<Impl>::SetCoefficientsInternal(RealD zolo_hi,std::vector<Coeff_t> & gamma,RealD b,RealD c)
|
||||||
{
|
{
|
||||||
int Ls=this->Ls;
|
int Ls=this->Ls;
|
||||||
|
|
||||||
|
@ -43,9 +43,9 @@ void
|
|||||||
CayleyFermion5D<Impl>::M5D(const FermionField &psi_i,
|
CayleyFermion5D<Impl>::M5D(const FermionField &psi_i,
|
||||||
const FermionField &phi_i,
|
const FermionField &phi_i,
|
||||||
FermionField &chi_i,
|
FermionField &chi_i,
|
||||||
Vector<Coeff_t> &lower,
|
std::vector<Coeff_t> &lower,
|
||||||
Vector<Coeff_t> &diag,
|
std::vector<Coeff_t> &diag,
|
||||||
Vector<Coeff_t> &upper)
|
std::vector<Coeff_t> &upper)
|
||||||
{
|
{
|
||||||
|
|
||||||
chi_i.Checkerboard()=psi_i.Checkerboard();
|
chi_i.Checkerboard()=psi_i.Checkerboard();
|
||||||
@ -55,12 +55,16 @@ CayleyFermion5D<Impl>::M5D(const FermionField &psi_i,
|
|||||||
autoView(chi , chi_i,AcceleratorWrite);
|
autoView(chi , chi_i,AcceleratorWrite);
|
||||||
assert(phi.Checkerboard() == psi.Checkerboard());
|
assert(phi.Checkerboard() == psi.Checkerboard());
|
||||||
|
|
||||||
auto pdiag = &diag[0];
|
|
||||||
auto pupper = &upper[0];
|
|
||||||
auto plower = &lower[0];
|
|
||||||
|
|
||||||
int Ls =this->Ls;
|
int Ls =this->Ls;
|
||||||
|
|
||||||
|
static deviceVector<Coeff_t> d_diag(Ls) ; acceleratorCopyToDevice(&diag[0] ,&d_diag[0],Ls*sizeof(Coeff_t));
|
||||||
|
static deviceVector<Coeff_t> d_upper(Ls); acceleratorCopyToDevice(&upper[0],&d_upper[0],Ls*sizeof(Coeff_t));
|
||||||
|
static deviceVector<Coeff_t> d_lower(Ls); acceleratorCopyToDevice(&lower[0],&d_lower[0],Ls*sizeof(Coeff_t));
|
||||||
|
|
||||||
|
auto pdiag = &d_diag[0];
|
||||||
|
auto pupper = &d_upper[0];
|
||||||
|
auto plower = &d_lower[0];
|
||||||
|
|
||||||
// 10 = 3 complex mult + 2 complex add
|
// 10 = 3 complex mult + 2 complex add
|
||||||
// Flops = 10.0*(Nc*Ns) *Ls*vol (/2 for red black counting)
|
// Flops = 10.0*(Nc*Ns) *Ls*vol (/2 for red black counting)
|
||||||
uint64_t nloop = grid->oSites();
|
uint64_t nloop = grid->oSites();
|
||||||
@ -82,9 +86,9 @@ void
|
|||||||
CayleyFermion5D<Impl>::M5Ddag(const FermionField &psi_i,
|
CayleyFermion5D<Impl>::M5Ddag(const FermionField &psi_i,
|
||||||
const FermionField &phi_i,
|
const FermionField &phi_i,
|
||||||
FermionField &chi_i,
|
FermionField &chi_i,
|
||||||
Vector<Coeff_t> &lower,
|
std::vector<Coeff_t> &lower,
|
||||||
Vector<Coeff_t> &diag,
|
std::vector<Coeff_t> &diag,
|
||||||
Vector<Coeff_t> &upper)
|
std::vector<Coeff_t> &upper)
|
||||||
{
|
{
|
||||||
chi_i.Checkerboard()=psi_i.Checkerboard();
|
chi_i.Checkerboard()=psi_i.Checkerboard();
|
||||||
GridBase *grid=psi_i.Grid();
|
GridBase *grid=psi_i.Grid();
|
||||||
@ -93,12 +97,16 @@ CayleyFermion5D<Impl>::M5Ddag(const FermionField &psi_i,
|
|||||||
autoView(chi , chi_i,AcceleratorWrite);
|
autoView(chi , chi_i,AcceleratorWrite);
|
||||||
assert(phi.Checkerboard() == psi.Checkerboard());
|
assert(phi.Checkerboard() == psi.Checkerboard());
|
||||||
|
|
||||||
auto pdiag = &diag[0];
|
|
||||||
auto pupper = &upper[0];
|
|
||||||
auto plower = &lower[0];
|
|
||||||
|
|
||||||
int Ls=this->Ls;
|
int Ls=this->Ls;
|
||||||
|
|
||||||
|
static deviceVector<Coeff_t> d_diag(Ls) ; acceleratorCopyToDevice(&diag[0] ,&d_diag[0],Ls*sizeof(Coeff_t));
|
||||||
|
static deviceVector<Coeff_t> d_upper(Ls); acceleratorCopyToDevice(&upper[0],&d_upper[0],Ls*sizeof(Coeff_t));
|
||||||
|
static deviceVector<Coeff_t> d_lower(Ls); acceleratorCopyToDevice(&lower[0],&d_lower[0],Ls*sizeof(Coeff_t));
|
||||||
|
|
||||||
|
auto pdiag = &d_diag[0];
|
||||||
|
auto pupper = &d_upper[0];
|
||||||
|
auto plower = &d_lower[0];
|
||||||
|
|
||||||
// Flops = 6.0*(Nc*Ns) *Ls*vol
|
// Flops = 6.0*(Nc*Ns) *Ls*vol
|
||||||
uint64_t nloop = grid->oSites();
|
uint64_t nloop = grid->oSites();
|
||||||
accelerator_for(sss,nloop,Simd::Nsimd(),{
|
accelerator_for(sss,nloop,Simd::Nsimd(),{
|
||||||
@ -126,11 +134,17 @@ CayleyFermion5D<Impl>::MooeeInv (const FermionField &psi_i, FermionField &chi
|
|||||||
|
|
||||||
int Ls=this->Ls;
|
int Ls=this->Ls;
|
||||||
|
|
||||||
auto plee = & lee [0];
|
static deviceVector<Coeff_t> d_lee(Ls); acceleratorCopyToDevice(&lee[0],&d_lee[0],Ls*sizeof(Coeff_t));
|
||||||
auto pdee = & dee [0];
|
static deviceVector<Coeff_t> d_dee(Ls); acceleratorCopyToDevice(&dee[0],&d_dee[0],Ls*sizeof(Coeff_t));
|
||||||
auto puee = & uee [0];
|
static deviceVector<Coeff_t> d_uee(Ls); acceleratorCopyToDevice(&uee[0],&d_uee[0],Ls*sizeof(Coeff_t));
|
||||||
auto pleem = & leem[0];
|
static deviceVector<Coeff_t> d_leem(Ls); acceleratorCopyToDevice(&leem[0],&d_leem[0],Ls*sizeof(Coeff_t));
|
||||||
auto pueem = & ueem[0];
|
static deviceVector<Coeff_t> d_ueem(Ls); acceleratorCopyToDevice(&ueem[0],&d_ueem[0],Ls*sizeof(Coeff_t));
|
||||||
|
|
||||||
|
auto plee = & d_lee [0];
|
||||||
|
auto pdee = & d_dee [0];
|
||||||
|
auto puee = & d_uee [0];
|
||||||
|
auto pleem = & d_leem[0];
|
||||||
|
auto pueem = & d_ueem[0];
|
||||||
|
|
||||||
uint64_t nloop = grid->oSites()/Ls;
|
uint64_t nloop = grid->oSites()/Ls;
|
||||||
accelerator_for(sss,nloop,Simd::Nsimd(),{
|
accelerator_for(sss,nloop,Simd::Nsimd(),{
|
||||||
@ -182,11 +196,17 @@ CayleyFermion5D<Impl>::MooeeInvDag (const FermionField &psi_i, FermionField &chi
|
|||||||
autoView(psi , psi_i,AcceleratorRead);
|
autoView(psi , psi_i,AcceleratorRead);
|
||||||
autoView(chi , chi_i,AcceleratorWrite);
|
autoView(chi , chi_i,AcceleratorWrite);
|
||||||
|
|
||||||
auto plee = & lee [0];
|
static deviceVector<Coeff_t> d_lee(Ls); acceleratorCopyToDevice(&lee[0],&d_lee[0],Ls*sizeof(Coeff_t));
|
||||||
auto pdee = & dee [0];
|
static deviceVector<Coeff_t> d_dee(Ls); acceleratorCopyToDevice(&dee[0],&d_dee[0],Ls*sizeof(Coeff_t));
|
||||||
auto puee = & uee [0];
|
static deviceVector<Coeff_t> d_uee(Ls); acceleratorCopyToDevice(&uee[0],&d_uee[0],Ls*sizeof(Coeff_t));
|
||||||
auto pleem = & leem[0];
|
static deviceVector<Coeff_t> d_leem(Ls); acceleratorCopyToDevice(&leem[0],&d_leem[0],Ls*sizeof(Coeff_t));
|
||||||
auto pueem = & ueem[0];
|
static deviceVector<Coeff_t> d_ueem(Ls); acceleratorCopyToDevice(&ueem[0],&d_ueem[0],Ls*sizeof(Coeff_t));
|
||||||
|
|
||||||
|
auto plee = & d_lee [0];
|
||||||
|
auto pdee = & d_dee [0];
|
||||||
|
auto puee = & d_uee [0];
|
||||||
|
auto pleem = & d_leem[0];
|
||||||
|
auto pueem = & d_ueem[0];
|
||||||
|
|
||||||
assert(psi.Checkerboard() == psi.Checkerboard());
|
assert(psi.Checkerboard() == psi.Checkerboard());
|
||||||
|
|
||||||
|
@ -41,7 +41,7 @@ NAMESPACE_BEGIN(Grid);
|
|||||||
// Pplus backwards..
|
// Pplus backwards..
|
||||||
template<class Impl>
|
template<class Impl>
|
||||||
void DomainWallEOFAFermion<Impl>::M5D(const FermionField& psi_i, const FermionField& phi_i,FermionField& chi_i,
|
void DomainWallEOFAFermion<Impl>::M5D(const FermionField& psi_i, const FermionField& phi_i,FermionField& chi_i,
|
||||||
Vector<Coeff_t>& lower, Vector<Coeff_t>& diag, Vector<Coeff_t>& upper)
|
std::vector<Coeff_t>& lower, std::vector<Coeff_t>& diag, std::vector<Coeff_t>& upper)
|
||||||
{
|
{
|
||||||
chi_i.Checkerboard() = psi_i.Checkerboard();
|
chi_i.Checkerboard() = psi_i.Checkerboard();
|
||||||
int Ls = this->Ls;
|
int Ls = this->Ls;
|
||||||
@ -50,9 +50,15 @@ void DomainWallEOFAFermion<Impl>::M5D(const FermionField& psi_i, const FermionFi
|
|||||||
autoView( psi , psi_i, AcceleratorRead);
|
autoView( psi , psi_i, AcceleratorRead);
|
||||||
autoView( chi , chi_i, AcceleratorWrite);
|
autoView( chi , chi_i, AcceleratorWrite);
|
||||||
assert(phi.Checkerboard() == psi.Checkerboard());
|
assert(phi.Checkerboard() == psi.Checkerboard());
|
||||||
auto pdiag = &diag[0];
|
|
||||||
auto pupper = &upper[0];
|
static deviceVector<Coeff_t> d_diag(Ls); acceleratorCopyToDevice(&diag[0],&d_diag[0],Ls*sizeof(Coeff_t));
|
||||||
auto plower = &lower[0];
|
static deviceVector<Coeff_t> d_upper(Ls);acceleratorCopyToDevice(&upper[0],&d_upper[0],Ls*sizeof(Coeff_t));
|
||||||
|
static deviceVector<Coeff_t> d_lower(Ls);acceleratorCopyToDevice(&lower[0],&d_lower[0],Ls*sizeof(Coeff_t));
|
||||||
|
|
||||||
|
auto pdiag = &d_diag[0];
|
||||||
|
auto pupper = &d_upper[0];
|
||||||
|
auto plower = &d_lower[0];
|
||||||
|
|
||||||
// Flops = 6.0*(Nc*Ns) *Ls*vol
|
// Flops = 6.0*(Nc*Ns) *Ls*vol
|
||||||
|
|
||||||
auto nloop=grid->oSites()/Ls;
|
auto nloop=grid->oSites()/Ls;
|
||||||
@ -73,7 +79,7 @@ void DomainWallEOFAFermion<Impl>::M5D(const FermionField& psi_i, const FermionFi
|
|||||||
|
|
||||||
template<class Impl>
|
template<class Impl>
|
||||||
void DomainWallEOFAFermion<Impl>::M5Ddag(const FermionField& psi_i, const FermionField& phi_i, FermionField& chi_i,
|
void DomainWallEOFAFermion<Impl>::M5Ddag(const FermionField& psi_i, const FermionField& phi_i, FermionField& chi_i,
|
||||||
Vector<Coeff_t>& lower, Vector<Coeff_t>& diag, Vector<Coeff_t>& upper)
|
std::vector<Coeff_t>& lower, std::vector<Coeff_t>& diag, std::vector<Coeff_t>& upper)
|
||||||
{
|
{
|
||||||
chi_i.Checkerboard() = psi_i.Checkerboard();
|
chi_i.Checkerboard() = psi_i.Checkerboard();
|
||||||
GridBase* grid = psi_i.Grid();
|
GridBase* grid = psi_i.Grid();
|
||||||
@ -83,9 +89,14 @@ void DomainWallEOFAFermion<Impl>::M5Ddag(const FermionField& psi_i, const Fermio
|
|||||||
autoView( phi , phi_i, AcceleratorRead);
|
autoView( phi , phi_i, AcceleratorRead);
|
||||||
autoView( chi , chi_i, AcceleratorWrite);
|
autoView( chi , chi_i, AcceleratorWrite);
|
||||||
assert(phi.Checkerboard() == psi.Checkerboard());
|
assert(phi.Checkerboard() == psi.Checkerboard());
|
||||||
auto pdiag = &diag[0];
|
|
||||||
auto pupper = &upper[0];
|
static deviceVector<Coeff_t> d_diag(Ls); acceleratorCopyToDevice(&diag[0],&d_diag[0],Ls*sizeof(Coeff_t));
|
||||||
auto plower = &lower[0];
|
static deviceVector<Coeff_t> d_upper(Ls);acceleratorCopyToDevice(&upper[0],&d_upper[0],Ls*sizeof(Coeff_t));
|
||||||
|
static deviceVector<Coeff_t> d_lower(Ls);acceleratorCopyToDevice(&lower[0],&d_lower[0],Ls*sizeof(Coeff_t));
|
||||||
|
|
||||||
|
auto pdiag = &d_diag[0];
|
||||||
|
auto pupper = &d_upper[0];
|
||||||
|
auto plower = &d_lower[0];
|
||||||
|
|
||||||
// Flops = 6.0*(Nc*Ns) *Ls*vol
|
// Flops = 6.0*(Nc*Ns) *Ls*vol
|
||||||
|
|
||||||
@ -114,13 +125,18 @@ void DomainWallEOFAFermion<Impl>::MooeeInv(const FermionField& psi_i, FermionFie
|
|||||||
autoView( chi, chi_i, AcceleratorWrite);
|
autoView( chi, chi_i, AcceleratorWrite);
|
||||||
int Ls = this->Ls;
|
int Ls = this->Ls;
|
||||||
|
|
||||||
auto plee = & this->lee[0];
|
static deviceVector<Coeff_t> d_lee(Ls); acceleratorCopyToDevice(&this->lee[0],&d_lee[0],Ls*sizeof(Coeff_t));
|
||||||
auto pdee = & this->dee[0];
|
static deviceVector<Coeff_t> d_dee(Ls); acceleratorCopyToDevice(&this->dee[0],&d_dee[0],Ls*sizeof(Coeff_t));
|
||||||
auto puee = & this->uee[0];
|
static deviceVector<Coeff_t> d_uee(Ls); acceleratorCopyToDevice(&this->uee[0],&d_uee[0],Ls*sizeof(Coeff_t));
|
||||||
|
static deviceVector<Coeff_t> d_leem(Ls); acceleratorCopyToDevice(&this->leem[0],&d_leem[0],Ls*sizeof(Coeff_t));
|
||||||
auto pleem = & this->leem[0];
|
static deviceVector<Coeff_t> d_ueem(Ls); acceleratorCopyToDevice(&this->ueem[0],&d_ueem[0],Ls*sizeof(Coeff_t));
|
||||||
auto pueem = & this->ueem[0];
|
|
||||||
|
|
||||||
|
auto plee = & d_lee [0];
|
||||||
|
auto pdee = & d_dee [0];
|
||||||
|
auto puee = & d_uee [0];
|
||||||
|
auto pleem = & d_leem[0];
|
||||||
|
auto pueem = & d_ueem[0];
|
||||||
|
|
||||||
uint64_t nloop=grid->oSites()/Ls;
|
uint64_t nloop=grid->oSites()/Ls;
|
||||||
accelerator_for(sss,nloop,Simd::Nsimd(),{
|
accelerator_for(sss,nloop,Simd::Nsimd(),{
|
||||||
uint64_t ss=sss*Ls;
|
uint64_t ss=sss*Ls;
|
||||||
|
@ -131,9 +131,9 @@ void DomainWallEOFAFermion<Impl>::M5D(const FermionField& psi, FermionField& chi
|
|||||||
else{ shiftm = -shift*(mq3-mq2); }
|
else{ shiftm = -shift*(mq3-mq2); }
|
||||||
}
|
}
|
||||||
|
|
||||||
Vector<Coeff_t> diag(Ls,1.0);
|
std::vector<Coeff_t> diag(Ls,1.0);
|
||||||
Vector<Coeff_t> upper(Ls,-1.0); upper[Ls-1] = mq1 + shiftm;
|
std::vector<Coeff_t> upper(Ls,-1.0); upper[Ls-1] = mq1 + shiftm;
|
||||||
Vector<Coeff_t> lower(Ls,-1.0); lower[0] = mq1 + shiftp;
|
std::vector<Coeff_t> lower(Ls,-1.0); lower[0] = mq1 + shiftp;
|
||||||
|
|
||||||
#if(0)
|
#if(0)
|
||||||
std::cout << GridLogMessage << "DomainWallEOFAFermion::M5D(FF&,FF&):" << std::endl;
|
std::cout << GridLogMessage << "DomainWallEOFAFermion::M5D(FF&,FF&):" << std::endl;
|
||||||
@ -168,9 +168,9 @@ void DomainWallEOFAFermion<Impl>::M5Ddag(const FermionField& psi, FermionField&
|
|||||||
else{ shiftm = -shift*(mq3-mq2); }
|
else{ shiftm = -shift*(mq3-mq2); }
|
||||||
}
|
}
|
||||||
|
|
||||||
Vector<Coeff_t> diag(Ls,1.0);
|
std::vector<Coeff_t> diag(Ls,1.0);
|
||||||
Vector<Coeff_t> upper(Ls,-1.0); upper[Ls-1] = mq1 + shiftp;
|
std::vector<Coeff_t> upper(Ls,-1.0); upper[Ls-1] = mq1 + shiftp;
|
||||||
Vector<Coeff_t> lower(Ls,-1.0); lower[0] = mq1 + shiftm;
|
std::vector<Coeff_t> lower(Ls,-1.0); lower[0] = mq1 + shiftm;
|
||||||
|
|
||||||
this->M5Ddag(psi, chi, chi, lower, diag, upper);
|
this->M5Ddag(psi, chi, chi, lower, diag, upper);
|
||||||
}
|
}
|
||||||
@ -181,9 +181,9 @@ void DomainWallEOFAFermion<Impl>::Mooee(const FermionField& psi, FermionField& c
|
|||||||
{
|
{
|
||||||
int Ls = this->Ls;
|
int Ls = this->Ls;
|
||||||
|
|
||||||
Vector<Coeff_t> diag = this->bee;
|
std::vector<Coeff_t> diag = this->bee;
|
||||||
Vector<Coeff_t> upper(Ls);
|
std::vector<Coeff_t> upper(Ls);
|
||||||
Vector<Coeff_t> lower(Ls);
|
std::vector<Coeff_t> lower(Ls);
|
||||||
|
|
||||||
for(int s=0; s<Ls; s++){
|
for(int s=0; s<Ls; s++){
|
||||||
upper[s] = -this->cee[s];
|
upper[s] = -this->cee[s];
|
||||||
@ -200,9 +200,9 @@ void DomainWallEOFAFermion<Impl>::MooeeDag(const FermionField& psi, FermionField
|
|||||||
{
|
{
|
||||||
int Ls = this->Ls;
|
int Ls = this->Ls;
|
||||||
|
|
||||||
Vector<Coeff_t> diag = this->bee;
|
std::vector<Coeff_t> diag = this->bee;
|
||||||
Vector<Coeff_t> upper(Ls);
|
std::vector<Coeff_t> upper(Ls);
|
||||||
Vector<Coeff_t> lower(Ls);
|
std::vector<Coeff_t> lower(Ls);
|
||||||
|
|
||||||
for(int s=0; s<Ls; s++){
|
for(int s=0; s<Ls; s++){
|
||||||
upper[s] = -this->cee[s];
|
upper[s] = -this->cee[s];
|
||||||
@ -218,7 +218,7 @@ void DomainWallEOFAFermion<Impl>::MooeeDag(const FermionField& psi, FermionField
|
|||||||
|
|
||||||
//Zolo
|
//Zolo
|
||||||
template<class Impl>
|
template<class Impl>
|
||||||
void DomainWallEOFAFermion<Impl>::SetCoefficientsInternal(RealD zolo_hi, Vector<Coeff_t>& gamma, RealD b, RealD c)
|
void DomainWallEOFAFermion<Impl>::SetCoefficientsInternal(RealD zolo_hi, std::vector<Coeff_t>& gamma, RealD b, RealD c)
|
||||||
{
|
{
|
||||||
int Ls = this->Ls;
|
int Ls = this->Ls;
|
||||||
int pm = this->pm;
|
int pm = this->pm;
|
||||||
|
@ -61,8 +61,6 @@ ImprovedStaggeredFermion5D<Impl>::ImprovedStaggeredFermion5D(GridCartesian
|
|||||||
UUUmu(&FourDimGrid),
|
UUUmu(&FourDimGrid),
|
||||||
UUUmuEven(&FourDimRedBlackGrid),
|
UUUmuEven(&FourDimRedBlackGrid),
|
||||||
UUUmuOdd(&FourDimRedBlackGrid),
|
UUUmuOdd(&FourDimRedBlackGrid),
|
||||||
Lebesgue(&FourDimGrid),
|
|
||||||
LebesgueEvenOdd(&FourDimRedBlackGrid),
|
|
||||||
_tmp(&FiveDimRedBlackGrid)
|
_tmp(&FiveDimRedBlackGrid)
|
||||||
{
|
{
|
||||||
|
|
||||||
@ -277,18 +275,18 @@ void ImprovedStaggeredFermion5D<Impl>::DhopDerivOE(GaugeField &mat,
|
|||||||
|
|
||||||
/*CHANGE */
|
/*CHANGE */
|
||||||
template<class Impl>
|
template<class Impl>
|
||||||
void ImprovedStaggeredFermion5D<Impl>::DhopInternal(StencilImpl & st, LebesgueOrder &lo,
|
void ImprovedStaggeredFermion5D<Impl>::DhopInternal(StencilImpl & st,
|
||||||
DoubledGaugeField & U,DoubledGaugeField & UUU,
|
DoubledGaugeField & U,DoubledGaugeField & UUU,
|
||||||
const FermionField &in, FermionField &out,int dag)
|
const FermionField &in, FermionField &out,int dag)
|
||||||
{
|
{
|
||||||
if ( StaggeredKernelsStatic::Comms == StaggeredKernelsStatic::CommsAndCompute )
|
if ( StaggeredKernelsStatic::Comms == StaggeredKernelsStatic::CommsAndCompute )
|
||||||
DhopInternalOverlappedComms(st,lo,U,UUU,in,out,dag);
|
DhopInternalOverlappedComms(st,U,UUU,in,out,dag);
|
||||||
else
|
else
|
||||||
DhopInternalSerialComms(st,lo,U,UUU,in,out,dag);
|
DhopInternalSerialComms(st,U,UUU,in,out,dag);
|
||||||
}
|
}
|
||||||
|
|
||||||
template<class Impl>
|
template<class Impl>
|
||||||
void ImprovedStaggeredFermion5D<Impl>::DhopInternalOverlappedComms(StencilImpl & st, LebesgueOrder &lo,
|
void ImprovedStaggeredFermion5D<Impl>::DhopInternalOverlappedComms(StencilImpl & st,
|
||||||
DoubledGaugeField & U,DoubledGaugeField & UUU,
|
DoubledGaugeField & U,DoubledGaugeField & UUU,
|
||||||
const FermionField &in, FermionField &out,int dag)
|
const FermionField &in, FermionField &out,int dag)
|
||||||
{
|
{
|
||||||
@ -313,7 +311,7 @@ void ImprovedStaggeredFermion5D<Impl>::DhopInternalOverlappedComms(StencilImpl &
|
|||||||
{
|
{
|
||||||
int interior=1;
|
int interior=1;
|
||||||
int exterior=0;
|
int exterior=0;
|
||||||
Kernels::DhopImproved(st,lo,U,UUU,in,out,dag,interior,exterior);
|
Kernels::DhopImproved(st,U,UUU,in,out,dag,interior,exterior);
|
||||||
}
|
}
|
||||||
|
|
||||||
st.CommsMerge(compressor);
|
st.CommsMerge(compressor);
|
||||||
@ -323,12 +321,12 @@ void ImprovedStaggeredFermion5D<Impl>::DhopInternalOverlappedComms(StencilImpl &
|
|||||||
{
|
{
|
||||||
int interior=0;
|
int interior=0;
|
||||||
int exterior=1;
|
int exterior=1;
|
||||||
Kernels::DhopImproved(st,lo,U,UUU,in,out,dag,interior,exterior);
|
Kernels::DhopImproved(st,U,UUU,in,out,dag,interior,exterior);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
template<class Impl>
|
template<class Impl>
|
||||||
void ImprovedStaggeredFermion5D<Impl>::DhopInternalSerialComms(StencilImpl & st, LebesgueOrder &lo,
|
void ImprovedStaggeredFermion5D<Impl>::DhopInternalSerialComms(StencilImpl & st,
|
||||||
DoubledGaugeField & U,DoubledGaugeField & UUU,
|
DoubledGaugeField & U,DoubledGaugeField & UUU,
|
||||||
const FermionField &in, FermionField &out,int dag)
|
const FermionField &in, FermionField &out,int dag)
|
||||||
{
|
{
|
||||||
@ -341,7 +339,7 @@ void ImprovedStaggeredFermion5D<Impl>::DhopInternalSerialComms(StencilImpl & st,
|
|||||||
{
|
{
|
||||||
int interior=1;
|
int interior=1;
|
||||||
int exterior=1;
|
int exterior=1;
|
||||||
Kernels::DhopImproved(st,lo,U,UUU,in,out,dag,interior,exterior);
|
Kernels::DhopImproved(st,U,UUU,in,out,dag,interior,exterior);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
/*CHANGE END*/
|
/*CHANGE END*/
|
||||||
@ -357,7 +355,7 @@ void ImprovedStaggeredFermion5D<Impl>::DhopOE(const FermionField &in, FermionFie
|
|||||||
assert(in.Checkerboard()==Even);
|
assert(in.Checkerboard()==Even);
|
||||||
out.Checkerboard() = Odd;
|
out.Checkerboard() = Odd;
|
||||||
|
|
||||||
DhopInternal(StencilEven,LebesgueEvenOdd,UmuOdd,UUUmuOdd,in,out,dag);
|
DhopInternal(StencilEven,UmuOdd,UUUmuOdd,in,out,dag);
|
||||||
}
|
}
|
||||||
template<class Impl>
|
template<class Impl>
|
||||||
void ImprovedStaggeredFermion5D<Impl>::DhopEO(const FermionField &in, FermionField &out,int dag)
|
void ImprovedStaggeredFermion5D<Impl>::DhopEO(const FermionField &in, FermionField &out,int dag)
|
||||||
@ -368,7 +366,7 @@ void ImprovedStaggeredFermion5D<Impl>::DhopEO(const FermionField &in, FermionFie
|
|||||||
assert(in.Checkerboard()==Odd);
|
assert(in.Checkerboard()==Odd);
|
||||||
out.Checkerboard() = Even;
|
out.Checkerboard() = Even;
|
||||||
|
|
||||||
DhopInternal(StencilOdd,LebesgueEvenOdd,UmuEven,UUUmuEven,in,out,dag);
|
DhopInternal(StencilOdd,UmuEven,UUUmuEven,in,out,dag);
|
||||||
}
|
}
|
||||||
template<class Impl>
|
template<class Impl>
|
||||||
void ImprovedStaggeredFermion5D<Impl>::Dhop(const FermionField &in, FermionField &out,int dag)
|
void ImprovedStaggeredFermion5D<Impl>::Dhop(const FermionField &in, FermionField &out,int dag)
|
||||||
@ -378,7 +376,7 @@ void ImprovedStaggeredFermion5D<Impl>::Dhop(const FermionField &in, FermionField
|
|||||||
|
|
||||||
out.Checkerboard() = in.Checkerboard();
|
out.Checkerboard() = in.Checkerboard();
|
||||||
|
|
||||||
DhopInternal(Stencil,Lebesgue,Umu,UUUmu,in,out,dag);
|
DhopInternal(Stencil,Umu,UUUmu,in,out,dag);
|
||||||
}
|
}
|
||||||
|
|
||||||
/////////////////////////////////////////////////////////////////////////
|
/////////////////////////////////////////////////////////////////////////
|
||||||
|
@ -48,8 +48,6 @@ ImprovedStaggeredFermion<Impl>::ImprovedStaggeredFermion(GridCartesian &Fgrid, G
|
|||||||
StencilEven(&Hgrid, npoint, Even, directions, displacements,p), // source is Even
|
StencilEven(&Hgrid, npoint, Even, directions, displacements,p), // source is Even
|
||||||
StencilOdd(&Hgrid, npoint, Odd, directions, displacements,p), // source is Odd
|
StencilOdd(&Hgrid, npoint, Odd, directions, displacements,p), // source is Odd
|
||||||
mass(_mass),
|
mass(_mass),
|
||||||
Lebesgue(_grid),
|
|
||||||
LebesgueEvenOdd(_cbgrid),
|
|
||||||
Umu(&Fgrid),
|
Umu(&Fgrid),
|
||||||
UmuEven(&Hgrid),
|
UmuEven(&Hgrid),
|
||||||
UmuOdd(&Hgrid),
|
UmuOdd(&Hgrid),
|
||||||
@ -339,7 +337,7 @@ void ImprovedStaggeredFermion<Impl>::Dhop(const FermionField &in, FermionField &
|
|||||||
|
|
||||||
out.Checkerboard() = in.Checkerboard();
|
out.Checkerboard() = in.Checkerboard();
|
||||||
|
|
||||||
DhopInternal(Stencil, Lebesgue, Umu, UUUmu, in, out, dag);
|
DhopInternal(Stencil, Umu, UUUmu, in, out, dag);
|
||||||
}
|
}
|
||||||
|
|
||||||
template <class Impl>
|
template <class Impl>
|
||||||
@ -351,7 +349,7 @@ void ImprovedStaggeredFermion<Impl>::DhopOE(const FermionField &in, FermionField
|
|||||||
assert(in.Checkerboard() == Even);
|
assert(in.Checkerboard() == Even);
|
||||||
out.Checkerboard() = Odd;
|
out.Checkerboard() = Odd;
|
||||||
|
|
||||||
DhopInternal(StencilEven, LebesgueEvenOdd, UmuOdd, UUUmuOdd, in, out, dag);
|
DhopInternal(StencilEven, UmuOdd, UUUmuOdd, in, out, dag);
|
||||||
}
|
}
|
||||||
|
|
||||||
template <class Impl>
|
template <class Impl>
|
||||||
@ -363,7 +361,7 @@ void ImprovedStaggeredFermion<Impl>::DhopEO(const FermionField &in, FermionField
|
|||||||
assert(in.Checkerboard() == Odd);
|
assert(in.Checkerboard() == Odd);
|
||||||
out.Checkerboard() = Even;
|
out.Checkerboard() = Even;
|
||||||
|
|
||||||
DhopInternal(StencilOdd, LebesgueEvenOdd, UmuEven, UUUmuEven, in, out, dag);
|
DhopInternal(StencilOdd, UmuEven, UUUmuEven, in, out, dag);
|
||||||
}
|
}
|
||||||
|
|
||||||
template <class Impl>
|
template <class Impl>
|
||||||
@ -394,19 +392,19 @@ void ImprovedStaggeredFermion<Impl>::DhopDir(const FermionField &in, FermionFiel
|
|||||||
|
|
||||||
|
|
||||||
template <class Impl>
|
template <class Impl>
|
||||||
void ImprovedStaggeredFermion<Impl>::DhopInternal(StencilImpl &st, LebesgueOrder &lo,
|
void ImprovedStaggeredFermion<Impl>::DhopInternal(StencilImpl &st,
|
||||||
DoubledGaugeField &U,
|
DoubledGaugeField &U,
|
||||||
DoubledGaugeField &UUU,
|
DoubledGaugeField &UUU,
|
||||||
const FermionField &in,
|
const FermionField &in,
|
||||||
FermionField &out, int dag)
|
FermionField &out, int dag)
|
||||||
{
|
{
|
||||||
if ( StaggeredKernelsStatic::Comms == StaggeredKernelsStatic::CommsAndCompute )
|
if ( StaggeredKernelsStatic::Comms == StaggeredKernelsStatic::CommsAndCompute )
|
||||||
DhopInternalOverlappedComms(st,lo,U,UUU,in,out,dag);
|
DhopInternalOverlappedComms(st,U,UUU,in,out,dag);
|
||||||
else
|
else
|
||||||
DhopInternalSerialComms(st,lo,U,UUU,in,out,dag);
|
DhopInternalSerialComms(st,U,UUU,in,out,dag);
|
||||||
}
|
}
|
||||||
template <class Impl>
|
template <class Impl>
|
||||||
void ImprovedStaggeredFermion<Impl>::DhopInternalOverlappedComms(StencilImpl &st, LebesgueOrder &lo,
|
void ImprovedStaggeredFermion<Impl>::DhopInternalOverlappedComms(StencilImpl &st,
|
||||||
DoubledGaugeField &U,
|
DoubledGaugeField &U,
|
||||||
DoubledGaugeField &UUU,
|
DoubledGaugeField &UUU,
|
||||||
const FermionField &in,
|
const FermionField &in,
|
||||||
@ -429,7 +427,7 @@ void ImprovedStaggeredFermion<Impl>::DhopInternalOverlappedComms(StencilImpl &st
|
|||||||
{
|
{
|
||||||
int interior=1;
|
int interior=1;
|
||||||
int exterior=0;
|
int exterior=0;
|
||||||
Kernels::DhopImproved(st,lo,U,UUU,in,out,dag,interior,exterior);
|
Kernels::DhopImproved(st,U,UUU,in,out,dag,interior,exterior);
|
||||||
}
|
}
|
||||||
|
|
||||||
st.CommunicateComplete(requests);
|
st.CommunicateComplete(requests);
|
||||||
@ -440,13 +438,13 @@ void ImprovedStaggeredFermion<Impl>::DhopInternalOverlappedComms(StencilImpl &st
|
|||||||
{
|
{
|
||||||
int interior=0;
|
int interior=0;
|
||||||
int exterior=1;
|
int exterior=1;
|
||||||
Kernels::DhopImproved(st,lo,U,UUU,in,out,dag,interior,exterior);
|
Kernels::DhopImproved(st,U,UUU,in,out,dag,interior,exterior);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
template <class Impl>
|
template <class Impl>
|
||||||
void ImprovedStaggeredFermion<Impl>::DhopInternalSerialComms(StencilImpl &st, LebesgueOrder &lo,
|
void ImprovedStaggeredFermion<Impl>::DhopInternalSerialComms(StencilImpl &st,
|
||||||
DoubledGaugeField &U,
|
DoubledGaugeField &U,
|
||||||
DoubledGaugeField &UUU,
|
DoubledGaugeField &UUU,
|
||||||
const FermionField &in,
|
const FermionField &in,
|
||||||
@ -460,7 +458,7 @@ void ImprovedStaggeredFermion<Impl>::DhopInternalSerialComms(StencilImpl &st, Le
|
|||||||
{
|
{
|
||||||
int interior=1;
|
int interior=1;
|
||||||
int exterior=1;
|
int exterior=1;
|
||||||
Kernels::DhopImproved(st,lo,U,UUU,in,out,dag,interior,exterior);
|
Kernels::DhopImproved(st,U,UUU,in,out,dag,interior,exterior);
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -39,7 +39,7 @@ NAMESPACE_BEGIN(Grid);
|
|||||||
|
|
||||||
template<class Impl>
|
template<class Impl>
|
||||||
void MobiusEOFAFermion<Impl>::M5D(const FermionField &psi_i, const FermionField &phi_i, FermionField &chi_i,
|
void MobiusEOFAFermion<Impl>::M5D(const FermionField &psi_i, const FermionField &phi_i, FermionField &chi_i,
|
||||||
Vector<Coeff_t> &lower, Vector<Coeff_t> &diag, Vector<Coeff_t> &upper)
|
std::vector<Coeff_t> &lower, std::vector<Coeff_t> &diag, std::vector<Coeff_t> &upper)
|
||||||
{
|
{
|
||||||
chi_i.Checkerboard() = psi_i.Checkerboard();
|
chi_i.Checkerboard() = psi_i.Checkerboard();
|
||||||
GridBase *grid = psi_i.Grid();
|
GridBase *grid = psi_i.Grid();
|
||||||
@ -50,9 +50,13 @@ void MobiusEOFAFermion<Impl>::M5D(const FermionField &psi_i, const FermionField
|
|||||||
|
|
||||||
assert(phi.Checkerboard() == psi.Checkerboard());
|
assert(phi.Checkerboard() == psi.Checkerboard());
|
||||||
|
|
||||||
auto pdiag = &diag[0];
|
static deviceVector<Coeff_t> d_diag(Ls); acceleratorCopyToDevice(&diag[0],&d_diag[0],Ls*sizeof(Coeff_t));
|
||||||
auto pupper = &upper[0];
|
static deviceVector<Coeff_t> d_upper(Ls);acceleratorCopyToDevice(&upper[0],&d_upper[0],Ls*sizeof(Coeff_t));
|
||||||
auto plower = &lower[0];
|
static deviceVector<Coeff_t> d_lower(Ls);acceleratorCopyToDevice(&lower[0],&d_lower[0],Ls*sizeof(Coeff_t));
|
||||||
|
|
||||||
|
auto pdiag = &d_diag[0];
|
||||||
|
auto pupper = &d_upper[0];
|
||||||
|
auto plower = &d_lower[0];
|
||||||
|
|
||||||
// Flops = 6.0*(Nc*Ns) *Ls*vol
|
// Flops = 6.0*(Nc*Ns) *Ls*vol
|
||||||
int nloop = grid->oSites()/Ls;
|
int nloop = grid->oSites()/Ls;
|
||||||
@ -74,8 +78,8 @@ void MobiusEOFAFermion<Impl>::M5D(const FermionField &psi_i, const FermionField
|
|||||||
|
|
||||||
template<class Impl>
|
template<class Impl>
|
||||||
void MobiusEOFAFermion<Impl>::M5D_shift(const FermionField &psi_i, const FermionField &phi_i, FermionField &chi_i,
|
void MobiusEOFAFermion<Impl>::M5D_shift(const FermionField &psi_i, const FermionField &phi_i, FermionField &chi_i,
|
||||||
Vector<Coeff_t> &lower, Vector<Coeff_t> &diag, Vector<Coeff_t> &upper,
|
std::vector<Coeff_t> &lower, std::vector<Coeff_t> &diag, std::vector<Coeff_t> &upper,
|
||||||
Vector<Coeff_t> &shift_coeffs)
|
std::vector<Coeff_t> &shift_coeffs)
|
||||||
{
|
{
|
||||||
chi_i.Checkerboard() = psi_i.Checkerboard();
|
chi_i.Checkerboard() = psi_i.Checkerboard();
|
||||||
GridBase *grid = psi_i.Grid();
|
GridBase *grid = psi_i.Grid();
|
||||||
@ -86,13 +90,18 @@ void MobiusEOFAFermion<Impl>::M5D_shift(const FermionField &psi_i, const Fermion
|
|||||||
|
|
||||||
auto pm = this->pm;
|
auto pm = this->pm;
|
||||||
int shift_s = (pm == 1) ? (Ls-1) : 0; // s-component modified by shift operator
|
int shift_s = (pm == 1) ? (Ls-1) : 0; // s-component modified by shift operator
|
||||||
|
|
||||||
assert(phi.Checkerboard() == psi.Checkerboard());
|
assert(phi.Checkerboard() == psi.Checkerboard());
|
||||||
|
|
||||||
auto pdiag = &diag[0];
|
static deviceVector<Coeff_t> d_diag(Ls); acceleratorCopyToDevice(&diag[0],&d_diag[0],Ls*sizeof(Coeff_t));
|
||||||
auto pupper = &upper[0];
|
static deviceVector<Coeff_t> d_upper(Ls);acceleratorCopyToDevice(&upper[0],&d_upper[0],Ls*sizeof(Coeff_t));
|
||||||
auto plower = &lower[0];
|
static deviceVector<Coeff_t> d_lower(Ls);acceleratorCopyToDevice(&lower[0],&d_lower[0],Ls*sizeof(Coeff_t));
|
||||||
auto pshift_coeffs = &shift_coeffs[0];
|
static deviceVector<Coeff_t> d_shift_coeffs(Ls);acceleratorCopyToDevice(&shift_coeffs[0],&d_shift_coeffs[0],Ls*sizeof(Coeff_t));
|
||||||
|
|
||||||
|
auto pdiag = &d_diag[0];
|
||||||
|
auto pupper = &d_upper[0];
|
||||||
|
auto plower = &d_lower[0];
|
||||||
|
auto pshift_coeffs = &d_shift_coeffs[0];
|
||||||
|
|
||||||
// Flops = 6.0*(Nc*Ns) *Ls*vol
|
// Flops = 6.0*(Nc*Ns) *Ls*vol
|
||||||
int nloop = grid->oSites()/Ls;
|
int nloop = grid->oSites()/Ls;
|
||||||
@ -119,7 +128,7 @@ void MobiusEOFAFermion<Impl>::M5D_shift(const FermionField &psi_i, const Fermion
|
|||||||
|
|
||||||
template<class Impl>
|
template<class Impl>
|
||||||
void MobiusEOFAFermion<Impl>::M5Ddag(const FermionField &psi_i, const FermionField &phi_i, FermionField &chi_i,
|
void MobiusEOFAFermion<Impl>::M5Ddag(const FermionField &psi_i, const FermionField &phi_i, FermionField &chi_i,
|
||||||
Vector<Coeff_t> &lower, Vector<Coeff_t> &diag, Vector<Coeff_t> &upper)
|
std::vector<Coeff_t> &lower, std::vector<Coeff_t> &diag, std::vector<Coeff_t> &upper)
|
||||||
{
|
{
|
||||||
chi_i.Checkerboard() = psi_i.Checkerboard();
|
chi_i.Checkerboard() = psi_i.Checkerboard();
|
||||||
GridBase *grid = psi_i.Grid();
|
GridBase *grid = psi_i.Grid();
|
||||||
@ -130,9 +139,13 @@ void MobiusEOFAFermion<Impl>::M5Ddag(const FermionField &psi_i, const FermionFie
|
|||||||
|
|
||||||
assert(phi.Checkerboard() == psi.Checkerboard());
|
assert(phi.Checkerboard() == psi.Checkerboard());
|
||||||
|
|
||||||
auto pdiag = &diag[0];
|
static deviceVector<Coeff_t> d_diag(Ls); acceleratorCopyToDevice(&diag[0],&d_diag[0],Ls*sizeof(Coeff_t));
|
||||||
auto pupper = &upper[0];
|
static deviceVector<Coeff_t> d_upper(Ls);acceleratorCopyToDevice(&upper[0],&d_upper[0],Ls*sizeof(Coeff_t));
|
||||||
auto plower = &lower[0];
|
static deviceVector<Coeff_t> d_lower(Ls);acceleratorCopyToDevice(&lower[0],&d_lower[0],Ls*sizeof(Coeff_t));
|
||||||
|
|
||||||
|
auto pdiag = &d_diag[0];
|
||||||
|
auto pupper = &d_upper[0];
|
||||||
|
auto plower = &d_lower[0];
|
||||||
|
|
||||||
// Flops = 6.0*(Nc*Ns) *Ls*vol
|
// Flops = 6.0*(Nc*Ns) *Ls*vol
|
||||||
int nloop = grid->oSites()/Ls;
|
int nloop = grid->oSites()/Ls;
|
||||||
@ -154,8 +167,8 @@ void MobiusEOFAFermion<Impl>::M5Ddag(const FermionField &psi_i, const FermionFie
|
|||||||
|
|
||||||
template<class Impl>
|
template<class Impl>
|
||||||
void MobiusEOFAFermion<Impl>::M5Ddag_shift(const FermionField &psi_i, const FermionField &phi_i, FermionField &chi_i,
|
void MobiusEOFAFermion<Impl>::M5Ddag_shift(const FermionField &psi_i, const FermionField &phi_i, FermionField &chi_i,
|
||||||
Vector<Coeff_t> &lower, Vector<Coeff_t> &diag, Vector<Coeff_t> &upper,
|
std::vector<Coeff_t> &lower, std::vector<Coeff_t> &diag, std::vector<Coeff_t> &upper,
|
||||||
Vector<Coeff_t> &shift_coeffs)
|
std::vector<Coeff_t> &shift_coeffs)
|
||||||
{
|
{
|
||||||
chi_i.Checkerboard() = psi_i.Checkerboard();
|
chi_i.Checkerboard() = psi_i.Checkerboard();
|
||||||
GridBase *grid = psi_i.Grid();
|
GridBase *grid = psi_i.Grid();
|
||||||
@ -167,10 +180,15 @@ void MobiusEOFAFermion<Impl>::M5Ddag_shift(const FermionField &psi_i, const Ferm
|
|||||||
|
|
||||||
assert(phi.Checkerboard() == psi.Checkerboard());
|
assert(phi.Checkerboard() == psi.Checkerboard());
|
||||||
|
|
||||||
auto pdiag = &diag[0];
|
static deviceVector<Coeff_t> d_diag(Ls); acceleratorCopyToDevice(&diag[0],&d_diag[0],Ls*sizeof(Coeff_t));
|
||||||
auto pupper = &upper[0];
|
static deviceVector<Coeff_t> d_upper(Ls);acceleratorCopyToDevice(&upper[0],&d_upper[0],Ls*sizeof(Coeff_t));
|
||||||
auto plower = &lower[0];
|
static deviceVector<Coeff_t> d_lower(Ls);acceleratorCopyToDevice(&lower[0],&d_lower[0],Ls*sizeof(Coeff_t));
|
||||||
auto pshift_coeffs = &shift_coeffs[0];
|
static deviceVector<Coeff_t> d_shift_coeffs(Ls);acceleratorCopyToDevice(&shift_coeffs[0],&d_shift_coeffs[0],Ls*sizeof(Coeff_t));
|
||||||
|
|
||||||
|
auto pdiag = &d_diag[0];
|
||||||
|
auto pupper = &d_upper[0];
|
||||||
|
auto plower = &d_lower[0];
|
||||||
|
auto pshift_coeffs = &d_shift_coeffs[0];
|
||||||
|
|
||||||
// Flops = 6.0*(Nc*Ns) *Ls*vol
|
// Flops = 6.0*(Nc*Ns) *Ls*vol
|
||||||
auto pm = this->pm;
|
auto pm = this->pm;
|
||||||
@ -212,11 +230,17 @@ void MobiusEOFAFermion<Impl>::MooeeInv(const FermionField &psi_i, FermionField &
|
|||||||
autoView(psi , psi_i, AcceleratorRead);
|
autoView(psi , psi_i, AcceleratorRead);
|
||||||
autoView(chi , chi_i, AcceleratorWrite);
|
autoView(chi , chi_i, AcceleratorWrite);
|
||||||
|
|
||||||
auto plee = & this->lee [0];
|
static deviceVector<Coeff_t> d_lee(Ls); acceleratorCopyToDevice(&this->lee[0],&d_lee[0],Ls*sizeof(Coeff_t));
|
||||||
auto pdee = & this->dee [0];
|
static deviceVector<Coeff_t> d_dee(Ls); acceleratorCopyToDevice(&this->dee[0],&d_dee[0],Ls*sizeof(Coeff_t));
|
||||||
auto puee = & this->uee [0];
|
static deviceVector<Coeff_t> d_uee(Ls); acceleratorCopyToDevice(&this->uee[0],&d_uee[0],Ls*sizeof(Coeff_t));
|
||||||
auto pleem= & this->leem[0];
|
static deviceVector<Coeff_t> d_leem(Ls); acceleratorCopyToDevice(&this->leem[0],&d_leem[0],Ls*sizeof(Coeff_t));
|
||||||
auto pueem= & this->ueem[0];
|
static deviceVector<Coeff_t> d_ueem(Ls); acceleratorCopyToDevice(&this->ueem[0],&d_ueem[0],Ls*sizeof(Coeff_t));
|
||||||
|
|
||||||
|
auto plee = & d_lee [0];
|
||||||
|
auto pdee = & d_dee [0];
|
||||||
|
auto puee = & d_uee [0];
|
||||||
|
auto pleem = & d_leem[0];
|
||||||
|
auto pueem = & d_ueem[0];
|
||||||
|
|
||||||
if(this->shift != 0.0){ MooeeInv_shift(psi_i,chi_i); return; }
|
if(this->shift != 0.0){ MooeeInv_shift(psi_i,chi_i); return; }
|
||||||
|
|
||||||
@ -268,14 +292,24 @@ void MobiusEOFAFermion<Impl>::MooeeInv_shift(const FermionField &psi_i, FermionF
|
|||||||
autoView(psi , psi_i, AcceleratorRead);
|
autoView(psi , psi_i, AcceleratorRead);
|
||||||
autoView(chi , chi_i, AcceleratorWrite);
|
autoView(chi , chi_i, AcceleratorWrite);
|
||||||
|
|
||||||
|
// Move into object and constructor
|
||||||
|
static deviceVector<Coeff_t> d_lee(Ls); acceleratorCopyToDevice(&this->lee[0],&d_lee[0],Ls*sizeof(Coeff_t));
|
||||||
|
static deviceVector<Coeff_t> d_dee(Ls); acceleratorCopyToDevice(&this->dee[0],&d_dee[0],Ls*sizeof(Coeff_t));
|
||||||
|
static deviceVector<Coeff_t> d_uee(Ls); acceleratorCopyToDevice(&this->uee[0],&d_uee[0],Ls*sizeof(Coeff_t));
|
||||||
|
static deviceVector<Coeff_t> d_leem(Ls); acceleratorCopyToDevice(&this->leem[0],&d_leem[0],Ls*sizeof(Coeff_t));
|
||||||
|
static deviceVector<Coeff_t> d_ueem(Ls); acceleratorCopyToDevice(&this->ueem[0],&d_ueem[0],Ls*sizeof(Coeff_t));
|
||||||
|
|
||||||
auto pm = this->pm;
|
auto pm = this->pm;
|
||||||
auto plee = & this->lee [0];
|
auto plee = & d_lee [0];
|
||||||
auto pdee = & this->dee [0];
|
auto pdee = & d_dee [0];
|
||||||
auto puee = & this->uee [0];
|
auto puee = & d_uee [0];
|
||||||
auto pleem= & this->leem[0];
|
auto pleem = & d_leem[0];
|
||||||
auto pueem= & this->ueem[0];
|
auto pueem = & d_ueem[0];
|
||||||
auto pMooeeInv_shift_lc = &MooeeInv_shift_lc[0];
|
|
||||||
auto pMooeeInv_shift_norm = &MooeeInv_shift_norm[0];
|
static deviceVector<Coeff_t> d_MooeeInv_shift_lc(Ls); acceleratorCopyToDevice(&MooeeInv_shift_lc[0],&d_MooeeInv_shift_lc[0],Ls*sizeof(Coeff_t));
|
||||||
|
static deviceVector<Coeff_t> d_MooeeInv_shift_norm(Ls); acceleratorCopyToDevice(&MooeeInv_shift_norm[0],&d_MooeeInv_shift_norm[0],Ls*sizeof(Coeff_t));
|
||||||
|
auto pMooeeInv_shift_lc = &d_MooeeInv_shift_lc[0];
|
||||||
|
auto pMooeeInv_shift_norm = &d_MooeeInv_shift_norm[0];
|
||||||
|
|
||||||
int nloop = grid->oSites()/Ls;
|
int nloop = grid->oSites()/Ls;
|
||||||
accelerator_for(sss,nloop,Simd::Nsimd(),{
|
accelerator_for(sss,nloop,Simd::Nsimd(),{
|
||||||
@ -333,11 +367,17 @@ void MobiusEOFAFermion<Impl>::MooeeInvDag(const FermionField &psi_i, FermionFiel
|
|||||||
autoView(psi , psi_i, AcceleratorRead);
|
autoView(psi , psi_i, AcceleratorRead);
|
||||||
autoView(chi , chi_i, AcceleratorWrite);
|
autoView(chi , chi_i, AcceleratorWrite);
|
||||||
|
|
||||||
auto plee = & this->lee [0];
|
static deviceVector<Coeff_t> d_lee(Ls); acceleratorCopyToDevice(&this->lee[0],&d_lee[0],Ls*sizeof(Coeff_t));
|
||||||
auto pdee = & this->dee [0];
|
static deviceVector<Coeff_t> d_dee(Ls); acceleratorCopyToDevice(&this->dee[0],&d_dee[0],Ls*sizeof(Coeff_t));
|
||||||
auto puee = & this->uee [0];
|
static deviceVector<Coeff_t> d_uee(Ls); acceleratorCopyToDevice(&this->uee[0],&d_uee[0],Ls*sizeof(Coeff_t));
|
||||||
auto pleem= & this->leem[0];
|
static deviceVector<Coeff_t> d_leem(Ls); acceleratorCopyToDevice(&this->leem[0],&d_leem[0],Ls*sizeof(Coeff_t));
|
||||||
auto pueem= & this->ueem[0];
|
static deviceVector<Coeff_t> d_ueem(Ls); acceleratorCopyToDevice(&this->ueem[0],&d_ueem[0],Ls*sizeof(Coeff_t));
|
||||||
|
|
||||||
|
auto plee = & d_lee [0];
|
||||||
|
auto pdee = & d_dee [0];
|
||||||
|
auto puee = & d_uee [0];
|
||||||
|
auto pleem = & d_leem[0];
|
||||||
|
auto pueem = & d_ueem[0];
|
||||||
|
|
||||||
int nloop = grid->oSites()/Ls;
|
int nloop = grid->oSites()/Ls;
|
||||||
accelerator_for(sss,nloop,Simd::Nsimd(),{
|
accelerator_for(sss,nloop,Simd::Nsimd(),{
|
||||||
@ -386,14 +426,28 @@ void MobiusEOFAFermion<Impl>::MooeeInvDag_shift(const FermionField &psi_i, Fermi
|
|||||||
autoView(chi , chi_i, AcceleratorWrite);
|
autoView(chi , chi_i, AcceleratorWrite);
|
||||||
int Ls = this->Ls;
|
int Ls = this->Ls;
|
||||||
|
|
||||||
|
static deviceVector<Coeff_t> d_lee(Ls); acceleratorCopyToDevice(&this->lee[0],&d_lee[0],Ls*sizeof(Coeff_t));
|
||||||
|
static deviceVector<Coeff_t> d_dee(Ls); acceleratorCopyToDevice(&this->dee[0],&d_dee[0],Ls*sizeof(Coeff_t));
|
||||||
|
static deviceVector<Coeff_t> d_uee(Ls); acceleratorCopyToDevice(&this->uee[0],&d_uee[0],Ls*sizeof(Coeff_t));
|
||||||
|
static deviceVector<Coeff_t> d_leem(Ls); acceleratorCopyToDevice(&this->leem[0],&d_leem[0],Ls*sizeof(Coeff_t));
|
||||||
|
static deviceVector<Coeff_t> d_ueem(Ls); acceleratorCopyToDevice(&this->ueem[0],&d_ueem[0],Ls*sizeof(Coeff_t));
|
||||||
|
|
||||||
auto pm = this->pm;
|
auto pm = this->pm;
|
||||||
auto plee = & this->lee [0];
|
auto plee = & d_lee [0];
|
||||||
auto pdee = & this->dee [0];
|
auto pdee = & d_dee [0];
|
||||||
auto puee = & this->uee [0];
|
auto puee = & d_uee [0];
|
||||||
auto pleem= & this->leem[0];
|
auto pleem = & d_leem[0];
|
||||||
auto pueem= & this->ueem[0];
|
auto pueem = & d_ueem[0];
|
||||||
auto pMooeeInvDag_shift_lc = &MooeeInvDag_shift_lc[0];
|
|
||||||
auto pMooeeInvDag_shift_norm = &MooeeInvDag_shift_norm[0];
|
static deviceVector<Coeff_t> d_MooeeInvDag_shift_lc(Ls);
|
||||||
|
static deviceVector<Coeff_t> d_MooeeInvDag_shift_norm(Ls);
|
||||||
|
acceleratorCopyToDevice(&MooeeInvDag_shift_lc[0],&d_MooeeInvDag_shift_lc[0],Ls*sizeof(Coeff_t));
|
||||||
|
acceleratorCopyToDevice(&MooeeInvDag_shift_norm[0],&d_MooeeInvDag_shift_norm[0],Ls*sizeof(Coeff_t));
|
||||||
|
auto pMooeeInvDag_shift_lc = &d_MooeeInvDag_shift_lc[0];
|
||||||
|
auto pMooeeInvDag_shift_norm = &d_MooeeInvDag_shift_norm[0];
|
||||||
|
|
||||||
|
// auto pMooeeInvDag_shift_lc = &MooeeInvDag_shift_lc[0];
|
||||||
|
// auto pMooeeInvDag_shift_norm = &MooeeInvDag_shift_norm[0];
|
||||||
|
|
||||||
int nloop = grid->oSites()/Ls;
|
int nloop = grid->oSites()/Ls;
|
||||||
accelerator_for(sss,nloop,Simd::Nsimd(),{
|
accelerator_for(sss,nloop,Simd::Nsimd(),{
|
||||||
|
@ -196,9 +196,9 @@ void MobiusEOFAFermion<Impl>::M5D(const FermionField& psi, FermionField& chi)
|
|||||||
{
|
{
|
||||||
int Ls = this->Ls;
|
int Ls = this->Ls;
|
||||||
|
|
||||||
Vector<Coeff_t> diag(Ls,1.0);
|
std::vector<Coeff_t> diag(Ls,1.0);
|
||||||
Vector<Coeff_t> upper(Ls,-1.0); upper[Ls-1] = this->mq1;
|
std::vector<Coeff_t> upper(Ls,-1.0); upper[Ls-1] = this->mq1;
|
||||||
Vector<Coeff_t> lower(Ls,-1.0); lower[0] = this->mq1;
|
std::vector<Coeff_t> lower(Ls,-1.0); lower[0] = this->mq1;
|
||||||
|
|
||||||
// no shift term
|
// no shift term
|
||||||
if(this->shift == 0.0){ this->M5D(psi, chi, chi, lower, diag, upper); }
|
if(this->shift == 0.0){ this->M5D(psi, chi, chi, lower, diag, upper); }
|
||||||
@ -212,9 +212,9 @@ void MobiusEOFAFermion<Impl>::M5Ddag(const FermionField& psi, FermionField& chi)
|
|||||||
{
|
{
|
||||||
int Ls = this->Ls;
|
int Ls = this->Ls;
|
||||||
|
|
||||||
Vector<Coeff_t> diag(Ls,1.0);
|
std::vector<Coeff_t> diag(Ls,1.0);
|
||||||
Vector<Coeff_t> upper(Ls,-1.0); upper[Ls-1] = this->mq1;
|
std::vector<Coeff_t> upper(Ls,-1.0); upper[Ls-1] = this->mq1;
|
||||||
Vector<Coeff_t> lower(Ls,-1.0); lower[0] = this->mq1;
|
std::vector<Coeff_t> lower(Ls,-1.0); lower[0] = this->mq1;
|
||||||
|
|
||||||
// no shift term
|
// no shift term
|
||||||
if(this->shift == 0.0){ this->M5Ddag(psi, chi, chi, lower, diag, upper); }
|
if(this->shift == 0.0){ this->M5Ddag(psi, chi, chi, lower, diag, upper); }
|
||||||
@ -230,9 +230,9 @@ void MobiusEOFAFermion<Impl>::Mooee(const FermionField& psi, FermionField& chi)
|
|||||||
int Ls = this->Ls;
|
int Ls = this->Ls;
|
||||||
|
|
||||||
// coefficients of Mooee
|
// coefficients of Mooee
|
||||||
Vector<Coeff_t> diag = this->bee;
|
std::vector<Coeff_t> diag = this->bee;
|
||||||
Vector<Coeff_t> upper(Ls);
|
std::vector<Coeff_t> upper(Ls);
|
||||||
Vector<Coeff_t> lower(Ls);
|
std::vector<Coeff_t> lower(Ls);
|
||||||
for(int s=0; s<Ls; s++){
|
for(int s=0; s<Ls; s++){
|
||||||
upper[s] = -this->cee[s];
|
upper[s] = -this->cee[s];
|
||||||
lower[s] = -this->cee[s];
|
lower[s] = -this->cee[s];
|
||||||
@ -253,9 +253,9 @@ void MobiusEOFAFermion<Impl>::MooeeDag(const FermionField& psi, FermionField& ch
|
|||||||
int Ls = this->Ls;
|
int Ls = this->Ls;
|
||||||
|
|
||||||
// coefficients of MooeeDag
|
// coefficients of MooeeDag
|
||||||
Vector<Coeff_t> diag = this->bee;
|
std::vector<Coeff_t> diag = this->bee;
|
||||||
Vector<Coeff_t> upper(Ls);
|
std::vector<Coeff_t> upper(Ls);
|
||||||
Vector<Coeff_t> lower(Ls);
|
std::vector<Coeff_t> lower(Ls);
|
||||||
for(int s=0; s<Ls; s++){
|
for(int s=0; s<Ls; s++){
|
||||||
if(s==0) {
|
if(s==0) {
|
||||||
upper[s] = -this->cee[s+1];
|
upper[s] = -this->cee[s+1];
|
||||||
@ -314,10 +314,10 @@ void MobiusEOFAFermion<Impl>::SetCoefficientsPrecondShiftOps()
|
|||||||
// Tridiagonal solve for MooeeInvDag_shift_lc
|
// Tridiagonal solve for MooeeInvDag_shift_lc
|
||||||
{
|
{
|
||||||
Coeff_t m(0.0);
|
Coeff_t m(0.0);
|
||||||
Vector<Coeff_t> d = Mooee_shift;
|
std::vector<Coeff_t> d = Mooee_shift;
|
||||||
Vector<Coeff_t> u(Ls,0.0);
|
std::vector<Coeff_t> u(Ls,0.0);
|
||||||
Vector<Coeff_t> y(Ls,0.0);
|
std::vector<Coeff_t> y(Ls,0.0);
|
||||||
Vector<Coeff_t> q(Ls,0.0);
|
std::vector<Coeff_t> q(Ls,0.0);
|
||||||
if(pm == 1){ u[0] = 1.0; }
|
if(pm == 1){ u[0] = 1.0; }
|
||||||
else{ u[Ls-1] = 1.0; }
|
else{ u[Ls-1] = 1.0; }
|
||||||
|
|
||||||
|
@ -48,8 +48,6 @@ NaiveStaggeredFermion<Impl>::NaiveStaggeredFermion(GridCartesian &Fgrid, GridRed
|
|||||||
StencilEven(&Hgrid, npoint, Even, directions, displacements,p), // source is Even
|
StencilEven(&Hgrid, npoint, Even, directions, displacements,p), // source is Even
|
||||||
StencilOdd(&Hgrid, npoint, Odd, directions, displacements,p), // source is Odd
|
StencilOdd(&Hgrid, npoint, Odd, directions, displacements,p), // source is Odd
|
||||||
mass(_mass),
|
mass(_mass),
|
||||||
Lebesgue(_grid),
|
|
||||||
LebesgueEvenOdd(_cbgrid),
|
|
||||||
Umu(&Fgrid),
|
Umu(&Fgrid),
|
||||||
UmuEven(&Hgrid),
|
UmuEven(&Hgrid),
|
||||||
UmuOdd(&Hgrid),
|
UmuOdd(&Hgrid),
|
||||||
@ -268,7 +266,7 @@ void NaiveStaggeredFermion<Impl>::Dhop(const FermionField &in, FermionField &out
|
|||||||
|
|
||||||
out.Checkerboard() = in.Checkerboard();
|
out.Checkerboard() = in.Checkerboard();
|
||||||
|
|
||||||
DhopInternal(Stencil, Lebesgue, Umu, in, out, dag);
|
DhopInternal(Stencil, Umu, in, out, dag);
|
||||||
}
|
}
|
||||||
|
|
||||||
template <class Impl>
|
template <class Impl>
|
||||||
@ -280,7 +278,7 @@ void NaiveStaggeredFermion<Impl>::DhopOE(const FermionField &in, FermionField &o
|
|||||||
assert(in.Checkerboard() == Even);
|
assert(in.Checkerboard() == Even);
|
||||||
out.Checkerboard() = Odd;
|
out.Checkerboard() = Odd;
|
||||||
|
|
||||||
DhopInternal(StencilEven, LebesgueEvenOdd, UmuOdd, in, out, dag);
|
DhopInternal(StencilEven, UmuOdd, in, out, dag);
|
||||||
}
|
}
|
||||||
|
|
||||||
template <class Impl>
|
template <class Impl>
|
||||||
@ -292,7 +290,7 @@ void NaiveStaggeredFermion<Impl>::DhopEO(const FermionField &in, FermionField &o
|
|||||||
assert(in.Checkerboard() == Odd);
|
assert(in.Checkerboard() == Odd);
|
||||||
out.Checkerboard() = Even;
|
out.Checkerboard() = Even;
|
||||||
|
|
||||||
DhopInternal(StencilOdd, LebesgueEvenOdd, UmuEven, in, out, dag);
|
DhopInternal(StencilOdd, UmuEven, in, out, dag);
|
||||||
}
|
}
|
||||||
|
|
||||||
template <class Impl>
|
template <class Impl>
|
||||||
@ -323,18 +321,18 @@ void NaiveStaggeredFermion<Impl>::DhopDir(const FermionField &in, FermionField &
|
|||||||
|
|
||||||
|
|
||||||
template <class Impl>
|
template <class Impl>
|
||||||
void NaiveStaggeredFermion<Impl>::DhopInternal(StencilImpl &st, LebesgueOrder &lo,
|
void NaiveStaggeredFermion<Impl>::DhopInternal(StencilImpl &st,
|
||||||
DoubledGaugeField &U,
|
DoubledGaugeField &U,
|
||||||
const FermionField &in,
|
const FermionField &in,
|
||||||
FermionField &out, int dag)
|
FermionField &out, int dag)
|
||||||
{
|
{
|
||||||
if ( StaggeredKernelsStatic::Comms == StaggeredKernelsStatic::CommsAndCompute )
|
if ( StaggeredKernelsStatic::Comms == StaggeredKernelsStatic::CommsAndCompute )
|
||||||
DhopInternalOverlappedComms(st,lo,U,in,out,dag);
|
DhopInternalOverlappedComms(st,U,in,out,dag);
|
||||||
else
|
else
|
||||||
DhopInternalSerialComms(st,lo,U,in,out,dag);
|
DhopInternalSerialComms(st,U,in,out,dag);
|
||||||
}
|
}
|
||||||
template <class Impl>
|
template <class Impl>
|
||||||
void NaiveStaggeredFermion<Impl>::DhopInternalOverlappedComms(StencilImpl &st, LebesgueOrder &lo,
|
void NaiveStaggeredFermion<Impl>::DhopInternalOverlappedComms(StencilImpl &st,
|
||||||
DoubledGaugeField &U,
|
DoubledGaugeField &U,
|
||||||
const FermionField &in,
|
const FermionField &in,
|
||||||
FermionField &out, int dag)
|
FermionField &out, int dag)
|
||||||
@ -356,7 +354,7 @@ void NaiveStaggeredFermion<Impl>::DhopInternalOverlappedComms(StencilImpl &st, L
|
|||||||
{
|
{
|
||||||
int interior=1;
|
int interior=1;
|
||||||
int exterior=0;
|
int exterior=0;
|
||||||
Kernels::DhopNaive(st,lo,U,in,out,dag,interior,exterior);
|
Kernels::DhopNaive(st,U,in,out,dag,interior,exterior);
|
||||||
}
|
}
|
||||||
|
|
||||||
st.CommunicateComplete(requests);
|
st.CommunicateComplete(requests);
|
||||||
@ -367,12 +365,12 @@ void NaiveStaggeredFermion<Impl>::DhopInternalOverlappedComms(StencilImpl &st, L
|
|||||||
{
|
{
|
||||||
int interior=0;
|
int interior=0;
|
||||||
int exterior=1;
|
int exterior=1;
|
||||||
Kernels::DhopNaive(st,lo,U,in,out,dag,interior,exterior);
|
Kernels::DhopNaive(st,U,in,out,dag,interior,exterior);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
template <class Impl>
|
template <class Impl>
|
||||||
void NaiveStaggeredFermion<Impl>::DhopInternalSerialComms(StencilImpl &st, LebesgueOrder &lo,
|
void NaiveStaggeredFermion<Impl>::DhopInternalSerialComms(StencilImpl &st,
|
||||||
DoubledGaugeField &U,
|
DoubledGaugeField &U,
|
||||||
const FermionField &in,
|
const FermionField &in,
|
||||||
FermionField &out, int dag)
|
FermionField &out, int dag)
|
||||||
@ -385,7 +383,7 @@ void NaiveStaggeredFermion<Impl>::DhopInternalSerialComms(StencilImpl &st, Lebes
|
|||||||
{
|
{
|
||||||
int interior=1;
|
int interior=1;
|
||||||
int exterior=1;
|
int exterior=1;
|
||||||
Kernels::DhopNaive(st,lo,U,in,out,dag,interior,exterior);
|
Kernels::DhopNaive(st,U,in,out,dag,interior,exterior);
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -375,23 +375,6 @@ void StaggeredKernels<Impl>::DhopSiteHandExt(StencilView &st,
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
|
||||||
#define DHOP_SITE_HAND_INSTANTIATE(IMPL) \
|
|
||||||
template void StaggeredKernels<IMPL>::DhopSiteHand(StencilImpl &st, LebesgueOrder &lo, \
|
|
||||||
DoubledGaugeFieldView &U,DoubledGaugeFieldView &UUU, \
|
|
||||||
SiteSpinor *buf, int LLs, int sU, \
|
|
||||||
const FermionFieldView &in, FermionFieldView &out, int dag); \
|
|
||||||
\
|
|
||||||
template void StaggeredKernels<IMPL>::DhopSiteHandInt(StencilImpl &st, LebesgueOrder &lo, \
|
|
||||||
DoubledGaugeFieldView &U,DoubledGaugeFieldView &UUU, \
|
|
||||||
SiteSpinor *buf, int LLs, int sU, \
|
|
||||||
const FermionFieldView &in, FermionFieldView &out, int dag); \
|
|
||||||
\
|
|
||||||
template void StaggeredKernels<IMPL>::DhopSiteHandExt(StencilImpl &st, LebesgueOrder &lo, \
|
|
||||||
DoubledGaugeFieldView &U,DoubledGaugeFieldView &UUU, \
|
|
||||||
SiteSpinor *buf, int LLs, int sU, \
|
|
||||||
const FermionFieldView &in, FermionFieldView &out, int dag); \
|
|
||||||
*/
|
|
||||||
#undef LOAD_CHI
|
#undef LOAD_CHI
|
||||||
#undef HAND_DECLARATIONS
|
#undef HAND_DECLARATIONS
|
||||||
|
|
||||||
|
@ -256,7 +256,7 @@ void StaggeredKernels<Impl>::DhopDirKernel(StencilImpl &st, DoubledGaugeFieldVie
|
|||||||
});
|
});
|
||||||
|
|
||||||
template <class Impl>
|
template <class Impl>
|
||||||
void StaggeredKernels<Impl>::DhopImproved(StencilImpl &st, LebesgueOrder &lo,
|
void StaggeredKernels<Impl>::DhopImproved(StencilImpl &st,
|
||||||
DoubledGaugeField &U, DoubledGaugeField &UUU,
|
DoubledGaugeField &U, DoubledGaugeField &UUU,
|
||||||
const FermionField &in, FermionField &out, int dag, int interior,int exterior)
|
const FermionField &in, FermionField &out, int dag, int interior,int exterior)
|
||||||
{
|
{
|
||||||
@ -294,7 +294,7 @@ void StaggeredKernels<Impl>::DhopImproved(StencilImpl &st, LebesgueOrder &lo,
|
|||||||
assert(0 && " Kernel optimisation case not covered ");
|
assert(0 && " Kernel optimisation case not covered ");
|
||||||
}
|
}
|
||||||
template <class Impl>
|
template <class Impl>
|
||||||
void StaggeredKernels<Impl>::DhopNaive(StencilImpl &st, LebesgueOrder &lo,
|
void StaggeredKernels<Impl>::DhopNaive(StencilImpl &st,
|
||||||
DoubledGaugeField &U,
|
DoubledGaugeField &U,
|
||||||
const FermionField &in, FermionField &out, int dag, int interior,int exterior)
|
const FermionField &in, FermionField &out, int dag, int interior,int exterior)
|
||||||
{
|
{
|
||||||
|
@ -58,15 +58,9 @@ WilsonFermion5D<Impl>::WilsonFermion5D(GaugeField &_Umu,
|
|||||||
Umu(_FourDimGrid),
|
Umu(_FourDimGrid),
|
||||||
UmuEven(_FourDimRedBlackGrid),
|
UmuEven(_FourDimRedBlackGrid),
|
||||||
UmuOdd (_FourDimRedBlackGrid),
|
UmuOdd (_FourDimRedBlackGrid),
|
||||||
Lebesgue(_FourDimGrid),
|
|
||||||
LebesgueEvenOdd(_FourDimRedBlackGrid),
|
|
||||||
_tmp(&FiveDimRedBlackGrid),
|
_tmp(&FiveDimRedBlackGrid),
|
||||||
Dirichlet(0)
|
Dirichlet(0)
|
||||||
{
|
{
|
||||||
Stencil.lo = &Lebesgue;
|
|
||||||
StencilEven.lo = &LebesgueEvenOdd;
|
|
||||||
StencilOdd.lo = &LebesgueEvenOdd;
|
|
||||||
|
|
||||||
// some assertions
|
// some assertions
|
||||||
assert(FiveDimGrid._ndimension==5);
|
assert(FiveDimGrid._ndimension==5);
|
||||||
assert(FourDimGrid._ndimension==4);
|
assert(FourDimGrid._ndimension==4);
|
||||||
@ -305,19 +299,19 @@ void WilsonFermion5D<Impl>::DhopDerivOE(GaugeField &mat,
|
|||||||
}
|
}
|
||||||
|
|
||||||
template<class Impl>
|
template<class Impl>
|
||||||
void WilsonFermion5D<Impl>::DhopInternal(StencilImpl & st, LebesgueOrder &lo,
|
void WilsonFermion5D<Impl>::DhopInternal(StencilImpl & st,
|
||||||
DoubledGaugeField & U,
|
DoubledGaugeField & U,
|
||||||
const FermionField &in, FermionField &out,int dag)
|
const FermionField &in, FermionField &out,int dag)
|
||||||
{
|
{
|
||||||
if ( WilsonKernelsStatic::Comms == WilsonKernelsStatic::CommsAndCompute )
|
if ( WilsonKernelsStatic::Comms == WilsonKernelsStatic::CommsAndCompute )
|
||||||
DhopInternalOverlappedComms(st,lo,U,in,out,dag);
|
DhopInternalOverlappedComms(st,U,in,out,dag);
|
||||||
else
|
else
|
||||||
DhopInternalSerialComms(st,lo,U,in,out,dag);
|
DhopInternalSerialComms(st,U,in,out,dag);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
template<class Impl>
|
template<class Impl>
|
||||||
void WilsonFermion5D<Impl>::DhopInternalOverlappedComms(StencilImpl & st, LebesgueOrder &lo,
|
void WilsonFermion5D<Impl>::DhopInternalOverlappedComms(StencilImpl & st,
|
||||||
DoubledGaugeField & U,
|
DoubledGaugeField & U,
|
||||||
const FermionField &in, FermionField &out,int dag)
|
const FermionField &in, FermionField &out,int dag)
|
||||||
{
|
{
|
||||||
@ -331,10 +325,12 @@ void WilsonFermion5D<Impl>::DhopInternalOverlappedComms(StencilImpl & st, Lebesg
|
|||||||
// Start comms // Gather intranode and extra node differentiated??
|
// Start comms // Gather intranode and extra node differentiated??
|
||||||
/////////////////////////////
|
/////////////////////////////
|
||||||
{
|
{
|
||||||
|
std::cout << " WilsonFermion5D gather " <<std::endl;
|
||||||
GRID_TRACE("Gather");
|
GRID_TRACE("Gather");
|
||||||
st.HaloExchangeOptGather(in,compressor); // Put the barrier in the routine
|
st.HaloExchangeOptGather(in,compressor); // Put the barrier in the routine
|
||||||
}
|
}
|
||||||
|
|
||||||
|
std::cout << " WilsonFermion5D Communicate Begin " <<std::endl;
|
||||||
std::vector<std::vector<CommsRequest_t> > requests;
|
std::vector<std::vector<CommsRequest_t> > requests;
|
||||||
auto id=traceStart("Communicate overlapped");
|
auto id=traceStart("Communicate overlapped");
|
||||||
st.CommunicateBegin(requests);
|
st.CommunicateBegin(requests);
|
||||||
@ -343,6 +339,7 @@ void WilsonFermion5D<Impl>::DhopInternalOverlappedComms(StencilImpl & st, Lebesg
|
|||||||
// Overlap with comms
|
// Overlap with comms
|
||||||
/////////////////////////////
|
/////////////////////////////
|
||||||
{
|
{
|
||||||
|
std::cout << " WilsonFermion5D Comms merge " <<std::endl;
|
||||||
GRID_TRACE("MergeSHM");
|
GRID_TRACE("MergeSHM");
|
||||||
st.CommsMergeSHM(compressor);// Could do this inside parallel region overlapped with comms
|
st.CommsMergeSHM(compressor);// Could do this inside parallel region overlapped with comms
|
||||||
}
|
}
|
||||||
@ -350,6 +347,7 @@ void WilsonFermion5D<Impl>::DhopInternalOverlappedComms(StencilImpl & st, Lebesg
|
|||||||
/////////////////////////////
|
/////////////////////////////
|
||||||
// do the compute interior
|
// do the compute interior
|
||||||
/////////////////////////////
|
/////////////////////////////
|
||||||
|
std::cout << " WilsonFermion5D Interior " <<std::endl;
|
||||||
int Opt = WilsonKernelsStatic::Opt; // Why pass this. Kernels should know
|
int Opt = WilsonKernelsStatic::Opt; // Why pass this. Kernels should know
|
||||||
if (dag == DaggerYes) {
|
if (dag == DaggerYes) {
|
||||||
GRID_TRACE("DhopDagInterior");
|
GRID_TRACE("DhopDagInterior");
|
||||||
@ -362,6 +360,7 @@ void WilsonFermion5D<Impl>::DhopInternalOverlappedComms(StencilImpl & st, Lebesg
|
|||||||
/////////////////////////////
|
/////////////////////////////
|
||||||
// Complete comms
|
// Complete comms
|
||||||
/////////////////////////////
|
/////////////////////////////
|
||||||
|
std::cout << " WilsonFermion5D Comms Complete " <<std::endl;
|
||||||
st.CommunicateComplete(requests);
|
st.CommunicateComplete(requests);
|
||||||
traceStop(id);
|
traceStop(id);
|
||||||
|
|
||||||
@ -369,11 +368,13 @@ void WilsonFermion5D<Impl>::DhopInternalOverlappedComms(StencilImpl & st, Lebesg
|
|||||||
// do the compute exterior
|
// do the compute exterior
|
||||||
/////////////////////////////
|
/////////////////////////////
|
||||||
{
|
{
|
||||||
|
std::cout << " WilsonFermion5D Comms Merge " <<std::endl;
|
||||||
GRID_TRACE("Merge");
|
GRID_TRACE("Merge");
|
||||||
st.CommsMerge(compressor);
|
st.CommsMerge(compressor);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
std::cout << " WilsonFermion5D Exterior " <<std::endl;
|
||||||
if (dag == DaggerYes) {
|
if (dag == DaggerYes) {
|
||||||
GRID_TRACE("DhopDagExterior");
|
GRID_TRACE("DhopDagExterior");
|
||||||
Kernels::DhopDagKernel(Opt,st,U,st.CommBuf(),LLs,U.oSites(),in,out,0,1);
|
Kernels::DhopDagKernel(Opt,st,U,st.CommBuf(),LLs,U.oSites(),in,out,0,1);
|
||||||
@ -381,11 +382,12 @@ void WilsonFermion5D<Impl>::DhopInternalOverlappedComms(StencilImpl & st, Lebesg
|
|||||||
GRID_TRACE("DhopExterior");
|
GRID_TRACE("DhopExterior");
|
||||||
Kernels::DhopKernel (Opt,st,U,st.CommBuf(),LLs,U.oSites(),in,out,0,1);
|
Kernels::DhopKernel (Opt,st,U,st.CommBuf(),LLs,U.oSites(),in,out,0,1);
|
||||||
}
|
}
|
||||||
|
std::cout << " WilsonFermion5D Done " <<std::endl;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
template<class Impl>
|
template<class Impl>
|
||||||
void WilsonFermion5D<Impl>::DhopInternalSerialComms(StencilImpl & st, LebesgueOrder &lo,
|
void WilsonFermion5D<Impl>::DhopInternalSerialComms(StencilImpl & st,
|
||||||
DoubledGaugeField & U,
|
DoubledGaugeField & U,
|
||||||
const FermionField &in,
|
const FermionField &in,
|
||||||
FermionField &out,int dag)
|
FermionField &out,int dag)
|
||||||
@ -395,11 +397,13 @@ void WilsonFermion5D<Impl>::DhopInternalSerialComms(StencilImpl & st, LebesgueOr
|
|||||||
|
|
||||||
int LLs = in.Grid()->_rdimensions[0];
|
int LLs = in.Grid()->_rdimensions[0];
|
||||||
|
|
||||||
|
std::cout << " WilsonFermion5D Halo exch " <<std::endl;
|
||||||
{
|
{
|
||||||
GRID_TRACE("HaloExchange");
|
GRID_TRACE("HaloExchange");
|
||||||
st.HaloExchangeOpt(in,compressor);
|
st.HaloExchangeOpt(in,compressor);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
std::cout << " WilsonFermion5D Dhop " <<std::endl;
|
||||||
int Opt = WilsonKernelsStatic::Opt;
|
int Opt = WilsonKernelsStatic::Opt;
|
||||||
if (dag == DaggerYes) {
|
if (dag == DaggerYes) {
|
||||||
GRID_TRACE("DhopDag");
|
GRID_TRACE("DhopDag");
|
||||||
@ -408,6 +412,7 @@ void WilsonFermion5D<Impl>::DhopInternalSerialComms(StencilImpl & st, LebesgueOr
|
|||||||
GRID_TRACE("Dhop");
|
GRID_TRACE("Dhop");
|
||||||
Kernels::DhopKernel(Opt,st,U,st.CommBuf(),LLs,U.oSites(),in,out);
|
Kernels::DhopKernel(Opt,st,U,st.CommBuf(),LLs,U.oSites(),in,out);
|
||||||
}
|
}
|
||||||
|
std::cout << " WilsonFermion5D Done " <<std::endl;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
@ -420,7 +425,7 @@ void WilsonFermion5D<Impl>::DhopOE(const FermionField &in, FermionField &out,int
|
|||||||
assert(in.Checkerboard()==Even);
|
assert(in.Checkerboard()==Even);
|
||||||
out.Checkerboard() = Odd;
|
out.Checkerboard() = Odd;
|
||||||
|
|
||||||
DhopInternal(StencilEven,LebesgueEvenOdd,UmuOdd,in,out,dag);
|
DhopInternal(StencilEven,UmuOdd,in,out,dag);
|
||||||
}
|
}
|
||||||
template<class Impl>
|
template<class Impl>
|
||||||
void WilsonFermion5D<Impl>::DhopEO(const FermionField &in, FermionField &out,int dag)
|
void WilsonFermion5D<Impl>::DhopEO(const FermionField &in, FermionField &out,int dag)
|
||||||
@ -431,7 +436,7 @@ void WilsonFermion5D<Impl>::DhopEO(const FermionField &in, FermionField &out,int
|
|||||||
assert(in.Checkerboard()==Odd);
|
assert(in.Checkerboard()==Odd);
|
||||||
out.Checkerboard() = Even;
|
out.Checkerboard() = Even;
|
||||||
|
|
||||||
DhopInternal(StencilOdd,LebesgueEvenOdd,UmuEven,in,out,dag);
|
DhopInternal(StencilOdd,UmuEven,in,out,dag);
|
||||||
}
|
}
|
||||||
template<class Impl>
|
template<class Impl>
|
||||||
void WilsonFermion5D<Impl>::Dhop(const FermionField &in, FermionField &out,int dag)
|
void WilsonFermion5D<Impl>::Dhop(const FermionField &in, FermionField &out,int dag)
|
||||||
@ -441,7 +446,7 @@ void WilsonFermion5D<Impl>::Dhop(const FermionField &in, FermionField &out,int d
|
|||||||
|
|
||||||
out.Checkerboard() = in.Checkerboard();
|
out.Checkerboard() = in.Checkerboard();
|
||||||
|
|
||||||
DhopInternal(Stencil,Lebesgue,Umu,in,out,dag);
|
DhopInternal(Stencil,Umu,in,out,dag);
|
||||||
}
|
}
|
||||||
template<class Impl>
|
template<class Impl>
|
||||||
void WilsonFermion5D<Impl>::DW(const FermionField &in, FermionField &out,int dag)
|
void WilsonFermion5D<Impl>::DW(const FermionField &in, FermionField &out,int dag)
|
||||||
|
@ -52,17 +52,12 @@ WilsonFermion<Impl>::WilsonFermion(GaugeField &_Umu, GridCartesian &Fgrid,
|
|||||||
StencilEven(&Hgrid, npoint, Even, directions,displacements,p), // source is Even
|
StencilEven(&Hgrid, npoint, Even, directions,displacements,p), // source is Even
|
||||||
StencilOdd(&Hgrid, npoint, Odd, directions,displacements,p), // source is Odd
|
StencilOdd(&Hgrid, npoint, Odd, directions,displacements,p), // source is Odd
|
||||||
mass(_mass),
|
mass(_mass),
|
||||||
Lebesgue(_grid),
|
|
||||||
LebesgueEvenOdd(_cbgrid),
|
|
||||||
Umu(&Fgrid),
|
Umu(&Fgrid),
|
||||||
UmuEven(&Hgrid),
|
UmuEven(&Hgrid),
|
||||||
UmuOdd(&Hgrid),
|
UmuOdd(&Hgrid),
|
||||||
_tmp(&Hgrid),
|
_tmp(&Hgrid),
|
||||||
anisotropyCoeff(anis)
|
anisotropyCoeff(anis)
|
||||||
{
|
{
|
||||||
Stencil.lo = &Lebesgue;
|
|
||||||
StencilEven.lo = &LebesgueEvenOdd;
|
|
||||||
StencilOdd.lo = &LebesgueEvenOdd;
|
|
||||||
// Allocate the required comms buffer
|
// Allocate the required comms buffer
|
||||||
ImportGauge(_Umu);
|
ImportGauge(_Umu);
|
||||||
if (anisotropyCoeff.isAnisotropic){
|
if (anisotropyCoeff.isAnisotropic){
|
||||||
@ -314,7 +309,7 @@ void WilsonFermion<Impl>::Dhop(const FermionField &in, FermionField &out, int da
|
|||||||
|
|
||||||
out.Checkerboard() = in.Checkerboard();
|
out.Checkerboard() = in.Checkerboard();
|
||||||
|
|
||||||
DhopInternal(Stencil, Lebesgue, Umu, in, out, dag);
|
DhopInternal(Stencil, Umu, in, out, dag);
|
||||||
}
|
}
|
||||||
|
|
||||||
template <class Impl>
|
template <class Impl>
|
||||||
@ -326,7 +321,7 @@ void WilsonFermion<Impl>::DhopOE(const FermionField &in, FermionField &out, int
|
|||||||
assert(in.Checkerboard() == Even);
|
assert(in.Checkerboard() == Even);
|
||||||
out.Checkerboard() = Odd;
|
out.Checkerboard() = Odd;
|
||||||
|
|
||||||
DhopInternal(StencilEven, LebesgueEvenOdd, UmuOdd, in, out, dag);
|
DhopInternal(StencilEven, UmuOdd, in, out, dag);
|
||||||
}
|
}
|
||||||
|
|
||||||
template <class Impl>
|
template <class Impl>
|
||||||
@ -338,7 +333,7 @@ void WilsonFermion<Impl>::DhopEO(const FermionField &in, FermionField &out,int d
|
|||||||
assert(in.Checkerboard() == Odd);
|
assert(in.Checkerboard() == Odd);
|
||||||
out.Checkerboard() = Even;
|
out.Checkerboard() = Even;
|
||||||
|
|
||||||
DhopInternal(StencilOdd, LebesgueEvenOdd, UmuEven, in, out, dag);
|
DhopInternal(StencilOdd, UmuEven, in, out, dag);
|
||||||
}
|
}
|
||||||
|
|
||||||
template <class Impl>
|
template <class Impl>
|
||||||
@ -391,21 +386,21 @@ void WilsonFermion<Impl>::DhopDirCalc(const FermionField &in, FermionField &out,
|
|||||||
};
|
};
|
||||||
|
|
||||||
template <class Impl>
|
template <class Impl>
|
||||||
void WilsonFermion<Impl>::DhopInternal(StencilImpl &st, LebesgueOrder &lo,
|
void WilsonFermion<Impl>::DhopInternal(StencilImpl &st,
|
||||||
DoubledGaugeField &U,
|
DoubledGaugeField &U,
|
||||||
const FermionField &in,
|
const FermionField &in,
|
||||||
FermionField &out, int dag)
|
FermionField &out, int dag)
|
||||||
{
|
{
|
||||||
#ifdef GRID_OMP
|
#ifdef GRID_OMP
|
||||||
if ( WilsonKernelsStatic::Comms == WilsonKernelsStatic::CommsAndCompute )
|
if ( WilsonKernelsStatic::Comms == WilsonKernelsStatic::CommsAndCompute )
|
||||||
DhopInternalOverlappedComms(st,lo,U,in,out,dag);
|
DhopInternalOverlappedComms(st,U,in,out,dag);
|
||||||
else
|
else
|
||||||
#endif
|
#endif
|
||||||
DhopInternalSerial(st,lo,U,in,out,dag);
|
DhopInternalSerial(st,U,in,out,dag);
|
||||||
}
|
}
|
||||||
|
|
||||||
template <class Impl>
|
template <class Impl>
|
||||||
void WilsonFermion<Impl>::DhopInternalOverlappedComms(StencilImpl &st, LebesgueOrder &lo,
|
void WilsonFermion<Impl>::DhopInternalOverlappedComms(StencilImpl &st,
|
||||||
DoubledGaugeField &U,
|
DoubledGaugeField &U,
|
||||||
const FermionField &in,
|
const FermionField &in,
|
||||||
FermionField &out, int dag)
|
FermionField &out, int dag)
|
||||||
@ -474,10 +469,10 @@ void WilsonFermion<Impl>::DhopInternalOverlappedComms(StencilImpl &st, LebesgueO
|
|||||||
|
|
||||||
|
|
||||||
template <class Impl>
|
template <class Impl>
|
||||||
void WilsonFermion<Impl>::DhopInternalSerial(StencilImpl &st, LebesgueOrder &lo,
|
void WilsonFermion<Impl>::DhopInternalSerial(StencilImpl &st,
|
||||||
DoubledGaugeField &U,
|
DoubledGaugeField &U,
|
||||||
const FermionField &in,
|
const FermionField &in,
|
||||||
FermionField &out, int dag)
|
FermionField &out, int dag)
|
||||||
{
|
{
|
||||||
GRID_TRACE("DhopSerial");
|
GRID_TRACE("DhopSerial");
|
||||||
assert((dag == DaggerNo) || (dag == DaggerYes));
|
assert((dag == DaggerNo) || (dag == DaggerYes));
|
||||||
|
@ -40,11 +40,11 @@ Author: paboyle <paboyle@ph.ed.ac.uk>
|
|||||||
/// Switch off the 5d vectorised code optimisations
|
/// Switch off the 5d vectorised code optimisations
|
||||||
#undef DWFVEC5D
|
#undef DWFVEC5D
|
||||||
|
|
||||||
static Vector<vComplexF> signsF;
|
static std::vector<vComplexF> signsF;
|
||||||
|
|
||||||
template<typename vtype>
|
template<typename vtype>
|
||||||
int setupSigns(Vector<vtype>& signs ){
|
int setupSigns(std::vector<vtype>& signs ){
|
||||||
Vector<vtype> bother(2);
|
std::vector<vtype> bother(2);
|
||||||
signs = bother;
|
signs = bother;
|
||||||
vrsign(signs[0]);
|
vrsign(signs[0]);
|
||||||
visign(signs[1]);
|
visign(signs[1]);
|
||||||
@ -364,7 +364,7 @@ WilsonKernels<ZDomainWallVec5dImplF>::AsmDhopSiteDagExt(StencilView &st, Doubled
|
|||||||
|
|
||||||
#include <simd/Intel512double.h>
|
#include <simd/Intel512double.h>
|
||||||
|
|
||||||
static Vector<vComplexD> signsD;
|
static std::vector<vComplexD> signsD;
|
||||||
static int signInitD = setupSigns(signsD);
|
static int signInitD = setupSigns(signsD);
|
||||||
|
|
||||||
#define MAYBEPERM(A,perm) if (perm) { A ; }
|
#define MAYBEPERM(A,perm) if (perm) { A ; }
|
||||||
|
@ -434,7 +434,7 @@ void WilsonKernels<Impl>::DhopDirKernel( StencilImpl &st, DoubledGaugeField &U,S
|
|||||||
|
|
||||||
#define ASM_CALL(A) \
|
#define ASM_CALL(A) \
|
||||||
thread_for( sss, Nsite, { \
|
thread_for( sss, Nsite, { \
|
||||||
int ss = st.lo->Reorder(sss); \
|
int ss = sss; /*st.lo->Reorder(sss);*/ \
|
||||||
int sU = ss; \
|
int sU = ss; \
|
||||||
int sF = ss*Ls; \
|
int sF = ss*Ls; \
|
||||||
WilsonKernels<Impl>::A(st_v,U_v,buf,sF,sU,Ls,1,in_v,out_v); \
|
WilsonKernels<Impl>::A(st_v,U_v,buf,sF,sU,Ls,1,in_v,out_v); \
|
||||||
|
@ -40,7 +40,7 @@ public:
|
|||||||
U = Zero();
|
U = Zero();
|
||||||
LatticeColourMatrix tmp(Uin.Grid());
|
LatticeColourMatrix tmp(Uin.Grid());
|
||||||
|
|
||||||
Vector<typename SU<ncolour>::Matrix> ta(Dimension);
|
std::vector<typename SU<ncolour>::Matrix> ta(Dimension);
|
||||||
|
|
||||||
// Debug lines
|
// Debug lines
|
||||||
// LatticeMatrix uno(Uin.Grid());
|
// LatticeMatrix uno(Uin.Grid());
|
||||||
|
@ -43,7 +43,7 @@ public:
|
|||||||
U = Zero();
|
U = Zero();
|
||||||
LatticeColourMatrix tmp(Uin.Grid());
|
LatticeColourMatrix tmp(Uin.Grid());
|
||||||
|
|
||||||
Vector<typename GaugeGroup<ncolour,group_name>::Matrix> eij(Dimension);
|
std::vector<typename GaugeGroup<ncolour,group_name>::Matrix> eij(Dimension);
|
||||||
|
|
||||||
for (int a = 0; a < Dimension; a++)
|
for (int a = 0; a < Dimension; a++)
|
||||||
GaugeGroupTwoIndex<ncolour, S, group_name>::base(a, eij[a]);
|
GaugeGroupTwoIndex<ncolour, S, group_name>::base(a, eij[a]);
|
||||||
|
@ -158,12 +158,12 @@ void A2Autils<FImpl>::MesonField(TensorType &mat,
|
|||||||
int MFrvol = rd*Lblock*Rblock*Nmom;
|
int MFrvol = rd*Lblock*Rblock*Nmom;
|
||||||
int MFlvol = ld*Lblock*Rblock*Nmom;
|
int MFlvol = ld*Lblock*Rblock*Nmom;
|
||||||
|
|
||||||
Vector<SpinMatrix_v > lvSum(MFrvol);
|
std::vector<SpinMatrix_v > lvSum(MFrvol);
|
||||||
thread_for( r, MFrvol,{
|
thread_for( r, MFrvol,{
|
||||||
lvSum[r] = Zero();
|
lvSum[r] = Zero();
|
||||||
});
|
});
|
||||||
|
|
||||||
Vector<SpinMatrix_s > lsSum(MFlvol);
|
std::vector<SpinMatrix_s > lsSum(MFlvol);
|
||||||
thread_for(r,MFlvol,{
|
thread_for(r,MFlvol,{
|
||||||
lsSum[r]=scalar_type(0.0);
|
lsSum[r]=scalar_type(0.0);
|
||||||
});
|
});
|
||||||
@ -346,12 +346,12 @@ void A2Autils<FImpl>::PionFieldXX(Eigen::Tensor<ComplexD,3> &mat,
|
|||||||
int MFrvol = rd*Lblock*Rblock;
|
int MFrvol = rd*Lblock*Rblock;
|
||||||
int MFlvol = ld*Lblock*Rblock;
|
int MFlvol = ld*Lblock*Rblock;
|
||||||
|
|
||||||
Vector<vector_type > lvSum(MFrvol);
|
std::vector<vector_type > lvSum(MFrvol);
|
||||||
thread_for(r,MFrvol,{
|
thread_for(r,MFrvol,{
|
||||||
lvSum[r] = Zero();
|
lvSum[r] = Zero();
|
||||||
});
|
});
|
||||||
|
|
||||||
Vector<scalar_type > lsSum(MFlvol);
|
std::vector<scalar_type > lsSum(MFlvol);
|
||||||
thread_for(r,MFlvol,{
|
thread_for(r,MFlvol,{
|
||||||
lsSum[r]=scalar_type(0.0);
|
lsSum[r]=scalar_type(0.0);
|
||||||
});
|
});
|
||||||
@ -493,12 +493,12 @@ void A2Autils<FImpl>::PionFieldWVmom(Eigen::Tensor<ComplexD,4> &mat,
|
|||||||
int MFrvol = rd*Lblock*Rblock*Nmom;
|
int MFrvol = rd*Lblock*Rblock*Nmom;
|
||||||
int MFlvol = ld*Lblock*Rblock*Nmom;
|
int MFlvol = ld*Lblock*Rblock*Nmom;
|
||||||
|
|
||||||
Vector<vector_type > lvSum(MFrvol);
|
std::vector<vector_type > lvSum(MFrvol);
|
||||||
thread_for(r,MFrvol,{
|
thread_for(r,MFrvol,{
|
||||||
lvSum[r] = Zero();
|
lvSum[r] = Zero();
|
||||||
});
|
});
|
||||||
|
|
||||||
Vector<scalar_type > lsSum(MFlvol);
|
std::vector<scalar_type > lsSum(MFlvol);
|
||||||
thread_for(r,MFlvol,{
|
thread_for(r,MFlvol,{
|
||||||
lsSum[r]=scalar_type(0.0);
|
lsSum[r]=scalar_type(0.0);
|
||||||
});
|
});
|
||||||
@ -700,13 +700,13 @@ void A2Autils<FImpl>::AslashField(TensorType &mat,
|
|||||||
int MFrvol = rd*Lblock*Rblock*Nem;
|
int MFrvol = rd*Lblock*Rblock*Nem;
|
||||||
int MFlvol = ld*Lblock*Rblock*Nem;
|
int MFlvol = ld*Lblock*Rblock*Nem;
|
||||||
|
|
||||||
Vector<vector_type> lvSum(MFrvol);
|
std::vector<vector_type> lvSum(MFrvol);
|
||||||
thread_for(r,MFrvol,
|
thread_for(r,MFrvol,
|
||||||
{
|
{
|
||||||
lvSum[r] = Zero();
|
lvSum[r] = Zero();
|
||||||
});
|
});
|
||||||
|
|
||||||
Vector<scalar_type> lsSum(MFlvol);
|
std::vector<scalar_type> lsSum(MFlvol);
|
||||||
thread_for(r,MFlvol,
|
thread_for(r,MFlvol,
|
||||||
{
|
{
|
||||||
lsSum[r] = scalar_type(0.0);
|
lsSum[r] = scalar_type(0.0);
|
||||||
|
@ -971,7 +971,9 @@ void BaryonUtils<FImpl>::BaryonGamma3pt(
|
|||||||
autoView( vq_ti , q_ti , AcceleratorRead);
|
autoView( vq_ti , q_ti , AcceleratorRead);
|
||||||
autoView( vq_tf , q_tf , AcceleratorRead);
|
autoView( vq_tf , q_tf , AcceleratorRead);
|
||||||
|
|
||||||
Vector<mobj> my_Dq_spec{Dq_spec1,Dq_spec2};
|
deviceVector<mobj> my_Dq_spec(2);
|
||||||
|
acceleratorPut(my_Dq_spec[0],Dq_spec1);
|
||||||
|
acceleratorPut(my_Dq_spec[1],Dq_spec2);
|
||||||
mobj * Dq_spec_p = &my_Dq_spec[0];
|
mobj * Dq_spec_p = &my_Dq_spec[0];
|
||||||
|
|
||||||
if (group == 1) {
|
if (group == 1) {
|
||||||
@ -1300,7 +1302,8 @@ void BaryonUtils<FImpl>::SigmaToNucleonEye(const PropagatorField &qq_loop,
|
|||||||
autoView( vd_tf , qd_tf , AcceleratorRead);
|
autoView( vd_tf , qd_tf , AcceleratorRead);
|
||||||
autoView( vs_ti , qs_ti , AcceleratorRead);
|
autoView( vs_ti , qs_ti , AcceleratorRead);
|
||||||
|
|
||||||
Vector<mobj> my_Dq_spec{Du_spec};
|
deviceVector<mobj> my_Dq_spec(1);
|
||||||
|
acceleratorPut(my_Dq_spec[0],Du_spec);
|
||||||
mobj * Dq_spec_p = &my_Dq_spec[0];
|
mobj * Dq_spec_p = &my_Dq_spec[0];
|
||||||
|
|
||||||
if(op == "Q1"){
|
if(op == "Q1"){
|
||||||
@ -1353,7 +1356,8 @@ void BaryonUtils<FImpl>::SigmaToNucleonNonEye(const PropagatorField &qq_ti,
|
|||||||
autoView( vd_tf , qd_tf , AcceleratorRead );
|
autoView( vd_tf , qd_tf , AcceleratorRead );
|
||||||
autoView( vs_ti , qs_ti , AcceleratorRead );
|
autoView( vs_ti , qs_ti , AcceleratorRead );
|
||||||
|
|
||||||
Vector<mobj> my_Dq_spec{Du_spec};
|
deviceVector<mobj> my_Dq_spec(1);
|
||||||
|
acceleratorPut(my_Dq_spec[0],Du_spec);
|
||||||
mobj * Dq_spec_p = &my_Dq_spec[0];
|
mobj * Dq_spec_p = &my_Dq_spec[0];
|
||||||
|
|
||||||
if(op == "Q1"){
|
if(op == "Q1"){
|
||||||
@ -1544,7 +1548,9 @@ void BaryonUtils<FImpl>::XiToSigmaEye(const PropagatorField &qq_loop,
|
|||||||
autoView( vd_tf , qd_tf , AcceleratorRead);
|
autoView( vd_tf , qd_tf , AcceleratorRead);
|
||||||
autoView( vs_ti , qs_ti , AcceleratorRead);
|
autoView( vs_ti , qs_ti , AcceleratorRead);
|
||||||
|
|
||||||
Vector<mobj> my_Dq_spec{Dd_spec,Ds_spec};
|
deviceVector<mobj> my_Dq_spec(2);
|
||||||
|
acceleratorPut(my_Dq_spec[0],Dd_spec);
|
||||||
|
acceleratorPut(my_Dq_spec[0],Ds_spec);
|
||||||
mobj * Dq_spec_p = &my_Dq_spec[0];
|
mobj * Dq_spec_p = &my_Dq_spec[0];
|
||||||
|
|
||||||
if(op == "Q1"){
|
if(op == "Q1"){
|
||||||
|
@ -62,7 +62,7 @@ public:
|
|||||||
// returns i(T_Adj)^index necessary for the projectors
|
// returns i(T_Adj)^index necessary for the projectors
|
||||||
// see definitions above
|
// see definitions above
|
||||||
iAdjTa = Zero();
|
iAdjTa = Zero();
|
||||||
Vector<iSUnMatrix<cplx> > ta(ncolour * ncolour - 1);
|
iSUnMatrix<cplx> ta[ncolour * ncolour - 1];
|
||||||
iSUnMatrix<cplx> tmp;
|
iSUnMatrix<cplx> tmp;
|
||||||
|
|
||||||
// FIXME not very efficient to get all the generators everytime
|
// FIXME not very efficient to get all the generators everytime
|
||||||
|
@ -72,7 +72,7 @@ public:
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Resident in managed memory
|
// Resident in managed memory
|
||||||
Vector<GeneralStencilEntry> _entries;
|
deviceVector<GeneralStencilEntry> _entries;
|
||||||
|
|
||||||
GeneralLocalStencil(GridBase *grid, const std::vector<Coordinate> &shifts)
|
GeneralLocalStencil(GridBase *grid, const std::vector<Coordinate> &shifts)
|
||||||
{
|
{
|
||||||
@ -141,7 +141,7 @@ public:
|
|||||||
////////////////////////////////////////////////
|
////////////////////////////////////////////////
|
||||||
// Store in look up table
|
// Store in look up table
|
||||||
////////////////////////////////////////////////
|
////////////////////////////////////////////////
|
||||||
this->_entries[lex] = SE;
|
acceleratorPut(this->_entries[lex],SE);
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
@ -19,7 +19,7 @@ public:
|
|||||||
static int PartialCompressionFactor(GridBase *grid) {return 1;};
|
static int PartialCompressionFactor(GridBase *grid) {return 1;};
|
||||||
// Decompress is after merge so ok
|
// Decompress is after merge so ok
|
||||||
template<class vobj,class cobj,class compressor>
|
template<class vobj,class cobj,class compressor>
|
||||||
static void Gather_plane_simple (commVector<std::pair<int,int> >& table,
|
static void Gather_plane_simple (deviceVector<std::pair<int,int> >& table,
|
||||||
const Lattice<vobj> &rhs,
|
const Lattice<vobj> &rhs,
|
||||||
cobj *buffer,
|
cobj *buffer,
|
||||||
compressor &compress,
|
compressor &compress,
|
||||||
@ -35,7 +35,7 @@ public:
|
|||||||
rhs_v.ViewClose();
|
rhs_v.ViewClose();
|
||||||
}
|
}
|
||||||
template<class vobj,class cobj,class compressor>
|
template<class vobj,class cobj,class compressor>
|
||||||
static void Gather_plane_exchange(commVector<std::pair<int,int> >& table,const Lattice<vobj> &rhs,
|
static void Gather_plane_exchange(deviceVector<std::pair<int,int> >& table,const Lattice<vobj> &rhs,
|
||||||
std::vector<cobj *> pointers,int dimension,int plane,int cbmask,
|
std::vector<cobj *> pointers,int dimension,int plane,int cbmask,
|
||||||
compressor &compress,int type,int partial)
|
compressor &compress,int type,int partial)
|
||||||
{
|
{
|
||||||
@ -83,25 +83,6 @@ public:
|
|||||||
// Wilson compressor will add alternate policies for Dirichlet
|
// Wilson compressor will add alternate policies for Dirichlet
|
||||||
// and possibly partial Dirichlet for DWF
|
// and possibly partial Dirichlet for DWF
|
||||||
////////////////////////////////////
|
////////////////////////////////////
|
||||||
/*
|
|
||||||
class FaceGatherDirichlet
|
|
||||||
{
|
|
||||||
// If it's dirichlet we don't assemble comms buffers
|
|
||||||
//
|
|
||||||
// Rely on zeroes in gauge field to drive the correct result
|
|
||||||
// NAN propgagation: field will locally wrap, so fermion should NOT contain NAN and just permute
|
|
||||||
template<class vobj,class cobj,class compressor>
|
|
||||||
static void Gather_plane_simple (commVector<std::pair<int,int> >& table,const Lattice<vobj> &rhs,cobj *buffer,compressor &compress, int off,int so){};
|
|
||||||
template<class vobj,class cobj,class compressor>
|
|
||||||
static void Gather_plane_exchange(commVector<std::pair<int,int> >& table,const Lattice<vobj> &rhs,
|
|
||||||
Vector<cobj *> pointers,int dimension,int plane,int cbmask,
|
|
||||||
compressor &compress,int type) {}
|
|
||||||
template<class decompressor,class Merger>
|
|
||||||
static void Merge(decompressor decompress,Merge &mm) { }
|
|
||||||
template<class decompressor,class Decompression>
|
|
||||||
static void Decompress(decompressor decompress,Decompression &dd) {}
|
|
||||||
};
|
|
||||||
*/
|
|
||||||
|
|
||||||
template<class vobj,class FaceGather>
|
template<class vobj,class FaceGather>
|
||||||
class SimpleCompressorGather : public FaceGather {
|
class SimpleCompressorGather : public FaceGather {
|
||||||
|
@ -31,7 +31,6 @@
|
|||||||
#define STENCIL_MAX (16)
|
#define STENCIL_MAX (16)
|
||||||
|
|
||||||
#include <Grid/stencil/SimpleCompressor.h> // subdir aggregate
|
#include <Grid/stencil/SimpleCompressor.h> // subdir aggregate
|
||||||
#include <Grid/stencil/Lebesgue.h> // subdir aggregate
|
|
||||||
#include <Grid/stencil/GeneralLocalStencil.h>
|
#include <Grid/stencil/GeneralLocalStencil.h>
|
||||||
|
|
||||||
//////////////////////////////////////////////////////////////////////////////////////////
|
//////////////////////////////////////////////////////////////////////////////////////////
|
||||||
@ -256,7 +255,6 @@ protected:
|
|||||||
GridBase * _grid;
|
GridBase * _grid;
|
||||||
public:
|
public:
|
||||||
GridBase *Grid(void) const { return _grid; }
|
GridBase *Grid(void) const { return _grid; }
|
||||||
LebesgueOrder *lo;
|
|
||||||
|
|
||||||
////////////////////////////////////////////////////////////////////////
|
////////////////////////////////////////////////////////////////////////
|
||||||
// Needed to conveniently communicate gparity parameters into GPU memory
|
// Needed to conveniently communicate gparity parameters into GPU memory
|
||||||
@ -273,11 +271,11 @@ public:
|
|||||||
int face_table_computed;
|
int face_table_computed;
|
||||||
int partialDirichlet;
|
int partialDirichlet;
|
||||||
int fullDirichlet;
|
int fullDirichlet;
|
||||||
std::vector<commVector<std::pair<int,int> > > face_table ;
|
std::vector<deviceVector<std::pair<int,int> > > face_table ;
|
||||||
Vector<int> surface_list;
|
deviceVector<int> surface_list;
|
||||||
|
|
||||||
stencilVector<StencilEntry> _entries; // Resident in managed memory
|
std::vector<StencilEntry> _entries; // Resident in host memory
|
||||||
commVector<StencilEntry> _entries_device; // Resident in device memory
|
deviceVector<StencilEntry> _entries_device; // Resident in device memory
|
||||||
std::vector<Packet> Packets;
|
std::vector<Packet> Packets;
|
||||||
std::vector<Merge> Mergers;
|
std::vector<Merge> Mergers;
|
||||||
std::vector<Merge> MergersSHM;
|
std::vector<Merge> MergersSHM;
|
||||||
@ -370,7 +368,6 @@ public:
|
|||||||
// accelerator_barrier(); // All kernels should ALREADY be complete
|
// accelerator_barrier(); // All kernels should ALREADY be complete
|
||||||
// _grid->StencilBarrier(); // Everyone is here, so noone running slow and still using receive buffer
|
// _grid->StencilBarrier(); // Everyone is here, so noone running slow and still using receive buffer
|
||||||
// But the HaloGather had a barrier too.
|
// But the HaloGather had a barrier too.
|
||||||
#ifdef ACCELERATOR_AWARE_MPI
|
|
||||||
for(int i=0;i<Packets.size();i++){
|
for(int i=0;i<Packets.size();i++){
|
||||||
_grid->StencilSendToRecvFromBegin(MpiReqs,
|
_grid->StencilSendToRecvFromBegin(MpiReqs,
|
||||||
Packets[i].send_buf,
|
Packets[i].send_buf,
|
||||||
@ -379,23 +376,6 @@ public:
|
|||||||
Packets[i].from_rank,Packets[i].do_recv,
|
Packets[i].from_rank,Packets[i].do_recv,
|
||||||
Packets[i].xbytes,Packets[i].rbytes,i);
|
Packets[i].xbytes,Packets[i].rbytes,i);
|
||||||
}
|
}
|
||||||
#else
|
|
||||||
#warning "Using COPY VIA HOST BUFFERS IN STENCIL"
|
|
||||||
for(int i=0;i<Packets.size();i++){
|
|
||||||
// Introduce a host buffer with a cheap slab allocator and zero cost wipe all
|
|
||||||
Packets[i].host_send_buf = _grid->HostBufferMalloc(Packets[i].xbytes);
|
|
||||||
Packets[i].host_recv_buf = _grid->HostBufferMalloc(Packets[i].rbytes);
|
|
||||||
if ( Packets[i].do_send ) {
|
|
||||||
acceleratorCopyFromDevice(Packets[i].send_buf, Packets[i].host_send_buf,Packets[i].xbytes);
|
|
||||||
}
|
|
||||||
_grid->StencilSendToRecvFromBegin(MpiReqs,
|
|
||||||
Packets[i].host_send_buf,
|
|
||||||
Packets[i].to_rank,Packets[i].do_send,
|
|
||||||
Packets[i].host_recv_buf,
|
|
||||||
Packets[i].from_rank,Packets[i].do_recv,
|
|
||||||
Packets[i].xbytes,Packets[i].rbytes,i);
|
|
||||||
}
|
|
||||||
#endif
|
|
||||||
// Get comms started then run checksums
|
// Get comms started then run checksums
|
||||||
// Having this PRIOR to the dslash seems to make Sunspot work... (!)
|
// Having this PRIOR to the dslash seems to make Sunspot work... (!)
|
||||||
for(int i=0;i<Packets.size();i++){
|
for(int i=0;i<Packets.size();i++){
|
||||||
@ -413,15 +393,6 @@ public:
|
|||||||
// acceleratorCopySynchronise() is in the StencilSendToRecvFromComplete
|
// acceleratorCopySynchronise() is in the StencilSendToRecvFromComplete
|
||||||
// accelerator_barrier();
|
// accelerator_barrier();
|
||||||
_grid->StencilBarrier();
|
_grid->StencilBarrier();
|
||||||
#ifndef ACCELERATOR_AWARE_MPI
|
|
||||||
#warning "Using COPY VIA HOST BUFFERS IN STENCIL"
|
|
||||||
for(int i=0;i<Packets.size();i++){
|
|
||||||
if ( Packets[i].do_recv ) {
|
|
||||||
acceleratorCopyToDevice(Packets[i].host_recv_buf, Packets[i].recv_buf,Packets[i].rbytes);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
_grid->HostBufferFreeAll();
|
|
||||||
#endif
|
|
||||||
// run any checksums
|
// run any checksums
|
||||||
for(int i=0;i<Packets.size();i++){
|
for(int i=0;i<Packets.size();i++){
|
||||||
if ( Packets[i].do_recv )
|
if ( Packets[i].do_recv )
|
||||||
@ -668,7 +639,7 @@ public:
|
|||||||
for(int point=0;point<this->_npoints;point++){
|
for(int point=0;point<this->_npoints;point++){
|
||||||
this->same_node[point] = this->SameNode(point);
|
this->same_node[point] = this->SameNode(point);
|
||||||
}
|
}
|
||||||
|
int32_t surface_list_size=0;
|
||||||
for(int site = 0 ;site< vol4;site++){
|
for(int site = 0 ;site< vol4;site++){
|
||||||
int local = 1;
|
int local = 1;
|
||||||
for(int point=0;point<this->_npoints;point++){
|
for(int point=0;point<this->_npoints;point++){
|
||||||
@ -678,11 +649,28 @@ public:
|
|||||||
}
|
}
|
||||||
if(local == 0) {
|
if(local == 0) {
|
||||||
for(int s=0;s<Ls;s++){
|
for(int s=0;s<Ls;s++){
|
||||||
surface_list.push_back(site*Ls+s);
|
surface_list_size++;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
//std::cout << "BuildSurfaceList size is "<<surface_list.size()<<std::endl;
|
surface_list.resize(surface_list_size);
|
||||||
|
int32_t ss=0;
|
||||||
|
for(int site = 0 ;site< vol4;site++){
|
||||||
|
int local = 1;
|
||||||
|
for(int point=0;point<this->_npoints;point++){
|
||||||
|
if( (!this->GetNodeLocal(site*Ls,point)) && (!this->same_node[point]) ){
|
||||||
|
local = 0;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if(local == 0) {
|
||||||
|
for(int s=0;s<Ls;s++){
|
||||||
|
int idx=site*Ls+s;
|
||||||
|
acceleratorPut(surface_list[ss],idx);
|
||||||
|
ss++;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
std::cout << "BuildSurfaceList size is "<<surface_list.size()<<std::endl;
|
||||||
}
|
}
|
||||||
/// Introduce a block structure and switch off comms on boundaries
|
/// Introduce a block structure and switch off comms on boundaries
|
||||||
void DirichletBlock(const Coordinate &dirichlet_block)
|
void DirichletBlock(const Coordinate &dirichlet_block)
|
||||||
|
@ -207,10 +207,10 @@ cl::sycl::queue *theCopyAccelerator;
|
|||||||
void acceleratorInit(void)
|
void acceleratorInit(void)
|
||||||
{
|
{
|
||||||
int nDevices = 1;
|
int nDevices = 1;
|
||||||
cl::sycl::gpu_selector selector;
|
// cl::sycl::gpu_selector selector;
|
||||||
cl::sycl::device selectedDevice { selector };
|
// cl::sycl::device selectedDevice { selector };
|
||||||
theGridAccelerator = new sycl::queue (selectedDevice);
|
theGridAccelerator = new sycl::queue (sycl::gpu_selector_v);
|
||||||
theCopyAccelerator = new sycl::queue (selectedDevice);
|
theCopyAccelerator = new sycl::queue (sycl::gpu_selector_v);
|
||||||
// theCopyAccelerator = theGridAccelerator; // Should proceed concurrenlty anyway.
|
// theCopyAccelerator = theGridAccelerator; // Should proceed concurrenlty anyway.
|
||||||
|
|
||||||
#ifdef GRID_SYCL_LEVEL_ZERO_IPC
|
#ifdef GRID_SYCL_LEVEL_ZERO_IPC
|
||||||
|
@ -464,16 +464,12 @@ void Grid_init(int *argc,char ***argv)
|
|||||||
std::cout<<GridLogMessage<<std::endl;
|
std::cout<<GridLogMessage<<std::endl;
|
||||||
std::cout<<GridLogMessage<<"Performance:"<<std::endl;
|
std::cout<<GridLogMessage<<"Performance:"<<std::endl;
|
||||||
std::cout<<GridLogMessage<<std::endl;
|
std::cout<<GridLogMessage<<std::endl;
|
||||||
std::cout<<GridLogMessage<<" --comms-concurrent : Asynchronous MPI calls; several dirs at a time "<<std::endl;
|
|
||||||
std::cout<<GridLogMessage<<" --comms-sequential : Synchronous MPI calls; one dirs at a time "<<std::endl;
|
|
||||||
std::cout<<GridLogMessage<<" --comms-overlap : Overlap comms with compute "<<std::endl;
|
std::cout<<GridLogMessage<<" --comms-overlap : Overlap comms with compute "<<std::endl;
|
||||||
std::cout<<GridLogMessage<<std::endl;
|
std::cout<<GridLogMessage<<std::endl;
|
||||||
std::cout<<GridLogMessage<<" --dslash-generic: Wilson kernel for generic Nc"<<std::endl;
|
std::cout<<GridLogMessage<<" --dslash-generic: Wilson kernel for generic Nc"<<std::endl;
|
||||||
std::cout<<GridLogMessage<<" --dslash-unroll : Wilson kernel for Nc=3"<<std::endl;
|
std::cout<<GridLogMessage<<" --dslash-unroll : Wilson kernel for Nc=3"<<std::endl;
|
||||||
std::cout<<GridLogMessage<<" --dslash-asm : Wilson kernel for AVX512"<<std::endl;
|
std::cout<<GridLogMessage<<" --dslash-asm : Wilson kernel for AVX512"<<std::endl;
|
||||||
std::cout<<GridLogMessage<<std::endl;
|
std::cout<<GridLogMessage<<std::endl;
|
||||||
std::cout<<GridLogMessage<<" --lebesgue : Cache oblivious Lebesgue curve/Morton order/Z-graph stencil looping"<<std::endl;
|
|
||||||
std::cout<<GridLogMessage<<" --cacheblocking n.m.o.p : Hypercuboidal cache blocking"<<std::endl;
|
|
||||||
std::cout<<GridLogMessage<<std::endl;
|
std::cout<<GridLogMessage<<std::endl;
|
||||||
exit(EXIT_SUCCESS);
|
exit(EXIT_SUCCESS);
|
||||||
}
|
}
|
||||||
@ -501,28 +497,8 @@ void Grid_init(int *argc,char ***argv)
|
|||||||
WilsonKernelsStatic::Comms = WilsonKernelsStatic::CommsThenCompute;
|
WilsonKernelsStatic::Comms = WilsonKernelsStatic::CommsThenCompute;
|
||||||
StaggeredKernelsStatic::Comms = StaggeredKernelsStatic::CommsThenCompute;
|
StaggeredKernelsStatic::Comms = StaggeredKernelsStatic::CommsThenCompute;
|
||||||
}
|
}
|
||||||
if( GridCmdOptionExists(*argv,*argv+*argc,"--comms-concurrent") ){
|
|
||||||
CartesianCommunicator::SetCommunicatorPolicy(CartesianCommunicator::CommunicatorPolicyConcurrent);
|
|
||||||
}
|
|
||||||
if( GridCmdOptionExists(*argv,*argv+*argc,"--comms-sequential") ){
|
|
||||||
CartesianCommunicator::SetCommunicatorPolicy(CartesianCommunicator::CommunicatorPolicySequential);
|
|
||||||
}
|
|
||||||
|
|
||||||
if( GridCmdOptionExists(*argv,*argv+*argc,"--lebesgue") ){
|
|
||||||
LebesgueOrder::UseLebesgueOrder=1;
|
|
||||||
}
|
|
||||||
CartesianCommunicator::nCommThreads = 1;
|
CartesianCommunicator::nCommThreads = 1;
|
||||||
#ifdef GRID_COMMS_THREADS
|
|
||||||
if( GridCmdOptionExists(*argv,*argv+*argc,"--comms-threads") ){
|
|
||||||
arg= GridCmdOptionPayload(*argv,*argv+*argc,"--comms-threads");
|
|
||||||
GridCmdOptionInt(arg,CartesianCommunicator::nCommThreads);
|
|
||||||
assert(CartesianCommunicator::nCommThreads > 0);
|
|
||||||
}
|
|
||||||
#endif
|
|
||||||
if( GridCmdOptionExists(*argv,*argv+*argc,"--cacheblocking") ){
|
|
||||||
arg= GridCmdOptionPayload(*argv,*argv+*argc,"--cacheblocking");
|
|
||||||
GridCmdOptionIntVector(arg,LebesgueOrder::Block);
|
|
||||||
}
|
|
||||||
if( GridCmdOptionExists(*argv,*argv+*argc,"--notimestamp") ){
|
if( GridCmdOptionExists(*argv,*argv+*argc,"--notimestamp") ){
|
||||||
GridLogTimestamp(0);
|
GridLogTimestamp(0);
|
||||||
} else {
|
} else {
|
||||||
|
@ -644,11 +644,6 @@ int main (int argc, char ** argv)
|
|||||||
Grid_init(&argc,&argv);
|
Grid_init(&argc,&argv);
|
||||||
|
|
||||||
CartesianCommunicator::SetCommunicatorPolicy(CartesianCommunicator::CommunicatorPolicySequential);
|
CartesianCommunicator::SetCommunicatorPolicy(CartesianCommunicator::CommunicatorPolicySequential);
|
||||||
#ifdef KNL
|
|
||||||
LebesgueOrder::Block = std::vector<int>({8,2,2,2});
|
|
||||||
#else
|
|
||||||
LebesgueOrder::Block = std::vector<int>({2,2,2,2});
|
|
||||||
#endif
|
|
||||||
Benchmark::Decomposition();
|
Benchmark::Decomposition();
|
||||||
|
|
||||||
int do_su4=1;
|
int do_su4=1;
|
||||||
|
@ -70,7 +70,7 @@ int main (int argc, char ** argv)
|
|||||||
pRNG.SeedFixedIntegers(std::vector<int>({56,17,89,101}));
|
pRNG.SeedFixedIntegers(std::vector<int>({56,17,89,101}));
|
||||||
|
|
||||||
std::vector<double> stop(threads);
|
std::vector<double> stop(threads);
|
||||||
Vector<Vec> sum(threads);
|
std::vector<Vec> sum(threads);
|
||||||
|
|
||||||
std::vector<LatticeVec> x(threads,&Grid);
|
std::vector<LatticeVec> x(threads,&Grid);
|
||||||
for(int t=0;t<threads;t++){
|
for(int t=0;t<threads;t++){
|
||||||
|
@ -78,9 +78,9 @@ int main (int argc, char ** argv)
|
|||||||
double t0,t1;
|
double t0,t1;
|
||||||
|
|
||||||
typedef typename DomainWallFermionD::Coeff_t Coeff_t;
|
typedef typename DomainWallFermionD::Coeff_t Coeff_t;
|
||||||
Vector<Coeff_t> diag = Dw.bs;
|
std::vector<Coeff_t> diag = Dw.bs;
|
||||||
Vector<Coeff_t> upper= Dw.cs;
|
std::vector<Coeff_t> upper= Dw.cs;
|
||||||
Vector<Coeff_t> lower= Dw.cs;
|
std::vector<Coeff_t> lower= Dw.cs;
|
||||||
upper[Ls-1]=-Dw.mass_minus*upper[Ls-1];
|
upper[Ls-1]=-Dw.mass_minus*upper[Ls-1];
|
||||||
lower[0] =-Dw.mass_plus*lower[0];
|
lower[0] =-Dw.mass_plus*lower[0];
|
||||||
|
|
||||||
|
@ -861,7 +861,7 @@ int main (int argc, char ** argv)
|
|||||||
}
|
}
|
||||||
|
|
||||||
CartesianCommunicator::SetCommunicatorPolicy(CartesianCommunicator::CommunicatorPolicySequential);
|
CartesianCommunicator::SetCommunicatorPolicy(CartesianCommunicator::CommunicatorPolicySequential);
|
||||||
LebesgueOrder::Block = std::vector<int>({2,2,2,2});
|
// LebesgueOrder::Block = std::vector<int>({2,2,2,2});
|
||||||
|
|
||||||
Benchmark::Decomposition();
|
Benchmark::Decomposition();
|
||||||
|
|
||||||
|
22
configure.ac
22
configure.ac
@ -225,18 +225,6 @@ case ${ac_SFW_FP16} in
|
|||||||
AC_MSG_ERROR(["SFW FP16 option not supported ${ac_SFW_FP16}"]);;
|
AC_MSG_ERROR(["SFW FP16 option not supported ${ac_SFW_FP16}"]);;
|
||||||
esac
|
esac
|
||||||
|
|
||||||
############### Default to accelerator cshift, but revert to host if UCX is buggy or other reasons
|
|
||||||
AC_ARG_ENABLE([accelerator-aware-mpi],
|
|
||||||
[AS_HELP_STRING([--enable-accelerator-aware-mpi=yes|no],[run mpi transfers from device])],
|
|
||||||
[ac_ACCELERATOR_AWARE_MPI=${enable_accelerator_aware_mpi}], [ac_ACCELERATOR_AWARE_MPI=yes])
|
|
||||||
|
|
||||||
case ${ac_ACCELERATOR_AWARE_MPI} in
|
|
||||||
yes)
|
|
||||||
AC_DEFINE([ACCELERATOR_CSHIFT],[1],[ Cshift runs on host])
|
|
||||||
AC_DEFINE([ACCELERATOR_AWARE_MPI],[1],[ Stencil can use device pointers]);;
|
|
||||||
*);;
|
|
||||||
esac
|
|
||||||
|
|
||||||
|
|
||||||
############### SYCL/CUDA/HIP/none
|
############### SYCL/CUDA/HIP/none
|
||||||
AC_ARG_ENABLE([accelerator],
|
AC_ARG_ENABLE([accelerator],
|
||||||
@ -664,16 +652,6 @@ case ${ac_SHM_FAST_PATH} in
|
|||||||
*) ;;
|
*) ;;
|
||||||
esac
|
esac
|
||||||
|
|
||||||
############### communication type selection
|
|
||||||
AC_ARG_ENABLE([comms-threads],[AS_HELP_STRING([--enable-comms-threads | --disable-comms-threads],[Use multiple threads in MPI calls])],[ac_COMMS_THREADS=${enable_comms_threads}],[ac_COMMS_THREADS=yes])
|
|
||||||
|
|
||||||
case ${ac_COMMS_THREADS} in
|
|
||||||
yes)
|
|
||||||
AC_DEFINE([GRID_COMMS_THREADING],[1],[GRID_COMMS_NONE] )
|
|
||||||
;;
|
|
||||||
*) ;;
|
|
||||||
esac
|
|
||||||
|
|
||||||
############### communication type selection
|
############### communication type selection
|
||||||
AC_ARG_ENABLE([comms],[AS_HELP_STRING([--enable-comms=none|mpi|mpi-auto],[Select communications])],[ac_COMMS=${enable_comms}],[ac_COMMS=none])
|
AC_ARG_ENABLE([comms],[AS_HELP_STRING([--enable-comms=none|mpi|mpi-auto],[Select communications])],[ac_COMMS=${enable_comms}],[ac_COMMS=none])
|
||||||
|
|
||||||
|
@ -1,6 +1,6 @@
|
|||||||
#!/bin/bash
|
#!/bin/bash
|
||||||
|
|
||||||
#PBS -q debug
|
#PBS -q EarlyAppAccess
|
||||||
#PBS -l select=1
|
#PBS -l select=1
|
||||||
#PBS -l walltime=00:20:00
|
#PBS -l walltime=00:20:00
|
||||||
#PBS -A LatticeQCD_aesp_CNDA
|
#PBS -A LatticeQCD_aesp_CNDA
|
||||||
@ -44,7 +44,7 @@ CMD="mpiexec -np 1 -ppn 1 -envall \
|
|||||||
./gpu_tile_compact.sh \
|
./gpu_tile_compact.sh \
|
||||||
./Benchmark_dwf_fp32 --mpi 1.1.1.1 --grid 16.32.32.32 \
|
./Benchmark_dwf_fp32 --mpi 1.1.1.1 --grid 16.32.32.32 \
|
||||||
--shm-mpi 0 --shm 2048 --device-mem 32000 --accelerator-threads 32 "
|
--shm-mpi 0 --shm 2048 --device-mem 32000 --accelerator-threads 32 "
|
||||||
#$CMD | tee 1tile.dwf
|
$CMD | tee 1tile.dwf
|
||||||
|
|
||||||
CMD="mpiexec -np 12 -ppn 12 -envall \
|
CMD="mpiexec -np 12 -ppn 12 -envall \
|
||||||
./gpu_tile_compact.sh \
|
./gpu_tile_compact.sh \
|
||||||
|
@ -1,6 +1,6 @@
|
|||||||
#!/bin/bash
|
#!/bin/bash
|
||||||
|
|
||||||
#PBS -q workq
|
#PBS -q EarlyAppAccess
|
||||||
#PBS -l select=2
|
#PBS -l select=2
|
||||||
#PBS -l walltime=00:20:00
|
#PBS -l walltime=00:20:00
|
||||||
#PBS -A LatticeQCD_aesp_CNDA
|
#PBS -A LatticeQCD_aesp_CNDA
|
||||||
@ -43,13 +43,13 @@ $CMD | tee 2node.comms
|
|||||||
CMD="mpiexec -np 24 -ppn 12 -envall \
|
CMD="mpiexec -np 24 -ppn 12 -envall \
|
||||||
./gpu_tile_compact.sh \
|
./gpu_tile_compact.sh \
|
||||||
./Benchmark_dwf_fp32 --mpi 2.2.2.3 --grid 32.32.64.48 \
|
./Benchmark_dwf_fp32 --mpi 2.2.2.3 --grid 32.32.64.48 \
|
||||||
--shm-mpi 0 --shm 2048 --device-mem 32000 --accelerator-threads 32 --comms-overlap"
|
--shm-mpi 1 --shm 2048 --device-mem 32000 --accelerator-threads 32 "
|
||||||
$CMD | tee 2node.32.32.64.48.dwf
|
$CMD | tee 2node.32.32.64.48.dwf
|
||||||
|
|
||||||
|
|
||||||
CMD="mpiexec -np 24 -ppn 12 -envall \
|
CMD="mpiexec -np 24 -ppn 12 -envall \
|
||||||
./gpu_tile_compact.sh \
|
./gpu_tile_compact.sh \
|
||||||
./Benchmark_dwf_fp32 --mpi 2.2.2.3 --grid 64.64.64.96 \
|
./Benchmark_dwf_fp32 --mpi 2.2.2.3 --grid 64.64.64.96 \
|
||||||
--shm-mpi 0 --shm 2048 --device-mem 32000 --accelerator-threads 32 --comms-overlap"
|
--shm-mpi 1 --shm 2048 --device-mem 32000 --accelerator-threads 32 "
|
||||||
$CMD | tee 2node.64.64.64.96.dwf
|
$CMD | tee 2node.64.64.64.96.dwf
|
||||||
|
|
||||||
|
@ -1,40 +1,12 @@
|
|||||||
|
module load oneapi/release/2023.12.15.001
|
||||||
|
#module load intel_compute_runtime/release/821.35
|
||||||
source ~/spack/share/spack/setup-env.sh
|
source ~/spack/share/spack/setup-env.sh
|
||||||
spack load c-lime
|
spack load c-lime
|
||||||
|
spack load openssl
|
||||||
export CLIME=`spack find --paths c-lime | grep ^c-lime | awk '{print $2}' `
|
export CLIME=`spack find --paths c-lime | grep ^c-lime | awk '{print $2}' `
|
||||||
#spack load libefence
|
|
||||||
#export EFENCE=`spack find --paths libefence | grep ^libefence | awk '{print $2}' `
|
|
||||||
#export LD_LIBRARY_PATH=${EFENCE}/lib:$LD_LIBRARY_PATH
|
|
||||||
#spack load gperftools
|
|
||||||
export TCMALLOC=/home/paboyle/gperftools/install
|
|
||||||
export LD_LIBRARY_PATH=${TCMALLOC}/lib:$LD_LIBRARY_PATH
|
|
||||||
export INTELGT_AUTO_ATTACH_DISABLE=1
|
|
||||||
|
|
||||||
#export ONEAPI_DEVICE_SELECTOR=level_zero:0.0
|
|
||||||
#module load oneapi/release/2023.12.15.001
|
|
||||||
#module use /soft/modulefiles
|
|
||||||
#module load intel_compute_runtime/release/agama-devel-682.22
|
|
||||||
|
|
||||||
#export FI_CXI_DEFAULT_CQ_SIZE=131072
|
|
||||||
#export FI_CXI_CQ_FILL_PERCENT=20
|
|
||||||
#export SYCL_PROGRAM_COMPILE_OPTIONS="-ze-opt-large-register-file"
|
|
||||||
#export SYCL_PROGRAM_COMPILE_OPTIONS="-ze-intel-enable-auto-large-GRF-mode"
|
|
||||||
|
|
||||||
#
|
|
||||||
# -ftarget-register-alloc-mode=pvc:default
|
|
||||||
# -ftarget-register-alloc-mode=pvc:small
|
|
||||||
# -ftarget-register-alloc-mode=pvc:large
|
|
||||||
# -ftarget-register-alloc-mode=pvc:auto
|
|
||||||
#export MPIR_CVAR_CH4_OFI_ENABLE_HMEM=1
|
|
||||||
|
|
||||||
export HTTP_PROXY=http://proxy.alcf.anl.gov:3128
|
export HTTP_PROXY=http://proxy.alcf.anl.gov:3128
|
||||||
export HTTPS_PROXY=http://proxy.alcf.anl.gov:3128
|
export HTTPS_PROXY=http://proxy.alcf.anl.gov:3128
|
||||||
export http_proxy=http://proxy.alcf.anl.gov:3128
|
export http_proxy=http://proxy.alcf.anl.gov:3128
|
||||||
export https_proxy=http://proxy.alcf.anl.gov:3128
|
export https_proxy=http://proxy.alcf.anl.gov:3128
|
||||||
git config --global http.proxy http://proxy.alcf.anl.gov:3128
|
git config --global http.proxy http://proxy.alcf.anl.gov:3128
|
||||||
|
|
||||||
#source ~/spack/share/spack/setup-env.sh
|
|
||||||
#spack load gperftools
|
|
||||||
#export TCMALLOC=`spack find --paths gperftools | grep ^gperftools | awk '{print $2}' `
|
|
||||||
#export LD_LIBRARY_PATH=${TCMALLOC}/lib:$LD_LIBRARY_PATH
|
|
||||||
|
|
||||||
export SYCL_PROGRAM_COMPILE_OPTIONS="-ze-opt-large-register-file"
|
export SYCL_PROGRAM_COMPILE_OPTIONS="-ze-opt-large-register-file"
|
||||||
|
@ -1,6 +1,6 @@
|
|||||||
#!/bin/bash
|
#!/bin/bash
|
||||||
|
|
||||||
#PBS -l select=16
|
#PBS -l select=32
|
||||||
#PBS -q EarlyAppAccess
|
#PBS -q EarlyAppAccess
|
||||||
#PBS -A LatticeQCD_aesp_CNDA
|
#PBS -A LatticeQCD_aesp_CNDA
|
||||||
#PBS -l walltime=02:00:00
|
#PBS -l walltime=02:00:00
|
||||||
@ -15,7 +15,7 @@
|
|||||||
|
|
||||||
# 56 cores / 6 threads ~9
|
# 56 cores / 6 threads ~9
|
||||||
export OMP_NUM_THREADS=6
|
export OMP_NUM_THREADS=6
|
||||||
export MPIR_CVAR_CH4_OFI_ENABLE_GPU_PIPELINE=1
|
#export MPIR_CVAR_CH4_OFI_ENABLE_GPU_PIPELINE=1
|
||||||
#export MPIR_CVAR_CH4_OFI_GPU_PIPELINE_D2H_ENGINE_TYPE=0
|
#export MPIR_CVAR_CH4_OFI_GPU_PIPELINE_D2H_ENGINE_TYPE=0
|
||||||
#export MPIR_CVAR_CH4_OFI_GPU_PIPELINE_H2D_ENGINE_TYPE=0
|
#export MPIR_CVAR_CH4_OFI_GPU_PIPELINE_H2D_ENGINE_TYPE=0
|
||||||
#export MPIR_CVAR_CH4_OFI_GPU_PIPELINE_BUFFER_SZ=1048576
|
#export MPIR_CVAR_CH4_OFI_GPU_PIPELINE_BUFFER_SZ=1048576
|
||||||
@ -24,14 +24,14 @@ export MPIR_CVAR_CH4_OFI_ENABLE_GPU_PIPELINE=1
|
|||||||
#export MPIR_CVAR_CH4_OFI_GPU_PIPELINE_MAX_NUM_BUFFERS=16
|
#export MPIR_CVAR_CH4_OFI_GPU_PIPELINE_MAX_NUM_BUFFERS=16
|
||||||
#export MPIR_CVAR_GPU_USE_IMMEDIATE_COMMAND_LIST=1
|
#export MPIR_CVAR_GPU_USE_IMMEDIATE_COMMAND_LIST=1
|
||||||
|
|
||||||
export SYCL_PI_LEVEL_ZERO_USE_IMMEDIATE_COMMANDLISTS=1
|
#export SYCL_PI_LEVEL_ZERO_USE_IMMEDIATE_COMMANDLISTS=1
|
||||||
export SYCL_PI_LEVEL_ZERO_USE_COPY_ENGINE=1
|
export SYCL_PI_LEVEL_ZERO_USE_COPY_ENGINE=1
|
||||||
export SYCL_PI_LEVEL_ZERO_USE_COPY_ENGINE_FOR_D2D_COPY=1
|
export SYCL_PI_LEVEL_ZERO_USE_COPY_ENGINE_FOR_D2D_COPY=1
|
||||||
export SYCL_PROGRAM_COMPILE_OPTIONS="-ze-opt-large-register-file"
|
export SYCL_PROGRAM_COMPILE_OPTIONS="-ze-opt-large-register-file"
|
||||||
|
|
||||||
export GRID_PRINT_ENTIRE_LOG=0
|
export GRID_PRINT_ENTIRE_LOG=0
|
||||||
export GRID_CHECKSUM_RECV_BUF=0
|
export GRID_CHECKSUM_RECV_BUF=1
|
||||||
export GRID_CHECKSUM_SEND_BUF=0
|
export GRID_CHECKSUM_SEND_BUF=1
|
||||||
|
|
||||||
export MPICH_OFI_NIC_POLICY=GPU
|
export MPICH_OFI_NIC_POLICY=GPU
|
||||||
|
|
||||||
@ -51,10 +51,10 @@ cd $DIR
|
|||||||
|
|
||||||
cp $PBS_NODEFILE nodefile
|
cp $PBS_NODEFILE nodefile
|
||||||
|
|
||||||
CMD="mpiexec -np 192 -ppn 12 -envall --hostfile nodefile \
|
CMD="mpiexec -np 384 -ppn 12 -envall --hostfile nodefile \
|
||||||
../gpu_tile_compact.sh \
|
../gpu_tile_compact.sh \
|
||||||
../Test_dwf_mixedcg_prec --mpi 4.4.4.3 --grid 128.128.128.96 \
|
../Test_dwf_mixedcg_prec --mpi 4.4.4.6 --grid 128.128.128.96 \
|
||||||
--shm-mpi 0 --shm 4096 --device-mem 32000 --accelerator-threads 32 --seconds 6000 --debug-stdout --log Message --comms-overlap"
|
--shm-mpi 1 --comms-overlap --shm 4096 --device-mem 32000 --accelerator-threads 32 --seconds 6000 --debug-stdout --log Message --debug-signals"
|
||||||
|
|
||||||
echo $CMD > command-line
|
echo $CMD > command-line
|
||||||
env > environment
|
env > environment
|
||||||
|
@ -88,6 +88,7 @@ int main (int argc, char ** argv)
|
|||||||
Ctilde=C;
|
Ctilde=C;
|
||||||
std::cout<<" Benchmarking FFT of LatticeComplex "<<std::endl;
|
std::cout<<" Benchmarking FFT of LatticeComplex "<<std::endl;
|
||||||
theFFT.FFT_dim(Ctilde,Ctilde,0,FFT::forward); std::cout << theFFT.MFlops()<<" Mflops "<<std::endl;
|
theFFT.FFT_dim(Ctilde,Ctilde,0,FFT::forward); std::cout << theFFT.MFlops()<<" Mflops "<<std::endl;
|
||||||
|
std::cout<<" FFT done "<<std::endl;
|
||||||
theFFT.FFT_dim(Ctilde,Ctilde,1,FFT::forward); std::cout << theFFT.MFlops()<<" Mflops "<<std::endl;
|
theFFT.FFT_dim(Ctilde,Ctilde,1,FFT::forward); std::cout << theFFT.MFlops()<<" Mflops "<<std::endl;
|
||||||
theFFT.FFT_dim(Ctilde,Ctilde,2,FFT::forward); std::cout << theFFT.MFlops()<<" Mflops "<<std::endl;
|
theFFT.FFT_dim(Ctilde,Ctilde,2,FFT::forward); std::cout << theFFT.MFlops()<<" Mflops "<<std::endl;
|
||||||
theFFT.FFT_dim(Ctilde,Ctilde,3,FFT::forward); std::cout << theFFT.MFlops()<<" Mflops "<<std::endl;
|
theFFT.FFT_dim(Ctilde,Ctilde,3,FFT::forward); std::cout << theFFT.MFlops()<<" Mflops "<<std::endl;
|
||||||
|
@ -54,6 +54,7 @@ static_assert(same_vComplex == 1, "Dirac Operators must have same underlying SIM
|
|||||||
|
|
||||||
int main (int argc, char ** argv)
|
int main (int argc, char ** argv)
|
||||||
{
|
{
|
||||||
|
#ifdef ENABLE_GPARITY
|
||||||
int nu = 0;
|
int nu = 0;
|
||||||
int tbc_aprd = 0; //use antiperiodic BCs in the time direction?
|
int tbc_aprd = 0; //use antiperiodic BCs in the time direction?
|
||||||
|
|
||||||
@ -325,4 +326,5 @@ int main (int argc, char ** argv)
|
|||||||
|
|
||||||
|
|
||||||
Grid_finalize();
|
Grid_finalize();
|
||||||
|
#endif
|
||||||
}
|
}
|
||||||
|
@ -30,6 +30,7 @@ See the full license in the file "LICENSE" in the top level distribution directo
|
|||||||
|
|
||||||
using namespace Grid;
|
using namespace Grid;
|
||||||
|
|
||||||
|
#ifdef ENABLE_GPARITY
|
||||||
static constexpr double tolerance = 1.0e-6;
|
static constexpr double tolerance = 1.0e-6;
|
||||||
static std::array<GparityFlavourMatrix, GparityFlavour::nSigma> testAlgebra;
|
static std::array<GparityFlavourMatrix, GparityFlavour::nSigma> testAlgebra;
|
||||||
|
|
||||||
@ -148,11 +149,12 @@ void checkSigma(const GparityFlavour::Algebra a, GridSerialRNG &rng)
|
|||||||
test(m*g, m*testg);
|
test(m*g, m*testg);
|
||||||
std::cout << std::endl;
|
std::cout << std::endl;
|
||||||
}
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
int main(int argc, char *argv[])
|
int main(int argc, char *argv[])
|
||||||
{
|
{
|
||||||
Grid_init(&argc,&argv);
|
Grid_init(&argc,&argv);
|
||||||
|
#ifdef ENABLE_GPARITY
|
||||||
Coordinate latt_size = GridDefaultLatt();
|
Coordinate latt_size = GridDefaultLatt();
|
||||||
Coordinate simd_layout = GridDefaultSimd(4,vComplex::Nsimd());
|
Coordinate simd_layout = GridDefaultSimd(4,vComplex::Nsimd());
|
||||||
Coordinate mpi_layout = GridDefaultMpi();
|
Coordinate mpi_layout = GridDefaultMpi();
|
||||||
@ -170,7 +172,7 @@ int main(int argc, char *argv[])
|
|||||||
checkSigma(i, sRNG);
|
checkSigma(i, sRNG);
|
||||||
}
|
}
|
||||||
std::cout << GridLogMessage << std::endl;
|
std::cout << GridLogMessage << std::endl;
|
||||||
|
#endif
|
||||||
Grid_finalize();
|
Grid_finalize();
|
||||||
|
|
||||||
return EXIT_SUCCESS;
|
return EXIT_SUCCESS;
|
||||||
|
@ -35,7 +35,7 @@ using namespace Grid;
|
|||||||
int main (int argc, char ** argv)
|
int main (int argc, char ** argv)
|
||||||
{
|
{
|
||||||
Grid_init(&argc,&argv);
|
Grid_init(&argc,&argv);
|
||||||
|
#ifdef ENABLE_GPARITY
|
||||||
Coordinate latt_size = GridDefaultLatt();
|
Coordinate latt_size = GridDefaultLatt();
|
||||||
Coordinate simd_layout = GridDefaultSimd(Nd,vComplex::Nsimd());
|
Coordinate simd_layout = GridDefaultSimd(Nd,vComplex::Nsimd());
|
||||||
Coordinate mpi_layout = GridDefaultMpi();
|
Coordinate mpi_layout = GridDefaultMpi();
|
||||||
@ -216,6 +216,6 @@ int main (int argc, char ** argv)
|
|||||||
|
|
||||||
std::cout<<GridLogMessage <<"pDce - conj(cDpo) "<< pDco-conj(cDpo) <<std::endl;
|
std::cout<<GridLogMessage <<"pDce - conj(cDpo) "<< pDco-conj(cDpo) <<std::endl;
|
||||||
std::cout<<GridLogMessage <<"pDco - conj(cDpe) "<< pDce-conj(cDpe) <<std::endl;
|
std::cout<<GridLogMessage <<"pDco - conj(cDpe) "<< pDce-conj(cDpe) <<std::endl;
|
||||||
|
#endif
|
||||||
Grid_finalize();
|
Grid_finalize();
|
||||||
}
|
}
|
||||||
|
@ -93,7 +93,7 @@ void MemoryTest(GridCartesian * FGrid, int N)
|
|||||||
if ( dev ) {
|
if ( dev ) {
|
||||||
autoView(A_v,A[v],AcceleratorRead);
|
autoView(A_v,A[v],AcceleratorRead);
|
||||||
accelerator_for(ss,FGrid->oSites(),1,{
|
accelerator_for(ss,FGrid->oSites(),1,{
|
||||||
assert(B[v]==A_v[ss]()()().getlane(0));
|
// assert(B[v]==A_v[ss]()()().getlane(0));
|
||||||
});
|
});
|
||||||
// std::cout << "["<<v<<"] checked on GPU"<<B[v]<<std::endl;
|
// std::cout << "["<<v<<"] checked on GPU"<<B[v]<<std::endl;
|
||||||
} else {
|
} else {
|
||||||
|
@ -23,8 +23,8 @@ template<class vobj> inline void sliceSumCPU(const Grid::Lattice<vobj> &Data,std
|
|||||||
int ld=grid->_ldimensions[orthogdim];
|
int ld=grid->_ldimensions[orthogdim];
|
||||||
int rd=grid->_rdimensions[orthogdim];
|
int rd=grid->_rdimensions[orthogdim];
|
||||||
|
|
||||||
Vector<vobj> lvSum(rd); // will locally sum vectors first
|
std::vector<vobj> lvSum(rd); // will locally sum vectors first
|
||||||
Vector<sobj> lsSum(ld,Zero()); // sum across these down to scalars
|
std::vector<sobj> lsSum(ld,Zero()); // sum across these down to scalars
|
||||||
ExtractBuffer<sobj> extracted(Nsimd); // splitting the SIMD
|
ExtractBuffer<sobj> extracted(Nsimd); // splitting the SIMD
|
||||||
|
|
||||||
result.resize(fd); // And then global sum to return the same vector to every node
|
result.resize(fd); // And then global sum to return the same vector to every node
|
||||||
|
@ -87,8 +87,8 @@ static void run_generators_checks() {
|
|||||||
typedef typename Sp_TwoIndex<this_nc, S>::template iGroupMatrix<Complex> Matrix;
|
typedef typename Sp_TwoIndex<this_nc, S>::template iGroupMatrix<Complex> Matrix;
|
||||||
int sum = 0;
|
int sum = 0;
|
||||||
int sum_im = 0;
|
int sum_im = 0;
|
||||||
Vector<Matrix> ta_fund(this_algebra_dim);
|
std::vector<Matrix> ta_fund(this_algebra_dim);
|
||||||
Vector<Matrix> eij(this_irrep_dim);
|
std::vector<Matrix> eij(this_irrep_dim);
|
||||||
Matrix tmp_l;
|
Matrix tmp_l;
|
||||||
Matrix tmp_r;
|
Matrix tmp_r;
|
||||||
for (int n = 0; n < this_algebra_dim; n++)
|
for (int n = 0; n < this_algebra_dim; n++)
|
||||||
|
Loading…
Reference in New Issue
Block a user