diff --git a/lib/Grid_communicator.h b/lib/Grid_communicator.h index 661f39b8..cfa6e0a7 100644 --- a/lib/Grid_communicator.h +++ b/lib/Grid_communicator.h @@ -1,117 +1,6 @@ #ifndef GRID_COMMUNICATOR_H #define GRID_COMMUNICATOR_H -/////////////////////////////////// -// Processor layout information -/////////////////////////////////// -#ifdef GRID_COMMS_MPI -#include -#endif -namespace Grid { -class CartesianCommunicator { - public: - - // Communicator should know nothing of the physics grid, only processor grid. - - int _Nprocessors; // How many in all - std::vector _processors; // Which dimensions get relayed out over processors lanes. - int _processor; // linear processor rank - std::vector _processor_coor; // linear processor coordinate - unsigned long _ndimension; - -#ifdef GRID_COMMS_MPI - MPI_Comm communicator; - typedef MPI_Request CommsRequest_t; -#else - typedef int CommsRequest_t; -#endif - - // Constructor - CartesianCommunicator(std::vector &pdimensions_in); - - // Wraps MPI_Cart routines - void ShiftedRanks(int dim,int shift,int & source, int & dest); - int RankFromProcessorCoor(std::vector &coor); - void ProcessorCoorFromRank(int rank,std::vector &coor); - - ///////////////////////////////// - // Grid information queries - ///////////////////////////////// - int IsBoss(void) { return _processor==0; }; - int BossRank(void) { return 0; }; - int ThisRank(void) { return _processor; }; - const std::vector & ThisProcessorCoor(void) { return _processor_coor; }; - const std::vector & ProcessorGrid(void) { return _processors; }; - int ProcessorCount(void) { return _Nprocessors; }; - - //////////////////////////////////////////////////////////// - // Reduction - //////////////////////////////////////////////////////////// - void GlobalSum(RealF &); - void GlobalSumVector(RealF *,int N); - - void GlobalSum(RealD &); - void GlobalSumVector(RealD *,int N); - - void GlobalSum(uint32_t &); - - void GlobalSum(ComplexF &c) - { - GlobalSumVector((float *)&c,2); - } - void GlobalSumVector(ComplexF *c,int N) - { - GlobalSumVector((float *)c,2*N); - } - - void GlobalSum(ComplexD &c) - { - GlobalSumVector((double *)&c,2); - } - void GlobalSumVector(ComplexD *c,int N) - { - GlobalSumVector((double *)c,2*N); - } - - template void GlobalSum(obj &o){ - typedef typename obj::scalar_type scalar_type; - int words = sizeof(obj)/sizeof(scalar_type); - scalar_type * ptr = (scalar_type *)& o; - GlobalSumVector(ptr,words); - } - //////////////////////////////////////////////////////////// - // Face exchange, buffer swap in translational invariant way - //////////////////////////////////////////////////////////// - void SendToRecvFrom(void *xmit, - int xmit_to_rank, - void *recv, - int recv_from_rank, - int bytes); - void SendToRecvFromBegin(std::vector &list, - void *xmit, - int xmit_to_rank, - void *recv, - int recv_from_rank, - int bytes); - void SendToRecvFromComplete(std::vector &waitall); - - //////////////////////////////////////////////////////////// - // Barrier - //////////////////////////////////////////////////////////// - void Barrier(void); - - //////////////////////////////////////////////////////////// - // Broadcast a buffer and composite larger - //////////////////////////////////////////////////////////// - void Broadcast(int root,void* data, int bytes); - template void Broadcast(int root,obj &data) - { - Broadcast(root,(void *)&data,sizeof(data)); - }; - - static void BroadcastWorld(int root,void* data, int bytes); - -}; -} +#include #endif diff --git a/lib/Grid_lattice.h b/lib/Grid_lattice.h index 356a25b8..35664aee 100644 --- a/lib/Grid_lattice.h +++ b/lib/Grid_lattice.h @@ -1,203 +1,6 @@ #ifndef GRID_LATTICE_H #define GRID_LATTICE_H -namespace Grid { - -// TODO: -// mac,real,imag - -// Functionality: -// -=,+=,*=,() -// add,+,sub,-,mult,mac,* -// adj,conj -// real,imag -// transpose,transposeIndex -// trace,traceIndex -// peekIndex -// innerProduct,outerProduct, -// localNorm2 -// localInnerProduct - -extern int GridCshiftPermuteMap[4][16]; - -//////////////////////////////////////////////// -// Basic expressions used in Expression Template -//////////////////////////////////////////////// - -class LatticeBase {}; -class LatticeExpressionBase {}; - -template -class LatticeUnaryExpression : public std::pair > , public LatticeExpressionBase { - public: - LatticeUnaryExpression(const std::pair > &arg): std::pair >(arg) {}; -}; - -template -class LatticeBinaryExpression : public std::pair > , public LatticeExpressionBase { - public: - LatticeBinaryExpression(const std::pair > &arg): std::pair >(arg) {}; -}; - -template -class LatticeTrinaryExpression :public std::pair >, public LatticeExpressionBase { - public: - LatticeTrinaryExpression(const std::pair > &arg): std::pair >(arg) {}; -}; - -template -class Lattice : public LatticeBase -{ -public: - - GridBase *_grid; - int checkerboard; - std::vector > _odata; - -public: - typedef typename vobj::scalar_type scalar_type; - typedef typename vobj::vector_type vector_type; - typedef vobj vector_object; - - //////////////////////////////////////////////////////////////////////////////// - // Expression Template closure support - //////////////////////////////////////////////////////////////////////////////// - template inline Lattice & operator=(const LatticeUnaryExpression &expr) - { -#pragma omp parallel for - for(int ss=0;ss<_grid->oSites();ss++){ - vobj tmp= eval(ss,expr); - vstream(_odata[ss] ,tmp); - } - return *this; - } - template inline Lattice & operator=(const LatticeBinaryExpression &expr) - { -#pragma omp parallel for - for(int ss=0;ss<_grid->oSites();ss++){ - vobj tmp= eval(ss,expr); - vstream(_odata[ss] ,tmp); - } - return *this; - } - template inline Lattice & operator=(const LatticeTrinaryExpression &expr) - { -#pragma omp parallel for - for(int ss=0;ss<_grid->oSites();ss++){ - vobj tmp= eval(ss,expr); - vstream(_odata[ss] ,tmp); - } - return *this; - } - //GridFromExpression is tricky to do - template - Lattice(const LatticeUnaryExpression & expr): _grid(nullptr){ - GridFromExpression(_grid,expr); - assert(_grid!=nullptr); - _odata.resize(_grid->oSites()); -#pragma omp parallel for - for(int ss=0;ss<_grid->oSites();ss++){ - _odata[ss] = eval(ss,expr); - } - }; - template - Lattice(const LatticeBinaryExpression & expr): _grid(nullptr){ - GridFromExpression(_grid,expr); - assert(_grid!=nullptr); - _odata.resize(_grid->oSites()); -#pragma omp parallel for - for(int ss=0;ss<_grid->oSites();ss++){ - _odata[ss] = eval(ss,expr); - } - }; - template - Lattice(const LatticeTrinaryExpression & expr): _grid(nullptr){ - GridFromExpression(_grid,expr); - assert(_grid!=nullptr); - _odata.resize(_grid->oSites()); -#pragma omp parallel for - for(int ss=0;ss<_grid->oSites();ss++){ - _odata[ss] = eval(ss,expr); - } - }; - - ////////////////////////////////////////////////////////////////// - // Constructor requires "grid" passed. - // what about a default grid? - ////////////////////////////////////////////////////////////////// - Lattice(GridBase *grid) : _grid(grid), _odata(_grid->oSites()) { - // _odata.reserve(_grid->oSites()); - // _odata.resize(_grid->oSites()); - assert((((uint64_t)&_odata[0])&0xF) ==0); - checkerboard=0; - } - - template inline Lattice & operator = (const sobj & r){ -#pragma omp parallel for - for(int ss=0;ss<_grid->oSites();ss++){ - this->_odata[ss]=r; - } - return *this; - } - template inline Lattice & operator = (const Lattice & r){ - conformable(*this,r); - std::cout<<"Lattice operator ="<oSites();ss++){ - this->_odata[ss]=r._odata[ss]; - } - return *this; - } - - // *=,+=,-= operators inherit behvour from correspond */+/- operation - template inline Lattice &operator *=(const T &r) { - *this = (*this)*r; - return *this; - } - - template inline Lattice &operator -=(const T &r) { - *this = (*this)-r; - return *this; - } - template inline Lattice &operator +=(const T &r) { - *this = (*this)+r; - return *this; - } - - inline friend Lattice operator / (const Lattice &lhs,const Lattice &rhs){ - conformable(lhs,rhs); - Lattice ret(lhs._grid); -#pragma omp parallel for - for(int ss=0;ssoSites();ss++){ - ret._odata[ss] = lhs._odata[ss]/rhs._odata[ss]; - } - return ret; - }; - }; // class Lattice -} - -#undef GRID_LATTICE_EXPRESSION_TEMPLATES - -#include - -#ifdef GRID_LATTICE_EXPRESSION_TEMPLATES -#include -#else -#include -#endif -#include - -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include - - +#include #endif diff --git a/lib/communicator/Grid_communicator_base.h b/lib/communicator/Grid_communicator_base.h new file mode 100644 index 00000000..47c1f525 --- /dev/null +++ b/lib/communicator/Grid_communicator_base.h @@ -0,0 +1,117 @@ +#ifndef GRID_COMMUNICATOR_BASE_H +#define GRID_COMMUNICATOR_BASE_H + +/////////////////////////////////// +// Processor layout information +/////////////////////////////////// +#ifdef GRID_COMMS_MPI +#include +#endif +namespace Grid { +class CartesianCommunicator { + public: + + // Communicator should know nothing of the physics grid, only processor grid. + + int _Nprocessors; // How many in all + std::vector _processors; // Which dimensions get relayed out over processors lanes. + int _processor; // linear processor rank + std::vector _processor_coor; // linear processor coordinate + unsigned long _ndimension; + +#ifdef GRID_COMMS_MPI + MPI_Comm communicator; + typedef MPI_Request CommsRequest_t; +#else + typedef int CommsRequest_t; +#endif + + // Constructor + CartesianCommunicator(std::vector &pdimensions_in); + + // Wraps MPI_Cart routines + void ShiftedRanks(int dim,int shift,int & source, int & dest); + int RankFromProcessorCoor(std::vector &coor); + void ProcessorCoorFromRank(int rank,std::vector &coor); + + ///////////////////////////////// + // Grid information queries + ///////////////////////////////// + int IsBoss(void) { return _processor==0; }; + int BossRank(void) { return 0; }; + int ThisRank(void) { return _processor; }; + const std::vector & ThisProcessorCoor(void) { return _processor_coor; }; + const std::vector & ProcessorGrid(void) { return _processors; }; + int ProcessorCount(void) { return _Nprocessors; }; + + //////////////////////////////////////////////////////////// + // Reduction + //////////////////////////////////////////////////////////// + void GlobalSum(RealF &); + void GlobalSumVector(RealF *,int N); + + void GlobalSum(RealD &); + void GlobalSumVector(RealD *,int N); + + void GlobalSum(uint32_t &); + + void GlobalSum(ComplexF &c) + { + GlobalSumVector((float *)&c,2); + } + void GlobalSumVector(ComplexF *c,int N) + { + GlobalSumVector((float *)c,2*N); + } + + void GlobalSum(ComplexD &c) + { + GlobalSumVector((double *)&c,2); + } + void GlobalSumVector(ComplexD *c,int N) + { + GlobalSumVector((double *)c,2*N); + } + + template void GlobalSum(obj &o){ + typedef typename obj::scalar_type scalar_type; + int words = sizeof(obj)/sizeof(scalar_type); + scalar_type * ptr = (scalar_type *)& o; + GlobalSumVector(ptr,words); + } + //////////////////////////////////////////////////////////// + // Face exchange, buffer swap in translational invariant way + //////////////////////////////////////////////////////////// + void SendToRecvFrom(void *xmit, + int xmit_to_rank, + void *recv, + int recv_from_rank, + int bytes); + void SendToRecvFromBegin(std::vector &list, + void *xmit, + int xmit_to_rank, + void *recv, + int recv_from_rank, + int bytes); + void SendToRecvFromComplete(std::vector &waitall); + + //////////////////////////////////////////////////////////// + // Barrier + //////////////////////////////////////////////////////////// + void Barrier(void); + + //////////////////////////////////////////////////////////// + // Broadcast a buffer and composite larger + //////////////////////////////////////////////////////////// + void Broadcast(int root,void* data, int bytes); + template void Broadcast(int root,obj &data) + { + Broadcast(root,(void *)&data,sizeof(data)); + }; + + static void BroadcastWorld(int root,void* data, int bytes); + +}; +} + +#endif diff --git a/lib/lattice/Grid_lattice_base.h b/lib/lattice/Grid_lattice_base.h new file mode 100644 index 00000000..248bf1f7 --- /dev/null +++ b/lib/lattice/Grid_lattice_base.h @@ -0,0 +1,203 @@ +#ifndef GRID_LATTICE_BASE_H +#define GRID_LATTICE_BASE_H + +namespace Grid { + +// TODO: +// mac,real,imag + +// Functionality: +// -=,+=,*=,() +// add,+,sub,-,mult,mac,* +// adj,conj +// real,imag +// transpose,transposeIndex +// trace,traceIndex +// peekIndex +// innerProduct,outerProduct, +// localNorm2 +// localInnerProduct + +extern int GridCshiftPermuteMap[4][16]; + +//////////////////////////////////////////////// +// Basic expressions used in Expression Template +//////////////////////////////////////////////// + +class LatticeBase {}; +class LatticeExpressionBase {}; + +template +class LatticeUnaryExpression : public std::pair > , public LatticeExpressionBase { + public: + LatticeUnaryExpression(const std::pair > &arg): std::pair >(arg) {}; +}; + +template +class LatticeBinaryExpression : public std::pair > , public LatticeExpressionBase { + public: + LatticeBinaryExpression(const std::pair > &arg): std::pair >(arg) {}; +}; + +template +class LatticeTrinaryExpression :public std::pair >, public LatticeExpressionBase { + public: + LatticeTrinaryExpression(const std::pair > &arg): std::pair >(arg) {}; +}; + +template +class Lattice : public LatticeBase +{ +public: + + GridBase *_grid; + int checkerboard; + std::vector > _odata; + +public: + typedef typename vobj::scalar_type scalar_type; + typedef typename vobj::vector_type vector_type; + typedef vobj vector_object; + + //////////////////////////////////////////////////////////////////////////////// + // Expression Template closure support + //////////////////////////////////////////////////////////////////////////////// + template inline Lattice & operator=(const LatticeUnaryExpression &expr) + { +#pragma omp parallel for + for(int ss=0;ss<_grid->oSites();ss++){ + vobj tmp= eval(ss,expr); + vstream(_odata[ss] ,tmp); + } + return *this; + } + template inline Lattice & operator=(const LatticeBinaryExpression &expr) + { +#pragma omp parallel for + for(int ss=0;ss<_grid->oSites();ss++){ + vobj tmp= eval(ss,expr); + vstream(_odata[ss] ,tmp); + } + return *this; + } + template inline Lattice & operator=(const LatticeTrinaryExpression &expr) + { +#pragma omp parallel for + for(int ss=0;ss<_grid->oSites();ss++){ + vobj tmp= eval(ss,expr); + vstream(_odata[ss] ,tmp); + } + return *this; + } + //GridFromExpression is tricky to do + template + Lattice(const LatticeUnaryExpression & expr): _grid(nullptr){ + GridFromExpression(_grid,expr); + assert(_grid!=nullptr); + _odata.resize(_grid->oSites()); +#pragma omp parallel for + for(int ss=0;ss<_grid->oSites();ss++){ + _odata[ss] = eval(ss,expr); + } + }; + template + Lattice(const LatticeBinaryExpression & expr): _grid(nullptr){ + GridFromExpression(_grid,expr); + assert(_grid!=nullptr); + _odata.resize(_grid->oSites()); +#pragma omp parallel for + for(int ss=0;ss<_grid->oSites();ss++){ + _odata[ss] = eval(ss,expr); + } + }; + template + Lattice(const LatticeTrinaryExpression & expr): _grid(nullptr){ + GridFromExpression(_grid,expr); + assert(_grid!=nullptr); + _odata.resize(_grid->oSites()); +#pragma omp parallel for + for(int ss=0;ss<_grid->oSites();ss++){ + _odata[ss] = eval(ss,expr); + } + }; + + ////////////////////////////////////////////////////////////////// + // Constructor requires "grid" passed. + // what about a default grid? + ////////////////////////////////////////////////////////////////// + Lattice(GridBase *grid) : _grid(grid), _odata(_grid->oSites()) { + // _odata.reserve(_grid->oSites()); + // _odata.resize(_grid->oSites()); + assert((((uint64_t)&_odata[0])&0xF) ==0); + checkerboard=0; + } + + template inline Lattice & operator = (const sobj & r){ +#pragma omp parallel for + for(int ss=0;ss<_grid->oSites();ss++){ + this->_odata[ss]=r; + } + return *this; + } + template inline Lattice & operator = (const Lattice & r){ + conformable(*this,r); + std::cout<<"Lattice operator ="<oSites();ss++){ + this->_odata[ss]=r._odata[ss]; + } + return *this; + } + + // *=,+=,-= operators inherit behvour from correspond */+/- operation + template inline Lattice &operator *=(const T &r) { + *this = (*this)*r; + return *this; + } + + template inline Lattice &operator -=(const T &r) { + *this = (*this)-r; + return *this; + } + template inline Lattice &operator +=(const T &r) { + *this = (*this)+r; + return *this; + } + + inline friend Lattice operator / (const Lattice &lhs,const Lattice &rhs){ + conformable(lhs,rhs); + Lattice ret(lhs._grid); +#pragma omp parallel for + for(int ss=0;ssoSites();ss++){ + ret._odata[ss] = lhs._odata[ss]/rhs._odata[ss]; + } + return ret; + }; + }; // class Lattice +} + +#undef GRID_LATTICE_EXPRESSION_TEMPLATES + +#include + +#ifdef GRID_LATTICE_EXPRESSION_TEMPLATES +#include +#else +#include +#endif +#include + +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include + + + +#endif diff --git a/lib/math/Grid_math_inner.h b/lib/math/Grid_math_inner.h index 1e44032c..bbb35605 100644 --- a/lib/math/Grid_math_inner.h +++ b/lib/math/Grid_math_inner.h @@ -6,7 +6,7 @@ namespace Grid { // innerProduct Vector x Vector -> Scalar // innerProduct Matrix x Matrix -> Scalar /////////////////////////////////////////////////////////////////////////////////////// - template inline RealD norm2l(const sobj &arg){ + template inline RealD norm2(const sobj &arg){ typedef typename sobj::scalar_type scalar; decltype(innerProduct(arg,arg)) nrm; nrm = innerProduct(arg,arg); diff --git a/tests/Grid_gamma.cc b/tests/Grid_gamma.cc index 9779ca07..ba298b19 100644 --- a/tests/Grid_gamma.cc +++ b/tests/Grid_gamma.cc @@ -97,7 +97,7 @@ int main (int argc, char ** argv) for(int mu=0;mu<6;mu++){ result = Gamma(g[mu])* ident * Gamma(g[mu]); result = result - ident; - double mag = TensorRemove(norm2l(result)); + double mag = TensorRemove(norm2(result)); std::cout << list[mu]<<" " << mag<