1
0
mirror of https://github.com/paboyle/Grid.git synced 2024-09-20 09:15:38 +01:00
Grid/tests/solver/Test_dwf_mrhs_cg_mpi.cc

239 lines
8.6 KiB
C++
Raw Normal View History

2017-10-30 00:25:31 +00:00
/*************************************************************************************
2017-10-09 23:20:58 +01:00
Grid physics library, www.github.com/paboyle/Grid
Source file: ./tests/Test_dwf_mrhs_cg.cc
Copyright (C) 2015
Author: Peter Boyle <paboyle@ph.ed.ac.uk>
This program is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 2 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License along
with this program; if not, write to the Free Software Foundation, Inc.,
51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
See the full license in the file "LICENSE" in the top level distribution directory
*************************************************************************************/
/* END LEGAL */
#include <Grid/Grid.h>
#include <Grid/algorithms/iterative/BlockConjugateGradient.h>
using namespace std;
using namespace Grid;
;
2017-10-09 23:20:58 +01:00
int main (int argc, char ** argv)
{
typedef typename DomainWallFermionR::FermionField FermionField;
typedef typename DomainWallFermionR::ComplexField ComplexField;
typename DomainWallFermionR::ImplParams params;
2018-11-06 12:49:05 +00:00
double stp=1.0e-5;
2017-10-09 23:20:58 +01:00
const int Ls=4;
Grid_init(&argc,&argv);
2018-02-24 22:18:33 +00:00
Coordinate latt_size = GridDefaultLatt();
Coordinate simd_layout = GridDefaultSimd(Nd,vComplex::Nsimd());
Coordinate mpi_layout = GridDefaultMpi();
Coordinate mpi_split (mpi_layout.size(),1);
2017-10-09 23:20:58 +01:00
2017-10-30 00:25:31 +00:00
GridCartesian * UGrid = SpaceTimeGrid::makeFourDimGrid(GridDefaultLatt(),
GridDefaultSimd(Nd,vComplex::Nsimd()),
GridDefaultMpi());
2017-10-09 23:20:58 +01:00
GridCartesian * FGrid = SpaceTimeGrid::makeFiveDimGrid(Ls,UGrid);
GridRedBlackCartesian * rbGrid = SpaceTimeGrid::makeFourDimRedBlackGrid(UGrid);
GridRedBlackCartesian * FrbGrid = SpaceTimeGrid::makeFiveDimRedBlackGrid(Ls,UGrid);
/////////////////////////////////////////////
// Split into 1^4 mpi communicators
/////////////////////////////////////////////
2017-10-30 00:25:31 +00:00
for(int i=0;i<argc;i++){
if(std::string(argv[i]) == "--split"){
for(int k=0;k<mpi_layout.size();k++){
std::stringstream ss;
ss << argv[i+1+k];
ss >> mpi_split[k];
}
break;
}
}
int nrhs = 1;
int me;
for(int i=0;i<mpi_layout.size();i++) nrhs *= (mpi_layout[i]/mpi_split[i]);
2018-01-08 11:29:20 +00:00
std::cout << GridLogMessage << "Creating split grids " <<std::endl;
2017-10-09 23:20:58 +01:00
GridCartesian * SGrid = new GridCartesian(GridDefaultLatt(),
GridDefaultSimd(Nd,vComplex::Nsimd()),
mpi_split,
2017-10-30 00:25:31 +00:00
*UGrid,me);
2018-01-08 11:29:20 +00:00
std::cout << GridLogMessage <<"Creating split ferm grids " <<std::endl;
2017-10-09 23:20:58 +01:00
GridCartesian * SFGrid = SpaceTimeGrid::makeFiveDimGrid(Ls,SGrid);
2018-01-08 11:29:20 +00:00
std::cout << GridLogMessage <<"Creating split rb grids " <<std::endl;
2017-10-09 23:20:58 +01:00
GridRedBlackCartesian * SrbGrid = SpaceTimeGrid::makeFourDimRedBlackGrid(SGrid);
2018-01-08 11:29:20 +00:00
std::cout << GridLogMessage <<"Creating split ferm rb grids " <<std::endl;
2017-10-09 23:20:58 +01:00
GridRedBlackCartesian * SFrbGrid = SpaceTimeGrid::makeFiveDimRedBlackGrid(Ls,SGrid);
2017-12-05 13:07:31 +00:00
std::cout << GridLogMessage << "Made the grids"<<std::endl;
2017-10-09 23:20:58 +01:00
///////////////////////////////////////////////
// Set up the problem as a 4d spreadout job
///////////////////////////////////////////////
std::vector<int> seeds({1,2,3,4});
std::vector<FermionField> src(nrhs,FGrid);
std::vector<FermionField> src_chk(nrhs,FGrid);
std::vector<FermionField> result(nrhs,FGrid);
FermionField tmp(FGrid);
2017-12-05 13:07:31 +00:00
std::cout << GridLogMessage << "Made the Fermion Fields"<<std::endl;
2017-10-09 23:20:58 +01:00
2018-01-27 23:46:02 +00:00
for(int s=0;s<nrhs;s++) result[s]=Zero();
2017-12-05 13:07:31 +00:00
#undef LEXICO_TEST
2017-10-30 00:25:31 +00:00
#ifdef LEXICO_TEST
{
2018-01-27 23:46:02 +00:00
LatticeFermion lex(FGrid); lex = Zero();
2017-10-30 00:25:31 +00:00
LatticeFermion ftmp(FGrid);
Integer stride =10000;
double nrm;
LatticeComplex coor(FGrid);
for(int d=0;d<5;d++){
LatticeCoordinate(coor,d);
ftmp = stride;
ftmp = ftmp * coor;
lex = lex + ftmp;
stride=stride/10;
}
for(int s=0;s<nrhs;s++) {
src[s]=lex;
ftmp = 1000*1000*s;
src[s] = src[s] + ftmp;
}
}
#else
2017-12-05 13:07:31 +00:00
GridParallelRNG pRNG5(FGrid); pRNG5.SeedFixedIntegers(seeds);
2017-10-30 00:25:31 +00:00
for(int s=0;s<nrhs;s++) {
random(pRNG5,src[s]);
2018-01-08 11:29:20 +00:00
tmp = 10.0*s;
2017-10-30 00:25:31 +00:00
src[s] = (src[s] * 0.1) + tmp;
std::cout << GridLogMessage << " src ["<<s<<"] "<<norm2(src[s])<<std::endl;
2017-10-30 00:25:31 +00:00
}
#endif
2017-12-05 13:07:31 +00:00
std::cout << GridLogMessage << "Intialised the Fermion Fields"<<std::endl;
LatticeGaugeField Umu(UGrid);
if(1) {
GridParallelRNG pRNG(UGrid );
std::cout << GridLogMessage << "Intialising 4D RNG "<<std::endl;
pRNG.SeedFixedIntegers(seeds);
std::cout << GridLogMessage << "Intialised 4D RNG "<<std::endl;
SU<Nc>::HotConfiguration(pRNG,Umu);
2018-01-08 11:29:20 +00:00
std::cout << GridLogMessage << "Intialised the HOT Gauge Field"<<std::endl;
2018-01-26 23:09:03 +00:00
// std::cout << " Site zero "<< Umu[0] <<std::endl;
2017-12-05 13:07:31 +00:00
} else {
SU<Nc>::ColdConfiguration(Umu);
2017-12-05 13:07:31 +00:00
std::cout << GridLogMessage << "Intialised the COLD Gauge Field"<<std::endl;
2017-10-30 00:25:31 +00:00
}
2017-10-09 23:20:58 +01:00
/////////////////
// MPI only sends
/////////////////
LatticeGaugeField s_Umu(SGrid);
FermionField s_src(SFGrid);
FermionField s_tmp(SFGrid);
FermionField s_res(SFGrid);
2017-12-05 13:07:31 +00:00
std::cout << GridLogMessage << "Made the split grid fields"<<std::endl;
2017-10-09 23:20:58 +01:00
///////////////////////////////////////////////////////////////
// split the source out using MPI instead of I/O
///////////////////////////////////////////////////////////////
Grid_split (Umu,s_Umu);
Grid_split (src,s_src);
std::cout << GridLogMessage << " split rank " <<me << " s_src "<<norm2(s_src)<<std::endl;
2017-10-30 00:25:31 +00:00
#ifdef LEXICO_TEST
FermionField s_src_tmp(SFGrid);
FermionField s_src_diff(SFGrid);
{
2018-01-27 23:46:02 +00:00
LatticeFermion lex(SFGrid); lex = Zero();
2017-10-30 00:25:31 +00:00
LatticeFermion ftmp(SFGrid);
Integer stride =10000;
double nrm;
LatticeComplex coor(SFGrid);
for(int d=0;d<5;d++){
LatticeCoordinate(coor,d);
ftmp = stride;
ftmp = ftmp * coor;
lex = lex + ftmp;
stride=stride/10;
}
s_src_tmp=lex;
ftmp = 1000*1000*me;
s_src_tmp = s_src_tmp + ftmp;
}
s_src_diff = s_src_tmp - s_src;
std::cout << GridLogMessage <<" LEXICO test: s_src_diff " << norm2(s_src_diff)<<std::endl;
2017-10-30 00:25:31 +00:00
#endif
2017-10-09 23:20:58 +01:00
///////////////////////////////////////////////////////////////
// Set up N-solvers as trivially parallel
///////////////////////////////////////////////////////////////
2017-12-05 13:07:31 +00:00
std::cout << GridLogMessage << " Building the solvers"<<std::endl;
2017-10-09 23:20:58 +01:00
RealD mass=0.01;
RealD M5=1.8;
DomainWallFermionR Dchk(Umu,*FGrid,*FrbGrid,*UGrid,*rbGrid,mass,M5);
DomainWallFermionR Ddwf(s_Umu,*SFGrid,*SFrbGrid,*SGrid,*SrbGrid,mass,M5);
std::cout << GridLogMessage << "****************************************************************** "<<std::endl;
std::cout << GridLogMessage << " Calling DWF CG "<<std::endl;
std::cout << GridLogMessage << "****************************************************************** "<<std::endl;
MdagMLinearOperator<DomainWallFermionR,FermionField> HermOp(Ddwf);
MdagMLinearOperator<DomainWallFermionR,FermionField> HermOpCk(Dchk);
2018-11-06 12:49:05 +00:00
ConjugateGradient<FermionField> CG((stp),10000);
2018-01-27 23:46:02 +00:00
s_res = Zero();
2017-10-09 23:20:58 +01:00
CG(HermOp,s_src,s_res);
std::cout << GridLogMessage << " split residual norm "<<norm2(s_res)<<std::endl;
2017-10-09 23:20:58 +01:00
/////////////////////////////////////////////////////////////
// Report how long they all took
/////////////////////////////////////////////////////////////
std::vector<uint32_t> iterations(nrhs,0);
iterations[me] = CG.IterationsToComplete;
for(int n=0;n<nrhs;n++){
UGrid->GlobalSum(iterations[n]);
std::cout << GridLogMessage<<" Rank "<<n<<" "<< iterations[n]<<" CG iterations"<<std::endl;
}
/////////////////////////////////////////////////////////////
// Gather and residual check on the results
/////////////////////////////////////////////////////////////
std::cout << GridLogMessage<< "Unsplitting the result"<<std::endl;
Grid_unsplit(result,s_res);
2017-10-30 00:25:31 +00:00
2017-10-09 23:20:58 +01:00
std::cout << GridLogMessage<< "Checking the residuals"<<std::endl;
for(int n=0;n<nrhs;n++){
std::cout << GridLogMessage<< " res["<<n<<"] norm "<<norm2(result[n])<<std::endl;
2017-10-09 23:20:58 +01:00
HermOpCk.HermOp(result[n],tmp); tmp = tmp - src[n];
2017-10-30 00:25:31 +00:00
std::cout << GridLogMessage<<" resid["<<n<<"] "<< norm2(tmp)/norm2(src[n])<<std::endl;
2017-10-09 23:20:58 +01:00
}
for(int s=0;s<nrhs;s++) result[s]=Zero();
2018-11-06 12:49:05 +00:00
int blockDim = 0;//not used for BlockCGVec
BlockConjugateGradient<FermionField> BCGV (BlockCGVec,blockDim,stp,10000);
BCGV.PrintInterval=10;
BCGV(HermOpCk,src,result);
2017-10-09 23:20:58 +01:00
Grid_finalize();
}