mirror of
https://github.com/paboyle/Grid.git
synced 2024-11-10 07:55:35 +00:00
dc814f30da
Number of IO MPI tasks can be varied by selecting which dimensions use parallel IO and which dimensions use Serial send to boss I/O. Thus can neck down from, say 1024 nodes = 4x4x8x8 to {1,8,32,64,128,256,1024} nodes doing the I/O. Interpolates nicely between ALL nodes write their data, a single boss per time-plane in processor space [old UKQCD fortran code did this], and a single node doing all I/O. Not sure I have the transfer sizes big enough and am not overly convinced fstream is guaranteed to not give buffer inconsistencies unless I set streambuf size to zero. Practically it has worked on 8 tasks, 2x1x2x2 writing /cloning NERSC configurations on my MacOS + OpenMPI and Clang environment. It is VERY easy to switch to pwrite at a later date, and also easy to send x-strips around from each node in order to gather bigger chunks at the syscall level. That would push us up to the circa 8x 18*4*8 == 4KB size write chunk, and by taking, say, x/y non parallel we get to 16MB contiguous chunks written in multi 4KB transactions per IOnode in 64^3 lattices for configuration I/O. I suspect this is fine for system performance.
102 lines
2.7 KiB
C++
102 lines
2.7 KiB
C++
#include <Grid.h>
|
|
|
|
using namespace std;
|
|
using namespace Grid;
|
|
using namespace Grid::QCD;
|
|
|
|
|
|
int main (int argc, char ** argv)
|
|
{
|
|
Grid_init(&argc,&argv);
|
|
|
|
|
|
std::vector<int> simd_layout = GridDefaultSimd(4,vComplex::Nsimd());
|
|
std::vector<int> mpi_layout = GridDefaultMpi();
|
|
std::vector<int> latt_size ({16,16,16,32});
|
|
std::vector<int> clatt_size ({4,4,4,8});
|
|
int orthodir=3;
|
|
int orthosz =latt_size[orthodir];
|
|
|
|
GridCartesian Fine(latt_size,simd_layout,mpi_layout);
|
|
GridCartesian Coarse(clatt_size,simd_layout,mpi_layout);
|
|
|
|
LatticeGaugeField Umu(&Fine);
|
|
LatticeGaugeField Umu_diff(&Fine);
|
|
LatticeGaugeField Umu_saved(&Fine);
|
|
|
|
std::vector<LatticeColourMatrix> U(4,&Fine);
|
|
|
|
NerscField header;
|
|
std::string file("./ckpoint_lat.4000");
|
|
NerscIO::readConfiguration(Umu,header,file);
|
|
|
|
for(int mu=0;mu<Nd;mu++){
|
|
U[mu] = PeekIndex<LorentzIndex>(Umu,mu);
|
|
}
|
|
|
|
// Painful ; fix syntactical niceness
|
|
LatticeComplex LinkTrace(&Fine);
|
|
LinkTrace=zero;
|
|
for(int mu=0;mu<Nd;mu++){
|
|
LinkTrace = LinkTrace + trace(U[mu]);
|
|
}
|
|
|
|
// (1+2+3)=6 = N(N-1)/2 terms
|
|
LatticeComplex Plaq(&Fine);
|
|
LatticeComplex cPlaq(&Coarse);
|
|
|
|
Plaq = zero;
|
|
#if 1
|
|
for(int mu=1;mu<Nd;mu++){
|
|
for(int nu=0;nu<mu;nu++){
|
|
Plaq = Plaq + trace(U[mu]*Cshift(U[nu],mu,1)*adj(Cshift(U[mu],nu,1))*adj(U[nu]));
|
|
}
|
|
}
|
|
#endif
|
|
double vol = Fine.gSites();
|
|
Complex PlaqScale(1.0/vol/6.0/3.0);
|
|
std::cout<<GridLogMessage <<"PlaqScale" << PlaqScale<<std::endl;
|
|
|
|
std::vector<TComplex> Plaq_T(orthosz);
|
|
sliceSum(Plaq,Plaq_T,Nd-1);
|
|
int Nt = Plaq_T.size();
|
|
|
|
TComplex Plaq_T_sum;
|
|
Plaq_T_sum=zero;
|
|
for(int t=0;t<Nt;t++){
|
|
Plaq_T_sum = Plaq_T_sum+Plaq_T[t];
|
|
Complex Pt=TensorRemove(Plaq_T[t]);
|
|
std::cout<<GridLogMessage << "sliced ["<<t<<"]" <<Pt*PlaqScale*Real(Nt)<<std::endl;
|
|
}
|
|
|
|
{
|
|
Complex Pt = TensorRemove(Plaq_T_sum);
|
|
std::cout<<GridLogMessage << "total " <<Pt*PlaqScale<<std::endl;
|
|
}
|
|
|
|
|
|
TComplex Tp = sum(Plaq);
|
|
Complex p = TensorRemove(Tp);
|
|
std::cout<<GridLogMessage << "calculated plaquettes " <<p*PlaqScale<<std::endl;
|
|
|
|
|
|
Complex LinkTraceScale(1.0/vol/4.0/3.0);
|
|
TComplex Tl = sum(LinkTrace);
|
|
Complex l = TensorRemove(Tl);
|
|
std::cout<<GridLogMessage << "calculated link trace " <<l*LinkTraceScale<<std::endl;
|
|
|
|
blockSum(cPlaq,Plaq);
|
|
TComplex TcP = sum(cPlaq);
|
|
Complex ll= TensorRemove(TcP);
|
|
std::cout<<GridLogMessage << "coarsened plaquettes sum to " <<ll*PlaqScale<<std::endl;
|
|
|
|
std::string clone2x3("./ckpoint_clone2x3.4000");
|
|
std::string clone3x3("./ckpoint_clone3x3.4000");
|
|
int precision32 = 0;
|
|
|
|
NerscIO::writeConfiguration(Umu,clone3x3,0,precision32);
|
|
NerscIO::writeConfiguration(Umu,clone2x3,1,precision32);
|
|
|
|
Grid_finalize();
|
|
}
|