1
0
mirror of https://github.com/paboyle/Grid.git synced 2024-11-14 01:35:36 +00:00

I'd forgotten that Intel '17 doesn't like auto var{value}; syntax

This commit is contained in:
Michael Marshall 2019-05-01 20:45:16 +01:00
parent 311c35a15c
commit 62692b68b9

View File

@ -103,7 +103,7 @@ inline void SliceShare( GridBase * gridLowDim, GridBase * gridHighDim, void * Bu
//#ifndef USE_LOCAL_SLICES //#ifndef USE_LOCAL_SLICES
// assert(0); // Can't do this without MPI (should really test whether MPI is defined) // assert(0); // Can't do this without MPI (should really test whether MPI is defined)
//#else //#else
const auto MyRank{gridHighDim->ThisRank()}; const auto MyRank = gridHighDim->ThisRank();
std::vector<CommsRequest_t> reqs(0); std::vector<CommsRequest_t> reqs(0);
int MySlice{coor[dimSpreadOut]}; int MySlice{coor[dimSpreadOut]};
char * const _buffer{(char *)Buffer}; char * const _buffer{(char *)Buffer};
@ -113,9 +113,9 @@ inline void SliceShare( GridBase * gridLowDim, GridBase * gridHighDim, void * Bu
int RecvSlice = ( MySlice - i + NumSlices ) % NumSlices; int RecvSlice = ( MySlice - i + NumSlices ) % NumSlices;
char * const RecvData{_buffer + RecvSlice * SliceSize}; char * const RecvData{_buffer + RecvSlice * SliceSize};
coor[dimSpreadOut] = SendSlice; coor[dimSpreadOut] = SendSlice;
const auto SendRank{gridHighDim->RankFromProcessorCoor(coor)}; const auto SendRank = gridHighDim->RankFromProcessorCoor(coor);
coor[dimSpreadOut] = RecvSlice; coor[dimSpreadOut] = RecvSlice;
const auto RecvRank{gridHighDim->RankFromProcessorCoor(coor)}; const auto RecvRank = gridHighDim->RankFromProcessorCoor(coor);
std::cout << GridLogMessage << "Send slice " << MySlice << " (" << MyRank << ") to " << SendSlice << " (" << SendRank std::cout << GridLogMessage << "Send slice " << MySlice << " (" << MyRank << ") to " << SendSlice << " (" << SendRank
<< "), receive slice from " << RecvSlice << " (" << RecvRank << ")" << std::endl; << "), receive slice from " << RecvSlice << " (" << RecvRank << ")" << std::endl;
gridHighDim->SendToRecvFromBegin(reqs,MyData,SendRank,RecvData,RecvRank,SliceSize); gridHighDim->SendToRecvFromBegin(reqs,MyData,SendRank,RecvData,RecvRank,SliceSize);
@ -389,7 +389,7 @@ void NamedTensor<Scalar_, NumIndices_, Endian_Scalar_Size>::WriteBinary(const st
assert((sizeof(Scalar_) % Endian_Scalar_Size) == 0 && "NamedTensor error: Scalar_ is not composed of Endian_Scalar_Size" ); assert((sizeof(Scalar_) % Endian_Scalar_Size) == 0 && "NamedTensor error: Scalar_ is not composed of Endian_Scalar_Size" );
// Size of the data (in bytes) // Size of the data (in bytes)
const uint32_t Scalar_Size{sizeof(Scalar_)}; const uint32_t Scalar_Size{sizeof(Scalar_)};
const auto NumElements{tensor.size()}; const auto NumElements = tensor.size();
const std::streamsize TotalDataSize{static_cast<std::streamsize>(NumElements * Scalar_Size)}; const std::streamsize TotalDataSize{static_cast<std::streamsize>(NumElements * Scalar_Size)};
uint64_t u64 = htobe64(static_cast<uint64_t>(TotalDataSize)); uint64_t u64 = htobe64(static_cast<uint64_t>(TotalDataSize));
w.write(reinterpret_cast<const char *>(&u64), sizeof(u64)); w.write(reinterpret_cast<const char *>(&u64), sizeof(u64));
@ -518,7 +518,7 @@ void NamedTensor<Scalar_, NumIndices_, Endian_Scalar_Size>::ReadBinary(const std
TotalDataSize = NumElements * Scalar_Size; TotalDataSize = NumElements * Scalar_Size;
} else { } else {
// dimensions together with names // dimensions together with names
const auto & TensorDims{tensor.dimensions()}; const auto & TensorDims = tensor.dimensions();
for( int d = 0; d < NumIndices_; d++ ) { for( int d = 0; d < NumIndices_; d++ ) {
// size of dimension // size of dimension
r.read(reinterpret_cast<char *>(&u16), sizeof(u16)); r.read(reinterpret_cast<char *>(&u16), sizeof(u16));
@ -665,11 +665,11 @@ inline void RotateEigen(std::vector<LatticeColourVector> & evec)
peekSite(cv0, evec[0], siteFirst); peekSite(cv0, evec[0], siteFirst);
auto & cplx0 = cv0()()(0); auto & cplx0 = cv0()()(0);
if( std::imag(cplx0) == 0 ) if( std::imag(cplx0) == 0 )
std::cout << GridLogMessage << "RotateEigen() : Site 0 : already meets phase convention" << std::endl; std::cout << GridLogMessage << "RotateEigen() : Site 0 : " << cplx0 << " => already meets phase convention" << std::endl;
else { else {
const auto cplx0_mag{std::abs(cplx0)}; const auto cplx0_mag = std::abs(cplx0);
const auto phase{std::conj(cplx0 / cplx0_mag)}; const auto phase = std::conj(cplx0 / cplx0_mag);
std::cout << GridLogMessage << "RotateEigen() : Site 0 : =" << cplx0_mag << " => phase=" << (std::arg(phase) / 3.14159265) << " pi" << std::endl; std::cout << GridLogMessage << "RotateEigen() : Site 0 : |" << cplx0 << "|=" << cplx0_mag << " => phase=" << (std::arg(phase) / 3.14159265) << " pi" << std::endl;
{ {
// TODO: Only really needed on the master slice // TODO: Only really needed on the master slice
for( int k = 0 ; k < evec.size() ; k++ ) for( int k = 0 ; k < evec.size() ; k++ )