2016-10-24 17:30:43 +01:00
|
|
|
/*************************************************************************************
|
|
|
|
|
|
|
|
Grid physics library, www.github.com/paboyle/Grid
|
|
|
|
|
|
|
|
Source file: ./lib/communicator/Communicator_none.cc
|
|
|
|
|
|
|
|
Copyright (C) 2015
|
|
|
|
|
|
|
|
Author: Peter Boyle <paboyle@ph.ed.ac.uk>
|
|
|
|
|
|
|
|
This program is free software; you can redistribute it and/or modify
|
|
|
|
it under the terms of the GNU General Public License as published by
|
|
|
|
the Free Software Foundation; either version 2 of the License, or
|
|
|
|
(at your option) any later version.
|
|
|
|
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
GNU General Public License for more details.
|
|
|
|
|
|
|
|
You should have received a copy of the GNU General Public License along
|
|
|
|
with this program; if not, write to the Free Software Foundation, Inc.,
|
|
|
|
51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
|
|
|
|
|
|
See the full license in the file "LICENSE" in the top level distribution directory
|
|
|
|
*************************************************************************************/
|
|
|
|
/* END LEGAL */
|
2017-02-22 18:09:33 +00:00
|
|
|
#include <Grid/GridCore.h>
|
2017-08-19 17:53:18 +01:00
|
|
|
#include <fcntl.h>
|
|
|
|
#include <unistd.h>
|
|
|
|
#include <limits.h>
|
|
|
|
#include <sys/mman.h>
|
2017-02-22 18:09:33 +00:00
|
|
|
|
2016-10-24 17:30:43 +01:00
|
|
|
namespace Grid {
|
|
|
|
|
|
|
|
///////////////////////////////////////////////////////////////
|
|
|
|
// Info that is setup once and indept of cartesian layout
|
|
|
|
///////////////////////////////////////////////////////////////
|
|
|
|
void * CartesianCommunicator::ShmCommBuf;
|
2017-08-31 10:42:35 +01:00
|
|
|
uint64_t CartesianCommunicator::MAX_MPI_SHM_BYTES = 1024LL*1024LL*1024LL;
|
2017-07-29 18:06:53 +01:00
|
|
|
CartesianCommunicator::CommunicatorPolicy_t
|
|
|
|
CartesianCommunicator::CommunicatorPolicy= CartesianCommunicator::CommunicatorPolicyConcurrent;
|
|
|
|
int CartesianCommunicator::nCommThreads = -1;
|
2017-08-20 01:10:50 +01:00
|
|
|
int CartesianCommunicator::Hugepages = 0;
|
2016-10-24 17:30:43 +01:00
|
|
|
|
|
|
|
/////////////////////////////////
|
|
|
|
// Alloc, free shmem region
|
|
|
|
/////////////////////////////////
|
|
|
|
void *CartesianCommunicator::ShmBufferMalloc(size_t bytes){
|
|
|
|
// bytes = (bytes+sizeof(vRealD))&(~(sizeof(vRealD)-1));// align up bytes
|
|
|
|
void *ptr = (void *)heap_top;
|
|
|
|
heap_top += bytes;
|
|
|
|
heap_bytes+= bytes;
|
2016-11-02 08:07:09 +00:00
|
|
|
if (heap_bytes >= MAX_MPI_SHM_BYTES) {
|
|
|
|
std::cout<< " ShmBufferMalloc exceeded shared heap size -- try increasing with --shm <MB> flag" <<std::endl;
|
|
|
|
std::cout<< " Parameter specified in units of MB (megabytes) " <<std::endl;
|
|
|
|
std::cout<< " Current value is " << (MAX_MPI_SHM_BYTES/(1024*1024)) <<std::endl;
|
|
|
|
assert(heap_bytes<MAX_MPI_SHM_BYTES);
|
|
|
|
}
|
2016-10-24 17:30:43 +01:00
|
|
|
return ptr;
|
|
|
|
}
|
2016-10-25 01:45:53 +01:00
|
|
|
void CartesianCommunicator::ShmBufferFreeAll(void) {
|
2016-10-24 17:30:43 +01:00
|
|
|
heap_top =(size_t)ShmBufferSelf();
|
|
|
|
heap_bytes=0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/////////////////////////////////
|
|
|
|
// Grid information queries
|
|
|
|
/////////////////////////////////
|
2017-06-22 08:14:34 +01:00
|
|
|
int CartesianCommunicator::Dimensions(void) { return _ndimension; };
|
2016-10-24 17:30:43 +01:00
|
|
|
int CartesianCommunicator::IsBoss(void) { return _processor==0; };
|
|
|
|
int CartesianCommunicator::BossRank(void) { return 0; };
|
|
|
|
int CartesianCommunicator::ThisRank(void) { return _processor; };
|
|
|
|
const std::vector<int> & CartesianCommunicator::ThisProcessorCoor(void) { return _processor_coor; };
|
|
|
|
const std::vector<int> & CartesianCommunicator::ProcessorGrid(void) { return _processors; };
|
|
|
|
int CartesianCommunicator::ProcessorCount(void) { return _Nprocessors; };
|
|
|
|
|
|
|
|
////////////////////////////////////////////////////////////////////////////////
|
|
|
|
// very VERY rarely (Log, serial RNG) we need world without a grid
|
|
|
|
////////////////////////////////////////////////////////////////////////////////
|
|
|
|
|
|
|
|
void CartesianCommunicator::GlobalSum(ComplexF &c)
|
|
|
|
{
|
|
|
|
GlobalSumVector((float *)&c,2);
|
|
|
|
}
|
|
|
|
void CartesianCommunicator::GlobalSumVector(ComplexF *c,int N)
|
|
|
|
{
|
|
|
|
GlobalSumVector((float *)c,2*N);
|
|
|
|
}
|
|
|
|
void CartesianCommunicator::GlobalSum(ComplexD &c)
|
|
|
|
{
|
|
|
|
GlobalSumVector((double *)&c,2);
|
|
|
|
}
|
|
|
|
void CartesianCommunicator::GlobalSumVector(ComplexD *c,int N)
|
|
|
|
{
|
|
|
|
GlobalSumVector((double *)c,2*N);
|
|
|
|
}
|
|
|
|
|
2017-10-02 23:14:56 +01:00
|
|
|
|
2017-10-30 00:16:12 +00:00
|
|
|
#if defined( GRID_COMMS_MPI) || defined (GRID_COMMS_MPIT) || defined (GRID_COMMS_MPI3)
|
2017-11-06 23:45:00 +00:00
|
|
|
void CartesianCommunicator::AllToAll(int dim,void *in,void *out,uint64_t words,uint64_t bytes)
|
|
|
|
{
|
|
|
|
std::vector<int> row(_ndimension,1);
|
|
|
|
assert(dim>=0 && dim<_ndimension);
|
|
|
|
|
|
|
|
// Split the communicator
|
|
|
|
row[dim] = _processors[dim];
|
2017-10-02 23:14:56 +01:00
|
|
|
|
2017-11-06 23:45:00 +00:00
|
|
|
int me;
|
|
|
|
CartesianCommunicator Comm(row,*this,me);
|
|
|
|
Comm.AllToAll(in,out,words,bytes);
|
|
|
|
}
|
|
|
|
void CartesianCommunicator::AllToAll(void *in,void *out,uint64_t words,uint64_t bytes)
|
|
|
|
{
|
|
|
|
// MPI is a pain and uses "int" arguments
|
|
|
|
// 64*64*64*128*16 == 500Million elements of data.
|
|
|
|
// When 24*4 bytes multiples get 50x 10^9 >>> 2x10^9 Y2K bug.
|
|
|
|
// (Turns up on 32^3 x 64 Gparity too)
|
|
|
|
MPI_Datatype object;
|
|
|
|
int iwords;
|
|
|
|
int ibytes;
|
|
|
|
iwords = words;
|
|
|
|
ibytes = bytes;
|
|
|
|
assert(words == iwords); // safe to cast to int ?
|
|
|
|
assert(bytes == ibytes); // safe to cast to int ?
|
|
|
|
MPI_Type_contiguous(ibytes,MPI_BYTE,&object);
|
|
|
|
MPI_Type_commit(&object);
|
|
|
|
MPI_Alltoall(in,iwords,object,out,iwords,object,communicator);
|
|
|
|
MPI_Type_free(&object);
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#if defined( GRID_COMMS_MPI) || defined (GRID_COMMS_MPIT)
|
2017-10-30 00:16:12 +00:00
|
|
|
CartesianCommunicator::CartesianCommunicator(const std::vector<int> &processors,const CartesianCommunicator &parent,int &srank)
|
2017-10-02 23:14:56 +01:00
|
|
|
{
|
|
|
|
_ndimension = processors.size();
|
2017-11-19 01:39:04 +00:00
|
|
|
|
|
|
|
int parent_ndimension = parent._ndimension; assert(_ndimension >= parent._ndimension);
|
|
|
|
std::vector<int> parent_processor_coor(_ndimension,0);
|
|
|
|
std::vector<int> parent_processors (_ndimension,1);
|
|
|
|
|
|
|
|
// Can make 5d grid from 4d etc...
|
|
|
|
int pad = _ndimension-parent_ndimension;
|
|
|
|
for(int d=0;d<parent_ndimension;d++){
|
|
|
|
parent_processor_coor[pad+d]=parent._processor_coor[d];
|
|
|
|
parent_processors [pad+d]=parent._processors[d];
|
|
|
|
}
|
|
|
|
|
2017-10-02 23:14:56 +01:00
|
|
|
//////////////////////////////////////////////////////////////////////////////////////////////////////
|
|
|
|
// split the communicator
|
|
|
|
//////////////////////////////////////////////////////////////////////////////////////////////////////
|
|
|
|
int Nparent;
|
|
|
|
MPI_Comm_size(parent.communicator,&Nparent);
|
|
|
|
|
|
|
|
int childsize=1;
|
|
|
|
for(int d=0;d<processors.size();d++) {
|
|
|
|
childsize *= processors[d];
|
|
|
|
}
|
|
|
|
int Nchild = Nparent/childsize;
|
|
|
|
assert (childsize * Nchild == Nparent);
|
|
|
|
|
2017-10-09 23:16:51 +01:00
|
|
|
std::vector<int> ccoor(_ndimension); // coor within subcommunicator
|
|
|
|
std::vector<int> scoor(_ndimension); // coor of split within parent
|
|
|
|
std::vector<int> ssize(_ndimension); // coor of split within parent
|
|
|
|
|
|
|
|
for(int d=0;d<_ndimension;d++){
|
2017-11-19 01:39:04 +00:00
|
|
|
ccoor[d] = parent_processor_coor[d] % processors[d];
|
|
|
|
scoor[d] = parent_processor_coor[d] / processors[d];
|
|
|
|
ssize[d] = parent_processors[d] / processors[d];
|
2017-10-09 23:16:51 +01:00
|
|
|
}
|
2017-10-30 00:16:12 +00:00
|
|
|
int crank; // rank within subcomm ; srank is rank of subcomm within blocks of subcomms
|
|
|
|
// Mpi uses the reverse Lexico convention to us
|
|
|
|
Lexicographic::IndexFromCoorReversed(ccoor,crank,processors);
|
|
|
|
Lexicographic::IndexFromCoorReversed(scoor,srank,ssize);
|
2017-10-02 23:14:56 +01:00
|
|
|
|
|
|
|
MPI_Comm comm_split;
|
|
|
|
if ( Nchild > 1 ) {
|
|
|
|
|
2017-10-30 00:16:12 +00:00
|
|
|
std::cout << GridLogMessage<<"Child communicator of "<< std::hex << parent.communicator << std::dec<<std::endl;
|
|
|
|
std::cout << GridLogMessage<<" parent grid["<< parent._ndimension<<"] ";
|
2017-11-19 01:39:04 +00:00
|
|
|
for(int d=0;d<parent._ndimension;d++) std::cout << parent._processors[d] << " ";
|
2017-10-30 00:16:12 +00:00
|
|
|
std::cout<<std::endl;
|
2017-10-02 23:14:56 +01:00
|
|
|
|
2017-10-30 00:16:12 +00:00
|
|
|
std::cout << GridLogMessage<<" child grid["<< _ndimension <<"] ";
|
|
|
|
for(int d=0;d<processors.size();d++) std::cout << processors[d] << " ";
|
|
|
|
std::cout<<std::endl;
|
|
|
|
|
2017-11-19 01:39:04 +00:00
|
|
|
std::cout << GridLogMessage<<" old rank "<< parent._processor<<" coor ["<< parent._ndimension <<"] ";
|
|
|
|
for(int d=0;d<parent._ndimension;d++) std::cout << parent._processor_coor[d] << " ";
|
2017-10-30 00:16:12 +00:00
|
|
|
std::cout<<std::endl;
|
|
|
|
|
2017-11-19 01:39:04 +00:00
|
|
|
std::cout << GridLogMessage<<" new split "<< srank<<" scoor ["<< _ndimension <<"] ";
|
|
|
|
for(int d=0;d<processors.size();d++) std::cout << scoor[d] << " ";
|
2017-10-30 00:16:12 +00:00
|
|
|
std::cout<<std::endl;
|
|
|
|
|
2017-11-19 01:39:04 +00:00
|
|
|
std::cout << GridLogMessage<<" new rank "<< crank<<" coor ["<< _ndimension <<"] ";
|
|
|
|
for(int d=0;d<processors.size();d++) std::cout << ccoor[d] << " ";
|
2017-10-30 00:16:12 +00:00
|
|
|
std::cout<<std::endl;
|
2017-10-02 23:14:56 +01:00
|
|
|
|
2017-10-09 23:16:51 +01:00
|
|
|
int ierr= MPI_Comm_split(parent.communicator,srank,crank,&comm_split);
|
2017-10-02 23:14:56 +01:00
|
|
|
assert(ierr==0);
|
|
|
|
//////////////////////////////////////////////////////////////////////////////////////////////////////
|
|
|
|
// Declare victory
|
|
|
|
//////////////////////////////////////////////////////////////////////////////////////////////////////
|
2017-10-30 00:16:12 +00:00
|
|
|
std::cout << GridLogMessage<<"Divided communicator "<< parent._Nprocessors<<" into "
|
|
|
|
<< Nchild <<" communicators with " << childsize << " ranks"<<std::endl;
|
2017-10-02 23:14:56 +01:00
|
|
|
} else {
|
|
|
|
comm_split=parent.communicator;
|
2017-10-30 00:16:12 +00:00
|
|
|
srank = 0;
|
2017-10-02 23:14:56 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////////////////////////////////////
|
|
|
|
// Set up from the new split communicator
|
|
|
|
//////////////////////////////////////////////////////////////////////////////////////////////////////
|
|
|
|
InitFromMPICommunicator(processors,comm_split);
|
2017-11-19 01:39:04 +00:00
|
|
|
|
|
|
|
std::cout << " ndim " <<_ndimension<<" " << parent._ndimension << std::endl;
|
|
|
|
for(int d=0;d<processors.size();d++){
|
|
|
|
std::cout << d<< " " << _processor_coor[d] <<" " << ccoor[d]<<std::endl;
|
|
|
|
}
|
|
|
|
for(int d=0;d<processors.size();d++){
|
|
|
|
assert(_processor_coor[d] == ccoor[d] );
|
|
|
|
}
|
|
|
|
|
2017-10-02 23:14:56 +01:00
|
|
|
}
|
2017-11-06 23:45:00 +00:00
|
|
|
|
2017-10-02 23:14:56 +01:00
|
|
|
//////////////////////////////////////////////////////////////////////////////////////////////////////
|
|
|
|
// Take an MPI_Comm and self assemble
|
|
|
|
//////////////////////////////////////////////////////////////////////////////////////////////////////
|
|
|
|
void CartesianCommunicator::InitFromMPICommunicator(const std::vector<int> &processors, MPI_Comm communicator_base)
|
|
|
|
{
|
|
|
|
_ndimension = processors.size();
|
|
|
|
_processor_coor.resize(_ndimension);
|
|
|
|
|
|
|
|
/////////////////////////////////
|
|
|
|
// Count the requested nodes
|
|
|
|
/////////////////////////////////
|
|
|
|
_Nprocessors=1;
|
|
|
|
_processors = processors;
|
|
|
|
for(int i=0;i<_ndimension;i++){
|
|
|
|
_Nprocessors*=_processors[i];
|
|
|
|
}
|
|
|
|
|
|
|
|
std::vector<int> periodic(_ndimension,1);
|
2017-10-30 00:16:12 +00:00
|
|
|
MPI_Cart_create(communicator_base, _ndimension,&_processors[0],&periodic[0],0,&communicator);
|
2017-10-02 23:14:56 +01:00
|
|
|
MPI_Comm_rank(communicator,&_processor);
|
|
|
|
MPI_Cart_coords(communicator,_processor,_ndimension,&_processor_coor[0]);
|
|
|
|
|
2017-10-30 00:16:12 +00:00
|
|
|
if ( communicator_base != communicator_world ) {
|
2017-11-06 23:45:00 +00:00
|
|
|
std::cout << "InitFromMPICommunicator Cartesian communicator created with a non-world communicator"<<std::endl;
|
2017-10-30 00:16:12 +00:00
|
|
|
|
|
|
|
std::cout << " new communicator rank "<<_processor<< " coor ["<<_ndimension<<"] ";
|
|
|
|
for(int d=0;d<_processors.size();d++){
|
|
|
|
std::cout << _processor_coor[d]<<" ";
|
|
|
|
}
|
|
|
|
std::cout << std::endl;
|
|
|
|
}
|
|
|
|
|
2017-10-02 23:14:56 +01:00
|
|
|
int Size;
|
|
|
|
MPI_Comm_size(communicator,&Size);
|
|
|
|
|
2017-11-06 23:45:00 +00:00
|
|
|
#if defined(GRID_COMMS_MPIT) || defined (GRID_COMMS_MPI3)
|
2017-10-02 23:14:56 +01:00
|
|
|
communicator_halo.resize (2*_ndimension);
|
|
|
|
for(int i=0;i<_ndimension*2;i++){
|
|
|
|
MPI_Comm_dup(communicator,&communicator_halo[i]);
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
assert(Size==_Nprocessors);
|
|
|
|
}
|
2017-11-06 23:45:00 +00:00
|
|
|
#endif
|
2017-10-02 23:14:56 +01:00
|
|
|
|
2017-11-06 23:45:00 +00:00
|
|
|
#if defined( GRID_COMMS_MPI) || defined (GRID_COMMS_MPIT)
|
2017-10-02 23:14:56 +01:00
|
|
|
CartesianCommunicator::CartesianCommunicator(const std::vector<int> &processors)
|
|
|
|
{
|
|
|
|
InitFromMPICommunicator(processors,communicator_world);
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
2017-06-24 23:10:24 +01:00
|
|
|
#if !defined( GRID_COMMS_MPI3)
|
2017-02-07 06:22:53 +00:00
|
|
|
int CartesianCommunicator::NodeCount(void) { return ProcessorCount();};
|
2017-05-30 23:39:16 +01:00
|
|
|
int CartesianCommunicator::RankCount(void) { return ProcessorCount();};
|
2017-06-24 23:10:24 +01:00
|
|
|
#endif
|
2017-11-06 23:45:00 +00:00
|
|
|
|
2017-06-24 23:10:24 +01:00
|
|
|
#if !defined( GRID_COMMS_MPI3) && !defined (GRID_COMMS_MPIT)
|
2017-08-20 01:37:07 +01:00
|
|
|
double CartesianCommunicator::StencilSendToRecvFrom( void *xmit,
|
|
|
|
int xmit_to_rank,
|
|
|
|
void *recv,
|
|
|
|
int recv_from_rank,
|
|
|
|
int bytes, int dir)
|
|
|
|
{
|
|
|
|
std::vector<CommsRequest_t> list;
|
|
|
|
// Discard the "dir"
|
|
|
|
SendToRecvFromBegin (list,xmit,xmit_to_rank,recv,recv_from_rank,bytes);
|
|
|
|
SendToRecvFromComplete(list);
|
|
|
|
return 2.0*bytes;
|
|
|
|
}
|
2017-02-20 22:47:40 +00:00
|
|
|
double CartesianCommunicator::StencilSendToRecvFromBegin(std::vector<CommsRequest_t> &list,
|
2017-06-24 23:10:24 +01:00
|
|
|
void *xmit,
|
|
|
|
int xmit_to_rank,
|
|
|
|
void *recv,
|
|
|
|
int recv_from_rank,
|
|
|
|
int bytes, int dir)
|
2016-10-24 17:30:43 +01:00
|
|
|
{
|
2017-06-24 23:10:24 +01:00
|
|
|
// Discard the "dir"
|
2016-10-24 17:30:43 +01:00
|
|
|
SendToRecvFromBegin(list,xmit,xmit_to_rank,recv,recv_from_rank,bytes);
|
2017-02-20 22:47:40 +00:00
|
|
|
return 2.0*bytes;
|
2016-10-24 17:30:43 +01:00
|
|
|
}
|
2017-06-24 23:10:24 +01:00
|
|
|
void CartesianCommunicator::StencilSendToRecvFromComplete(std::vector<CommsRequest_t> &waitall,int dir)
|
2016-10-24 17:30:43 +01:00
|
|
|
{
|
|
|
|
SendToRecvFromComplete(waitall);
|
|
|
|
}
|
2017-06-24 23:10:24 +01:00
|
|
|
#endif
|
|
|
|
|
|
|
|
#if !defined( GRID_COMMS_MPI3)
|
|
|
|
|
2016-10-24 23:45:31 +01:00
|
|
|
void CartesianCommunicator::StencilBarrier(void){};
|
2016-10-24 17:30:43 +01:00
|
|
|
|
|
|
|
commVector<uint8_t> CartesianCommunicator::ShmBufStorageVector;
|
|
|
|
|
|
|
|
void *CartesianCommunicator::ShmBufferSelf(void) { return ShmCommBuf; }
|
2016-10-24 23:45:31 +01:00
|
|
|
|
2016-10-24 17:30:43 +01:00
|
|
|
void *CartesianCommunicator::ShmBuffer(int rank) {
|
2016-10-24 23:45:31 +01:00
|
|
|
return NULL;
|
2016-10-24 17:30:43 +01:00
|
|
|
}
|
|
|
|
void *CartesianCommunicator::ShmBufferTranslate(int rank,void * local_p) {
|
2016-10-24 23:45:31 +01:00
|
|
|
return NULL;
|
2016-10-24 17:30:43 +01:00
|
|
|
}
|
|
|
|
void CartesianCommunicator::ShmInitGeneric(void){
|
2017-08-19 17:53:18 +01:00
|
|
|
#if 1
|
2017-10-02 12:41:02 +01:00
|
|
|
int mmap_flag =0;
|
|
|
|
#ifdef MAP_ANONYMOUS
|
|
|
|
mmap_flag = mmap_flag| MAP_SHARED | MAP_ANONYMOUS;
|
|
|
|
#endif
|
|
|
|
#ifdef MAP_ANON
|
|
|
|
mmap_flag = mmap_flag| MAP_SHARED | MAP_ANON;
|
|
|
|
#endif
|
2017-08-20 03:08:54 +01:00
|
|
|
#ifdef MAP_HUGETLB
|
2017-08-20 01:10:50 +01:00
|
|
|
if ( Hugepages ) mmap_flag |= MAP_HUGETLB;
|
2017-08-20 03:08:54 +01:00
|
|
|
#endif
|
2017-08-20 01:10:50 +01:00
|
|
|
ShmCommBuf =(void *) mmap(NULL, MAX_MPI_SHM_BYTES, PROT_READ | PROT_WRITE, mmap_flag, -1, 0);
|
2017-08-25 09:25:54 +01:00
|
|
|
if (ShmCommBuf == (void *)MAP_FAILED) {
|
|
|
|
perror("mmap failed ");
|
|
|
|
exit(EXIT_FAILURE);
|
|
|
|
}
|
2017-09-05 15:00:16 +01:00
|
|
|
#ifdef MADV_HUGEPAGE
|
2017-09-04 15:41:21 +01:00
|
|
|
if (!Hugepages ) madvise(ShmCommBuf,MAX_MPI_SHM_BYTES,MADV_HUGEPAGE);
|
2017-09-05 15:00:16 +01:00
|
|
|
#endif
|
2017-08-19 17:53:18 +01:00
|
|
|
#else
|
2016-10-24 17:30:43 +01:00
|
|
|
ShmBufStorageVector.resize(MAX_MPI_SHM_BYTES);
|
|
|
|
ShmCommBuf=(void *)&ShmBufStorageVector[0];
|
2017-08-19 17:53:18 +01:00
|
|
|
#endif
|
|
|
|
bzero(ShmCommBuf,MAX_MPI_SHM_BYTES);
|
2016-10-24 17:30:43 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
}
|
|
|
|
|