mirror of
				https://github.com/paboyle/Grid.git
				synced 2025-11-04 05:54:32 +00:00 
			
		
		
		
	Merge branch 'develop' into feature/dirichlet
This commit is contained in:
		@@ -55,6 +55,7 @@ NAMESPACE_CHECK(BiCGSTAB);
 | 
				
			|||||||
#include <Grid/algorithms/iterative/ConjugateGradientMultiShift.h>
 | 
					#include <Grid/algorithms/iterative/ConjugateGradientMultiShift.h>
 | 
				
			||||||
#include <Grid/algorithms/iterative/ConjugateGradientMixedPrec.h>
 | 
					#include <Grid/algorithms/iterative/ConjugateGradientMixedPrec.h>
 | 
				
			||||||
#include <Grid/algorithms/iterative/ConjugateGradientMultiShiftMixedPrec.h>
 | 
					#include <Grid/algorithms/iterative/ConjugateGradientMultiShiftMixedPrec.h>
 | 
				
			||||||
 | 
					#include <Grid/algorithms/iterative/ConjugateGradientMixedPrecBatched.h>
 | 
				
			||||||
#include <Grid/algorithms/iterative/BiCGSTABMixedPrec.h>
 | 
					#include <Grid/algorithms/iterative/BiCGSTABMixedPrec.h>
 | 
				
			||||||
#include <Grid/algorithms/iterative/BlockConjugateGradient.h>
 | 
					#include <Grid/algorithms/iterative/BlockConjugateGradient.h>
 | 
				
			||||||
#include <Grid/algorithms/iterative/ConjugateGradientReliableUpdate.h>
 | 
					#include <Grid/algorithms/iterative/ConjugateGradientReliableUpdate.h>
 | 
				
			||||||
 
 | 
				
			|||||||
							
								
								
									
										213
									
								
								Grid/algorithms/iterative/ConjugateGradientMixedPrecBatched.h
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										213
									
								
								Grid/algorithms/iterative/ConjugateGradientMixedPrecBatched.h
									
									
									
									
									
										Normal file
									
								
							@@ -0,0 +1,213 @@
 | 
				
			|||||||
 | 
					/*************************************************************************************
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    Grid physics library, www.github.com/paboyle/Grid 
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    Source file: ./lib/algorithms/iterative/ConjugateGradientMixedPrecBatched.h
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    Copyright (C) 2015
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    Author: Raoul Hodgson <raoul.hodgson@ed.ac.uk>
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    This program is free software; you can redistribute it and/or modify
 | 
				
			||||||
 | 
					    it under the terms of the GNU General Public License as published by
 | 
				
			||||||
 | 
					    the Free Software Foundation; either version 2 of the License, or
 | 
				
			||||||
 | 
					    (at your option) any later version.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    This program is distributed in the hope that it will be useful,
 | 
				
			||||||
 | 
					    but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
				
			||||||
 | 
					    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | 
				
			||||||
 | 
					    GNU General Public License for more details.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    You should have received a copy of the GNU General Public License along
 | 
				
			||||||
 | 
					    with this program; if not, write to the Free Software Foundation, Inc.,
 | 
				
			||||||
 | 
					    51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    See the full license in the file "LICENSE" in the top level distribution directory
 | 
				
			||||||
 | 
					*************************************************************************************/
 | 
				
			||||||
 | 
					/*  END LEGAL */
 | 
				
			||||||
 | 
					#ifndef GRID_CONJUGATE_GRADIENT_MIXED_PREC_BATCHED_H
 | 
				
			||||||
 | 
					#define GRID_CONJUGATE_GRADIENT_MIXED_PREC_BATCHED_H
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					NAMESPACE_BEGIN(Grid);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					//Mixed precision restarted defect correction CG
 | 
				
			||||||
 | 
					template<class FieldD,class FieldF, 
 | 
				
			||||||
 | 
					  typename std::enable_if< getPrecision<FieldD>::value == 2, int>::type = 0,
 | 
				
			||||||
 | 
					  typename std::enable_if< getPrecision<FieldF>::value == 1, int>::type = 0> 
 | 
				
			||||||
 | 
					class MixedPrecisionConjugateGradientBatched : public LinearFunction<FieldD> {
 | 
				
			||||||
 | 
					public:
 | 
				
			||||||
 | 
					  using LinearFunction<FieldD>::operator();
 | 
				
			||||||
 | 
					  RealD   Tolerance;
 | 
				
			||||||
 | 
					  RealD   InnerTolerance; //Initial tolerance for inner CG. Defaults to Tolerance but can be changed
 | 
				
			||||||
 | 
					  Integer MaxInnerIterations;
 | 
				
			||||||
 | 
					  Integer MaxOuterIterations;
 | 
				
			||||||
 | 
					  Integer MaxPatchupIterations;
 | 
				
			||||||
 | 
					  GridBase* SinglePrecGrid; //Grid for single-precision fields
 | 
				
			||||||
 | 
					  RealD OuterLoopNormMult; //Stop the outer loop and move to a final double prec solve when the residual is OuterLoopNormMult * Tolerance
 | 
				
			||||||
 | 
					  LinearOperatorBase<FieldF> &Linop_f;
 | 
				
			||||||
 | 
					  LinearOperatorBase<FieldD> &Linop_d;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					  //Option to speed up *inner single precision* solves using a LinearFunction that produces a guess
 | 
				
			||||||
 | 
					  LinearFunction<FieldF> *guesser;
 | 
				
			||||||
 | 
					  bool updateResidual;
 | 
				
			||||||
 | 
					  
 | 
				
			||||||
 | 
					  MixedPrecisionConjugateGradientBatched(RealD tol, 
 | 
				
			||||||
 | 
					          Integer maxinnerit, 
 | 
				
			||||||
 | 
					          Integer maxouterit, 
 | 
				
			||||||
 | 
					          Integer maxpatchit,
 | 
				
			||||||
 | 
					          GridBase* _sp_grid, 
 | 
				
			||||||
 | 
					          LinearOperatorBase<FieldF> &_Linop_f, 
 | 
				
			||||||
 | 
					          LinearOperatorBase<FieldD> &_Linop_d,
 | 
				
			||||||
 | 
					          bool _updateResidual=true) :
 | 
				
			||||||
 | 
					    Linop_f(_Linop_f), Linop_d(_Linop_d),
 | 
				
			||||||
 | 
					    Tolerance(tol), InnerTolerance(tol), MaxInnerIterations(maxinnerit), MaxOuterIterations(maxouterit), MaxPatchupIterations(maxpatchit), SinglePrecGrid(_sp_grid),
 | 
				
			||||||
 | 
					    OuterLoopNormMult(100.), guesser(NULL), updateResidual(_updateResidual) { };
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					  void useGuesser(LinearFunction<FieldF> &g){
 | 
				
			||||||
 | 
					    guesser = &g;
 | 
				
			||||||
 | 
					  }
 | 
				
			||||||
 | 
					  
 | 
				
			||||||
 | 
					  void operator() (const FieldD &src_d_in, FieldD &sol_d){
 | 
				
			||||||
 | 
					    std::vector<FieldD> srcs_d_in{src_d_in};
 | 
				
			||||||
 | 
					    std::vector<FieldD> sols_d{sol_d};
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    (*this)(srcs_d_in,sols_d);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    sol_d = sols_d[0];
 | 
				
			||||||
 | 
					  }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					  void operator() (const std::vector<FieldD> &src_d_in, std::vector<FieldD> &sol_d){
 | 
				
			||||||
 | 
					    assert(src_d_in.size() == sol_d.size());
 | 
				
			||||||
 | 
					    int NBatch = src_d_in.size();
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    std::cout << GridLogMessage << "NBatch = " << NBatch << std::endl;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    Integer TotalOuterIterations = 0; //Number of restarts
 | 
				
			||||||
 | 
					    std::vector<Integer> TotalInnerIterations(NBatch,0);     //Number of inner CG iterations
 | 
				
			||||||
 | 
					    std::vector<Integer> TotalFinalStepIterations(NBatch,0); //Number of CG iterations in final patch-up step
 | 
				
			||||||
 | 
					  
 | 
				
			||||||
 | 
					    GridStopWatch TotalTimer;
 | 
				
			||||||
 | 
					    TotalTimer.Start();
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    GridStopWatch InnerCGtimer;
 | 
				
			||||||
 | 
					    GridStopWatch PrecChangeTimer;
 | 
				
			||||||
 | 
					    
 | 
				
			||||||
 | 
					    int cb = src_d_in[0].Checkerboard();
 | 
				
			||||||
 | 
					    
 | 
				
			||||||
 | 
					    std::vector<RealD> src_norm;
 | 
				
			||||||
 | 
					    std::vector<RealD> norm;
 | 
				
			||||||
 | 
					    std::vector<RealD> stop;
 | 
				
			||||||
 | 
					    
 | 
				
			||||||
 | 
					    GridBase* DoublePrecGrid = src_d_in[0].Grid();
 | 
				
			||||||
 | 
					    FieldD tmp_d(DoublePrecGrid);
 | 
				
			||||||
 | 
					    tmp_d.Checkerboard() = cb;
 | 
				
			||||||
 | 
					    
 | 
				
			||||||
 | 
					    FieldD tmp2_d(DoublePrecGrid);
 | 
				
			||||||
 | 
					    tmp2_d.Checkerboard() = cb;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    std::vector<FieldD> src_d;
 | 
				
			||||||
 | 
					    std::vector<FieldF> src_f;
 | 
				
			||||||
 | 
					    std::vector<FieldF> sol_f;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    for (int i=0; i<NBatch; i++) {
 | 
				
			||||||
 | 
					      sol_d[i].Checkerboard() = cb;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					      src_norm.push_back(norm2(src_d_in[i]));
 | 
				
			||||||
 | 
					      norm.push_back(0.);
 | 
				
			||||||
 | 
					      stop.push_back(src_norm[i] * Tolerance*Tolerance);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					      src_d.push_back(src_d_in[i]); //source for next inner iteration, computed from residual during operation
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					      src_f.push_back(SinglePrecGrid);
 | 
				
			||||||
 | 
					      src_f[i].Checkerboard() = cb;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					      sol_f.push_back(SinglePrecGrid);
 | 
				
			||||||
 | 
					      sol_f[i].Checkerboard() = cb;
 | 
				
			||||||
 | 
					    }
 | 
				
			||||||
 | 
					    
 | 
				
			||||||
 | 
					    RealD inner_tol = InnerTolerance;
 | 
				
			||||||
 | 
					    
 | 
				
			||||||
 | 
					    ConjugateGradient<FieldF> CG_f(inner_tol, MaxInnerIterations);
 | 
				
			||||||
 | 
					    CG_f.ErrorOnNoConverge = false;
 | 
				
			||||||
 | 
					    
 | 
				
			||||||
 | 
					    Integer &outer_iter = TotalOuterIterations; //so it will be equal to the final iteration count
 | 
				
			||||||
 | 
					      
 | 
				
			||||||
 | 
					    for(outer_iter = 0; outer_iter < MaxOuterIterations; outer_iter++){
 | 
				
			||||||
 | 
					      std::cout << GridLogMessage << std::endl;
 | 
				
			||||||
 | 
					      std::cout << GridLogMessage << "Outer iteration " << outer_iter << std::endl;
 | 
				
			||||||
 | 
					      
 | 
				
			||||||
 | 
					      bool allConverged = true;
 | 
				
			||||||
 | 
					      
 | 
				
			||||||
 | 
					      for (int i=0; i<NBatch; i++) {
 | 
				
			||||||
 | 
					        //Compute double precision rsd and also new RHS vector.
 | 
				
			||||||
 | 
					        Linop_d.HermOp(sol_d[i], tmp_d);
 | 
				
			||||||
 | 
					        norm[i] = axpy_norm(src_d[i], -1., tmp_d, src_d_in[i]); //src_d is residual vector
 | 
				
			||||||
 | 
					        
 | 
				
			||||||
 | 
					        std::cout<<GridLogMessage<<"MixedPrecisionConjugateGradientBatched: Outer iteration " << outer_iter <<" solve " << i << " residual "<< norm[i] << " target "<< stop[i] <<std::endl;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        PrecChangeTimer.Start();
 | 
				
			||||||
 | 
					        precisionChange(src_f[i], src_d[i]);
 | 
				
			||||||
 | 
					        PrecChangeTimer.Stop();
 | 
				
			||||||
 | 
					        
 | 
				
			||||||
 | 
					        sol_f[i] = Zero();
 | 
				
			||||||
 | 
					      
 | 
				
			||||||
 | 
					        if(norm[i] > OuterLoopNormMult * stop[i]) {
 | 
				
			||||||
 | 
					          allConverged = false;
 | 
				
			||||||
 | 
					        }
 | 
				
			||||||
 | 
					      }
 | 
				
			||||||
 | 
					      if (allConverged) break;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					      if (updateResidual) {
 | 
				
			||||||
 | 
					        RealD normMax = *std::max_element(std::begin(norm), std::end(norm));
 | 
				
			||||||
 | 
					        RealD stopMax = *std::max_element(std::begin(stop), std::end(stop));
 | 
				
			||||||
 | 
					        while( normMax * inner_tol * inner_tol < stopMax) inner_tol *= 2;  // inner_tol = sqrt(stop/norm) ??
 | 
				
			||||||
 | 
					        CG_f.Tolerance = inner_tol;
 | 
				
			||||||
 | 
					      }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					      //Optionally improve inner solver guess (eg using known eigenvectors)
 | 
				
			||||||
 | 
					      if(guesser != NULL) {
 | 
				
			||||||
 | 
					        (*guesser)(src_f, sol_f);
 | 
				
			||||||
 | 
					      }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					      for (int i=0; i<NBatch; i++) {
 | 
				
			||||||
 | 
					        //Inner CG
 | 
				
			||||||
 | 
					        InnerCGtimer.Start();
 | 
				
			||||||
 | 
					        CG_f(Linop_f, src_f[i], sol_f[i]);
 | 
				
			||||||
 | 
					        InnerCGtimer.Stop();
 | 
				
			||||||
 | 
					        TotalInnerIterations[i] += CG_f.IterationsToComplete;
 | 
				
			||||||
 | 
					        
 | 
				
			||||||
 | 
					        //Convert sol back to double and add to double prec solution
 | 
				
			||||||
 | 
					        PrecChangeTimer.Start();
 | 
				
			||||||
 | 
					        precisionChange(tmp_d, sol_f[i]);
 | 
				
			||||||
 | 
					        PrecChangeTimer.Stop();
 | 
				
			||||||
 | 
					        
 | 
				
			||||||
 | 
					        axpy(sol_d[i], 1.0, tmp_d, sol_d[i]);
 | 
				
			||||||
 | 
					      }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    }
 | 
				
			||||||
 | 
					    
 | 
				
			||||||
 | 
					    //Final trial CG
 | 
				
			||||||
 | 
					    std::cout << GridLogMessage << std::endl;
 | 
				
			||||||
 | 
					    std::cout<<GridLogMessage<<"MixedPrecisionConjugateGradientBatched: Starting final patch-up double-precision solve"<<std::endl;
 | 
				
			||||||
 | 
					    
 | 
				
			||||||
 | 
					    for (int i=0; i<NBatch; i++) {
 | 
				
			||||||
 | 
					      ConjugateGradient<FieldD> CG_d(Tolerance, MaxPatchupIterations);
 | 
				
			||||||
 | 
					      CG_d(Linop_d, src_d_in[i], sol_d[i]);
 | 
				
			||||||
 | 
					      TotalFinalStepIterations[i] += CG_d.IterationsToComplete;
 | 
				
			||||||
 | 
					    }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    TotalTimer.Stop();
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    std::cout << GridLogMessage << std::endl;
 | 
				
			||||||
 | 
					    for (int i=0; i<NBatch; i++) {
 | 
				
			||||||
 | 
					      std::cout<<GridLogMessage<<"MixedPrecisionConjugateGradientBatched: solve " << i << " Inner CG iterations " << TotalInnerIterations[i] << " Restarts " << TotalOuterIterations << " Final CG iterations " << TotalFinalStepIterations[i] << std::endl;
 | 
				
			||||||
 | 
					    }
 | 
				
			||||||
 | 
					    std::cout << GridLogMessage << std::endl;
 | 
				
			||||||
 | 
					    std::cout<<GridLogMessage<<"MixedPrecisionConjugateGradientBatched: Total time " << TotalTimer.Elapsed() << " Precision change " << PrecChangeTimer.Elapsed() << " Inner CG total " << InnerCGtimer.Elapsed() << std::endl;
 | 
				
			||||||
 | 
					    
 | 
				
			||||||
 | 
					  }
 | 
				
			||||||
 | 
					};
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					NAMESPACE_END(Grid);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					#endif
 | 
				
			||||||
@@ -4,11 +4,14 @@ NAMESPACE_BEGIN(Grid);
 | 
				
			|||||||
 | 
					
 | 
				
			||||||
/*Allocation types, saying which pointer cache should be used*/
 | 
					/*Allocation types, saying which pointer cache should be used*/
 | 
				
			||||||
#define Cpu      (0)
 | 
					#define Cpu      (0)
 | 
				
			||||||
#define CpuSmall (1)
 | 
					#define CpuHuge  (1)
 | 
				
			||||||
#define Acc      (2)
 | 
					#define CpuSmall (2)
 | 
				
			||||||
#define AccSmall (3)
 | 
					#define Acc      (3)
 | 
				
			||||||
#define Shared   (4)
 | 
					#define AccHuge  (4)
 | 
				
			||||||
#define SharedSmall (5)
 | 
					#define AccSmall (5)
 | 
				
			||||||
 | 
					#define Shared   (6)
 | 
				
			||||||
 | 
					#define SharedHuge  (7)
 | 
				
			||||||
 | 
					#define SharedSmall (8)
 | 
				
			||||||
#undef GRID_MM_VERBOSE 
 | 
					#undef GRID_MM_VERBOSE 
 | 
				
			||||||
uint64_t total_shared;
 | 
					uint64_t total_shared;
 | 
				
			||||||
uint64_t total_device;
 | 
					uint64_t total_device;
 | 
				
			||||||
@@ -35,12 +38,15 @@ void MemoryManager::PrintBytes(void)
 | 
				
			|||||||
  
 | 
					  
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					uint64_t MemoryManager::DeviceCacheBytes() { return CacheBytes[Acc] + CacheBytes[AccHuge] + CacheBytes[AccSmall]; }
 | 
				
			||||||
 | 
					uint64_t MemoryManager::HostCacheBytes()   { return CacheBytes[Cpu] + CacheBytes[CpuHuge] + CacheBytes[CpuSmall]; }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
//////////////////////////////////////////////////////////////////////
 | 
					//////////////////////////////////////////////////////////////////////
 | 
				
			||||||
// Data tables for recently freed pooiniter caches
 | 
					// Data tables for recently freed pooiniter caches
 | 
				
			||||||
//////////////////////////////////////////////////////////////////////
 | 
					//////////////////////////////////////////////////////////////////////
 | 
				
			||||||
MemoryManager::AllocationCacheEntry MemoryManager::Entries[MemoryManager::NallocType][MemoryManager::NallocCacheMax];
 | 
					MemoryManager::AllocationCacheEntry MemoryManager::Entries[MemoryManager::NallocType][MemoryManager::NallocCacheMax];
 | 
				
			||||||
int MemoryManager::Victim[MemoryManager::NallocType];
 | 
					int MemoryManager::Victim[MemoryManager::NallocType];
 | 
				
			||||||
int MemoryManager::Ncache[MemoryManager::NallocType] = { 2, 8, 8, 16, 8, 16 };
 | 
					int MemoryManager::Ncache[MemoryManager::NallocType] = { 2, 0, 8, 8, 0, 16, 8, 0, 16 };
 | 
				
			||||||
uint64_t MemoryManager::CacheBytes[MemoryManager::NallocType];
 | 
					uint64_t MemoryManager::CacheBytes[MemoryManager::NallocType];
 | 
				
			||||||
//////////////////////////////////////////////////////////////////////
 | 
					//////////////////////////////////////////////////////////////////////
 | 
				
			||||||
// Actual allocation and deallocation utils
 | 
					// Actual allocation and deallocation utils
 | 
				
			||||||
@@ -170,6 +176,16 @@ void MemoryManager::Init(void)
 | 
				
			|||||||
    }
 | 
					    }
 | 
				
			||||||
  }
 | 
					  }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					  str= getenv("GRID_ALLOC_NCACHE_HUGE");
 | 
				
			||||||
 | 
					  if ( str ) {
 | 
				
			||||||
 | 
					    Nc = atoi(str);
 | 
				
			||||||
 | 
					    if ( (Nc>=0) && (Nc < NallocCacheMax)) {
 | 
				
			||||||
 | 
					      Ncache[CpuHuge]=Nc;
 | 
				
			||||||
 | 
					      Ncache[AccHuge]=Nc;
 | 
				
			||||||
 | 
					      Ncache[SharedHuge]=Nc;
 | 
				
			||||||
 | 
					    }
 | 
				
			||||||
 | 
					  }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
  str= getenv("GRID_ALLOC_NCACHE_SMALL");
 | 
					  str= getenv("GRID_ALLOC_NCACHE_SMALL");
 | 
				
			||||||
  if ( str ) {
 | 
					  if ( str ) {
 | 
				
			||||||
    Nc = atoi(str);
 | 
					    Nc = atoi(str);
 | 
				
			||||||
@@ -190,7 +206,9 @@ void MemoryManager::InitMessage(void) {
 | 
				
			|||||||
  
 | 
					  
 | 
				
			||||||
  std::cout << GridLogMessage<< "MemoryManager::Init() setting up"<<std::endl;
 | 
					  std::cout << GridLogMessage<< "MemoryManager::Init() setting up"<<std::endl;
 | 
				
			||||||
#ifdef ALLOCATION_CACHE
 | 
					#ifdef ALLOCATION_CACHE
 | 
				
			||||||
  std::cout << GridLogMessage<< "MemoryManager::Init() cache pool for recent allocations: SMALL "<<Ncache[CpuSmall]<<" LARGE "<<Ncache[Cpu]<<std::endl;
 | 
					  std::cout << GridLogMessage<< "MemoryManager::Init() cache pool for recent host   allocations: SMALL "<<Ncache[CpuSmall]<<" LARGE "<<Ncache[Cpu]<<" HUGE "<<Ncache[CpuHuge]<<std::endl;
 | 
				
			||||||
 | 
					  std::cout << GridLogMessage<< "MemoryManager::Init() cache pool for recent device allocations: SMALL "<<Ncache[AccSmall]<<" LARGE "<<Ncache[Acc]<<" Huge "<<Ncache[AccHuge]<<std::endl;
 | 
				
			||||||
 | 
					  std::cout << GridLogMessage<< "MemoryManager::Init() cache pool for recent shared allocations: SMALL "<<Ncache[SharedSmall]<<" LARGE "<<Ncache[Shared]<<" Huge "<<Ncache[SharedHuge]<<std::endl;
 | 
				
			||||||
#endif
 | 
					#endif
 | 
				
			||||||
  
 | 
					  
 | 
				
			||||||
#ifdef GRID_UVM
 | 
					#ifdef GRID_UVM
 | 
				
			||||||
@@ -222,8 +240,11 @@ void MemoryManager::InitMessage(void) {
 | 
				
			|||||||
void *MemoryManager::Insert(void *ptr,size_t bytes,int type) 
 | 
					void *MemoryManager::Insert(void *ptr,size_t bytes,int type) 
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
#ifdef ALLOCATION_CACHE
 | 
					#ifdef ALLOCATION_CACHE
 | 
				
			||||||
  bool small = (bytes < GRID_ALLOC_SMALL_LIMIT);
 | 
					  int cache;
 | 
				
			||||||
  int cache = type + small;
 | 
					  if      (bytes < GRID_ALLOC_SMALL_LIMIT) cache = type + 2;
 | 
				
			||||||
 | 
					  else if (bytes >= GRID_ALLOC_HUGE_LIMIT) cache = type + 1;
 | 
				
			||||||
 | 
					  else                                     cache = type;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
  return Insert(ptr,bytes,Entries[cache],Ncache[cache],Victim[cache],CacheBytes[cache]);  
 | 
					  return Insert(ptr,bytes,Entries[cache],Ncache[cache],Victim[cache],CacheBytes[cache]);  
 | 
				
			||||||
#else
 | 
					#else
 | 
				
			||||||
  return ptr;
 | 
					  return ptr;
 | 
				
			||||||
@@ -232,11 +253,12 @@ void *MemoryManager::Insert(void *ptr,size_t bytes,int type)
 | 
				
			|||||||
 | 
					
 | 
				
			||||||
void *MemoryManager::Insert(void *ptr,size_t bytes,AllocationCacheEntry *entries,int ncache,int &victim, uint64_t &cacheBytes) 
 | 
					void *MemoryManager::Insert(void *ptr,size_t bytes,AllocationCacheEntry *entries,int ncache,int &victim, uint64_t &cacheBytes) 
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
  assert(ncache>0);
 | 
					 | 
				
			||||||
#ifdef GRID_OMP
 | 
					#ifdef GRID_OMP
 | 
				
			||||||
  assert(omp_in_parallel()==0);
 | 
					  assert(omp_in_parallel()==0);
 | 
				
			||||||
#endif 
 | 
					#endif 
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					  if (ncache == 0) return ptr;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
  void * ret = NULL;
 | 
					  void * ret = NULL;
 | 
				
			||||||
  int v = -1;
 | 
					  int v = -1;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
@@ -271,8 +293,11 @@ void *MemoryManager::Insert(void *ptr,size_t bytes,AllocationCacheEntry *entries
 | 
				
			|||||||
void *MemoryManager::Lookup(size_t bytes,int type)
 | 
					void *MemoryManager::Lookup(size_t bytes,int type)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
#ifdef ALLOCATION_CACHE
 | 
					#ifdef ALLOCATION_CACHE
 | 
				
			||||||
  bool small = (bytes < GRID_ALLOC_SMALL_LIMIT);
 | 
					  int cache;
 | 
				
			||||||
  int cache = type+small;
 | 
					  if      (bytes < GRID_ALLOC_SMALL_LIMIT) cache = type + 2;
 | 
				
			||||||
 | 
					  else if (bytes >= GRID_ALLOC_HUGE_LIMIT) cache = type + 1;
 | 
				
			||||||
 | 
					  else                                     cache = type;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
  return Lookup(bytes,Entries[cache],Ncache[cache],CacheBytes[cache]);
 | 
					  return Lookup(bytes,Entries[cache],Ncache[cache],CacheBytes[cache]);
 | 
				
			||||||
#else
 | 
					#else
 | 
				
			||||||
  return NULL;
 | 
					  return NULL;
 | 
				
			||||||
@@ -281,7 +306,6 @@ void *MemoryManager::Lookup(size_t bytes,int type)
 | 
				
			|||||||
 | 
					
 | 
				
			||||||
void *MemoryManager::Lookup(size_t bytes,AllocationCacheEntry *entries,int ncache,uint64_t & cacheBytes) 
 | 
					void *MemoryManager::Lookup(size_t bytes,AllocationCacheEntry *entries,int ncache,uint64_t & cacheBytes) 
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
  assert(ncache>0);
 | 
					 | 
				
			||||||
#ifdef GRID_OMP
 | 
					#ifdef GRID_OMP
 | 
				
			||||||
  assert(omp_in_parallel()==0);
 | 
					  assert(omp_in_parallel()==0);
 | 
				
			||||||
#endif 
 | 
					#endif 
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -35,6 +35,7 @@ NAMESPACE_BEGIN(Grid);
 | 
				
			|||||||
// Move control to configure.ac and Config.h?
 | 
					// Move control to configure.ac and Config.h?
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#define GRID_ALLOC_SMALL_LIMIT (4096)
 | 
					#define GRID_ALLOC_SMALL_LIMIT (4096)
 | 
				
			||||||
 | 
					#define GRID_ALLOC_HUGE_LIMIT  (2147483648)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#define STRINGIFY(x) #x
 | 
					#define STRINGIFY(x) #x
 | 
				
			||||||
#define TOSTRING(x) STRINGIFY(x)
 | 
					#define TOSTRING(x) STRINGIFY(x)
 | 
				
			||||||
@@ -70,6 +71,21 @@ enum ViewMode {
 | 
				
			|||||||
  CpuWriteDiscard = 0x10 // same for now
 | 
					  CpuWriteDiscard = 0x10 // same for now
 | 
				
			||||||
};
 | 
					};
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					struct MemoryStatus {
 | 
				
			||||||
 | 
					  uint64_t     DeviceBytes;
 | 
				
			||||||
 | 
					  uint64_t     DeviceLRUBytes;
 | 
				
			||||||
 | 
					  uint64_t     DeviceMaxBytes;
 | 
				
			||||||
 | 
					  uint64_t     HostToDeviceBytes;
 | 
				
			||||||
 | 
					  uint64_t     DeviceToHostBytes;
 | 
				
			||||||
 | 
					  uint64_t     HostToDeviceXfer;
 | 
				
			||||||
 | 
					  uint64_t     DeviceToHostXfer;
 | 
				
			||||||
 | 
					  uint64_t     DeviceEvictions;
 | 
				
			||||||
 | 
					  uint64_t     DeviceDestroy;
 | 
				
			||||||
 | 
					  uint64_t     DeviceAllocCacheBytes;
 | 
				
			||||||
 | 
					  uint64_t     HostAllocCacheBytes;
 | 
				
			||||||
 | 
					};
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
class MemoryManager {
 | 
					class MemoryManager {
 | 
				
			||||||
private:
 | 
					private:
 | 
				
			||||||
 | 
					
 | 
				
			||||||
@@ -83,7 +99,7 @@ private:
 | 
				
			|||||||
  } AllocationCacheEntry;
 | 
					  } AllocationCacheEntry;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
  static const int NallocCacheMax=128; 
 | 
					  static const int NallocCacheMax=128; 
 | 
				
			||||||
  static const int NallocType=6;
 | 
					  static const int NallocType=9;
 | 
				
			||||||
  static AllocationCacheEntry Entries[NallocType][NallocCacheMax];
 | 
					  static AllocationCacheEntry Entries[NallocType][NallocCacheMax];
 | 
				
			||||||
  static int Victim[NallocType];
 | 
					  static int Victim[NallocType];
 | 
				
			||||||
  static int Ncache[NallocType];
 | 
					  static int Ncache[NallocType];
 | 
				
			||||||
@@ -121,7 +137,26 @@ private:
 | 
				
			|||||||
  static uint64_t     DeviceToHostXfer;
 | 
					  static uint64_t     DeviceToHostXfer;
 | 
				
			||||||
  static uint64_t     DeviceEvictions;
 | 
					  static uint64_t     DeviceEvictions;
 | 
				
			||||||
  static uint64_t     DeviceDestroy;
 | 
					  static uint64_t     DeviceDestroy;
 | 
				
			||||||
 
 | 
					  
 | 
				
			||||||
 | 
					  static uint64_t     DeviceCacheBytes();
 | 
				
			||||||
 | 
					  static uint64_t     HostCacheBytes();
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					  static MemoryStatus GetFootprint(void) {
 | 
				
			||||||
 | 
					    MemoryStatus stat;
 | 
				
			||||||
 | 
					    stat.DeviceBytes       = DeviceBytes;
 | 
				
			||||||
 | 
					    stat.DeviceLRUBytes    = DeviceLRUBytes;
 | 
				
			||||||
 | 
					    stat.DeviceMaxBytes    = DeviceMaxBytes;
 | 
				
			||||||
 | 
					    stat.HostToDeviceBytes = HostToDeviceBytes;
 | 
				
			||||||
 | 
					    stat.DeviceToHostBytes = DeviceToHostBytes;
 | 
				
			||||||
 | 
					    stat.HostToDeviceXfer  = HostToDeviceXfer;
 | 
				
			||||||
 | 
					    stat.DeviceToHostXfer  = DeviceToHostXfer;
 | 
				
			||||||
 | 
					    stat.DeviceEvictions   = DeviceEvictions;
 | 
				
			||||||
 | 
					    stat.DeviceDestroy     = DeviceDestroy;
 | 
				
			||||||
 | 
					    stat.DeviceAllocCacheBytes = DeviceCacheBytes();
 | 
				
			||||||
 | 
					    stat.HostAllocCacheBytes   = HostCacheBytes();
 | 
				
			||||||
 | 
					    return stat;
 | 
				
			||||||
 | 
					  };
 | 
				
			||||||
 | 
					  
 | 
				
			||||||
 private:
 | 
					 private:
 | 
				
			||||||
#ifndef GRID_UVM
 | 
					#ifndef GRID_UVM
 | 
				
			||||||
  //////////////////////////////////////////////////////////////////////
 | 
					  //////////////////////////////////////////////////////////////////////
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -400,9 +400,6 @@ double CartesianCommunicator::StencilSendToRecvFromBegin(std::vector<CommsReques
 | 
				
			|||||||
}
 | 
					}
 | 
				
			||||||
void CartesianCommunicator::StencilSendToRecvFromComplete(std::vector<CommsRequest_t> &list,int dir)
 | 
					void CartesianCommunicator::StencilSendToRecvFromComplete(std::vector<CommsRequest_t> &list,int dir)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
  acceleratorCopySynchronise();
 | 
					 | 
				
			||||||
  StencilBarrier();// Synch shared memory on a single nodes
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
  int nreq=list.size();
 | 
					  int nreq=list.size();
 | 
				
			||||||
 | 
					
 | 
				
			||||||
  if (nreq==0) return;
 | 
					  if (nreq==0) return;
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -37,9 +37,11 @@ Author: Christoph Lehner <christoph@lhnr.de>
 | 
				
			|||||||
#ifdef GRID_HIP
 | 
					#ifdef GRID_HIP
 | 
				
			||||||
#include <hip/hip_runtime_api.h>
 | 
					#include <hip/hip_runtime_api.h>
 | 
				
			||||||
#endif
 | 
					#endif
 | 
				
			||||||
#ifdef GRID_SYCl
 | 
					#ifdef GRID_SYCL
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#endif
 | 
					#endif
 | 
				
			||||||
 | 
					#define GRID_SYCL_LEVEL_ZERO_IPC
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
NAMESPACE_BEGIN(Grid); 
 | 
					NAMESPACE_BEGIN(Grid); 
 | 
				
			||||||
#define header "SharedMemoryMpi: "
 | 
					#define header "SharedMemoryMpi: "
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -297,6 +297,30 @@ template<class vobj> void Scatter_plane_merge(Lattice<vobj> &rhs,ExtractPointerA
 | 
				
			|||||||
  }
 | 
					  }
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					#if (defined(GRID_CUDA) || defined(GRID_HIP)) && defined(ACCELERATOR_CSHIFT)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					template <typename T>
 | 
				
			||||||
 | 
					T iDivUp(T a, T b) // Round a / b to nearest higher integer value
 | 
				
			||||||
 | 
					{ return (a % b != 0) ? (a / b + 1) : (a / b); }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					template <typename T>
 | 
				
			||||||
 | 
					__global__ void populate_Cshift_table(T* vector, T lo, T ro, T e1, T e2, T stride)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
					    int idx = blockIdx.x*blockDim.x + threadIdx.x;
 | 
				
			||||||
 | 
					    if (idx >= e1*e2) return;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    int n, b, o;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    n = idx / e2;
 | 
				
			||||||
 | 
					    b = idx % e2;
 | 
				
			||||||
 | 
					    o = n*stride + b;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    vector[2*idx + 0] = lo + o;
 | 
				
			||||||
 | 
					    vector[2*idx + 1] = ro + o;
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					#endif
 | 
				
			||||||
 | 
					
 | 
				
			||||||
//////////////////////////////////////////////////////
 | 
					//////////////////////////////////////////////////////
 | 
				
			||||||
// local to node block strided copies
 | 
					// local to node block strided copies
 | 
				
			||||||
//////////////////////////////////////////////////////
 | 
					//////////////////////////////////////////////////////
 | 
				
			||||||
@@ -321,12 +345,20 @@ template<class vobj> void Copy_plane(Lattice<vobj>& lhs,const Lattice<vobj> &rhs
 | 
				
			|||||||
  int ent=0;
 | 
					  int ent=0;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
  if(cbmask == 0x3 ){
 | 
					  if(cbmask == 0x3 ){
 | 
				
			||||||
 | 
					#if (defined(GRID_CUDA) || defined(GRID_HIP)) && defined(ACCELERATOR_CSHIFT)
 | 
				
			||||||
 | 
					    ent = e1*e2;
 | 
				
			||||||
 | 
					    dim3 blockSize(acceleratorThreads());
 | 
				
			||||||
 | 
					    dim3 gridSize(iDivUp((unsigned int)ent, blockSize.x));
 | 
				
			||||||
 | 
					    populate_Cshift_table<<<gridSize, blockSize>>>(&Cshift_table[0].first, lo, ro, e1, e2, stride);
 | 
				
			||||||
 | 
					    accelerator_barrier();
 | 
				
			||||||
 | 
					#else
 | 
				
			||||||
    for(int n=0;n<e1;n++){
 | 
					    for(int n=0;n<e1;n++){
 | 
				
			||||||
      for(int b=0;b<e2;b++){
 | 
					      for(int b=0;b<e2;b++){
 | 
				
			||||||
        int o =n*stride+b;
 | 
					        int o =n*stride+b;
 | 
				
			||||||
	Cshift_table[ent++] = std::pair<int,int>(lo+o,ro+o);
 | 
						Cshift_table[ent++] = std::pair<int,int>(lo+o,ro+o);
 | 
				
			||||||
      }
 | 
					      }
 | 
				
			||||||
    }
 | 
					    }
 | 
				
			||||||
 | 
					#endif
 | 
				
			||||||
  } else { 
 | 
					  } else { 
 | 
				
			||||||
    for(int n=0;n<e1;n++){
 | 
					    for(int n=0;n<e1;n++){
 | 
				
			||||||
      for(int b=0;b<e2;b++){
 | 
					      for(int b=0;b<e2;b++){
 | 
				
			||||||
@@ -377,11 +409,19 @@ template<class vobj> void Copy_plane_permute(Lattice<vobj>& lhs,const Lattice<vo
 | 
				
			|||||||
  int ent=0;
 | 
					  int ent=0;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
  if ( cbmask == 0x3 ) {
 | 
					  if ( cbmask == 0x3 ) {
 | 
				
			||||||
 | 
					#if (defined(GRID_CUDA) || defined(GRID_HIP)) && defined(ACCELERATOR_CSHIFT)
 | 
				
			||||||
 | 
					    ent = e1*e2;
 | 
				
			||||||
 | 
					    dim3 blockSize(acceleratorThreads());
 | 
				
			||||||
 | 
					    dim3 gridSize(iDivUp((unsigned int)ent, blockSize.x));
 | 
				
			||||||
 | 
					    populate_Cshift_table<<<gridSize, blockSize>>>(&Cshift_table[0].first, lo, ro, e1, e2, stride);
 | 
				
			||||||
 | 
					    accelerator_barrier();
 | 
				
			||||||
 | 
					#else
 | 
				
			||||||
    for(int n=0;n<e1;n++){
 | 
					    for(int n=0;n<e1;n++){
 | 
				
			||||||
    for(int b=0;b<e2;b++){
 | 
					    for(int b=0;b<e2;b++){
 | 
				
			||||||
      int o  =n*stride;
 | 
					      int o  =n*stride;
 | 
				
			||||||
      Cshift_table[ent++] = std::pair<int,int>(lo+o+b,ro+o+b);
 | 
					      Cshift_table[ent++] = std::pair<int,int>(lo+o+b,ro+o+b);
 | 
				
			||||||
    }}
 | 
					    }}
 | 
				
			||||||
 | 
					#endif
 | 
				
			||||||
  } else {
 | 
					  } else {
 | 
				
			||||||
    for(int n=0;n<e1;n++){
 | 
					    for(int n=0;n<e1;n++){
 | 
				
			||||||
    for(int b=0;b<e2;b++){
 | 
					    for(int b=0;b<e2;b++){
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -153,33 +153,44 @@ inline typename vobj::scalar_objectD sumD_large(const vobj *arg, Integer osites)
 | 
				
			|||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
template<class vobj>
 | 
					template<class vobj>
 | 
				
			||||||
inline typename vobj::scalar_object sum(const Lattice<vobj> &arg)
 | 
					inline typename vobj::scalar_object rankSum(const Lattice<vobj> &arg)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
  Integer osites = arg.Grid()->oSites();
 | 
					  Integer osites = arg.Grid()->oSites();
 | 
				
			||||||
#if defined(GRID_CUDA)||defined(GRID_HIP)||defined(GRID_SYCL)
 | 
					#if defined(GRID_CUDA)||defined(GRID_HIP)||defined(GRID_SYCL)
 | 
				
			||||||
  typename vobj::scalar_object ssum;
 | 
					 | 
				
			||||||
  autoView( arg_v, arg, AcceleratorRead);
 | 
					  autoView( arg_v, arg, AcceleratorRead);
 | 
				
			||||||
  ssum= sum_gpu(&arg_v[0],osites);
 | 
					  return sum_gpu(&arg_v[0],osites);
 | 
				
			||||||
#else
 | 
					#else
 | 
				
			||||||
  autoView(arg_v, arg, CpuRead);
 | 
					  autoView(arg_v, arg, CpuRead);
 | 
				
			||||||
  auto ssum= sum_cpu(&arg_v[0],osites);
 | 
					  return sum_cpu(&arg_v[0],osites);
 | 
				
			||||||
#endif  
 | 
					#endif  
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					template<class vobj>
 | 
				
			||||||
 | 
					inline typename vobj::scalar_object sum(const Lattice<vobj> &arg)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
					  auto ssum = rankSum(arg);
 | 
				
			||||||
  arg.Grid()->GlobalSum(ssum);
 | 
					  arg.Grid()->GlobalSum(ssum);
 | 
				
			||||||
  return ssum;
 | 
					  return ssum;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
template<class vobj>
 | 
					template<class vobj>
 | 
				
			||||||
inline typename vobj::scalar_object sum_large(const Lattice<vobj> &arg)
 | 
					inline typename vobj::scalar_object rankSumLarge(const Lattice<vobj> &arg)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
#if defined(GRID_CUDA)||defined(GRID_HIP)||defined(GRID_SYCL)
 | 
					#if defined(GRID_CUDA)||defined(GRID_HIP)||defined(GRID_SYCL)
 | 
				
			||||||
  autoView( arg_v, arg, AcceleratorRead);
 | 
					  autoView( arg_v, arg, AcceleratorRead);
 | 
				
			||||||
  Integer osites = arg.Grid()->oSites();
 | 
					  Integer osites = arg.Grid()->oSites();
 | 
				
			||||||
  auto ssum= sum_gpu_large(&arg_v[0],osites);
 | 
					  return sum_gpu_large(&arg_v[0],osites);
 | 
				
			||||||
#else
 | 
					#else
 | 
				
			||||||
  autoView(arg_v, arg, CpuRead);
 | 
					  autoView(arg_v, arg, CpuRead);
 | 
				
			||||||
  Integer osites = arg.Grid()->oSites();
 | 
					  Integer osites = arg.Grid()->oSites();
 | 
				
			||||||
  auto ssum= sum_cpu(&arg_v[0],osites);
 | 
					  return sum_cpu(&arg_v[0],osites);
 | 
				
			||||||
#endif
 | 
					#endif
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					template<class vobj>
 | 
				
			||||||
 | 
					inline typename vobj::scalar_object sum_large(const Lattice<vobj> &arg)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
					  auto ssum = rankSumLarge(arg);
 | 
				
			||||||
  arg.Grid()->GlobalSum(ssum);
 | 
					  arg.Grid()->GlobalSum(ssum);
 | 
				
			||||||
  return ssum;
 | 
					  return ssum;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -288,7 +288,36 @@ inline void blockProject(Lattice<iVector<CComplex,nbasis > > &coarseData,
 | 
				
			|||||||
    blockZAXPY(fineDataRed,ip,Basis[v],fineDataRed); 
 | 
					    blockZAXPY(fineDataRed,ip,Basis[v],fineDataRed); 
 | 
				
			||||||
  }
 | 
					  }
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					template<class vobj,class CComplex,int nbasis,class VLattice>
 | 
				
			||||||
 | 
					inline void batchBlockProject(std::vector<Lattice<iVector<CComplex,nbasis>>> &coarseData,
 | 
				
			||||||
 | 
					                               const std::vector<Lattice<vobj>> &fineData,
 | 
				
			||||||
 | 
					                               const VLattice &Basis)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
					  int NBatch = fineData.size();
 | 
				
			||||||
 | 
					  assert(coarseData.size() == NBatch);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					  GridBase * fine  = fineData[0].Grid();
 | 
				
			||||||
 | 
					  GridBase * coarse= coarseData[0].Grid();
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					  Lattice<iScalar<CComplex>> ip(coarse);
 | 
				
			||||||
 | 
					  std::vector<Lattice<vobj>> fineDataCopy = fineData;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					  autoView(ip_, ip, AcceleratorWrite);
 | 
				
			||||||
 | 
					  for(int v=0;v<nbasis;v++) {
 | 
				
			||||||
 | 
					    for (int k=0; k<NBatch; k++) {
 | 
				
			||||||
 | 
					      autoView( coarseData_ , coarseData[k], AcceleratorWrite);
 | 
				
			||||||
 | 
					      blockInnerProductD(ip,Basis[v],fineDataCopy[k]); // ip = <basis|fine>
 | 
				
			||||||
 | 
					      accelerator_for( sc, coarse->oSites(), vobj::Nsimd(), {
 | 
				
			||||||
 | 
					        convertType(coarseData_[sc](v),ip_[sc]);
 | 
				
			||||||
 | 
					      });
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					      // improve numerical stability of projection
 | 
				
			||||||
 | 
					      // |fine> = |fine> - <basis|fine> |basis>
 | 
				
			||||||
 | 
					      ip=-ip;
 | 
				
			||||||
 | 
					      blockZAXPY(fineDataCopy[k],ip,Basis[v],fineDataCopy[k]); 
 | 
				
			||||||
 | 
					    }
 | 
				
			||||||
 | 
					  }
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
template<class vobj,class vobj2,class CComplex>
 | 
					template<class vobj,class vobj2,class CComplex>
 | 
				
			||||||
  inline void blockZAXPY(Lattice<vobj> &fineZ,
 | 
					  inline void blockZAXPY(Lattice<vobj> &fineZ,
 | 
				
			||||||
@@ -590,6 +619,26 @@ inline void blockPromote(const Lattice<iVector<CComplex,nbasis > > &coarseData,
 | 
				
			|||||||
}
 | 
					}
 | 
				
			||||||
#endif
 | 
					#endif
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					template<class vobj,class CComplex,int nbasis,class VLattice>
 | 
				
			||||||
 | 
					inline void batchBlockPromote(const std::vector<Lattice<iVector<CComplex,nbasis>>> &coarseData,
 | 
				
			||||||
 | 
					                               std::vector<Lattice<vobj>> &fineData,
 | 
				
			||||||
 | 
					                               const VLattice &Basis)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
					  int NBatch = coarseData.size();
 | 
				
			||||||
 | 
					  assert(fineData.size() == NBatch);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					  GridBase * fine   = fineData[0].Grid();
 | 
				
			||||||
 | 
					  GridBase * coarse = coarseData[0].Grid();
 | 
				
			||||||
 | 
					  for (int k=0; k<NBatch; k++)
 | 
				
			||||||
 | 
					    fineData[k]=Zero();
 | 
				
			||||||
 | 
					  for (int i=0;i<nbasis;i++) {
 | 
				
			||||||
 | 
					    for (int k=0; k<NBatch; k++) {
 | 
				
			||||||
 | 
					      Lattice<iScalar<CComplex>> ip = PeekIndex<0>(coarseData[k],i);
 | 
				
			||||||
 | 
					      blockZAXPY(fineData[k],ip,Basis[i],fineData[k]);
 | 
				
			||||||
 | 
					    }
 | 
				
			||||||
 | 
					  }
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
// Useful for precision conversion, or indeed anything where an operator= does a conversion on scalars.
 | 
					// Useful for precision conversion, or indeed anything where an operator= does a conversion on scalars.
 | 
				
			||||||
// Simd layouts need not match since we use peek/poke Local
 | 
					// Simd layouts need not match since we use peek/poke Local
 | 
				
			||||||
template<class vobj,class vvobj>
 | 
					template<class vobj,class vvobj>
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -434,7 +434,6 @@ public:
 | 
				
			|||||||
  ////////////////////////////////////////////////////////////////////////
 | 
					  ////////////////////////////////////////////////////////////////////////
 | 
				
			||||||
  void CommunicateBegin(std::vector<std::vector<CommsRequest_t> > &reqs)
 | 
					  void CommunicateBegin(std::vector<std::vector<CommsRequest_t> > &reqs)
 | 
				
			||||||
  {
 | 
					  {
 | 
				
			||||||
    accelerator_barrier();
 | 
					 | 
				
			||||||
    for(int i=0;i<Packets.size();i++){
 | 
					    for(int i=0;i<Packets.size();i++){
 | 
				
			||||||
      _grid->StencilSendToRecvFromBegin(MpiReqs,
 | 
					      _grid->StencilSendToRecvFromBegin(MpiReqs,
 | 
				
			||||||
					Packets[i].send_buf,
 | 
										Packets[i].send_buf,
 | 
				
			||||||
@@ -443,7 +442,6 @@ public:
 | 
				
			|||||||
					Packets[i].from_rank,Packets[i].do_recv,
 | 
										Packets[i].from_rank,Packets[i].do_recv,
 | 
				
			||||||
					Packets[i].xbytes,Packets[i].rbytes,i);
 | 
										Packets[i].xbytes,Packets[i].rbytes,i);
 | 
				
			||||||
    }
 | 
					    }
 | 
				
			||||||
    _grid->StencilBarrier();// Synch shared memory on a single nodes
 | 
					 | 
				
			||||||
  }
 | 
					  }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
  void CommunicateComplete(std::vector<std::vector<CommsRequest_t> > &reqs)
 | 
					  void CommunicateComplete(std::vector<std::vector<CommsRequest_t> > &reqs)
 | 
				
			||||||
@@ -452,6 +450,9 @@ public:
 | 
				
			|||||||
    if   ( this->partialDirichlet ) DslashLogPartial();
 | 
					    if   ( this->partialDirichlet ) DslashLogPartial();
 | 
				
			||||||
    else if ( this->fullDirichlet ) DslashLogDirichlet();
 | 
					    else if ( this->fullDirichlet ) DslashLogDirichlet();
 | 
				
			||||||
    else DslashLogFull();
 | 
					    else DslashLogFull();
 | 
				
			||||||
 | 
					    acceleratorCopySynchronise();
 | 
				
			||||||
 | 
					    // Everyone agrees we are all done
 | 
				
			||||||
 | 
					    _grid->StencilBarrier(); 
 | 
				
			||||||
  }
 | 
					  }
 | 
				
			||||||
  ////////////////////////////////////////////////////////////////////////
 | 
					  ////////////////////////////////////////////////////////////////////////
 | 
				
			||||||
  // Blocking send and receive. Either sequential or parallel.
 | 
					  // Blocking send and receive. Either sequential or parallel.
 | 
				
			||||||
@@ -529,7 +530,6 @@ public:
 | 
				
			|||||||
  {
 | 
					  {
 | 
				
			||||||
    _grid->StencilBarrier();// Synch shared memory on a single nodes
 | 
					    _grid->StencilBarrier();// Synch shared memory on a single nodes
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    // conformable(source.Grid(),_grid);
 | 
					 | 
				
			||||||
    assert(source.Grid()==_grid);
 | 
					    assert(source.Grid()==_grid);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    u_comm_offset=0;
 | 
					    u_comm_offset=0;
 | 
				
			||||||
@@ -655,8 +655,8 @@ public:
 | 
				
			|||||||
    CommsMerge(decompress,Mergers,Decompressions);
 | 
					    CommsMerge(decompress,Mergers,Decompressions);
 | 
				
			||||||
  }
 | 
					  }
 | 
				
			||||||
  template<class decompressor>  void CommsMergeSHM(decompressor decompress) {
 | 
					  template<class decompressor>  void CommsMergeSHM(decompressor decompress) {
 | 
				
			||||||
    _grid->StencilBarrier();// Synch shared memory on a single nodes
 | 
					    assert(MergersSHM.size()==0);
 | 
				
			||||||
    CommsMerge(decompress,MergersSHM,DecompressionsSHM);
 | 
					    assert(DecompressionsSHM.size()==0);
 | 
				
			||||||
  }
 | 
					  }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
  template<class decompressor>
 | 
					  template<class decompressor>
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -4,7 +4,7 @@
 | 
				
			|||||||
#SBATCH -p QZ1J-ICX-PVC
 | 
					#SBATCH -p QZ1J-ICX-PVC
 | 
				
			||||||
##SBATCH -p QZ1J-SPR-PVC-2C
 | 
					##SBATCH -p QZ1J-SPR-PVC-2C
 | 
				
			||||||
 | 
					
 | 
				
			||||||
source /nfs/site/home/paboylex/ATS/GridNew/Grid/systems/PVC-nightly/setup.sh
 | 
					#source /nfs/site/home/paboylex/ATS/GridNew/Grid/systems/PVC-nightly/setup.sh
 | 
				
			||||||
 | 
					
 | 
				
			||||||
export NT=8
 | 
					export NT=8
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -4,7 +4,7 @@
 | 
				
			|||||||
 | 
					
 | 
				
			||||||
#SBATCH -p QZ1J-ICX-PVC
 | 
					#SBATCH -p QZ1J-ICX-PVC
 | 
				
			||||||
 | 
					
 | 
				
			||||||
source /nfs/site/home/paboylex/ATS/GridNew/Grid/systems/PVC-nightly/setup.sh
 | 
					#source /nfs/site/home/paboylex/ATS/GridNew/Grid/systems/PVC-nightly/setup.sh
 | 
				
			||||||
 | 
					
 | 
				
			||||||
export NT=16
 | 
					export NT=16
 | 
				
			||||||
 | 
					
 | 
				
			||||||
@@ -19,11 +19,15 @@ export SYCL_DEVICE_FILTER=gpu,level_zero
 | 
				
			|||||||
export I_MPI_OFFLOAD_CELL=tile
 | 
					export I_MPI_OFFLOAD_CELL=tile
 | 
				
			||||||
export EnableImplicitScaling=0
 | 
					export EnableImplicitScaling=0
 | 
				
			||||||
export EnableWalkerPartition=0
 | 
					export EnableWalkerPartition=0
 | 
				
			||||||
export SYCL_PI_LEVEL_ZERO_DEVICE_SCOPE_EVENTS=1
 | 
					#export SYCL_PI_LEVEL_ZERO_DEVICE_SCOPE_EVENTS=1
 | 
				
			||||||
export SYCL_PI_LEVEL_ZERO_USE_IMMEDIATE_COMMANDLISTS=1
 | 
					#export SYCL_PI_LEVEL_ZERO_USE_IMMEDIATE_COMMANDLISTS=1
 | 
				
			||||||
export SYCL_PI_LEVEL_ZERO_USE_COPY_ENGINE=0
 | 
					export SYCL_PI_LEVEL_ZERO_USE_COPY_ENGINE=0
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#mpiexec -launcher ssh -n 1 -host localhost  ./wrap.sh ./Benchmark_dwf_fp32 --mpi 1.1.1.1 --grid 32.32.32.32 --accelerator-threads $NT --comms-sequential --shm-mpi 0 > 1tile.log
 | 
					for i in 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
 | 
				
			||||||
 | 
					do
 | 
				
			||||||
 | 
					mpiexec -launcher ssh -n 2 -host localhost  ./wrap.sh ./Benchmark_dwf_fp32 --mpi 1.1.1.2 --grid 32.32.32.64 --accelerator-threads $NT  --shm-mpi 0  --device-mem 32768 > 1.1.1.2.log$i
 | 
				
			||||||
 | 
					mpiexec -launcher ssh -n 2 -host localhost  ./wrap.sh ./Benchmark_dwf_fp32 --mpi 2.1.1.1 --grid 64.32.32.32 --accelerator-threads $NT  --shm-mpi 0  --device-mem 32768 > 2.1.1.1.log$i 
 | 
				
			||||||
 | 
					done
 | 
				
			||||||
 | 
					
 | 
				
			||||||
mpiexec -launcher ssh -n 2 -host localhost  ./wrap.sh ./Benchmark_dwf_fp32 --mpi 2.1.1.1 --grid 64.32.32.32 --accelerator-threads $NT --comms-sequential --shm-mpi 0
 | 
					mpiexec -launcher ssh -n 2 -host localhost  ./wrap.sh ./Benchmark_dwf_fp32 --mpi 2.1.1.1 --grid 64.32.32.32 --accelerator-threads $NT --comms-sequential --shm-mpi 0
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -5,10 +5,5 @@ export ZE_AFFINITY_MASK=0.$MPI_LOCALRANKID
 | 
				
			|||||||
echo Ranke $MPI_LOCALRANKID ZE_AFFINITY_MASK is $ZE_AFFINITY_MASK
 | 
					echo Ranke $MPI_LOCALRANKID ZE_AFFINITY_MASK is $ZE_AFFINITY_MASK
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#if [ $MPI_LOCALRANKID = "0" ] 
 | 
					 | 
				
			||||||
#then
 | 
					 | 
				
			||||||
#  ~psteinbr/build_pti/ze_tracer -c $@
 | 
					 | 
				
			||||||
#  onetrace --chrome-kernel-timeline $@
 | 
					 | 
				
			||||||
#else
 | 
					 | 
				
			||||||
  $@
 | 
					  $@
 | 
				
			||||||
#fi
 | 
					
 | 
				
			||||||
 
 | 
				
			|||||||
		Reference in New Issue
	
	Block a user