mirror of
https://github.com/paboyle/Grid.git
synced 2024-11-09 23:45:36 +00:00
RNG seed change safer for large volumes; this is a long term solution
This commit is contained in:
parent
91cf5ee312
commit
7019916294
@ -152,6 +152,7 @@ public:
|
|||||||
#ifdef RNG_FAST_DISCARD
|
#ifdef RNG_FAST_DISCARD
|
||||||
static void Skip(RngEngine &eng,uint64_t site)
|
static void Skip(RngEngine &eng,uint64_t site)
|
||||||
{
|
{
|
||||||
|
#if 0
|
||||||
/////////////////////////////////////////////////////////////////////////////////////
|
/////////////////////////////////////////////////////////////////////////////////////
|
||||||
// Skip by 2^40 elements between successive lattice sites
|
// Skip by 2^40 elements between successive lattice sites
|
||||||
// This goes by 10^12.
|
// This goes by 10^12.
|
||||||
@ -162,9 +163,9 @@ public:
|
|||||||
// tens of seconds per trajectory so this is clean in all reasonable cases,
|
// tens of seconds per trajectory so this is clean in all reasonable cases,
|
||||||
// and margin of safety is orders of magnitude.
|
// and margin of safety is orders of magnitude.
|
||||||
// We could hack Sitmo to skip in the higher order words of state if necessary
|
// We could hack Sitmo to skip in the higher order words of state if necessary
|
||||||
//
|
//
|
||||||
// Replace with 2^30 ; avoid problem on large volumes
|
// Replace with 2^30 ; avoid problem on large volumes
|
||||||
//
|
//
|
||||||
/////////////////////////////////////////////////////////////////////////////////////
|
/////////////////////////////////////////////////////////////////////////////////////
|
||||||
// uint64_t skip = site+1; // Old init Skipped then drew. Checked compat with faster init
|
// uint64_t skip = site+1; // Old init Skipped then drew. Checked compat with faster init
|
||||||
const int shift = 30;
|
const int shift = 30;
|
||||||
@ -179,6 +180,9 @@ public:
|
|||||||
assert((skip >> shift)==site); // check for overflow
|
assert((skip >> shift)==site); // check for overflow
|
||||||
|
|
||||||
eng.discard(skip);
|
eng.discard(skip);
|
||||||
|
#else
|
||||||
|
eng.discardhi(site);
|
||||||
|
#endif
|
||||||
// std::cout << " Engine " <<site << " state " <<eng<<std::endl;
|
// std::cout << " Engine " <<site << " state " <<eng<<std::endl;
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
@ -218,6 +218,10 @@ public:
|
|||||||
// -------------------------------------------------
|
// -------------------------------------------------
|
||||||
// misc
|
// misc
|
||||||
// -------------------------------------------------
|
// -------------------------------------------------
|
||||||
|
void discardhi(uint64_t z) {
|
||||||
|
_s[3] += z;
|
||||||
|
encrypt_counter();
|
||||||
|
}
|
||||||
|
|
||||||
// req: 26.5.1.4 Random number engine requirements, p.908 table 117, row 9
|
// req: 26.5.1.4 Random number engine requirements, p.908 table 117, row 9
|
||||||
// Advances e’s state ei to ei+z by any means equivalent to z
|
// Advances e’s state ei to ei+z by any means equivalent to z
|
||||||
|
@ -1,51 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
|
|
||||||
## qsub -q EarlyAppAccess -A Aurora_Deployment -I -l select=1 -l walltime=60:00
|
|
||||||
#PBS -q EarlyAppAccess
|
|
||||||
#PBS -l select=1
|
|
||||||
#PBS -l walltime=01:00:00
|
|
||||||
#PBS -A LatticeQCD_aesp_CNDA
|
|
||||||
|
|
||||||
HDIR=/home/paboyle/
|
|
||||||
#module use /soft/testing/modulefiles/
|
|
||||||
#module load intel-UMD23.05.25593.11/23.05.25593.11
|
|
||||||
#module load tools/pti-gpu
|
|
||||||
#export LD_LIBRARY_PATH=$HDIR/tools/lib64:$LD_LIBRARY_PATH
|
|
||||||
#export PATH=$HDIR/tools/bin:$PATH
|
|
||||||
|
|
||||||
export TZ='/usr/share/zoneinfo/US/Central'
|
|
||||||
export OMP_PROC_BIND=spread
|
|
||||||
export OMP_NUM_THREADS=3
|
|
||||||
unset OMP_PLACES
|
|
||||||
|
|
||||||
cd $PBS_O_WORKDIR
|
|
||||||
|
|
||||||
source ../sourceme.sh
|
|
||||||
|
|
||||||
echo Jobid: $PBS_JOBID
|
|
||||||
echo Running on host `hostname`
|
|
||||||
echo Running on nodes `cat $PBS_NODEFILE`
|
|
||||||
|
|
||||||
echo NODES
|
|
||||||
cat $PBS_NODEFILE
|
|
||||||
NNODES=`wc -l < $PBS_NODEFILE`
|
|
||||||
NRANKS=12 # Number of MPI ranks per node
|
|
||||||
NDEPTH=4 # Number of hardware threads per rank, spacing between MPI ranks on a node
|
|
||||||
NTHREADS=$OMP_NUM_THREADS # Number of OMP threads per rank, given to OMP_NUM_THREADS
|
|
||||||
|
|
||||||
NTOTRANKS=$(( NNODES * NRANKS ))
|
|
||||||
|
|
||||||
echo "NUM_NODES=${NNODES} TOTAL_RANKS=${NTOTRANKS} RANKS_PER_NODE=${NRANKS} THREADS_PER_RANK=${OMP_NUM_THREADS}"
|
|
||||||
echo "OMP_PROC_BIND=$OMP_PROC_BIND OMP_PLACES=$OMP_PLACES"
|
|
||||||
|
|
||||||
|
|
||||||
#CMD="mpiexec -np ${NTOTRANKS} -ppn ${NRANKS} -d ${NDEPTH} --cpu-bind=depth -envall \
|
|
||||||
# ./gpu_tile_compact.sh \
|
|
||||||
# ./Benchmark_dwf_fp32 --mpi 1.1.2.6 --grid 16.32.64.192 --comms-overlap \
|
|
||||||
# --shm-mpi 0 --shm 2048 --device-mem 32000 --accelerator-threads 32"
|
|
||||||
|
|
||||||
CMD="mpiexec -np ${NTOTRANKS} -ppn ${NRANKS} -d ${NDEPTH} --cpu-bind=depth -envall \
|
|
||||||
./gpu_tile_compact.sh \
|
|
||||||
./Benchmark_memory_bandwidth --mpi 1.1.2.6 --grid 32.24.32.192 \
|
|
||||||
--shm-mpi 1 --shm 2048 --device-mem 32000 --accelerator-threads 32"
|
|
||||||
$CMD
|
|
@ -1,95 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
|
|
||||||
## qsub -q EarlyAppAccess -A Aurora_Deployment -I -l select=1 -l walltime=60:00
|
|
||||||
|
|
||||||
#PBS -q EarlyAppAccess
|
|
||||||
#PBS -l select=2
|
|
||||||
#PBS -l walltime=01:00:00
|
|
||||||
#PBS -A LatticeQCD_aesp_CNDA
|
|
||||||
#export OMP_PROC_BIND=spread
|
|
||||||
#export OMP_NUM_THREADS=3
|
|
||||||
#unset OMP_PLACES
|
|
||||||
|
|
||||||
cd $PBS_O_WORKDIR
|
|
||||||
|
|
||||||
source ../sourceme.sh
|
|
||||||
|
|
||||||
#echo Jobid: $PBS_JOBID
|
|
||||||
#echo Running on host `hostname`
|
|
||||||
#echo Running on nodes `cat $PBS_NODEFILE`
|
|
||||||
|
|
||||||
#echo NODES
|
|
||||||
#cat $PBS_NODEFILE
|
|
||||||
NNODES=`wc -l < $PBS_NODEFILE`
|
|
||||||
NRANKS=12 # Number of MPI ranks per node
|
|
||||||
NDEPTH=3 # Number of hardware threads per rank, spacing between MPI ranks on a node
|
|
||||||
NTHREADS=$OMP_NUM_THREADS # Number of OMP threads per rank, given to OMP_NUM_THREADS
|
|
||||||
|
|
||||||
NTOTRANKS=$(( NNODES * NRANKS ))
|
|
||||||
|
|
||||||
|
|
||||||
CMD="mpiexec -np 2 -ppn 1 -d ${NDEPTH} -envall \
|
|
||||||
./gpu_tile_compact.sh \
|
|
||||||
./Benchmark_comms_host_device --mpi 1.1.1.2 --grid 32.24.32.192 \
|
|
||||||
--shm-mpi 1 --shm 2048 --device-mem 32000 --accelerator-threads 32"
|
|
||||||
|
|
||||||
export MPIR_CVAR_CH4_OFI_ENABLE_HMEM=0
|
|
||||||
#$CMD | tee 1-to-1.comms.hmem0
|
|
||||||
export MPIR_CVAR_CH4_OFI_ENABLE_HMEM=1
|
|
||||||
#$CMD | tee 1-to-1.comms.hmem1
|
|
||||||
|
|
||||||
|
|
||||||
CMD="mpiexec -np 4 -ppn 2 -d ${NDEPTH} --cpu-bind=depth -envall \
|
|
||||||
./gpu_tile_compact.sh \
|
|
||||||
./Benchmark_comms_host_device --mpi 2.2.1.1 --grid 32.24.32.96 \
|
|
||||||
--shm-mpi 1 --shm 2048 --device-mem 32000 --accelerator-threads 32"
|
|
||||||
export MPIR_CVAR_CH4_OFI_ENABLE_HMEM=1
|
|
||||||
#$CMD | tee 2-to-2.comms.hmem1
|
|
||||||
|
|
||||||
export MPIR_CVAR_CH4_OFI_ENABLE_HMEM=0
|
|
||||||
$CMD | tee 2-to-2.comms.hmem0
|
|
||||||
|
|
||||||
CMD="mpiexec -np 6 -ppn 3 -d ${NDEPTH} --cpu-bind=depth -envall \
|
|
||||||
./gpu_tile_compact.sh \
|
|
||||||
./Benchmark_comms_host_device --mpi 3.2.1.1 --grid 32.24.32.96 \
|
|
||||||
--shm-mpi 1 --shm 2048 --device-mem 32000 --accelerator-threads 32"
|
|
||||||
export MPIR_CVAR_CH4_OFI_ENABLE_HMEM=1
|
|
||||||
#$CMD | tee 3-to-3.comms.hmem1
|
|
||||||
|
|
||||||
export MPIR_CVAR_CH4_OFI_ENABLE_HMEM=0
|
|
||||||
#$CMD | tee 3-to-3.comms.hmem0
|
|
||||||
|
|
||||||
|
|
||||||
CMD="mpiexec -np 8 -ppn 4 -d ${NDEPTH} --cpu-bind=depth -envall \
|
|
||||||
./gpu_tile_compact4a.sh \
|
|
||||||
./Benchmark_comms_host_device --mpi 2.2.2.1 --grid 32.24.32.96 \
|
|
||||||
--shm-mpi 1 --shm 2048 --device-mem 32000 --accelerator-threads 32"
|
|
||||||
export MPIR_CVAR_CH4_OFI_ENABLE_HMEM=1
|
|
||||||
#$CMD | tee 4-to-4.comms.hmem1.nic-affinity
|
|
||||||
|
|
||||||
export MPIR_CVAR_CH4_OFI_ENABLE_HMEM=0
|
|
||||||
$CMD | tee 4-to-4.comms.hmem0
|
|
||||||
|
|
||||||
#mpiexec -np 1 --ppn 1 -d 1 numactl -H | tee numa.log
|
|
||||||
|
|
||||||
CMD="mpiexec -np 12 -ppn 6 -d ${NDEPTH} --cpu-bind=depth -envall \
|
|
||||||
./gpu_tile_compact.sh \
|
|
||||||
./Benchmark_comms_host_device --mpi 3.2.2.1 --grid 32.24.32.96 \
|
|
||||||
--shm-mpi 1 --shm 2048 --device-mem 32000 --accelerator-threads 32"
|
|
||||||
export MPIR_CVAR_CH4_OFI_ENABLE_HMEM=1
|
|
||||||
#$CMD | tee 6-to-6.comms.hmem1
|
|
||||||
|
|
||||||
export MPIR_CVAR_CH4_OFI_ENABLE_HMEM=0
|
|
||||||
$CMD | tee 6-to-6.comms.hmem0
|
|
||||||
|
|
||||||
|
|
||||||
CMD="mpiexec -np ${NTOTRANKS} -ppn ${NRANKS} -d ${NDEPTH} --cpu-bind=depth -envall \
|
|
||||||
./gpu_tile_compact.sh \
|
|
||||||
./Benchmark_comms_host_device --mpi 3.2.2.2 --grid 32.24.32.192 \
|
|
||||||
--shm-mpi 1 --shm 2048 --device-mem 32000 --accelerator-threads 32"
|
|
||||||
|
|
||||||
export MPIR_CVAR_CH4_OFI_ENABLE_HMEM=1
|
|
||||||
#$CMD | tee 12-to-12.comms.hmem1
|
|
||||||
|
|
||||||
export MPIR_CVAR_CH4_OFI_ENABLE_HMEM=0
|
|
||||||
$CMD | tee 12-to-12.comms.hmem0
|
|
Loading…
Reference in New Issue
Block a user