mirror of
https://github.com/paboyle/Grid.git
synced 2025-04-09 21:50:45 +01:00
Update from chulwoo ; high level link for Grid.pdf in documentation
This commit is contained in:
parent
2c162577b5
commit
083af92ac2
1
documentation/Grid.pdf
Symbolic link
1
documentation/Grid.pdf
Symbolic link
@ -0,0 +1 @@
|
|||||||
|
_build/latex/Grid.pdf
|
@ -27,15 +27,21 @@ Grid itself is initialized with a call::
|
|||||||
|
|
||||||
Grid_init(&argc, &argv);
|
Grid_init(&argc, &argv);
|
||||||
|
|
||||||
.. todo:: CD: Where are the command-line arguments explained above?
|
Command line options include::
|
||||||
|
|
||||||
where `argc` and `argv` are constructed to simulate the command-line
|
|
||||||
options described above. At a minimum one must provide the `--grid`
|
|
||||||
and `--mpi` parameters. The latter specifies the grid of processors
|
|
||||||
(MPI ranks).
|
|
||||||
|
|
||||||
The following Grid procedures are useful for verifying that Grid is
|
--mpi n.n.n.n : default MPI decomposition
|
||||||
properly initialized.
|
--threads n : default number of OMP threads
|
||||||
|
--grid n.n.n.n : default Grid size
|
||||||
|
|
||||||
|
where `argc` and `argv` are constructed to simulate the command-line
|
||||||
|
options described above. At a minimum one usually provides the
|
||||||
|
`--grid` and `--mpi` parameters. The former specifies the lattice
|
||||||
|
dimensions and the latter specifies the grid of processors (MPI
|
||||||
|
ranks). If these parameters are not specified with the `Grid_init`
|
||||||
|
call, they need to be supplied later when creating Grid fields.
|
||||||
|
|
||||||
|
The following Grid procedures are useful for verifying that Grid
|
||||||
|
"default" values are properly initialized.
|
||||||
|
|
||||||
============================================================= ===========================================================================================================
|
============================================================= ===========================================================================================================
|
||||||
Grid procedure returns
|
Grid procedure returns
|
||||||
@ -52,10 +58,9 @@ MPI coordination
|
|||||||
|
|
||||||
Grid wants to use its own numbering of MPI ranks and its own
|
Grid wants to use its own numbering of MPI ranks and its own
|
||||||
assignment of the lattice coordinates with each rank. Obviously, the
|
assignment of the lattice coordinates with each rank. Obviously, the
|
||||||
calling program and Grid must agree on these conventions. It is
|
calling program and Grid must agree on these conventions. One should
|
||||||
convenient to use Grid's Cartesian communicator class to discover the
|
use Grid's Cartesian communicator class to discover the processor
|
||||||
processor assignments. For a four-dimensional processor grid one can
|
assignments. For a four-dimensional processor grid one can define::
|
||||||
define::
|
|
||||||
|
|
||||||
static Grid::CartesianCommunicator *grid_cart = NULL;
|
static Grid::CartesianCommunicator *grid_cart = NULL;
|
||||||
grid_cart = new Grid::CartesianCommunicator(processors);
|
grid_cart = new Grid::CartesianCommunicator(processors);
|
||||||
@ -96,14 +101,38 @@ index number of the subcommunicator. Once this is done,::
|
|||||||
|
|
||||||
returns a rank that agrees with Grid's `peRank`.
|
returns a rank that agrees with Grid's `peRank`.
|
||||||
|
|
||||||
|
QMP coordination
|
||||||
|
----------------
|
||||||
|
|
||||||
|
If the calling program uses the SciDAC QMP message-passing package, a
|
||||||
|
call to QMP_comm_split() instead can be used to reassign the ranks.
|
||||||
|
In the example below, `peGrid` gives the processor-grid dimensions,
|
||||||
|
usually set on the command line with `-qmp-geom`.
|
||||||
|
|
||||||
|
**Example**::
|
||||||
|
|
||||||
|
int NDIM = QMP_get_allocated_number_of_dimensions();
|
||||||
|
Grid::Grid_init(argc,argv);
|
||||||
|
FgridBase::grid_initted=true;
|
||||||
|
std::vector<int> processors;
|
||||||
|
for(int i=0;i<NDIM;i++) processors.push_back(peGrid[i]);
|
||||||
|
Grid::CartesianCommunicator grid_cart(processors);
|
||||||
|
std::vector<int> pePos(NDIM);
|
||||||
|
for(int i=NDIM-1;i>=0;i--)
|
||||||
|
pePos[i] = grid_cart._processor_coor[i];
|
||||||
|
int peRank = grid_cart->RankFromProcessorCoor(pePos);
|
||||||
|
QMP_comm_split(QMP_comm_get_default(),0,peRank,&qmp_comm);
|
||||||
|
QMP_comm_set_default(qmp_comm);
|
||||||
|
|
||||||
|
|
||||||
Mapping fields between Grid and user layouts
|
Mapping fields between Grid and user layouts
|
||||||
---------------------------------------------
|
---------------------------------------------
|
||||||
|
|
||||||
In order to map data between layouts, it is important to know
|
In order to map data between calling-program and Grid layouts, it is
|
||||||
how the lattice sites are distributed across the processor grid. A
|
important to know how the lattice sites are distributed across the
|
||||||
lattice site with coordinates `r[mu]` is assigned to the processor with
|
processor grid. A lattice site with coordinates `r[mu]` is assigned
|
||||||
processor coordinates `pePos[mu]` according to the rule::
|
to the processor with processor coordinates `pePos[mu]` according to
|
||||||
|
the rule::
|
||||||
|
|
||||||
pePos[mu] = r[mu]/dim[mu]
|
pePos[mu] = r[mu]/dim[mu]
|
||||||
|
|
||||||
@ -118,7 +147,7 @@ defined on the appropriate grid, whether it be a full lattice (4D
|
|||||||
`GridCartesian`), one of the checkerboards (4D
|
`GridCartesian`), one of the checkerboards (4D
|
||||||
`GridRedBlackCartesian`), a five-dimensional full grid (5D
|
`GridRedBlackCartesian`), a five-dimensional full grid (5D
|
||||||
`GridCartesian`), or a five-dimensional checkerboard (5D
|
`GridCartesian`), or a five-dimensional checkerboard (5D
|
||||||
`GridRedBlackCartesian`). For example, an improved staggered fermion
|
`GridRedBlackCartesian`). For example, an improved staggered-fermion
|
||||||
color-vector field `cv` on a single checkerboard would be constructed
|
color-vector field `cv` on a single checkerboard would be constructed
|
||||||
using
|
using
|
||||||
|
|
||||||
@ -133,12 +162,16 @@ using
|
|||||||
|
|
||||||
typename ImprovedStaggeredFermion::FermionField cv(RBGrid);
|
typename ImprovedStaggeredFermion::FermionField cv(RBGrid);
|
||||||
|
|
||||||
To map data within an MPI rank, the external code must iterate over
|
The example above assumes that the grid default values were set in the
|
||||||
the sites belonging to that rank (full or checkerboard as
|
`Grid_init` call. If not, they can be set at this point and passed
|
||||||
appropriate). To import data into Grid, the external data on a single
|
when `GridCartesian` is instantiated here. To map data within an MPI
|
||||||
site with coordinates `r` is first copied into the appropriate Grid
|
rank, the external code must iterate over the sites belonging to that
|
||||||
scalar object `s`. Then it is copied into the Grid lattice field `l`
|
rank (full or checkerboard as appropriate). Note that the site
|
||||||
with `pokeLocalSite`::
|
coordinates are specified relative to the origin of the lattice
|
||||||
|
subvolume on that rank. To import data into Grid, the external data on
|
||||||
|
a single site with coordinates `r` is first copied into the
|
||||||
|
appropriate Grid scalar object `s`. Then it is copied into the Grid
|
||||||
|
lattice field `l` with `pokeLocalSite`::
|
||||||
|
|
||||||
pokeLocalSite(const sobj &s, Lattice<vobj> &l, Coordinate &r);
|
pokeLocalSite(const sobj &s, Lattice<vobj> &l, Coordinate &r);
|
||||||
|
|
||||||
|
Loading…
x
Reference in New Issue
Block a user