mirror of
https://github.com/paboyle/Grid.git
synced 2025-04-04 11:15:55 +01:00
Added a new, much faster implementation of precision change that uses (optionally) a precomputed workspace containing pointer offsets that is device resident, such that all lattice copying occurs only on the device and no host<->device transfer is required, other than the pointer table. It also avoids the need to unpack and repack the fields using explicit lane copying. When this new precisionChange is called without a workspace, one will be computed on-the-fly; however it is still considerably faster than the original implementation. In the special case of using double2 and when the Grids are the same, calls to the new precisionChange will automatically use precisionChangeFast, such that there is a single API call for all precision changes. Reliable update and mixed-prec multishift have been modified to precompute precision change workspaces Renamed the original precisionChange as precisionChangeOrig Fixed incorrect pointer offset bug in copyLane Added a test and a benchmark for precisionChange Added a test for reliable update CG
263 lines
9.8 KiB
C++
263 lines
9.8 KiB
C++
/*************************************************************************************
|
|
|
|
Grid physics library, www.github.com/paboyle/Grid
|
|
|
|
Source file: ./lib/tensors/Tensor_extract_merge.h
|
|
|
|
Copyright (C) 2015
|
|
|
|
Author: Azusa Yamaguchi <ayamaguc@staffmail.ed.ac.uk>
|
|
Author: Peter Boyle <paboyle@ph.ed.ac.uk>
|
|
Author: neo <cossu@post.kek.jp>
|
|
Author: paboyle <paboyle@ph.ed.ac.uk>
|
|
Author: Christopher Kelly <ckelly@phys.columbia.edu>
|
|
|
|
This program is free software; you can redistribute it and/or modify
|
|
it under the terms of the GNU General Public License as published by
|
|
the Free Software Foundation; either version 2 of the License, or
|
|
(at your option) any later version.
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
GNU General Public License for more details.
|
|
|
|
You should have received a copy of the GNU General Public License along
|
|
with this program; if not, write to the Free Software Foundation, Inc.,
|
|
51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
|
|
See the full license in the file "LICENSE" in the top level distribution directory
|
|
*************************************************************************************/
|
|
/* END LEGAL */
|
|
#pragma once
|
|
|
|
#include <string.h>
|
|
|
|
//#pragma GCC optimize("no-strict-aliasing")
|
|
|
|
NAMESPACE_BEGIN(Grid);
|
|
|
|
/////////////////////////////////////////////////////////////////
|
|
// Generic extract/merge/permute
|
|
/////////////////////////////////////////////////////////////////
|
|
|
|
template<class __T> using ExtractPointerArray = AcceleratorVector<__T *,GRID_MAX_SIMD>;
|
|
template<class __T> using ExtractBuffer = AcceleratorVector<__T ,GRID_MAX_SIMD>;
|
|
|
|
//void extract(const vobj &vec,ExtractBuffer<typename vobj::scalar_object> &extracted);
|
|
//void extract(const vobj &vec,ExtractPointerArray<sobj> &extracted, int offset);
|
|
//void merge(vobj &vec,ExtractBuffer<typename vobj::scalar_object> &extracted)
|
|
//void merge(vobj &vec,ExtractPointerArray<typename vobj::scalar_object> &extracted)
|
|
|
|
////////////////////////////////////////////////////////////////////////
|
|
// Extract to contiguous array scalar object
|
|
////////////////////////////////////////////////////////////////////////
|
|
template<class vobj,class sobj> accelerator
|
|
void extract(const vobj &vec,ExtractBuffer<sobj> &extracted)
|
|
{
|
|
typedef typename GridTypeMapper<sobj>::scalar_type sobj_scalar_type;
|
|
typedef typename GridTypeMapper<vobj>::scalar_type scalar_type;
|
|
typedef typename GridTypeMapper<vobj>::vector_type vector_type;
|
|
|
|
const int words=sizeof(vobj)/sizeof(vector_type);
|
|
const int Nsimd=vector_type::Nsimd();
|
|
const int Nextr=extracted.size();
|
|
vector_type * vp = (vector_type *)&vec;
|
|
const int s=Nsimd/Nextr;
|
|
sobj_scalar_type *sp = (sobj_scalar_type *) &extracted[0];
|
|
sobj_scalar_type stmp;
|
|
for(int w=0;w<words;w++){
|
|
for(int i=0;i<Nextr;i++){
|
|
stmp = vp[w].getlane(i*s);
|
|
sp[i*words+w] =stmp;
|
|
// memcpy((char *)&sp[i*words+w],(char *)&stmp,sizeof(stmp));
|
|
}
|
|
}
|
|
/*
|
|
scalar_type *vp = (scalar_type *)&vec;
|
|
scalar_type vtmp;
|
|
sobj_scalar_type stmp;
|
|
for(int w=0;w<words;w++){
|
|
for(int i=0;i<Nextr;i++){
|
|
memcpy((char *)&vtmp,(char *)&vp[w*Nsimd+i*s],sizeof(vtmp));
|
|
stmp = vtmp;
|
|
memcpy((char *)&sp[i*words+w],(char *)&stmp,sizeof(stmp));
|
|
}
|
|
}
|
|
*/
|
|
|
|
return;
|
|
}
|
|
|
|
////////////////////////////////////////////////////////////////////////
|
|
// Merge a contiguous array of scalar objects
|
|
////////////////////////////////////////////////////////////////////////
|
|
template<class vobj,class sobj> accelerator
|
|
void merge(vobj &vec,ExtractBuffer<sobj> &extracted)
|
|
{
|
|
typedef typename GridTypeMapper<sobj>::scalar_type sobj_scalar_type;
|
|
typedef typename GridTypeMapper<vobj>::scalar_type scalar_type;
|
|
typedef typename GridTypeMapper<vobj>::vector_type vector_type;
|
|
|
|
const int words=sizeof(vobj)/sizeof(vector_type);
|
|
const int Nsimd=vector_type::Nsimd();
|
|
const int Nextr = extracted.size();
|
|
const int s=Nsimd/Nextr;
|
|
|
|
sobj_scalar_type *sp = (sobj_scalar_type *)&extracted[0];
|
|
vector_type *vp = (vector_type *)&vec;
|
|
scalar_type vtmp;
|
|
sobj_scalar_type stmp;
|
|
for(int w=0;w<words;w++){
|
|
for(int i=0;i<Nextr;i++){
|
|
for(int ii=0;ii<s;ii++){
|
|
memcpy((char *)&stmp,(char *)&sp[i*words+w],sizeof(stmp));
|
|
vtmp = stmp;
|
|
vp[w].putlane(vtmp,i*s+ii);
|
|
// memcpy((char *)&vp[w*Nsimd+i*s+ii],(char *)&vtmp,sizeof(vtmp));
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
////////////////////////////////////////////////////////////////////////
|
|
// Extract/Insert a single lane
|
|
////////////////////////////////////////////////////////////////////////
|
|
template<class vobj> accelerator_inline
|
|
typename vobj::scalar_object extractLane(int lane, const vobj & __restrict__ vec)
|
|
{
|
|
typedef typename vobj::scalar_type scalar_type;
|
|
typedef typename vobj::scalar_object scalar_object;
|
|
typedef typename vobj::vector_type vector_type;
|
|
typedef typename ExtractTypeMap<scalar_type>::extract_type extract_type;
|
|
typedef scalar_type * pointer;
|
|
|
|
constexpr int words=sizeof(vobj)/sizeof(vector_type);
|
|
constexpr int Nsimd=vector_type::Nsimd();
|
|
|
|
scalar_object extracted;
|
|
pointer __restrict__ sp = (pointer)&extracted; // Type pun
|
|
vector_type *vp = (vector_type *)&vec;
|
|
for(int w=0;w<words;w++){
|
|
sp[w]=vp[w].getlane(lane);
|
|
}
|
|
return extracted;
|
|
}
|
|
|
|
template<class vobj> accelerator_inline
|
|
void insertLane(int lane, vobj & __restrict__ vec,const typename vobj::scalar_object & __restrict__ extracted)
|
|
{
|
|
typedef typename vobj::vector_type vector_type;
|
|
typedef typename vector_type::scalar_type scalar_type;
|
|
typedef typename ExtractTypeMap<scalar_type>::extract_type extract_type;
|
|
typedef scalar_type * pointer;
|
|
|
|
constexpr int words=sizeof(vobj)/sizeof(vector_type);
|
|
constexpr int Nsimd=vector_type::Nsimd();
|
|
|
|
pointer __restrict__ sp = (pointer)&extracted;
|
|
vector_type *vp = (vector_type *)&vec;
|
|
for(int w=0;w<words;w++){
|
|
vp[w].putlane(sp[w],lane);
|
|
}
|
|
}
|
|
|
|
////////////////////////////////////////////////////////////////////////
|
|
// Extract to a bunch of scalar object pointers of different scalar type, with offset. Useful for precision change
|
|
////////////////////////////////////////////////////////////////////////
|
|
template<class vobj, class sobj> accelerator
|
|
void extract(const vobj &vec,const ExtractPointerArray<sobj> &extracted, int offset)
|
|
{
|
|
typedef typename GridTypeMapper<sobj>::scalar_type sobj_scalar_type;
|
|
typedef typename GridTypeMapper<vobj>::scalar_type scalar_type;
|
|
typedef typename GridTypeMapper<vobj>::vector_type vector_type;
|
|
|
|
const int words=sizeof(vobj)/sizeof(vector_type);
|
|
const int Nsimd=vector_type::Nsimd();
|
|
const int Nextr=extracted.size();
|
|
const int s = Nsimd/Nextr;
|
|
|
|
vector_type * vp = (vector_type *)&vec;
|
|
scalar_type vtmp;
|
|
sobj_scalar_type stmp;
|
|
for(int w=0;w<words;w++){
|
|
for(int i=0;i<Nextr;i++){
|
|
sobj_scalar_type * pointer = (sobj_scalar_type *)& extracted[i][offset];
|
|
pointer[w] = vp[w].getlane(i*s);
|
|
}
|
|
}
|
|
}
|
|
|
|
////////////////////////////////////////////////////////////////////////
|
|
// Merge bunch of scalar object pointers of different scalar type, with offset. Useful for precision change
|
|
////////////////////////////////////////////////////////////////////////
|
|
template<class vobj, class sobj> accelerator
|
|
void merge(vobj &vec,const ExtractPointerArray<sobj> &extracted, int offset)
|
|
{
|
|
typedef typename GridTypeMapper<sobj>::scalar_type sobj_scalar_type;
|
|
typedef typename GridTypeMapper<vobj>::scalar_type scalar_type;
|
|
typedef typename GridTypeMapper<vobj>::vector_type vector_type;
|
|
|
|
const int words=sizeof(vobj)/sizeof(vector_type);
|
|
const int Nsimd=vector_type::Nsimd();
|
|
const int Nextr=extracted.size();
|
|
const int s = Nsimd/Nextr;
|
|
|
|
vector_type * vp = (vector_type *)&vec;
|
|
scalar_type vtmp;
|
|
sobj_scalar_type stmp;
|
|
for(int w=0;w<words;w++){
|
|
for(int i=0;i<Nextr;i++){
|
|
sobj_scalar_type * pointer = (sobj_scalar_type *)& extracted[i][offset];
|
|
for(int ii=0;ii<s;ii++){
|
|
vtmp=pointer[w];
|
|
vp[w].putlane(vtmp,i*s+ii);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////////////////
|
|
//Copy a single lane of a SIMD tensor type from one object to another
|
|
//Output object must be of the same tensor type but may be of a different precision (i.e. it can have a different root data type)
|
|
///////////////////////////////////////////////////////////////////////////////////
|
|
template<class vobjOut, class vobjIn>
|
|
accelerator_inline
|
|
void copyLane(vobjOut & __restrict__ vecOut, int lane_out, const vobjIn & __restrict__ vecIn, int lane_in)
|
|
{
|
|
static_assert( std::is_same<typename vobjOut::scalar_typeD, typename vobjIn::scalar_typeD>::value == 1, "copyLane: tensor types must be the same" ); //if tensor types are same the DoublePrecision type must be the same
|
|
|
|
typedef typename vobjOut::vector_type ovector_type;
|
|
typedef typename vobjIn::vector_type ivector_type;
|
|
constexpr int owords=sizeof(vobjOut)/sizeof(ovector_type);
|
|
constexpr int iwords=sizeof(vobjIn)/sizeof(ivector_type);
|
|
static_assert( owords == iwords, "copyLane: Expected number of vector words in input and output objects to be equal" );
|
|
|
|
typedef typename vobjOut::scalar_type oscalar_type;
|
|
typedef typename vobjIn::scalar_type iscalar_type;
|
|
typedef typename ExtractTypeMap<oscalar_type>::extract_type oextract_type;
|
|
typedef typename ExtractTypeMap<iscalar_type>::extract_type iextract_type;
|
|
|
|
typedef oextract_type * opointer;
|
|
typedef iextract_type * ipointer;
|
|
|
|
constexpr int oNsimd=ovector_type::Nsimd();
|
|
constexpr int iNsimd=ivector_type::Nsimd();
|
|
|
|
iscalar_type itmp;
|
|
oscalar_type otmp;
|
|
|
|
ovector_type * __restrict__ op = (ovector_type *)&vecOut;
|
|
ivector_type * __restrict__ ip = (ivector_type *)&vecIn;
|
|
for(int w=0;w<owords;w++){
|
|
itmp = ip[w].getlane(lane_in);
|
|
otmp = itmp; //potential precision change
|
|
op[w].putlane(otmp,lane_out);
|
|
}
|
|
}
|
|
|
|
|
|
NAMESPACE_END(Grid);
|
|
|