1
0
mirror of https://github.com/paboyle/Grid.git synced 2025-04-04 19:25:56 +01:00

Improved the prefetching when using cache blocking codes

This commit is contained in:
paboyle 2016-06-26 12:54:14 -07:00
parent 1445189361
commit 8fcefc021a
5 changed files with 208 additions and 162 deletions

View File

@ -1,4 +1,4 @@
/************************************************************************************* /*************************************************************************************
Grid physics library, www.github.com/paboyle/Grid Grid physics library, www.github.com/paboyle/Grid
@ -261,6 +261,9 @@
} }
}; };
inline uint64_t Touch(int ent) {
// _mm_prefetch((char *)&_entries[ent],_MM_HINT_T0);
}
inline uint64_t GetInfo(int &ptype,int &local,int &perm,int point,int ent,uint64_t base) { inline uint64_t GetInfo(int &ptype,int &local,int &perm,int point,int ent,uint64_t base) {
_mm_prefetch((char *)&_entries[ent+1],_MM_HINT_T0); _mm_prefetch((char *)&_entries[ent+1],_MM_HINT_T0);
local = _entries[ent]._is_local; local = _entries[ent]._is_local;

View File

@ -1,7 +1,9 @@
{ {
int locala,perma, ptypea; int locala,perma, ptypea;
int localb,permb, ptypeb; int localb,permb, ptypeb;
uint64_t basea, baseb; int localc,permc, ptypec;
uint64_t basea, baseb, basec;
const uint64_t plocal =(uint64_t) & in._odata[0]; const uint64_t plocal =(uint64_t) & in._odata[0];
// vComplexF isigns[2] = { signs[0], signs[1] }; // vComplexF isigns[2] = { signs[0], signs[1] };
@ -10,15 +12,22 @@
MASK_REGS; MASK_REGS;
for(int site=0;site<Ns;site++) { for(int site=0;site<Ns;site++) {
int sU=lo.Reorder(ssU); int sU =lo.Reorder(ssU);
for(int s=0;s<Ls;s++) { for(int s=0;s<Ls;s++) {
ss=sU*Ls+s; ss =sU*Ls+s;
//////////////////////////////// ////////////////////////////////
// Xp // Xp
//////////////////////////////// ////////////////////////////////
int ent=ss*8;// 2*Ndim int ent=ss*8;// 2*Ndim
basea = st.GetInfo(ptypea,locala,perma,Xp,ent,plocal); ent++; basea = st.GetInfo(ptypea,locala,perma,Xp,ent,plocal); ent++;
PREFETCH1_CHIMU(basea);
PF_GAUGE(Xp);
baseb = st.GetInfo(ptypeb,localb,permb,Yp,ent,plocal); ent++; baseb = st.GetInfo(ptypeb,localb,permb,Yp,ent,plocal); ent++;
PREFETCH_CHIMU(baseb);
basec = st.GetInfo(ptypec,localc,permc,Zp,ent,plocal); ent++;
PREFETCH_CHIMU(basec);
if ( locala ) { if ( locala ) {
LOAD64(%r10,isigns); LOAD64(%r10,isigns);
@ -36,7 +45,8 @@
//////////////////////////////// ////////////////////////////////
// Yp // Yp
//////////////////////////////// ////////////////////////////////
basea = st.GetInfo(ptypea,locala,perma,Zp,ent,plocal); ent++; basea = st.GetInfo(ptypea,locala,perma,Tp,ent,plocal); ent++;
PREFETCH_CHIMU(basea);
if ( localb ) { if ( localb ) {
LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit
YM_PROJMEM(baseb); YM_PROJMEM(baseb);
@ -45,7 +55,7 @@
LOAD_CHI(baseb); LOAD_CHI(baseb);
} }
{ {
MULT_2SPIN_DIR_PFYP(Yp,basea); MULT_2SPIN_DIR_PFYP(Yp,basec);
} }
LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit
YM_RECON_ACCUM; YM_RECON_ACCUM;
@ -53,16 +63,17 @@
//////////////////////////////// ////////////////////////////////
// Zp // Zp
//////////////////////////////// ////////////////////////////////
baseb = st.GetInfo(ptypeb,localb,permb,Tp,ent,plocal); ent++; baseb = st.GetInfo(ptypeb,localb,permb,Xm,ent,plocal); ent++;
if ( locala ) { PREFETCH_CHIMU(baseb);
if ( localc ) {
LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit
ZM_PROJMEM(basea); ZM_PROJMEM(basec);
MAYBEPERM(PERMUTE_DIR1,perma); MAYBEPERM(PERMUTE_DIR1,permc);
} else { } else {
LOAD_CHI(basea); LOAD_CHI(basec);
} }
{ {
MULT_2SPIN_DIR_PFZP(Zp,baseb); MULT_2SPIN_DIR_PFZP(Zp,basea);
} }
LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit
ZM_RECON_ACCUM; ZM_RECON_ACCUM;
@ -70,16 +81,17 @@
//////////////////////////////// ////////////////////////////////
// Tp // Tp
//////////////////////////////// ////////////////////////////////
basea = st.GetInfo(ptypea,locala,perma,Xm,ent,plocal); ent++; basec = st.GetInfo(ptypec,localc,permc,Ym,ent,plocal); ent++;
if ( localb ) { PREFETCH_CHIMU(basec);
if ( locala ) {
LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit
TM_PROJMEM(baseb); TM_PROJMEM(basea);
MAYBEPERM(PERMUTE_DIR0,permb); MAYBEPERM(PERMUTE_DIR0,perma);
} else { } else {
LOAD_CHI(baseb); LOAD_CHI(basea);
} }
{ {
MULT_2SPIN_DIR_PFTP(Tp,basea); MULT_2SPIN_DIR_PFTP(Tp,baseb);
} }
LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit
TM_RECON_ACCUM; TM_RECON_ACCUM;
@ -87,16 +99,17 @@
//////////////////////////////// ////////////////////////////////
// Xm // Xm
//////////////////////////////// ////////////////////////////////
baseb = st.GetInfo(ptypeb,localb,permb,Ym,ent,plocal); ent++; basea = st.GetInfo(ptypea,locala,perma,Zm,ent,plocal); ent++;
if ( locala ) { PREFETCH_CHIMU(basea);
if ( localb ) {
LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit
XP_PROJMEM(basea); XP_PROJMEM(baseb);
MAYBEPERM(PERMUTE_DIR3,perma); MAYBEPERM(PERMUTE_DIR3,permb);
} else { } else {
LOAD_CHI(basea); LOAD_CHI(baseb);
} }
{ {
MULT_2SPIN_DIR_PFXM(Xm,baseb); MULT_2SPIN_DIR_PFXM(Xm,basec);
} }
LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit
XP_RECON_ACCUM; XP_RECON_ACCUM;
@ -104,13 +117,14 @@
//////////////////////////////// ////////////////////////////////
// Ym // Ym
//////////////////////////////// ////////////////////////////////
basea = st.GetInfo(ptypea,locala,perma,Zm,ent,plocal); ent++; baseb = st.GetInfo(ptypeb,localb,permb,Tm,ent,plocal); ent++;
if ( localb ) { PREFETCH_CHIMU(baseb);
if ( localc ) {
LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit
YP_PROJMEM(baseb); YP_PROJMEM(basec);
MAYBEPERM(PERMUTE_DIR2,permb); MAYBEPERM(PERMUTE_DIR2,permc);
} else { } else {
LOAD_CHI(baseb); LOAD_CHI(basec);
} }
{ {
MULT_2SPIN_DIR_PFYM(Ym,basea); MULT_2SPIN_DIR_PFYM(Ym,basea);
@ -121,7 +135,8 @@
//////////////////////////////// ////////////////////////////////
// Zm // Zm
//////////////////////////////// ////////////////////////////////
baseb = st.GetInfo(ptypeb,localb,permb,Tm,ent,plocal); ent++; basec = (uint64_t)&out._odata[ss];
PREFETCH_CHIMU(basec);
if ( locala ) { if ( locala ) {
LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit
ZP_PROJMEM(basea); ZP_PROJMEM(basea);
@ -138,7 +153,8 @@
//////////////////////////////// ////////////////////////////////
// Tm // Tm
//////////////////////////////// ////////////////////////////////
basea = (uint64_t)&out._odata[ss]; // basea = st.GetInfo(ptypea,locala,perma,Xp,ent,plocal); ent++;
// PREFETCH_CHIMU(basea);
if ( localb ) { if ( localb ) {
LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit
TP_PROJMEM(baseb); TP_PROJMEM(baseb);
@ -146,16 +162,16 @@
} else { } else {
LOAD_CHI(baseb); LOAD_CHI(baseb);
} }
baseb = st.GetInfo(ptypeb,localb,permb,Xp,ent,plocal);
{ {
MULT_2SPIN_DIR_PFTM(Tm,basea); MULT_2SPIN_DIR_PFTM(Tm,basec);
} }
// baseb = st.GetInfo(ptypeb,localb,permb,Yp,ent,plocal); ent++;
LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit
TP_RECON_ACCUM; TP_RECON_ACCUM;
SAVE_RESULT(&out._odata[ss],baseb); SAVE_RESULT(&out._odata[ss],basec);
} }
ssU++; ssU++;
} }
} }

View File

@ -2,7 +2,6 @@
int locala,perma, ptypea; int locala,perma, ptypea;
int localb,permb, ptypeb; int localb,permb, ptypeb;
uint64_t basea, baseb; uint64_t basea, baseb;
uint64_t basex;
const uint64_t plocal =(uint64_t) & in._odata[0]; const uint64_t plocal =(uint64_t) & in._odata[0];
// vComplexF isigns[2] = { signs[0], signs[1] }; // vComplexF isigns[2] = { signs[0], signs[1] };
@ -19,9 +18,7 @@
//////////////////////////////// ////////////////////////////////
int ent=ss*8;// 2*Ndim int ent=ss*8;// 2*Ndim
basea = st.GetInfo(ptypea,locala,perma,Xp,ent,plocal); ent++; basea = st.GetInfo(ptypea,locala,perma,Xp,ent,plocal); ent++;
PREFETCH_CHIMU(basea);
baseb = st.GetInfo(ptypeb,localb,permb,Yp,ent,plocal); ent++; baseb = st.GetInfo(ptypeb,localb,permb,Yp,ent,plocal); ent++;
basex = basea;
if ( locala ) { if ( locala ) {
LOAD64(%r10,isigns); LOAD64(%r10,isigns);
@ -39,7 +36,7 @@
//////////////////////////////// ////////////////////////////////
// Yp // Yp
//////////////////////////////// ////////////////////////////////
basea = st.GetInfo(ptypea,locala,perma,Xp,ent,plocal); ent++; basea = st.GetInfo(ptypea,locala,perma,Zp,ent,plocal); ent++;
if ( localb ) { if ( localb ) {
LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit
YM_PROJMEM(baseb); YM_PROJMEM(baseb);
@ -56,7 +53,7 @@
//////////////////////////////// ////////////////////////////////
// Zp // Zp
//////////////////////////////// ////////////////////////////////
baseb = st.GetInfo(ptypeb,localb,permb,Yp,ent,plocal); ent++; baseb = st.GetInfo(ptypeb,localb,permb,Tp,ent,plocal); ent++;
if ( locala ) { if ( locala ) {
LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit
ZM_PROJMEM(basea); ZM_PROJMEM(basea);
@ -73,7 +70,7 @@
//////////////////////////////// ////////////////////////////////
// Tp // Tp
//////////////////////////////// ////////////////////////////////
basea = st.GetInfo(ptypea,locala,perma,Xp,ent,plocal); ent++; basea = st.GetInfo(ptypea,locala,perma,Xm,ent,plocal); ent++;
if ( localb ) { if ( localb ) {
LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit
TM_PROJMEM(baseb); TM_PROJMEM(baseb);
@ -90,7 +87,7 @@
//////////////////////////////// ////////////////////////////////
// Xm // Xm
//////////////////////////////// ////////////////////////////////
baseb = st.GetInfo(ptypeb,localb,permb,Yp,ent,plocal); ent++; baseb = st.GetInfo(ptypeb,localb,permb,Ym,ent,plocal); ent++;
if ( locala ) { if ( locala ) {
LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit
XP_PROJMEM(basea); XP_PROJMEM(basea);
@ -107,7 +104,7 @@
//////////////////////////////// ////////////////////////////////
// Ym // Ym
//////////////////////////////// ////////////////////////////////
basea = st.GetInfo(ptypea,locala,perma,Xp,ent,plocal); ent++; basea = st.GetInfo(ptypea,locala,perma,Zm,ent,plocal); ent++;
if ( localb ) { if ( localb ) {
LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit
YP_PROJMEM(baseb); YP_PROJMEM(baseb);
@ -124,7 +121,7 @@
//////////////////////////////// ////////////////////////////////
// Zm // Zm
//////////////////////////////// ////////////////////////////////
baseb = st.GetInfo(ptypeb,localb,permb,Yp,ent,plocal); ent++; baseb = st.GetInfo(ptypeb,localb,permb,Tm,ent,plocal); ent++;
if ( locala ) { if ( locala ) {
LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit
ZP_PROJMEM(basea); ZP_PROJMEM(basea);
@ -149,13 +146,14 @@
} else { } else {
LOAD_CHI(baseb); LOAD_CHI(baseb);
} }
baseb = st.GetInfo(ptypeb,localb,permb,Xp,ent,plocal);
{ {
MULT_2SPIN_DIR_PFTM(Tm,basea); MULT_2SPIN_DIR_PFTM(Tm,basea);
} }
LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit LOAD64(%r10,isigns); // times i => shuffle and xor the real part sign bit
TP_RECON_ACCUM; TP_RECON_ACCUM;
SAVE_RESULT(&out._odata[ss]); SAVE_RESULT(&out._odata[ss],baseb);
} }
ssU++; ssU++;

View File

@ -31,9 +31,6 @@ Author: paboyle <paboyle@ph.ed.ac.uk>
//////////////////////////////////////////////////////////////////////////////////////////////////// ////////////////////////////////////////////////////////////////////////////////////////////////////
// Peformance options // Peformance options
//////////////////////////////////////////////////////////////////////////////////////////////////// ////////////////////////////////////////////////////////////////////////////////////////////////////
#define AVX512_PF_L1
#undef AVX512_PF_L2_LINEAR
#undef AVX512_PF_L2_TABLE
#undef AVX512_PF_L2_WRITE #undef AVX512_PF_L2_WRITE
//////////////////////////////////////////////////////////////////////////////////////////////////// ////////////////////////////////////////////////////////////////////////////////////////////////////
@ -45,7 +42,7 @@ Author: paboyle <paboyle@ph.ed.ac.uk>
"mov $0x5555, %%eax \n"\ "mov $0x5555, %%eax \n"\
"kmovw %%eax, %%k7 \n" : : : "%eax"); "kmovw %%eax, %%k7 \n" : : : "%eax");
//#define label(B) __asm__ ( __func__ __LINE__ #B ":\n" ); //#define label(B) __asm__ ( __func__ _LINE__ #B ":\n" );
#define VZEROf(A) "vpxorq " #A "," #A "," #A ";\n" #define VZEROf(A) "vpxorq " #A "," #A "," #A ";\n"
#define VZEROd(A) "vpxorq " #A "," #A "," #A ";\n" #define VZEROd(A) "vpxorq " #A "," #A "," #A ";\n"
@ -96,30 +93,13 @@ Author: paboyle <paboyle@ph.ed.ac.uk>
#define VMOVf(A,DEST) "vmovaps " #A ", " #DEST ";\n" #define VMOVf(A,DEST) "vmovaps " #A ", " #DEST ";\n"
#define VMOVd(A,DEST) "vmovapd " #A ", " #DEST ";\n" #define VMOVd(A,DEST) "vmovapd " #A ", " #DEST ";\n"
#ifdef AVX512_PF_L1 #define VPREFETCH1(O,A) "prefetcht0 "#O"*64("#A");\n"
#define VPREFETCHG(O,A) "prefetcht0 "#O"*64("#A");\n"
#else
#define VPREFETCHG(O,A)
#endif
#ifdef AVX512_PF_L2_LINEAR
#define VPREFETCH2(O,A) "prefetcht1 "#O"*64("#A");\n" #define VPREFETCH2(O,A) "prefetcht1 "#O"*64("#A");\n"
#else
#define VPREFETCH2(O,A)
#endif
#ifdef AVX512_PF_L2_TABLE
#define VPREFETCHP(O,A) "prefetcht1 "#O"*64("#A");\n"
#else
#define VPREFETCHP(O,A)
#endif
#ifdef AVX512_PF_L2_WRITE #ifdef AVX512_PF_L2_WRITE
#define VPREFETCHW(O,A) "prefetchwt1 "#O"*64("#A");\n" #define VPREFETCHW(O,A) "prefetchwt1 "#O"*64("#A");\n"
#else #else
#define VPREFETCHW(O,A) #define VPREFETCHW(O,A)
#endif #endif
#define VPREFETCHNTA(O,A) #define VPREFETCHNTA(O,A)
#define VPREFETCH(O,A) #define VPREFETCH(O,A)

View File

@ -169,23 +169,6 @@ Author: paboyle <paboyle@ph.ed.ac.uk>
VSTORE(5,%r8,Chi_12) \ VSTORE(5,%r8,Chi_12) \
); );
#define SAVE_RESULTi(PTR,pf) \
LOAD64(%r8,PTR) \
LOAD64(%r9,pf) \
__asm__ ( \
VSTORE(0,%r8,result_00) VPREFETCHG(0,%r9) \
VSTORE(1,%r8,result_01) VPREFETCHG(1,%r9) \
VSTORE(2,%r8,result_02) VPREFETCHG(2,%r9) \
VSTORE(3,%r8,result_10) VPREFETCHG(3,%r9) \
VSTORE(4,%r8,result_11) VPREFETCHG(4,%r9) \
VSTORE(5,%r8,result_12) VPREFETCHG(5,%r9) \
VSTORE(6,%r8,result_20) VPREFETCHG(6,%r9) \
VSTORE(7,%r8,result_21) VPREFETCHG(7,%r9) \
VSTORE(8,%r8,result_22) VPREFETCHG(8,%r9) \
VSTORE(9,%r8,result_30) VPREFETCHG(9,%r9) \
VSTORE(10,%r8,result_31) VPREFETCHG(10,%r9) \
VSTORE(11,%r8,result_32) VPREFETCHG(11,%r9) \
);
#define MULT_2SPIN_DIR_PFXP(A,p) MULT_2SPIN_PFXP(&U._odata[sU](A),p) #define MULT_2SPIN_DIR_PFXP(A,p) MULT_2SPIN_PFXP(&U._odata[sU](A),p)
#define MULT_2SPIN_DIR_PFYP(A,p) MULT_2SPIN_PFYP(&U._odata[sU](A),p) #define MULT_2SPIN_DIR_PFYP(A,p) MULT_2SPIN_PFYP(&U._odata[sU](A),p)
@ -560,24 +543,89 @@ Author: paboyle <paboyle@ph.ed.ac.uk>
VSUB(UChi_02,result_22,result_22)\ VSUB(UChi_02,result_22,result_22)\
VSUB(UChi_12,result_32,result_32) ); VSUB(UChi_12,result_32,result_32) );
#define PREFETCH_CHIMU(A) #define AVX512_PF_L1
/* #define AVX512_PF_L2_GAUGE
LOAD64(%r9,A) \ #define AVX512_PF_L2_TABLE
__asm__ ( \ #undef AVX512_PF_L2_LINEAR
VPREFETCHG(0,%r9)\
VPREFETCHG(1,%r9)\ #ifdef AVX512_PF_L2_TABLE
VPREFETCHG(2,%r9)\ #define VPREFETCH_P1(A,B) VPREFETCH1(A,B)
VPREFETCHG(3,%r9)\ #define VPREFETCH_P2(A,B) VPREFETCH1(A,B)
VPREFETCHG(4,%r9)\ #else
VPREFETCHG(5,%r9)\ #define VPREFETCH_P1(A,B)
VPREFETCHG(6,%r9)\ #define VPREFETCH_P2(A,B)
VPREFETCHG(7,%r9)\ #endif
VPREFETCHG(8,%r9)\ #ifdef AVX512_PF_L2_LINEAR
VPREFETCHG(9,%r9)\ #define VPREFETCH_M1(A,B)
VPREFETCHG(10,%r9)\ #define VPREFETCH_M2(A,B)
VPREFETCHG(11,%r9)); #else
*/ #define VPREFETCH_M1(A,B) VPREFETCH1(A,B)
#define PERMUTE_DIR0 __asm__ ( \ #define VPREFETCH_M2(A,B) VPREFETCH2(A,B)
#endif
#ifdef AVX512_PF_L2_GAUGE
#define VPREFETCH_G1(A,B) VPREFETCH1(A,B)
#define VPREFETCH_G2(A,B) VPREFETCH2(A,B)
#else
#endif
#define PF_GAUGE(A) \
LOAD64(%r8,&U._odata[sU](A)) \
__asm__ ( \
VPREFETCH_G1(0,%r8) VPREFETCH_G1(1,%r8) \
VPREFETCH_G1(2,%r8) VPREFETCH_G1(3,%r8) \
);
#define SAVE_RESULTi(PTR,pf) \
LOAD64(%r8,PTR) \
LOAD64(%r9,pf) \
__asm__ ( \
VSTORE(0,%r8,result_00) VPREFETCH_M1(0,%r9) \
VSTORE(1,%r8,result_01) VPREFETCH_M1(1,%r9) \
VSTORE(2,%r8,result_02) VPREFETCH_M1(2,%r9) \
VSTORE(3,%r8,result_10) VPREFETCH_M1(3,%r9) \
VSTORE(4,%r8,result_11) VPREFETCH_M1(4,%r9) \
VSTORE(5,%r8,result_12) VPREFETCH_M1(5,%r9) \
VSTORE(6,%r8,result_20) VPREFETCH_M1(6,%r9) \
VSTORE(7,%r8,result_21) VPREFETCH_M1(7,%r9) \
VSTORE(8,%r8,result_22) VPREFETCH_M1(8,%r9) \
VSTORE(9,%r8,result_30) VPREFETCH_M1(9,%r9) \
VSTORE(10,%r8,result_31) VPREFETCH_M1(10,%r9) \
VSTORE(11,%r8,result_32) VPREFETCH_M1(11,%r9) \
);
#define PREFETCH_CHIMU(A) \
LOAD64(%r9,A) \
__asm__ ( \
VPREFETCH_P2(0,%r9) \
VPREFETCH_P2(1,%r9) \
VPREFETCH_P2(2,%r9) \
VPREFETCH_P2(3,%r9) \
VPREFETCH_P2(4,%r9) \
VPREFETCH_P2(5,%r9) \
VPREFETCH_P2(6,%r9) \
VPREFETCH_P2(7,%r9) \
VPREFETCH_P2(8,%r9) \
VPREFETCH_P2(9,%r9) \
VPREFETCH_P2(10,%r9) \
VPREFETCH_P2(11,%r9));
#define PREFETCH1_CHIMU(A) \
LOAD64(%r9,A) \
__asm__ ( \
VPREFETCH_P1(0,%r9) \
VPREFETCH_P1(1,%r9) \
VPREFETCH_P1(2,%r9) \
VPREFETCH_P1(3,%r9) \
VPREFETCH_P1(4,%r9) \
VPREFETCH_P1(5,%r9) \
VPREFETCH_P1(6,%r9) \
VPREFETCH_P1(7,%r9) \
VPREFETCH_P1(8,%r9) \
VPREFETCH_P1(9,%r9) \
VPREFETCH_P1(10,%r9) \
VPREFETCH_P1(11,%r9));
#define PERMUTE_DIR0 __asm__ ( \
VPERM0(Chi_00,Chi_00) \ VPERM0(Chi_00,Chi_00) \
VPERM0(Chi_01,Chi_01) \ VPERM0(Chi_01,Chi_01) \
VPERM0(Chi_02,Chi_02) \ VPERM0(Chi_02,Chi_02) \
@ -614,14 +662,15 @@ Author: paboyle <paboyle@ph.ed.ac.uk>
LOAD64(%r8,ptr) \ LOAD64(%r8,ptr) \
LOAD64(%r9,pf) \ LOAD64(%r9,pf) \
__asm__ ( \ __asm__ ( \
VPREFETCH2(9,%r8) VPREFETCH2(10,%r8) \ VPREFETCH_G2(9,%r8) \
VPREFETCH2(11,%r8) \ VPREFETCH_G2(10,%r8) \
VPREFETCH2(12,%r8) \ VPREFETCH_G2(11,%r8) \
VPREFETCH2(13,%r8) \ VPREFETCH_G2(12,%r8) \
VPREFETCH2(14,%r8) \ VPREFETCH_G2(13,%r8) \
VPREFETCH2(15,%r8) \ VPREFETCH_G2(14,%r8) \
VPREFETCH2(16,%r8) \ VPREFETCH_G2(15,%r8) \
VPREFETCH2(17,%r8) \ VPREFETCH_G2(16,%r8) \
VPREFETCH_G2(17,%r8) \
VSHUF(Chi_00,T1) \ VSHUF(Chi_00,T1) \
VMOVIDUP(0,%r8,Z0 ) \ VMOVIDUP(0,%r8,Z0 ) \
VMOVIDUP(3,%r8,Z1 ) \ VMOVIDUP(3,%r8,Z1 ) \
@ -633,10 +682,10 @@ Author: paboyle <paboyle@ph.ed.ac.uk>
VMUL(Z1,T2,UChi_11) VMOVIDUP(1,%r8,Z0 ) \ VMUL(Z1,T2,UChi_11) VMOVIDUP(1,%r8,Z0 ) \
VMUL(Z2,T1,UChi_02) VMOVIDUP(4,%r8,Z1 ) \ VMUL(Z2,T1,UChi_02) VMOVIDUP(4,%r8,Z1 ) \
VMUL(Z2,T2,UChi_12) VMOVIDUP(7,%r8,Z2 ) \ VMUL(Z2,T2,UChi_12) VMOVIDUP(7,%r8,Z2 ) \
VPREFETCHG(0,%r9) \ VPREFETCH_M1(0,%r9) \
VPREFETCHG(1,%r9) \ VPREFETCH_M1(1,%r9) \
VPREFETCHG(2,%r9) \ VPREFETCH_M1(2,%r9) \
VPREFETCHG(3,%r9) \ VPREFETCH_M1(3,%r9) \
/*18*/ \ /*18*/ \
VMADDSUB(Z3,Chi_00,UChi_00) VSHUF(Chi_01,T1) \ VMADDSUB(Z3,Chi_00,UChi_00) VSHUF(Chi_01,T1) \
VMADDSUB(Z3,Chi_10,UChi_10) \ VMADDSUB(Z3,Chi_10,UChi_10) \
@ -644,10 +693,10 @@ Author: paboyle <paboyle@ph.ed.ac.uk>
VMADDSUB(Z4,Chi_10,UChi_11) VSHUF(Chi_11,T2) \ VMADDSUB(Z4,Chi_10,UChi_11) VSHUF(Chi_11,T2) \
VMADDSUB(Z5,Chi_00,UChi_02) VMOVRDUP(4,%r8,Z4 ) \ VMADDSUB(Z5,Chi_00,UChi_02) VMOVRDUP(4,%r8,Z4 ) \
VMADDSUB(Z5,Chi_10,UChi_12) \ VMADDSUB(Z5,Chi_10,UChi_12) \
VPREFETCHG(4,%r9) \ VPREFETCH_M1(4,%r9) \
VPREFETCHG(5,%r9) \ VPREFETCH_M1(5,%r9) \
VPREFETCHG(6,%r9) \ VPREFETCH_M1(6,%r9) \
VPREFETCHG(7,%r9) \ VPREFETCH_M1(7,%r9) \
/*28*/ \ /*28*/ \
VMADDSUB(Z0,T1,UChi_00) VMOVRDUP(7,%r8,Z5 ) \ VMADDSUB(Z0,T1,UChi_00) VMOVRDUP(7,%r8,Z5 ) \
VMADDSUB(Z0,T2,UChi_10) \ VMADDSUB(Z0,T2,UChi_10) \
@ -674,15 +723,15 @@ Author: paboyle <paboyle@ph.ed.ac.uk>
VMADDSUB(Z4,Chi_11,UChi_11) VSHUF(Chi_12,T2) \ VMADDSUB(Z4,Chi_11,UChi_11) VSHUF(Chi_12,T2) \
VMADDSUB(Z5,Chi_01,UChi_02) VMOVRDUP(5,%r8,Z4 ) \ VMADDSUB(Z5,Chi_01,UChi_02) VMOVRDUP(5,%r8,Z4 ) \
VMADDSUB(Z5,Chi_11,UChi_12) \ VMADDSUB(Z5,Chi_11,UChi_12) \
VPREFETCHG(9,%r8) \ VPREFETCH_M1(9,%r8) \
VPREFETCHG(10,%r8) \ VPREFETCH_M1(10,%r8) \
VPREFETCHG(11,%r8) \ VPREFETCH_M1(11,%r8) \
VPREFETCHG(12,%r8) \ VPREFETCH_M1(12,%r8) \
VPREFETCHG(13,%r8) \ VPREFETCH_M1(13,%r8) \
VPREFETCHG(14,%r8) \ VPREFETCH_M1(14,%r8) \
VPREFETCHG(15,%r8) \ VPREFETCH_M1(15,%r8) \
VPREFETCHG(16,%r8) \ VPREFETCH_M1(16,%r8) \
VPREFETCHG(17,%r8) \ VPREFETCH_M1(17,%r8) \
/*48*/ \ /*48*/ \
VMADDSUB(Z0,T1,UChi_00) VMOVRDUP(8,%r8,Z5 ) \ VMADDSUB(Z0,T1,UChi_00) VMOVRDUP(8,%r8,Z5 ) \
VMADDSUB(Z0,T2,UChi_10) \ VMADDSUB(Z0,T2,UChi_10) \
@ -690,10 +739,10 @@ Author: paboyle <paboyle@ph.ed.ac.uk>
VMADDSUB(Z1,T2,UChi_11) \ VMADDSUB(Z1,T2,UChi_11) \
VMADDSUB(Z2,T1,UChi_02) \ VMADDSUB(Z2,T1,UChi_02) \
VMADDSUB(Z2,T2,UChi_12) \ VMADDSUB(Z2,T2,UChi_12) \
VPREFETCHG(8,%r9) \ VPREFETCH_M1(8,%r9) \
VPREFETCHG(9,%r9) \ VPREFETCH_M1(9,%r9) \
VPREFETCHG(10,%r9) \ VPREFETCH_M1(10,%r9) \
VPREFETCHG(11,%r9) \ VPREFETCH_M1(11,%r9) \
/*55*/ \ /*55*/ \
VMADDSUB(Z3,Chi_02,UChi_00) \ VMADDSUB(Z3,Chi_02,UChi_00) \
VMADDSUB(Z3,Chi_12,UChi_10) \ VMADDSUB(Z3,Chi_12,UChi_10) \
@ -712,56 +761,56 @@ Author: paboyle <paboyle@ph.ed.ac.uk>
VMULIDUP(0,%r8,T1,UChi_00) VMULIDUP(0,%r8,T2,UChi_10) \ VMULIDUP(0,%r8,T1,UChi_00) VMULIDUP(0,%r8,T2,UChi_10) \
VMULIDUP(3,%r8,T1,UChi_01) VMULIDUP(3,%r8,T2,UChi_11) \ VMULIDUP(3,%r8,T1,UChi_01) VMULIDUP(3,%r8,T2,UChi_11) \
VMULIDUP(6,%r8,T1,UChi_02) VMULIDUP(6,%r8,T2,UChi_12) \ VMULIDUP(6,%r8,T1,UChi_02) VMULIDUP(6,%r8,T2,UChi_12) \
VPREFETCHG(0,%r9) \ VPREFETCH_M1(0,%r9) \
VPREFETCHG(1,%r9) \ VPREFETCH_M1(1,%r9) \
VPREFETCHG(2,%r9) \ VPREFETCH_M1(2,%r9) \
VPREFETCHG(3,%r9) \ VPREFETCH_M1(3,%r9) \
/*8*/ \ /*8*/ \
VSHUF(Chi_01,T1) VSHUF(Chi_11,T2) \ VSHUF(Chi_01,T1) VSHUF(Chi_11,T2) \
VMADDSUBRDUP(0,%r8,Chi_00,UChi_00) VMADDSUBRDUP(0,%r8,Chi_10,UChi_10) \ VMADDSUBRDUP(0,%r8,Chi_00,UChi_00) VMADDSUBRDUP(0,%r8,Chi_10,UChi_10) \
VMADDSUBRDUP(3,%r8,Chi_00,UChi_01) VMADDSUBRDUP(3,%r8,Chi_10,UChi_11) \ VMADDSUBRDUP(3,%r8,Chi_00,UChi_01) VMADDSUBRDUP(3,%r8,Chi_10,UChi_11) \
VMADDSUBRDUP(6,%r8,Chi_00,UChi_02) VMADDSUBRDUP(6,%r8,Chi_10,UChi_12) \ VMADDSUBRDUP(6,%r8,Chi_00,UChi_02) VMADDSUBRDUP(6,%r8,Chi_10,UChi_12) \
VPREFETCHG(4,%r9) \ VPREFETCH_M1(4,%r9) \
VPREFETCHG(5,%r9) \ VPREFETCH_M1(5,%r9) \
VPREFETCHG(6,%r9) \ VPREFETCH_M1(6,%r9) \
VPREFETCHG(7,%r9) \ VPREFETCH_M1(7,%r9) \
/*16*/ \ /*16*/ \
VMADDSUBIDUP(1,%r8,T1,UChi_00) VMADDSUBIDUP(1,%r8,T2,UChi_10) \ VMADDSUBIDUP(1,%r8,T1,UChi_00) VMADDSUBIDUP(1,%r8,T2,UChi_10) \
VMADDSUBIDUP(4,%r8,T1,UChi_01) VMADDSUBIDUP(4,%r8,T2,UChi_11) \ VMADDSUBIDUP(4,%r8,T1,UChi_01) VMADDSUBIDUP(4,%r8,T2,UChi_11) \
VMADDSUBIDUP(7,%r8,T1,UChi_02) VMADDSUBIDUP(7,%r8,T2,UChi_12) \ VMADDSUBIDUP(7,%r8,T1,UChi_02) VMADDSUBIDUP(7,%r8,T2,UChi_12) \
VPREFETCHG(8,%r9) \ VPREFETCH_M1(8,%r9) \
VPREFETCHG(9,%r9) \ VPREFETCH_M1(9,%r9) \
VPREFETCHG(10,%r9) \ VPREFETCH_M1(10,%r9) \
VPREFETCHG(11,%r9) \ VPREFETCH_M1(11,%r9) \
/*22*/ \ /*22*/ \
VSHUF(Chi_02,T1) VSHUF(Chi_12,T2) \ VSHUF(Chi_02,T1) VSHUF(Chi_12,T2) \
VMADDSUBRDUP(1,%r8,Chi_01,UChi_00) VMADDSUBRDUP(1,%r8,Chi_11,UChi_10) \ VMADDSUBRDUP(1,%r8,Chi_01,UChi_00) VMADDSUBRDUP(1,%r8,Chi_11,UChi_10) \
VMADDSUBRDUP(4,%r8,Chi_01,UChi_01) VMADDSUBRDUP(4,%r8,Chi_11,UChi_11) \ VMADDSUBRDUP(4,%r8,Chi_01,UChi_01) VMADDSUBRDUP(4,%r8,Chi_11,UChi_11) \
VMADDSUBRDUP(7,%r8,Chi_01,UChi_02) VMADDSUBRDUP(7,%r8,Chi_11,UChi_12) \ VMADDSUBRDUP(7,%r8,Chi_01,UChi_02) VMADDSUBRDUP(7,%r8,Chi_11,UChi_12) \
VPREFETCH2(12,%r9) \ VPREFETCH_M2(12,%r9) \
VPREFETCH2(13,%r9) \ VPREFETCH_M2(13,%r9) \
VPREFETCH2(14,%r9) \ VPREFETCH_M2(14,%r9) \
VPREFETCH2(15,%r9) \ VPREFETCH_M2(15,%r9) \
/*30*/ \ /*30*/ \
VMADDSUBIDUP(2,%r8,T1,UChi_00) VMADDSUBIDUP(2,%r8,T2,UChi_10) \ VMADDSUBIDUP(2,%r8,T1,UChi_00) VMADDSUBIDUP(2,%r8,T2,UChi_10) \
VMADDSUBIDUP(5,%r8,T1,UChi_01) VMADDSUBIDUP(5,%r8,T2,UChi_11) \ VMADDSUBIDUP(5,%r8,T1,UChi_01) VMADDSUBIDUP(5,%r8,T2,UChi_11) \
VPREFETCH2(16,%r9) \ VPREFETCH_M2(16,%r9) \
VPREFETCH2(17,%r9) \ VPREFETCH_M2(17,%r9) \
VPREFETCH2(18,%r9) \ VPREFETCH_M2(18,%r9) \
VPREFETCH2(19,%r9) \ VPREFETCH_M2(19,%r9) \
VMADDSUBIDUP(8,%r8,T1,UChi_02) VMADDSUBIDUP(8,%r8,T2,UChi_12) \ VMADDSUBIDUP(8,%r8,T1,UChi_02) VMADDSUBIDUP(8,%r8,T2,UChi_12) \
/*36*/ \ /*36*/ \
VMADDSUBRDUP(2,%r8,Chi_02,UChi_00) VMADDSUBRDUP(2,%r8,Chi_12,UChi_10) \ VMADDSUBRDUP(2,%r8,Chi_02,UChi_00) VMADDSUBRDUP(2,%r8,Chi_12,UChi_10) \
VMADDSUBRDUP(5,%r8,Chi_02,UChi_01) VMADDSUBRDUP(5,%r8,Chi_12,UChi_11) \ VMADDSUBRDUP(5,%r8,Chi_02,UChi_01) VMADDSUBRDUP(5,%r8,Chi_12,UChi_11) \
VMADDSUBRDUP(8,%r8,Chi_02,UChi_02) VMADDSUBRDUP(8,%r8,Chi_12,UChi_12) \ VMADDSUBRDUP(8,%r8,Chi_02,UChi_02) VMADDSUBRDUP(8,%r8,Chi_12,UChi_12) \
VPREFETCH2(20,%r9) \ VPREFETCH_M2(20,%r9) \
VPREFETCH2(21,%r9) \ VPREFETCH_M2(21,%r9) \
VPREFETCH2(22,%r9) \ VPREFETCH_M2(22,%r9) \
VPREFETCH2(23,%r9) \ VPREFETCH_M2(23,%r9) \
VPREFETCHG(2,%r8) \ VPREFETCH_G1(2,%r8) \
VPREFETCHG(3,%r8) \ VPREFETCH_G1(3,%r8) \
VPREFETCH2(4,%r8) \ VPREFETCH_G2(4,%r8) \
VPREFETCH2(5,%r8) \ VPREFETCH_G2(5,%r8) \
/*42 insns*/ ); /*42 insns*/ );
#define MULT_ADDSUB_2SPIN_LSNOPF(ptr,pf) \ #define MULT_ADDSUB_2SPIN_LSNOPF(ptr,pf) \
@ -794,8 +843,8 @@ Author: paboyle <paboyle@ph.ed.ac.uk>
VMADDSUBRDUP(2,%r8,Chi_02,UChi_00) VMADDSUBRDUP(2,%r8,Chi_12,UChi_10) \ VMADDSUBRDUP(2,%r8,Chi_02,UChi_00) VMADDSUBRDUP(2,%r8,Chi_12,UChi_10) \
VMADDSUBRDUP(5,%r8,Chi_02,UChi_01) VMADDSUBRDUP(5,%r8,Chi_12,UChi_11) \ VMADDSUBRDUP(5,%r8,Chi_02,UChi_01) VMADDSUBRDUP(5,%r8,Chi_12,UChi_11) \
VMADDSUBRDUP(8,%r8,Chi_02,UChi_02) VMADDSUBRDUP(8,%r8,Chi_12,UChi_12) \ VMADDSUBRDUP(8,%r8,Chi_02,UChi_02) VMADDSUBRDUP(8,%r8,Chi_12,UChi_12) \
/* VPREFETCHG(2,%r8)*/ \ /* VPREFETCH1(2,%r8)*/ \
/* VPREFETCHG(3,%r8)*/ \ /* VPREFETCH1(3,%r8)*/ \
/*42 insns*/ ); /*42 insns*/ );