#ifndef GRID_EXTRACT_H #define GRID_EXTRACT_H ///////////////////////////////////////////////////////////////// // Generic extract/merge/permute ///////////////////////////////////////////////////////////////// namespace Grid{ //////////////////////////////////////////////////////////////////////////////////////////////// // Extract/merge a fundamental vector type, to pointer array with offset //////////////////////////////////////////////////////////////////////////////////////////////// template inline void extract(typename std::enable_if::value, const vsimd >::type * y, std::vector &extracted,int offset){ // FIXME: bounce off memory is painful int Nextr=extracted.size(); int Nsimd=vsimd::Nsimd(); int s=Nsimd/Nextr; scalar*buf = (scalar *)y; for(int i=0;i inline void merge(typename std::enable_if::value, vsimd >::type * y, std::vector &extracted,int offset){ int Nextr=extracted.size(); int Nsimd=vsimd::Nsimd(); int s=Nsimd/Nextr; // can have sparse occupation of simd vector if simd_layout does not fill it // replicate n-fold. Use to allow Integer masks to // predicate floating point of various width assignments and maintain conformable. scalar *buf =(scalar *) y; for(int i=0;i inline void extract(typename std::enable_if::value, const vsimd >::type &y,std::vector &extracted){ int Nextr=extracted.size(); int Nsimd=vsimd::Nsimd(); int s=Nsimd/Nextr; scalar *buf = (scalar *)&y; for(int i=0;i inline void merge(typename std::enable_if::value, vsimd >::type &y,std::vector &extracted){ int Nextr=extracted.size(); int Nsimd=vsimd::Nsimd(); int s=Nsimd/Nextr; scalar *buf = (scalar *)&y; for(int i=0;i inline void extract(const vobj &vec,std::vector &extracted) { typedef typename vobj::scalar_type scalar_type ; typedef typename vobj::vector_type vector_type ; const int Nsimd=vobj::vector_type::Nsimd(); int Nextr=extracted.size(); const int words=sizeof(vobj)/sizeof(vector_type); int s=Nsimd/Nextr; std::vector pointers(Nextr); for(int i=0;i(&vp[w],pointers,w); } } //////////////////////////////////////////////////////////////////////// // Extract to a bunch of scalar object pointers, with offset //////////////////////////////////////////////////////////////////////// template inline void extract(const vobj &vec,std::vector &extracted, int offset) { typedef typename vobj::scalar_type scalar_type ; typedef typename vobj::vector_type vector_type ; const int words=sizeof(vobj)/sizeof(vector_type); const int Nsimd=vobj::vector_type::Nsimd(); int Nextr=extracted.size(); int s = Nsimd/Nextr; std::vector pointers(Nsimd); for(int i=0;i(&vp[w],pointers,w); } } //////////////////////////////////////////////////////////////////////// // Merge a contiguous array of scalar objects //////////////////////////////////////////////////////////////////////// template inline void merge(vobj &vec,std::vector &extracted) { typedef typename vobj::scalar_type scalar_type ; typedef typename vobj::vector_type vector_type ; const int Nsimd=vobj::vector_type::Nsimd(); const int words=sizeof(vobj)/sizeof(vector_type); int Nextr = extracted.size(); int splat=Nsimd/Nextr; std::vector pointers(Nextr); for(int i=0;i(&vp[w],pointers,w); } } //////////////////////////////////////////////////////////////////////// // Merge a bunch of different scalar object pointers, with offset //////////////////////////////////////////////////////////////////////// template inline void merge(vobj &vec,std::vector &extracted,int offset) { typedef typename vobj::scalar_type scalar_type ; typedef typename vobj::vector_type vector_type ; const int Nsimd=vobj::vector_type::Nsimd(); const int words=sizeof(vobj)/sizeof(vector_type); int Nextr=extracted.size(); std::vector pointers(Nextr); for(int i=0;i(&vp[w],pointers,w); } } } #endif