search for: __cuda_builtin_blockdim_t

Displaying 2 results from an estimated 2 matches for "__cuda_builtin_blockdim_t".

2020 Aug 22
5
Looking for suggestions: Inferring GPU memory accesses
...stion: - Given a kernel function with M possible memory accesses. For how many of those M accesses we can statically infer its location given concrete values for the grid/block and executing thread? (Assume CUDA only for now) My initial idea is to replace all uses of dim-related values, e.g: __cuda_builtin_blockDim_t::__fetch_builtin_x() __cuda_builtin_gridDim_t::__fetch_builtin_x() and index related values, e.g: __cuda_builtin_blockIdx_t::__fetch_builtin_x() __cuda_builtin_threadIdx_t::__fetch_builtin_x() with ConstantInts. Then run constant folding on the result and check how many GEPs have cons...
2020 Aug 23
2
Looking for suggestions: Inferring GPU memory accesses
...of > > those M accesses we can statically infer its location given concrete > values > > for the grid/block and executing thread? > > > > (Assume CUDA only for now) > > > > My initial idea is to replace all uses of dim-related values, e.g: > > __cuda_builtin_blockDim_t::__fetch_builtin_x() > > __cuda_builtin_gridDim_t::__fetch_builtin_x() > > > > and index related values, e.g: > > __cuda_builtin_blockIdx_t::__fetch_builtin_x() > > __cuda_builtin_threadIdx_t::__fetch_builtin_x() > > > > with ConstantInts...