diff --git a/csrc/sampler.cu b/csrc/sampler.cu index d458f8e4c1d02..f7c091f1d4ee4 100644 --- a/csrc/sampler.cu +++ b/csrc/sampler.cu @@ -1,3 +1,4 @@ +#include "cuda_compat.h" #include "dispatch_utils.h" #include @@ -97,7 +98,9 @@ static inline __device__ bool isPartialMatch(float x, uint32_t pattern) { template __device__ void vectorized_process(size_t thread_rank, size_t num_threads, const T* in, idxT len, Func f) { - constexpr int WARP_SIZE = 32; + // Use dynamic WARP_SIZE from cuda_compat.h to support both + // Wave64 (MI300X/gfx942) and Wave32 (Strix Halo/gfx1151) architectures + constexpr int kWarpSize = WARP_SIZE; using WideT = float4; if constexpr (sizeof(T) >= sizeof(WideT)) { for (idxT i = thread_rank; i < len; i += num_threads) { @@ -132,8 +135,8 @@ __device__ void vectorized_process(size_t thread_rank, size_t num_threads, } } - static_assert(WARP_SIZE >= items_per_scalar); - // and because items_per_scalar > skip_cnt, WARP_SIZE > skip_cnt + static_assert(kWarpSize >= items_per_scalar); + // and because items_per_scalar > skip_cnt, kWarpSize > skip_cnt // no need to use loop if (thread_rank < skip_cnt) { f(in[thread_rank], thread_rank); @@ -142,7 +145,7 @@ __device__ void vectorized_process(size_t thread_rank, size_t num_threads, // len_cast * items_per_scalar + items_per_scalar > len - skip_cnt; // and so // len - (skip_cnt + len_cast * items_per_scalar) < items_per_scalar <= - // WARP_SIZE no need to use loop + // kWarpSize no need to use loop const idxT remain_i = skip_cnt + len_cast * items_per_scalar + thread_rank; if (remain_i < len) { f(in[remain_i], remain_i);