Files
llama.cpp/ggml/src/ggml-cuda
uvos 34c961b181 CUDA/HIP: Fix fattn-vec-* when device warp size is not 32 (#12315)
When fattn-wmma was ported over to warp64 various bits that also touch fattn-vec where converted to
selectable warp size, however the fattn-vec kernels dont work with 64 wide warps for now, so we need
to avoid launching them with parameters for warp64
2025-03-12 10:14:11 +01:00
..
2024-11-21 18:18:50 +01:00
2024-11-21 18:18:50 +01:00
2024-08-27 22:41:27 +03:00
2024-08-27 22:41:27 +03:00