Files
llama.cpp/ggml
Aman Gupta 009b709d6e CUDA: fuse adds, fuse add with rms norm (#15631)
* CUDA: fused add with rms_norm_mul

* Non-broadcast fuse works

* Add fused adds

* format

* Remove n_fuse from template params

* Address review comments

* Move template inside binbcast
2025-08-29 11:35:58 +08:00
..
2025-08-22 15:33:15 +02:00
2024-07-13 18:12:39 +02:00