Files
llama.cpp/ggml/src
Aman Gupta 9a3ea685b9 CUDA: Fix bug in topk-moe for gpt-oss (#16821)
* CUDA: Fix bug in topk-moe for gpt-oss

When using ggml_can_fuse_subgraph, the output nodes which are passed are wrong. This causes `test-backend-ops` to still fuse ndoes (because the nodes are not used elsewhere in the graph),
but it actually doesn't fuse in the actual gpt-oss

* fix for qwen3 too

* change ifndef to ifdef
2025-10-29 15:55:06 +08:00
..
2025-09-29 17:43:58 +03:00
2025-08-05 22:10:36 +03:00
2025-08-05 22:10:36 +03:00
2025-09-05 11:34:28 +02:00