mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-11-20 12:07:33 +00:00
cuda: fix rope fusion for gemma3 (#17378)
This commit is contained in:
@@ -3001,6 +3001,10 @@ static void update_cuda_graph_executable(ggml_backend_cuda_context * cuda_ctx) {
|
|||||||
static bool ggml_cuda_should_fuse_rope_set_rows(const ggml_tensor * rope,
|
static bool ggml_cuda_should_fuse_rope_set_rows(const ggml_tensor * rope,
|
||||||
const ggml_tensor * view,
|
const ggml_tensor * view,
|
||||||
const ggml_tensor * set_rows) {
|
const ggml_tensor * set_rows) {
|
||||||
|
|
||||||
|
if (rope->op != GGML_OP_ROPE || view->op != GGML_OP_VIEW || set_rows->op != GGML_OP_SET_ROWS) {
|
||||||
|
return false;
|
||||||
|
}
|
||||||
// ne3 not tested
|
// ne3 not tested
|
||||||
if (rope->src[0]->ne[3] != 1) {
|
if (rope->src[0]->ne[3] != 1) {
|
||||||
return false;
|
return false;
|
||||||
|
|||||||
Reference in New Issue
Block a user