mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-10-29 08:41:22 +00:00
metal : fix batch size for MUL_MAT_ID
This commit is contained in:
@@ -6743,8 +6743,9 @@ static bool ggml_backend_metal_device_supports_buft(ggml_backend_dev_t dev, ggml
|
|||||||
static int64_t get_op_batch_size(const struct ggml_tensor * op) {
|
static int64_t get_op_batch_size(const struct ggml_tensor * op) {
|
||||||
switch (op->op) {
|
switch (op->op) {
|
||||||
case GGML_OP_MUL_MAT:
|
case GGML_OP_MUL_MAT:
|
||||||
case GGML_OP_MUL_MAT_ID:
|
|
||||||
return op->ne[1];
|
return op->ne[1];
|
||||||
|
case GGML_OP_MUL_MAT_ID:
|
||||||
|
return op->ne[2];
|
||||||
default:
|
default:
|
||||||
return ggml_nrows(op);
|
return ggml_nrows(op);
|
||||||
}
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user