mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-11-12 10:47:01 +00:00
read arch from gguf.MODEL_ARCH
This commit is contained in:
@@ -4909,8 +4909,7 @@ class Mamba2Model(TextModel):
|
|||||||
# Fail early for models which don't have a block expansion factor of 2
|
# Fail early for models which don't have a block expansion factor of 2
|
||||||
# TODO: does this really matter?
|
# TODO: does this really matter?
|
||||||
# skip the assertion for FalconH1 Model
|
# skip the assertion for FalconH1 Model
|
||||||
architectures = self.hparams.get("architectures")
|
if self.model_arch != gguf.MODEL_ARCH.FALCON_H1:
|
||||||
if architectures is None or architectures[0] != "FalconH1ForCausalLM":
|
|
||||||
assert d_inner == 2 * d_model
|
assert d_inner == 2 * d_model
|
||||||
assert d_inner % head_dim == 0
|
assert d_inner % head_dim == 0
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user