mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-11-12 10:47:01 +00:00
read arch from gguf.MODEL_ARCH
This commit is contained in:
@@ -4909,8 +4909,7 @@ class Mamba2Model(TextModel):
|
||||
# Fail early for models which don't have a block expansion factor of 2
|
||||
# TODO: does this really matter?
|
||||
# skip the assertion for FalconH1 Model
|
||||
architectures = self.hparams.get("architectures")
|
||||
if architectures is None or architectures[0] != "FalconH1ForCausalLM":
|
||||
if self.model_arch != gguf.MODEL_ARCH.FALCON_H1:
|
||||
assert d_inner == 2 * d_model
|
||||
assert d_inner % head_dim == 0
|
||||
|
||||
|
||||
Reference in New Issue
Block a user