mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-11-06 09:46:50 +00:00
d_ssm -> d_inner;
This commit is contained in:
@@ -6674,7 +6674,8 @@ class FalconH1Model(Mamba2Model):
|
||||
|
||||
# Add Falcon Mamba2 specific configuration
|
||||
self.gguf_writer.add_uint32("falcon_h1.attention.head_dim", self.hparams["head_dim"])
|
||||
self.gguf_writer.add_uint32("falcon_h1.ssm.mamba_d_ssm", self.hparams["mamba_d_ssm"])
|
||||
self.gguf_writer.add_uint32("falcon_h1.ssm.mamba_d_inner", self.hparams["mamba_d_ssm"])
|
||||
self.gguf_writer.add_ssm_inner_size(self.hparams["mamba_d_ssm"])
|
||||
self.gguf_writer.add_uint32("falcon_h1.num_attention_heads", self.find_hparam(["num_attention_heads"]))
|
||||
self.gguf_writer.add_uint32("falcon_h1.num_key_value_heads",
|
||||
self.find_hparam(["num_key_value_heads"], optional=True) or
|
||||
|
||||
Reference in New Issue
Block a user