mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-11-06 09:46:50 +00:00
rm unused MAMBA_CHUNK_SIZE
This commit is contained in:
@@ -6672,7 +6672,6 @@ class FalconH1Model(Mamba2Model):
|
||||
|
||||
|
||||
# Add Falcon Mamba2 specific configuration
|
||||
self.gguf_writer.add_uint32("falcon_h1.ssm.mamba_chunk_size", self.hparams["mamba_chunk_size"])
|
||||
self.gguf_writer.add_uint32("falcon_h1.attention.head_dim", self.hparams["head_dim"])
|
||||
self.gguf_writer.add_uint32("falcon_h1.ssm.mamba_d_ssm", self.hparams["mamba_d_ssm"])
|
||||
self.gguf_writer.add_uint32("falcon_h1.num_attention_heads", self.find_hparam(["num_attention_heads"]))
|
||||
|
||||
Reference in New Issue
Block a user