mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-11-13 10:57:15 +00:00
injected mup
This commit is contained in:
@@ -163,18 +163,8 @@ enum llm_kv {
|
||||
LLM_KV_MAMBA_D_SSM,
|
||||
LLM_KV_N_LAYER,
|
||||
LLM_KV_FALCON_H1_USE_MLP,
|
||||
LLM_KV_FALCON_H1_ATTENTION_IN_MULTIPLIER,
|
||||
LLM_KV_FALCON_H1_ATTENTION_OUT_MULTIPLIER,
|
||||
LLM_KV_FALCON_H1_SSM_IN_MULTIPLIER,
|
||||
LLM_KV_FALCON_H1_SSM_OUT_MULTIPLIER,
|
||||
LLM_KV_FALCON_H1_MLP_GATE_MULTIPLIER,
|
||||
LLM_KV_FALCON_H1_MLP_DOWN_MULTIPLIER,
|
||||
LLM_KV_FALCON_H1_SSM_HAS_MUP,
|
||||
LLM_KV_FALCON_H1_MAMBA_NORM_BEFORE_GATE,
|
||||
LLM_KV_FALCON_H1_MAMBA_RMS_NORM,
|
||||
LLM_KV_FALCON_H1_KEY_MULTIPLIER,
|
||||
LLM_KV_FALCON_H1_LM_HEAD_MULTIPLIER,
|
||||
LLM_KV_FALCON_H1_EMBEDDING_MULTIPLIER,
|
||||
LLM_KV_FALCON_H1_MAMBA_CHUNK_SIZE,
|
||||
|
||||
LLM_KV_ROPE_DIMENSION_COUNT,
|
||||
@@ -410,7 +400,6 @@ enum llm_tensor {
|
||||
LLM_TENSOR_POS_NET_ATTN_K,
|
||||
LLM_TENSOR_POS_NET_ATTN_V,
|
||||
LLM_TENSOR_POS_NET_ATTN_OUT,
|
||||
LLM_TENSOR_SSM_MUP_VEC,
|
||||
LLM_TENSOR_FFN_PRE_NORM,
|
||||
LLM_TENSOR_FINAL_NORM,
|
||||
};
|
||||
|
||||
Reference in New Issue
Block a user