mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-11-03 09:22:01 +00:00
cleaning debug quant
This commit is contained in:
@@ -315,11 +315,6 @@ static ggml_type llama_tensor_get_type(quantize_state_impl & qs, ggml_type new_t
|
||||
else if (ftype == LLAMA_FTYPE_MOSTLY_IQ3_XXS) {
|
||||
new_type = GGML_TYPE_IQ2_S;
|
||||
}
|
||||
} else if (name.find("ssm_in.weight") != std::string::npos) {
|
||||
// For mamba-based models it's better to not quantize the ssm-proj layers
|
||||
if (ftype == LLAMA_FTYPE_MOSTLY_Q3_K_S) {
|
||||
new_type = GGML_TYPE_BF16;
|
||||
}
|
||||
} else if (name.find("attn_q.weight") != std::string::npos) {
|
||||
if (ftype == LLAMA_FTYPE_MOSTLY_IQ3_XS) {
|
||||
new_type = GGML_TYPE_IQ3_XXS;
|
||||
|
||||
Reference in New Issue
Block a user