mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-11-04 09:32:00 +00:00 
			
		
		
		
	remove todo
This commit is contained in:
		@@ -583,6 +583,7 @@ class TensorNameMap:
 | 
			
		||||
        ),
 | 
			
		||||
 | 
			
		||||
        MODEL_TENSOR.SSM_NORM: (
 | 
			
		||||
            "model.layers.{bid}.mamba.norm", # falcon-h1
 | 
			
		||||
            "backbone.layers.{bid}.mixer.norm",  # mamba2
 | 
			
		||||
        ),
 | 
			
		||||
 | 
			
		||||
@@ -1177,10 +1178,6 @@ class TensorNameMap:
 | 
			
		||||
            "resampler.attn.out_proj",
 | 
			
		||||
        ),
 | 
			
		||||
        
 | 
			
		||||
        MODEL_TENSOR.SSM_NORM: (
 | 
			
		||||
            "model.layers.{bid}.mamba.norm",
 | 
			
		||||
        ),
 | 
			
		||||
 | 
			
		||||
        MODEL_TENSOR.V_RESMPL_KV: (
 | 
			
		||||
            "resampler.kv_proj",
 | 
			
		||||
        ),
 | 
			
		||||
 
 | 
			
		||||
@@ -1955,8 +1955,7 @@ bool llm_arch_is_recurrent(const llm_arch & arch) {
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
bool llm_arch_is_hybrid(const llm_arch & arch) {
 | 
			
		||||
    // TODO: There are currently no hybrid models! Once there are, this will be
 | 
			
		||||
    //  the place to identify them
 | 
			
		||||
    // List all mamba-attention hybrid models here
 | 
			
		||||
    switch (arch) {
 | 
			
		||||
        case LLM_ARCH_FALCON_H1:
 | 
			
		||||
            return true;
 | 
			
		||||
 
 | 
			
		||||
		Reference in New Issue
	
	Block a user