mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-11-04 09:32:00 +00:00 
			
		
		
		
	gptneox-main.cpp : add tensor data layout
This commit is contained in:
		@@ -381,6 +381,8 @@ bool gpt_neox_model_load(const std::string & fname, gpt_neox_model & model, gpt2
 | 
				
			|||||||
        if (keyidx != -1) { fprintf(stdout, "%s: model architecture   = %s\n", __func__, gguf_get_val_str(ggufctx, keyidx)); }
 | 
					        if (keyidx != -1) { fprintf(stdout, "%s: model architecture   = %s\n", __func__, gguf_get_val_str(ggufctx, keyidx)); }
 | 
				
			||||||
        keyidx = gguf_find_key(ggufctx, "general.file_type");
 | 
					        keyidx = gguf_find_key(ggufctx, "general.file_type");
 | 
				
			||||||
        if (keyidx != -1) { fprintf(stdout, "%s: model file type      = %s\n", __func__, gguf_get_val_str(ggufctx, keyidx)); }
 | 
					        if (keyidx != -1) { fprintf(stdout, "%s: model file type      = %s\n", __func__, gguf_get_val_str(ggufctx, keyidx)); }
 | 
				
			||||||
 | 
					        keyidx = gguf_find_key(ggufctx, "gptneox.tensor_data_layout");
 | 
				
			||||||
 | 
					        if (keyidx != -1) { fprintf(stdout, "%s: model data layout    = %s\n", __func__, gguf_get_val_str(ggufctx, keyidx)); }
 | 
				
			||||||
        keyidx = gguf_find_key(ggufctx, "general.source.hugginface.repository");
 | 
					        keyidx = gguf_find_key(ggufctx, "general.source.hugginface.repository");
 | 
				
			||||||
        if (keyidx != -1) { fprintf(stdout, "%s: model source HF repo = %s\n", __func__, gguf_get_val_str(ggufctx, keyidx)); }
 | 
					        if (keyidx != -1) { fprintf(stdout, "%s: model source HF repo = %s\n", __func__, gguf_get_val_str(ggufctx, keyidx)); }
 | 
				
			||||||
    }
 | 
					    }
 | 
				
			||||||
 
 | 
				
			|||||||
		Reference in New Issue
	
	Block a user