mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	llama : check LLAMA_TRACE env for extra logging (#4929)
* llama : minor fix indent * llama : check LLAMA_TRACE env for extra logging ggml-ci
This commit is contained in:
		
							
								
								
									
										32
									
								
								llama.cpp
									
									
									
									
									
								
							
							
						
						
									
										32
									
								
								llama.cpp
									
									
									
									
									
								
							| @@ -2190,6 +2190,11 @@ struct llama_model_loader { | |||||||
|     LLM_KV      llm_kv    = LLM_KV(LLM_ARCH_UNKNOWN); |     LLM_KV      llm_kv    = LLM_KV(LLM_ARCH_UNKNOWN); | ||||||
|  |  | ||||||
|     llama_model_loader(const std::string & fname, bool use_mmap, const struct llama_model_kv_override * param_overrides_p) : file(fname.c_str(), "rb") { |     llama_model_loader(const std::string & fname, bool use_mmap, const struct llama_model_kv_override * param_overrides_p) : file(fname.c_str(), "rb") { | ||||||
|  |         int trace = 0; | ||||||
|  |         if (getenv("LLAMA_TRACE")) { | ||||||
|  |             trace = atoi(getenv("LLAMA_TRACE")); | ||||||
|  |         } | ||||||
|  |  | ||||||
|         struct gguf_init_params params = { |         struct gguf_init_params params = { | ||||||
|             /*.no_alloc = */ true, |             /*.no_alloc = */ true, | ||||||
|             /*.ctx      = */ &ctx_meta, |             /*.ctx      = */ &ctx_meta, | ||||||
| @@ -2242,11 +2247,10 @@ struct llama_model_loader { | |||||||
|                     type_max   = type; |                     type_max   = type; | ||||||
|                 } |                 } | ||||||
|  |  | ||||||
|                 // TODO: make runtime configurable |                 if (trace > 0) { | ||||||
| #if 0 |                     struct ggml_tensor * meta = ggml_get_tensor(ctx_meta, gguf_get_tensor_name(ctx_gguf, i)); | ||||||
|                 struct ggml_tensor * meta = ggml_get_tensor(ctx_meta, gguf_get_tensor_name(ctx_gguf, i)); |                     LLAMA_LOG_INFO("%s: - tensor %4d: %32s %-8s [ %s ]\n", __func__, i, ggml_get_name(meta), ggml_type_name(type), llama_format_tensor_shape(meta).c_str()); | ||||||
|                 LLAMA_LOG_INFO("%s: - tensor %4d: %32s %-8s [ %s ]\n", __func__, i, ggml_get_name(meta), ggml_type_name(type), llama_format_tensor_shape(meta).c_str()); |                 } | ||||||
| #endif |  | ||||||
|             } |             } | ||||||
|  |  | ||||||
|             switch (type_max) { |             switch (type_max) { | ||||||
| @@ -6451,15 +6455,15 @@ static uint8_t llama_token_to_byte(const llama_vocab& vocab, llama_token id) { | |||||||
| static llama_token llama_byte_to_token(const llama_vocab & vocab, uint8_t ch) { | static llama_token llama_byte_to_token(const llama_vocab & vocab, uint8_t ch) { | ||||||
|     static const char * hex = "0123456789ABCDEF"; |     static const char * hex = "0123456789ABCDEF"; | ||||||
|     switch (llama_vocab_get_type(vocab)) { |     switch (llama_vocab_get_type(vocab)) { | ||||||
|     case LLAMA_VOCAB_TYPE_SPM: { |         case LLAMA_VOCAB_TYPE_SPM: { | ||||||
|         const char buf[7] = { '<', '0', 'x', hex[ch >> 4], hex[ch & 15], '>', 0 }; |             const char buf[7] = { '<', '0', 'x', hex[ch >> 4], hex[ch & 15], '>', 0 }; | ||||||
|         return vocab.token_to_id.at(buf); |             return vocab.token_to_id.at(buf); | ||||||
|     } |         } | ||||||
|     case LLAMA_VOCAB_TYPE_BPE: { |         case LLAMA_VOCAB_TYPE_BPE: { | ||||||
|         return vocab.token_to_id.at(bytes_to_unicode_bpe(ch)); |             return vocab.token_to_id.at(bytes_to_unicode_bpe(ch)); | ||||||
|     } |         } | ||||||
|     default: |         default: | ||||||
|         GGML_ASSERT(false); |             GGML_ASSERT(false); | ||||||
|     } |     } | ||||||
| } | } | ||||||
|  |  | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user
	 Georgi Gerganov
					Georgi Gerganov