mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-11-03 09:22:01 +00:00 
			
		
		
		
	convert-falcon-hf-to-gguf.py : add tensor data layout
This commit is contained in:
		@@ -95,7 +95,8 @@ print("gguf: get model metadata")
 | 
			
		||||
block_count = hparams["n_layer"]
 | 
			
		||||
 | 
			
		||||
gguf_writer.add_name(last_dir)
 | 
			
		||||
gguf_writer.add_context_length(2048)
 | 
			
		||||
gguf_writer.add_context_length(2048) # not in config.json
 | 
			
		||||
gguf_writer.add_tensor_data_layout("jploski") # qkv tensor transform
 | 
			
		||||
gguf_writer.add_embedding_length(hparams["hidden_size"])
 | 
			
		||||
gguf_writer.add_feed_forward_length(4 * hparams["hidden_size"])
 | 
			
		||||
gguf_writer.add_block_count(block_count)
 | 
			
		||||
 
 | 
			
		||||
		Reference in New Issue
	
	Block a user