mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	correct llama_decode_ext
This commit is contained in:
		| @@ -1058,7 +1058,7 @@ struct common_init_result common_init_from_params(common_params & params) { | ||||
|         } | ||||
|         if (llama_model_has_decoder(model)) { | ||||
|             llama_batch_ext_ptr batch(llama_batch_ext_init_from_text(tmp.data(), std::min(tmp.size(), (size_t) params.n_batch), 0, 0)); | ||||
|             llama_encode_ext(lctx, batch.get()); | ||||
|             llama_decode_ext(lctx, batch.get()); | ||||
|         } | ||||
|         llama_kv_cache_clear(lctx); | ||||
|         llama_synchronize(lctx); | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Xuan Son Nguyen
					Xuan Son Nguyen