mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-11-03 09:22:01 +00:00 
			
		
		
		
	fix llama_batch_ext_init_from_text
This commit is contained in:
		@@ -91,8 +91,8 @@ int main(int argc, char ** argv){
 | 
			
		||||
 | 
			
		||||
    const auto t_enc_start = ggml_time_us();
 | 
			
		||||
 | 
			
		||||
    llama_batch_ext_ptr batch0(llama_batch_ext_init_from_text( inp.data(), n_input - 1, 0, 0));
 | 
			
		||||
    llama_batch_ext_ptr batch1(llama_batch_ext_init_from_text(&inp.back(),           1, 0, 0));
 | 
			
		||||
    llama_batch_ext_ptr batch0(llama_batch_ext_init_from_text( inp.data(), n_input - 1, 0, 0, true));
 | 
			
		||||
    llama_batch_ext_ptr batch1(llama_batch_ext_init_from_text(&inp.back(),           1, 0, 0, true));
 | 
			
		||||
    llama_decode_ext(ctx, batch0.get());
 | 
			
		||||
    llama_decode_ext(ctx, batch1.get());
 | 
			
		||||
 | 
			
		||||
 
 | 
			
		||||
		Reference in New Issue
	
	Block a user