mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	memory : fix broken batch splits for recurrent cache
This commit is contained in:
		| @@ -377,14 +377,18 @@ llama_memory_context_ptr llama_memory_recurrent::init_batch(llama_batch_allocr & | |||||||
|                 ubatch = balloc.split_equal(n_ubatch, false); |                 ubatch = balloc.split_equal(n_ubatch, false); | ||||||
|             } |             } | ||||||
|  |  | ||||||
|             if (balloc.get_n_used() < balloc.get_n_tokens()) { |             if (ubatch.n_tokens == 0) { | ||||||
|                 // failed to find a suitable split |  | ||||||
|                 break; |                 break; | ||||||
|             } |             } | ||||||
|  |  | ||||||
|             ubatches.push_back(std::move(ubatch)); // NOLINT |             ubatches.push_back(std::move(ubatch)); // NOLINT | ||||||
|         } |         } | ||||||
|  |  | ||||||
|  |         if (balloc.get_n_used() < balloc.get_n_tokens()) { | ||||||
|  |             // failed to find a suitable split | ||||||
|  |             break; | ||||||
|  |         } | ||||||
|  |  | ||||||
|         if (!prepare(ubatches)) { |         if (!prepare(ubatches)) { | ||||||
|             break; |             break; | ||||||
|         } |         } | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user
	 Francis Couture-Harpin
					Francis Couture-Harpin