mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-10-31 08:51:55 +00:00
memory : fix broken batch splits for recurrent cache
This commit is contained in:
@@ -377,14 +377,18 @@ llama_memory_context_ptr llama_memory_recurrent::init_batch(llama_batch_allocr &
|
|||||||
ubatch = balloc.split_equal(n_ubatch, false);
|
ubatch = balloc.split_equal(n_ubatch, false);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (balloc.get_n_used() < balloc.get_n_tokens()) {
|
if (ubatch.n_tokens == 0) {
|
||||||
// failed to find a suitable split
|
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
|
||||||
ubatches.push_back(std::move(ubatch)); // NOLINT
|
ubatches.push_back(std::move(ubatch)); // NOLINT
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (balloc.get_n_used() < balloc.get_n_tokens()) {
|
||||||
|
// failed to find a suitable split
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
if (!prepare(ubatches)) {
|
if (!prepare(ubatches)) {
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user