mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-11-06 09:46:50 +00:00
sampling : same for typical sampling
This commit is contained in:
@@ -909,7 +909,7 @@ static void llama_sampler_typical_apply(struct llama_sampler * smpl, llama_token
|
|||||||
cum_sum += cur_p->data[idx].p;
|
cum_sum += cur_p->data[idx].p;
|
||||||
|
|
||||||
// Check if the running sum is greater than typical or if we have kept at least min_keep tokens
|
// Check if the running sum is greater than typical or if we have kept at least min_keep tokens
|
||||||
if (cum_sum > ctx->p && i >= ctx->min_keep - 1) {
|
if (cum_sum > ctx->p && (ctx->min_keep == 0 || i >= ctx->min_keep - 1)) {
|
||||||
last_idx = i + 1;
|
last_idx = i + 1;
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user