mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-11-04 09:32:00 +00:00 
			
		
		
		
	sampling : same for typical sampling
This commit is contained in:
		@@ -909,7 +909,7 @@ static void llama_sampler_typical_apply(struct llama_sampler * smpl, llama_token
 | 
				
			|||||||
        cum_sum += cur_p->data[idx].p;
 | 
					        cum_sum += cur_p->data[idx].p;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
        // Check if the running sum is greater than typical or if we have kept at least min_keep tokens
 | 
					        // Check if the running sum is greater than typical or if we have kept at least min_keep tokens
 | 
				
			||||||
        if (cum_sum > ctx->p && i >= ctx->min_keep - 1) {
 | 
					        if (cum_sum > ctx->p && (ctx->min_keep == 0 || i >= ctx->min_keep - 1)) {
 | 
				
			||||||
            last_idx = i + 1;
 | 
					            last_idx = i + 1;
 | 
				
			||||||
            break;
 | 
					            break;
 | 
				
			||||||
        }
 | 
					        }
 | 
				
			||||||
 
 | 
				
			|||||||
		Reference in New Issue
	
	Block a user