mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-11-03 09:22:01 +00:00 
			
		
		
		
	main : fix off by one error for context shift (#6921)
This commit is contained in:
		@@ -544,7 +544,7 @@ int main(int argc, char ** argv) {
 | 
			
		||||
                // if we run out of context:
 | 
			
		||||
                // - take the n_keep first tokens from the original prompt (via n_past)
 | 
			
		||||
                // - take half of the last (n_ctx - n_keep) tokens and recompute the logits in batches
 | 
			
		||||
                if (n_past + (int) embd.size() + std::max<int>(0, guidance_offset) > n_ctx) {
 | 
			
		||||
                if (n_past + (int) embd.size() + std::max<int>(0, guidance_offset) >= n_ctx) {
 | 
			
		||||
                    if (params.n_predict == -2) {
 | 
			
		||||
                        LOG_TEE("\n\n%s: context full and n_predict == -%d => stopping\n", __func__, params.n_predict);
 | 
			
		||||
                        break;
 | 
			
		||||
 
 | 
			
		||||
		Reference in New Issue
	
	Block a user