mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-29 08:41:22 +00:00 
			
		
		
		
	perplexity : remove extra new lines after chunks (#9596)
This commit is contained in:
		| @@ -444,7 +444,6 @@ static results_perplexity perplexity_v2(llama_context * ctx, const gpt_params & | ||||
|             } | ||||
|             LOG("%.2f minutes\n", total_seconds / 60.0); | ||||
|         } | ||||
|         LOG("\n"); | ||||
|  | ||||
|         //LOG_DBG("%s: using tokens %d...%d\n",__func__,params.n_ctx - params.ppl_stride + start, params.n_ctx + start); | ||||
|         for (int j = n_ctx - params.ppl_stride - 1; j < n_ctx - 1; ++j) { | ||||
| @@ -638,7 +637,6 @@ static results_perplexity perplexity(llama_context * ctx, const gpt_params & par | ||||
|             } | ||||
|             LOG("%.2f minutes\n", total_seconds / 60.0); | ||||
|         } | ||||
|         LOG("\n"); | ||||
|  | ||||
|         for (int seq = 0; seq < n_seq_batch; seq++) { | ||||
|             const float * all_logits = num_batches > 1 ? logits.data() : llama_get_logits_ith(ctx, seq*n_ctx + first); | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Georgi Gerganov
					Georgi Gerganov