mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	context : fix SWA-related warning for multiple sequences (#14045)
This commit is contained in:
		| @@ -123,7 +123,7 @@ llama_context::llama_context( | ||||
|                 __func__, n_ctx_per_seq, hparams.n_ctx_train); | ||||
|     } | ||||
|  | ||||
|     if (!params.swa_full && cparams.n_seq_max > 1) { | ||||
|     if (!params.swa_full && cparams.n_seq_max > 1 && hparams.is_swa_any()) { | ||||
|         LLAMA_LOG_WARN("%s: requested n_seq_max (%u) > 1, but swa_full is not enabled -- performance may be degraded: %s\n", | ||||
|                 __func__, cparams.n_seq_max, "https://github.com/ggml-org/llama.cpp/pull/13845#issuecomment-2924800573"); | ||||
|     } | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Georgi Gerganov
					Georgi Gerganov