mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-29 08:41:22 +00:00 
			
		
		
		
	llama-run : fix context size (#11094)
Set `n_ctx` equal to `n_batch` in `Opt` class. Now context size is a more reasonable 2048. Signed-off-by: Eric Curtin <ecurtin@redhat.com>
This commit is contained in:
		| @@ -83,6 +83,7 @@ class Opt { | ||||
|         } | ||||
|  | ||||
|         ctx_params.n_batch        = context_size >= 0 ? context_size : context_size_default; | ||||
|         ctx_params.n_ctx          = ctx_params.n_batch; | ||||
|         model_params.n_gpu_layers = ngl >= 0 ? ngl : ngl_default; | ||||
|         temperature               = temperature >= 0 ? temperature : temperature_default; | ||||
|  | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Eric Curtin
					Eric Curtin