mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	llama-run : fix context size (#11094)
Set `n_ctx` equal to `n_batch` in `Opt` class. Now context size is a more reasonable 2048. Signed-off-by: Eric Curtin <ecurtin@redhat.com>
This commit is contained in:
		| @@ -83,6 +83,7 @@ class Opt { | |||||||
|         } |         } | ||||||
|  |  | ||||||
|         ctx_params.n_batch        = context_size >= 0 ? context_size : context_size_default; |         ctx_params.n_batch        = context_size >= 0 ? context_size : context_size_default; | ||||||
|  |         ctx_params.n_ctx          = ctx_params.n_batch; | ||||||
|         model_params.n_gpu_layers = ngl >= 0 ? ngl : ngl_default; |         model_params.n_gpu_layers = ngl >= 0 ? ngl : ngl_default; | ||||||
|         temperature               = temperature >= 0 ? temperature : temperature_default; |         temperature               = temperature >= 0 ? temperature : temperature_default; | ||||||
|  |  | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user
	 Eric Curtin
					Eric Curtin