Document the new max GPU layers default in help (#15771)

This is a key change, just letting users know.

Signed-off-by: Eric Curtin <ericcurtin17@gmail.com>
This commit is contained in:
Eric Curtin
2025-09-04 10:49:44 +01:00
committed by GitHub
parent 0a1b3982cd
commit badb80cadb

View File

@@ -2466,7 +2466,7 @@ common_params_context common_params_parser_init(common_params & params, llama_ex
).set_examples({LLAMA_EXAMPLE_SPECULATIVE, LLAMA_EXAMPLE_SERVER}).set_env("LLAMA_ARG_N_CPU_MOE_DRAFT"));
add_opt(common_arg(
{"-ngl", "--gpu-layers", "--n-gpu-layers"}, "N",
"number of layers to store in VRAM",
string_format("max. number of layers to store in VRAM (default: %d)", params.n_gpu_layers),
[](common_params & params, int value) {
params.n_gpu_layers = value;
if (!llama_supports_gpu_offload()) {