Browse Source

Document the new max GPU layers default in help (#15771)

This is a key change, just letting users know.

Signed-off-by: Eric Curtin <ericcurtin17@gmail.com>
Eric Curtin 4 months ago
parent
commit
badb80cadb
1 changed files with 1 additions and 1 deletions
  1. 1 1
      common/arg.cpp

+ 1 - 1
common/arg.cpp

@@ -2466,7 +2466,7 @@ common_params_context common_params_parser_init(common_params & params, llama_ex
     ).set_examples({LLAMA_EXAMPLE_SPECULATIVE, LLAMA_EXAMPLE_SERVER}).set_env("LLAMA_ARG_N_CPU_MOE_DRAFT"));
     add_opt(common_arg(
         {"-ngl", "--gpu-layers", "--n-gpu-layers"}, "N",
-        "number of layers to store in VRAM",
+        string_format("max. number of layers to store in VRAM (default: %d)", params.n_gpu_layers),
         [](common_params & params, int value) {
             params.n_gpu_layers = value;
             if (!llama_supports_gpu_offload()) {