Explorar o código

server: update readme to mention n_past_max metric (#16436)

https://github.com/ggml-org/llama.cpp/pull/15361 added new metric
exported, but I've missed this doc.
Oleksandr Kuvshynov hai 3 meses
pai
achega
c5fef0fcea
Modificáronse 1 ficheiros con 1 adicións e 0 borrados
  1. 1 0
      tools/server/README.md

+ 1 - 0
tools/server/README.md

@@ -1045,6 +1045,7 @@ Available metrics:
 - `llamacpp:kv_cache_tokens`: KV-cache tokens.
 - `llamacpp:requests_processing`: Number of requests processing.
 - `llamacpp:requests_deferred`: Number of requests deferred.
+- `llamacpp:n_past_max`: High watermark of the context size observed.
 
 ### POST `/slots/{id_slot}?action=save`: Save the prompt cache of the specified slot to a file.