Georgi Gerganov 2adf8d83ac parallel : add option for different RNG seeds (#14757) 6 months ago
..
CMakeLists.txt 7cc2d2c889 ggml : move AMX to the CPU backend (#10570) 1 year ago
README.md dd665cc9d4 parallel : increase the variability of the prompt lengths (#13927) 7 months ago
parallel.cpp 2adf8d83ac parallel : add option for different RNG seeds (#14757) 6 months ago

README.md

llama.cpp/example/parallel

Simplified simulation of serving incoming requests in parallel

Example

Generate 128 client requests (-ns 128), simulating 8 concurrent clients (-np 8). The system prompt is shared (-pps), meaning that it is computed once at the start. The client requests consist of up to 10 junk questions (--junk 10) followed by the actual question.

llama-parallel -m model.gguf -np 8 -ns 128 --top-k 1 -pps --junk 10 -c 16384

[!NOTE] It's recommended to use base models with this example. Instruction tuned models might not be able to properly follow the custom chat template specified here, so the results might not be as expected.