g2mt 94933c8c2e server : implement universal assisted decoding (#12635) 5 月之前
..
batched-bench 225e7a1438 llama : add high-throughput mode (#14363) 6 月之前
cvector-generator 745aa5319b llama : deprecate llama_kv_self_ API (#14030) 7 月之前
export-lora 749e0d27f0 mtmd : fix 32-bit narrowing issue in export-lora and mtmd clip (#14503) 5 月之前
gguf-split e9b6350e61 scripts : make the shell scripts cross-platform (#14341) 6 月之前
imatrix d1aa0cc5d1 imatrix: add option to display importance score statistics for a given imatrix file (#12718) 5 月之前
llama-bench c556418b60 llama-bench : use local GPUs along with RPC servers (#14917) 5 月之前
main c82d48ec23 llama : fix `--reverse-prompt` crashing issue (#14794) 6 月之前
mtmd 00fa15fedc mtmd : add support for Voxtral (#14862) 5 月之前
perplexity 745aa5319b llama : deprecate llama_kv_self_ API (#14030) 7 月之前
quantize e9192bec56 quantize : fix using combined imatrix GGUFs (multiple datasets) (#14973) 5 月之前
rpc c508256db2 rpc : Fix build on OpenBSD (#13541) 7 月之前
run a457551332 cmake : do not search for curl libraries by ourselves (#14613) 6 月之前
server 94933c8c2e server : implement universal assisted decoding (#12635) 5 月之前
tokenize 1d36b3670b llama : move end-user examples to tools directory (#13249) 8 月之前
tts 53f925074d sync : vendor (#13901) 7 月之前
CMakeLists.txt 9b61acf060 mtmd : rename llava directory to mtmd (#13311) 8 月之前