| .. |
|
peg-parser
|
0a8026e768
common : introduce composable PEG parser combinators for chat parsing (#17136)
|
1 maand geleden |
|
.gitignore
|
0a8026e768
common : introduce composable PEG parser combinators for chat parsing (#17136)
|
1 maand geleden |
|
CMakeLists.txt
|
0a8026e768
common : introduce composable PEG parser combinators for chat parsing (#17136)
|
1 maand geleden |
|
get-model.cpp
|
413e7b0559
ci : add model tests + script wrapper (#4586)
|
2 jaren geleden |
|
get-model.h
|
413e7b0559
ci : add model tests + script wrapper (#4586)
|
2 jaren geleden |
|
run-json-schema-to-grammar.mjs
|
1d36b3670b
llama : move end-user examples to tools directory (#13249)
|
8 maanden geleden |
|
test-alloc.cpp
|
638d330246
ggml : fix graph reallocation with multiple chunks (#16396)
|
3 maanden geleden |
|
test-arg-parser.cpp
|
380b4c984e
common: support negated args (#17919)
|
1 maand geleden |
|
test-autorelease.cpp
|
afa8a9ec9b
llama : add `llama_vocab`, functions -> methods, naming (#11110)
|
1 jaar geleden |
|
test-backend-ops.cpp
|
07a10c1090
vulkan: Allow non-pow2 n_experts in topk_moe (#17872)
|
1 maand geleden |
|
test-barrier.cpp
|
e1f4921980
Fix race conditions in threadpool when dealing with dynamic/frequent n_threads changes (#17748)
|
1 maand geleden |
|
test-c.c
|
d4cdd9c1c3
ggml : remove kompute backend (#14501)
|
6 maanden geleden |
|
test-chat-parser.cpp
|
2c301e91ab
common : handle unicode during partial json parsing (#16526)
|
3 maanden geleden |
|
test-chat-peg-parser.cpp
|
0a8026e768
common : introduce composable PEG parser combinators for chat parsing (#17136)
|
1 maand geleden |
|
test-chat-template.cpp
|
c08002a198
chat : Granite Docling stopping (#16438)
|
3 maanden geleden |
|
test-chat.cpp
|
2fbe3b7bb7
common : add parser for ministral/mistral large 3/devstral 2 (#17713)
|
1 maand geleden |
|
test-double-float.cpp
|
370b1f7e7a
ggml : minor naming changes (#8433)
|
1 jaar geleden |
|
test-gbnf-validator.cpp
|
13b4548877
cmake : do not include ./src as public for libllama (#13062)
|
8 maanden geleden |
|
test-gguf.cpp
|
7675c555a1
gguf: fix failure on version == 0 (#13956)
|
7 maanden geleden |
|
test-grammar-integration.cpp
|
e39502e74b
llama : add token matching support to llama-grammar (#17816)
|
1 maand geleden |
|
test-grammar-llguidance.cpp
|
13b4548877
cmake : do not include ./src as public for libllama (#13062)
|
8 maanden geleden |
|
test-grammar-parser.cpp
|
e39502e74b
llama : add token matching support to llama-grammar (#17816)
|
1 maand geleden |
|
test-json-partial.cpp
|
2c301e91ab
common : handle unicode during partial json parsing (#16526)
|
3 maanden geleden |
|
test-json-schema-to-grammar.cpp
|
c4357dcc35
Server: Change Invalid Schema from Server Error (500) to User Error (400) (#17572)
|
1 maand geleden |
|
test-llama-grammar.cpp
|
e39502e74b
llama : add token matching support to llama-grammar (#17816)
|
1 maand geleden |
|
test-log.cpp
|
7eee341bee
common : use common_ prefix for common library functions (#9805)
|
1 jaar geleden |
|
test-lora-conversion-inference.sh
|
6c2131773c
cli: new CLI experience (#17824)
|
1 maand geleden |
|
test-model-load-cancel.cpp
|
47182dd03f
llama : update llama_model API names (#11063)
|
1 jaar geleden |
|
test-mtmd-c-api.c
|
27aa259532
mtmd : add C public API (#13184)
|
8 maanden geleden |
|
test-opt.cpp
|
bcbddcd54f
tests : fix test-opt with GGML_BACKEND_DL (#15599)
|
4 maanden geleden |
|
test-peg-parser.cpp
|
0a8026e768
common : introduce composable PEG parser combinators for chat parsing (#17136)
|
1 maand geleden |
|
test-quantize-fns.cpp
|
e128a1bf5b
tests : fix test-quantize-fns to init the CPU backend (#12306)
|
10 maanden geleden |
|
test-quantize-perf.cpp
|
bee378e098
ci: run the x64 and arm ci on the github machines instead (#16183)
|
3 maanden geleden |
|
test-quantize-stats.cpp
|
ec18edfcba
server: introduce API for serving / loading / unloading multiple models (#17470)
|
1 maand geleden |
|
test-regex-partial.cpp
|
3198405e98
`common`: add partial regex support (#12808)
|
8 maanden geleden |
|
test-rope.cpp
|
73460f6278
ggml-cpu: templateify ggml_compute_forward_rope_f32 and _f16 (#16805)
|
2 maanden geleden |
|
test-sampling.cpp
|
e92d53b29e
sampling : optimize samplers by reusing bucket sort (#15665)
|
4 maanden geleden |
|
test-thread-safety.cpp
|
cd5e3b5754
server : support unified cache across slots (#16736)
|
2 maanden geleden |
|
test-tokenizer-0.cpp
|
afa8a9ec9b
llama : add `llama_vocab`, functions -> methods, naming (#11110)
|
1 jaar geleden |
|
test-tokenizer-0.py
|
6fbd432211
py : logging and flake8 suppression refactoring (#7081)
|
1 jaar geleden |
|
test-tokenizer-0.sh
|
e9b6350e61
scripts : make the shell scripts cross-platform (#14341)
|
6 maanden geleden |
|
test-tokenizer-1-bpe.cpp
|
13b4548877
cmake : do not include ./src as public for libllama (#13062)
|
8 maanden geleden |
|
test-tokenizer-1-spm.cpp
|
13b4548877
cmake : do not include ./src as public for libllama (#13062)
|
8 maanden geleden |
|
test-tokenizer-random.py
|
70cd37dbbe
requirements : update transformers/torch for Embedding Gemma (#15828)
|
4 maanden geleden |
|
test-tokenizers-repo.sh
|
624207e676
devops: add s390x & ppc64le CI (#15925)
|
3 maanden geleden |