Olivier Chafik
|
7593639ce3
`main`: add --json-schema / -j flag (#6659)
|
1 year ago |
Clint Herron
|
57dd02c44b
Tests: Added integration tests for GBNF parser (#6472)
|
1 year ago |
Olivier Chafik
|
5b7b0ac8df
json-schema-to-grammar improvements (+ added to server) (#5978)
|
1 year ago |
Xuan Son Nguyen
|
11b12de39b
llama : add llama_chat_apply_template() (#5538)
|
1 year ago |
crasm
|
413e7b0559
ci : add model tests + script wrapper (#4586)
|
2 years ago |
Georgi Gerganov
|
c918fe8dca
metal : create autorelease pool during library build (#4970)
|
2 years ago |
Cuong Trinh Manh
|
97bbca6e85
cmake : fix ld warning duplicate libraries libllama.a (#4671)
|
2 years ago |
manikbhandari
|
ea5497df5d
gpt2 : Add gpt2 architecture integration (#4555)
|
2 years ago |
Georgi Gerganov
|
fe680e3d10
sync : ggml (new ops, tests, backend, etc.) (#4359)
|
2 years ago |
Galunid
|
36eed0c42c
stablelm : StableLM support (#3586)
|
2 years ago |
Galunid
|
daab3d7f45
Add more tokenizer tests (#3742)
|
2 years ago |
goerch
|
9e70cc0322
Add test for MPT tokenization (#3728)
|
2 years ago |
goerch
|
ff5a3f0c09
Work on the BPE tokenizer (#3252)
|
2 years ago |
Georgi Gerganov
|
ec893798b7
llama : custom attention mask + parallel decoding + no context swaps (#3228)
|
2 years ago |
goerch
|
71ca2fad7d
whisper : tokenizer fix + re-enable tokenizer test for LLaMa (#3096)
|
2 years ago |
Cebtenzzre
|
849408957c
tests : add a C compliance test (#2848)
|
2 years ago |
Georgi Gerganov
|
edd4c14817
llama : more tokenizer fixes (#2810)
|
2 years ago |
Georgi Gerganov
|
cf658adc83
llm : add Falcon support (#2717)
|
2 years ago |
Georgi Gerganov
|
6381d4e110
gguf : new file format with flexible meta data (beta) (#2398)
|
2 years ago |
drbh
|
7cf54e1f74
tests : adds simple llama grammar tests (#2618)
|
2 years ago |
drbh
|
ee77efea2a
test : add simple grammar parsing tests (#2594)
|
2 years ago |
Eve
|
81844fbcfd
tests : Fix compilation warnings (Linux/GCC) (#2451)
|
2 years ago |
wzy
|
b1f4290953
cmake : install targets (#2256)
|
2 years ago |
Qingyou Meng
|
1d656d6360
ggml : change ggml_graph_compute() API to not require context (#1999)
|
2 years ago |
xaedes
|
f954edda93
ggml : implement backward pass for llama + small training-llama-from-scratch example (#1360)
|
2 years ago |
Ivan Stepanov
|
dd7eff57d8
llama : new sampling algorithms (#1126)
|
2 years ago |
unbounded
|
5f939498d5
ggml : unit test for quantization functions (#953)
|
2 years ago |
Stephan Walter
|
436e561931
all : be more strict about converting float to double (#458)
|
2 years ago |
Georgi Gerganov
|
a316a425d0
Overhaul the examples structure
|
2 years ago |
Stephan Walter
|
69c92298a9
Deduplicate q4 quantization functions (#383)
|
2 years ago |