Brian
|
a2ac89d6ef
convert.py : add python logging instead of print() (#6511)
|
hai 1 ano |
Pedro Cuenca
|
b97bc3966e
llama : support Llama 3 HF conversion (#6745)
|
hai 1 ano |
slaren
|
65c64dc36f
convert.py : add consolidated.safetensors for mixtral 8x22b (#6587)
|
hai 1 ano |
Jared Van Bortel
|
1b67731e18
BERT tokenizer fixes (#6498)
|
hai 1 ano |
kunnis
|
cecd8d3c98
Comment explaining a decision (#6531)
|
hai 1 ano |
slaren
|
08a0c02060
ggml : mul_mat_id use the same tensor for all the experts (#6387)
|
hai 1 ano |
Jared Van Bortel
|
be55134a53
convert : refactor vocab selection logic (#6355)
|
hai 1 ano |
Romain D
|
3a6efdd03c
convert : use f32 outtype for bf16 tensors (#6106)
|
hai 1 ano |
Michael Podvitskiy
|
69ff61397d
llama : support models without vocabulary (#5798)
|
hai 1 ano |
Georgi Gerganov
|
1e35d619a6
convert : remove AWQ remnants (#5768)
|
hai 1 ano |
Jared Van Bortel
|
4d4d2366fc
convert : automatically fall back to HfVocab if tokenizer.model doesn't exist (#5821)
|
hai 1 ano |
John
|
aa23412989
llava : support v1.6 (#5267)
|
hai 1 ano |
Sang-Kil Park
|
f68664ac24
convert : fix TypeError on GPT-2 vocab.json (#5288)
|
hai 1 ano |
Georgi Gerganov
|
906cff55c2
py : handle byte tokens in `get_token_type` (#5341)
|
hai 1 ano |
Georgi Gerganov
|
14fef85e2d
py : fix except (#5194)
|
hai 1 ano |
Sang-Kil Park
|
e76627bcce
py : improve BPE tokenizer support (#5189)
|
hai 1 ano |
Jared Van Bortel
|
b43ebde3b0
convert : partially revert PR #4818 (#5041)
|
%!s(int64=2) %!d(string=hai) anos |
David Sommers
|
b46757735d
convert.py : fix llama/llama2 conversion due to vocab_size=-1 (#5019)
|
%!s(int64=2) %!d(string=hai) anos |
Georgi Gerganov
|
0f83e727af
py : fix whitespace
|
%!s(int64=2) %!d(string=hai) anos |
Georgi Gerganov
|
4f4bf35f46
py : fix missing added_tokens_dict for SPM and BPE vocabs (#4971)
|
%!s(int64=2) %!d(string=hai) anos |
Austin
|
6efb8eb30e
convert.py : fix vanilla LLaMA model conversion (#4818)
|
%!s(int64=2) %!d(string=hai) anos |
Nam D. Tran
|
f6793491b5
llama : add AWQ for llama, llama2, mpt, and mistral models (#4593)
|
%!s(int64=2) %!d(string=hai) anos |
wonjun Jang
|
f56d6077d0
Add byte token type when tokenizer.model is not exists (#4641)
|
%!s(int64=2) %!d(string=hai) anos |
wonjun Jang
|
873637afc7
convert : support loading vocab from fast tokenizer config (#3633)
|
%!s(int64=2) %!d(string=hai) anos |
slaren
|
799a1cb13b
llama : add Mixtral support (#4406)
|
%!s(int64=2) %!d(string=hai) anos |
Richard Kiss
|
9494d7c477
english : use `typos` to fix comments and logs (#4354)
|
%!s(int64=2) %!d(string=hai) anos |
slaren
|
f4d973cecb
convert.py : fix llama/llama2 conversion due to vocab_size=-1 (#4258)
|
%!s(int64=2) %!d(string=hai) anos |
crasm
|
3014b5415d
Update docs for yarn_ext_factor <0.0 as unspecified instead of NaN (#4189)
|
%!s(int64=2) %!d(string=hai) anos |
Galunid
|
f23c0359a3
ci : add flake8 to github actions (python linting) (#4129)
|
%!s(int64=2) %!d(string=hai) anos |
Don Mahurin
|
2ab0707acb
convert : use 'model' value if it exists. This allows karpathy/tinyllamas to load (#4089)
|
%!s(int64=2) %!d(string=hai) anos |