goerch
|
ff5a3f0c09
Work on the BPE tokenizer (#3252)
|
il y a 2 ans |
cebtenzzre
|
1c84003c08
convert : fix vocab size when not defined in hparams (#3421)
|
il y a 2 ans |
cebtenzzre
|
e78f0b0d05
cmake : increase minimum version for add_link_options (#3444)
|
il y a 2 ans |
shibe2
|
665018c749
CLBlast: Add broadcast support for matrix multiplication (#3402)
|
il y a 2 ans |
cebtenzzre
|
29a404a951
gguf : add BERT, MPT, and GPT-J arch info (#3408)
|
il y a 2 ans |
cebtenzzre
|
0fe321031a
gguf : general usability improvements (#3409)
|
il y a 2 ans |
cebtenzzre
|
9476b01226
cmake : make CUDA flags more similar to the Makefile (#3420)
|
il y a 2 ans |
xaedes
|
a03ce38455
finetune : fix #3404 (#3437)
|
il y a 2 ans |
Adrian
|
a847676984
metal : set log callback before initializing (#3427)
|
il y a 2 ans |
bandoti
|
095231dfd3
cmake : fix transient definitions in find pkg (#3411)
|
il y a 2 ans |
Kevin Ji
|
ea55295a74
docker : ignore Git files (#3314)
|
il y a 2 ans |
vvhg1
|
c97f01c362
infill : add new example + extend server API (#3296)
|
il y a 2 ans |
slaren
|
f5ef5cfb18
ggml-cuda : perform cublas mat mul of quantized types as f16 (#3412)
|
il y a 2 ans |
slaren
|
40e07a60f9
llama.cpp : add documentation about rope_freq_base and scale values (#3401)
|
il y a 2 ans |
Georgi Gerganov
|
bc34dd4f5b
train : fix KQ_pos allocation (#3392)
|
il y a 2 ans |
Cebtenzzre
|
2777a84be4
llama : quantize up to 31% faster on Linux and Windows with mmap (#3206)
|
il y a 2 ans |
BarfingLemurs
|
0a4a4a0982
readme : update hot topics + model links (#3399)
|
il y a 2 ans |
Andrew Duffy
|
569550df20
readme : add link to grammars app (#3388)
|
il y a 2 ans |
Jhen-Jie Hong
|
c71bf2c45c
swift : fix build on xcode 15 (#3387)
|
il y a 2 ans |
Cebtenzzre
|
bc39553c90
build : enable more non-default compiler warnings (#3200)
|
il y a 2 ans |
Hua Jiang
|
0ccfc62a96
ggml_tensor: update the structure comments. (#3283)
|
il y a 2 ans |
Qu Zongfu
|
7f1a0fe709
ggml : release the requested thread pool resource (#3292)
|
il y a 2 ans |
slaren
|
16bc66d947
llama.cpp : split llama_context_params into model and context params (#3301)
|
il y a 2 ans |
Eve
|
0512d66670
ci : multithreaded builds (#3311)
|
il y a 2 ans |
xaedes
|
0e76a8992c
train : finetune LORA (#2632)
|
il y a 2 ans |
Cebtenzzre
|
2db94d98ed
gguf : basic type checking in gguf_get_* (#3346)
|
il y a 2 ans |
Cebtenzzre
|
ecf90b1a51
gguf : make token scores and types optional (#3347)
|
il y a 2 ans |
Georgi Gerganov
|
2619109ad5
ci : disable freeBSD builds due to lack of VMs (#3381)
|
il y a 2 ans |
Georgi Gerganov
|
ec893798b7
llama : custom attention mask + parallel decoding + no context swaps (#3228)
|
il y a 2 ans |
Kevin Ji
|
45855b3f1c
docs : mark code as Bash (#3375)
|
il y a 2 ans |