Kerfuffle
|
7694adda8d
Fix for main example getting stuck when -n -2 and --interactive (#2767)
|
%!s(int64=2) %!d(string=hai) anos |
slaren
|
fea95c682d
fix convert.py for codellama, add llama 34B to the list of recognized models (#2768)
|
%!s(int64=2) %!d(string=hai) anos |
DannyDaemonic
|
ef955fbd23
Tag release with build number (#2732)
|
%!s(int64=2) %!d(string=hai) anos |
Georgi Gerganov
|
d67777c202
metal : add Q8_0 support (#2763)
|
%!s(int64=2) %!d(string=hai) anos |
Georgi Gerganov
|
c3e53b421a
llama : escape all U+2581 in a string (#2750)
|
%!s(int64=2) %!d(string=hai) anos |
Evan Jones
|
6e91a1b070
llama : fix grammar sometimes generating null char (#2756)
|
%!s(int64=2) %!d(string=hai) anos |
Georgi Gerganov
|
44d5462b5c
readme : fix link
|
%!s(int64=2) %!d(string=hai) anos |
Georgi Gerganov
|
c7868b0753
minor : fix trailing whitespace
|
%!s(int64=2) %!d(string=hai) anos |
Georgi Gerganov
|
79da24b58c
readme : update hot topics
|
%!s(int64=2) %!d(string=hai) anos |
Georgi Gerganov
|
cf658adc83
llm : add Falcon support (#2717)
|
%!s(int64=2) %!d(string=hai) anos |
Georgi Gerganov
|
a192860cfe
minor : fix trailing whitespace
|
%!s(int64=2) %!d(string=hai) anos |
Olivier Chafik
|
95385241a9
examples : restore the functionality to import llama2.c models (#2685)
|
%!s(int64=2) %!d(string=hai) anos |
slaren
|
335acd2ffd
fix convert-lora-to-ggml.py (#2738)
|
%!s(int64=2) %!d(string=hai) anos |
klosax
|
5290c38e6e
main : insert bos if no tokens (#2727)
|
%!s(int64=2) %!d(string=hai) anos |
akawrykow
|
cc34dbda96
gitignore : fix for windows (#2729)
|
%!s(int64=2) %!d(string=hai) anos |
Cebtenzzre
|
7c2227a197
chmod : make scripts executable (#2675)
|
%!s(int64=2) %!d(string=hai) anos |
JohnnyB
|
f19dca04ea
devops : RPM Specs (#2723)
|
%!s(int64=2) %!d(string=hai) anos |
Kawrakow
|
8207214b6a
Fix values shown in the quantize tool help (#2735)
|
%!s(int64=2) %!d(string=hai) anos |
Kawrakow
|
62959e740e
Strided perplexity (#2714)
|
%!s(int64=2) %!d(string=hai) anos |
IgnacioFDM
|
7f7ddd5002
Fix ggml to gguf conversion on Windows (#2733)
|
%!s(int64=2) %!d(string=hai) anos |
Xiao-Yong Jin
|
b8ad1b66b2
server : allow json array in prompt or content for direct token input (#2306)
|
%!s(int64=2) %!d(string=hai) anos |
Evan Jones
|
f5fe98d11b
docs : add grammar docs (#2701)
|
%!s(int64=2) %!d(string=hai) anos |
Kerfuffle
|
777f42ba18
Improve handling of special tokens in GGML to GGUF converter (#2725)
|
%!s(int64=2) %!d(string=hai) anos |
goerch
|
46ef5b5fcf
llama : fix whitespace escaping in tokenizer (#2724)
|
%!s(int64=2) %!d(string=hai) anos |
Johannes Gäßler
|
c63bb1d16a
CUDA: use mul_mat_q kernels by default (#2683)
|
%!s(int64=2) %!d(string=hai) anos |
Alex Petenchea
|
3b6cfe7c92
convert.py : clarifying error message (#2718)
|
%!s(int64=2) %!d(string=hai) anos |
Jiahao Li
|
800c9635b4
Fix CUDA softmax by subtracting max value before exp (#2665)
|
%!s(int64=2) %!d(string=hai) anos |
Georgi Gerganov
|
deb7dfca4b
gguf : add ftype meta info to the model (#2710)
|
%!s(int64=2) %!d(string=hai) anos |
Kawrakow
|
bac66994cf
Quantization imrovements for k_quants (#2707)
|
%!s(int64=2) %!d(string=hai) anos |
slaren
|
519c981f8b
embedding : evaluate prompt in batches (#2713)
|
%!s(int64=2) %!d(string=hai) anos |