Xiao-Yong Jin
|
6e7cca4047
llama : add custom RoPE (#2054)
|
%!s(int64=2) %!d(string=hai) anos |
Bach Le
|
c9c74b4e3f
llama : add classifier-free guidance (#2135)
|
%!s(int64=2) %!d(string=hai) anos |
Evan Miller
|
5656d10599
mpi : add support for distributed inference via MPI (#2099)
|
%!s(int64=2) %!d(string=hai) anos |
Judd
|
36680f6e40
convert : update for baichuan (#2081)
|
%!s(int64=2) %!d(string=hai) anos |
Howard Su
|
b8c8dda75f
Use unsigned for random seed (#2006)
|
%!s(int64=2) %!d(string=hai) anos |
zrm
|
b853d45601
ggml : add NUMA support (#1556)
|
%!s(int64=2) %!d(string=hai) anos |
Didzis Gosko
|
527b6fba1d
llama : make model stateless and context stateful (llama_state) (#1797)
|
%!s(int64=2) %!d(string=hai) anos |
Georgi Gerganov
|
4f9c43e3bd
minor : warning fixes
|
%!s(int64=2) %!d(string=hai) anos |
FrankHB
|
5b9ccaf104
Fixed possible macro redefinition (#1892)
|
%!s(int64=2) %!d(string=hai) anos |
Borislav Stanimirov
|
9cbf50c041
build : fix and ignore MSVC warnings (#1889)
|
%!s(int64=2) %!d(string=hai) anos |
Georgi Gerganov
|
2347e45e7b
llama : do a warm-up eval at start for better timings (#1824)
|
%!s(int64=2) %!d(string=hai) anos |
Kerfuffle
|
fa84c4b3e8
Fix issue where interactive mode crashes when input exceeds ctx size (#1789)
|
%!s(int64=2) %!d(string=hai) anos |
Willy Tarreau
|
35a84916fb
main: add the possibility to open the prompt cache read-only (#1640)
|
%!s(int64=2) %!d(string=hai) anos |
Georgi Gerganov
|
ecb217db4f
llama : Metal inference (#1642)
|
%!s(int64=2) %!d(string=hai) anos |
Evan Jones
|
136476e898
Fix prompt cache saving and chat-persistent rollover (#1678)
|
%!s(int64=2) %!d(string=hai) anos |
DannyDaemonic
|
248367605e
Work around for recalculating logits in cached prompts (Fixes #1585) (#1609)
|
%!s(int64=2) %!d(string=hai) anos |
Kerfuffle
|
66874d4fbc
Some improvements to loading the session with --prompt-cache (#1550)
|
%!s(int64=2) %!d(string=hai) anos |
Georgi Gerganov
|
ec2e10c444
llama : add llama_init_backend() API (close #1527)
|
%!s(int64=2) %!d(string=hai) anos |
Jason McCartney
|
7694b52b9a
main : make reverse prompt option act as a stop token in non-interactive mode (#1032)
|
%!s(int64=2) %!d(string=hai) anos |
DannyDaemonic
|
ee9654138a
Fixes #1511 lambda issue for w64devkit (mingw) (#1513)
|
%!s(int64=2) %!d(string=hai) anos |
András Salamon
|
9560655409
define default model path once, sync path with readme (#1366)
|
%!s(int64=2) %!d(string=hai) anos |
Georgi Gerganov
|
fb62f92433
llama : fix --mtest option (close #1414)
|
%!s(int64=2) %!d(string=hai) anos |
Evan Jones
|
cf348a60e0
main : add option to save full output to session (#1338)
|
%!s(int64=2) %!d(string=hai) anos |
DannyDaemonic
|
41654efea8
Interface improvements and `--multiline-input` (previously `--author-mode`) (#1040)
|
%!s(int64=2) %!d(string=hai) anos |
Georgi Gerganov
|
f9a6364912
llama : require first token to be BOS (#1303)
|
%!s(int64=2) %!d(string=hai) anos |
Jed Fox
|
3924088512
Remove default arguments from sampling functions (#1343)
|
%!s(int64=2) %!d(string=hai) anos |
44670
|
2edbdb0f99
main : add --in-suffix option (#1318)
|
%!s(int64=2) %!d(string=hai) anos |
Tomas
|
f647ce040f
fix #1224 reverse prompt and multi line (#1297)
|
%!s(int64=2) %!d(string=hai) anos |
DannyDaemonic
|
13b0c68ed7
Handle signals properly on Windows (#1123)
|
%!s(int64=2) %!d(string=hai) anos |
Ron Evans
|
67c77799e0
examples : add llama_init_from_gpt_params() common function (#1290)
|
%!s(int64=2) %!d(string=hai) anos |