Willy Tarreau 35a84916fb main: add the possibility to open the prompt cache read-only (#1640) 2 lat temu
..
baby-llama f954edda93 ggml : implement backward pass for llama + small training-llama-from-scratch example (#1360) 2 lat temu
benchmark ec2e10c444 llama : add llama_init_backend() API (close #1527) 2 lat temu
embedding ec2e10c444 llama : add llama_init_backend() API (close #1527) 2 lat temu
jeopardy 5fba3c016b examples : add Jeopardy example (#1168) 2 lat temu
main 35a84916fb main: add the possibility to open the prompt cache read-only (#1640) 2 lat temu
metal ecb217db4f llama : Metal inference (#1642) 2 lat temu
perplexity ec2e10c444 llama : add llama_init_backend() API (close #1527) 2 lat temu
quantize 99009e72f8 ggml : add SOTA 2,3,4,5,6 bit k-quantizations (#1684) 2 lat temu
quantize-stats 99009e72f8 ggml : add SOTA 2,3,4,5,6 bit k-quantizations (#1684) 2 lat temu
save-load-state dc271c52ed Remove unused n_parts parameter (#1509) 2 lat temu
server 17366df842 Multi GPU support, CUDA refactor, CUDA scratch buffer (#1703) 2 lat temu
CMakeLists.txt ecb217db4f llama : Metal inference (#1642) 2 lat temu
Miku.sh a8a2efdc81 examples : various prompt and example fixes (#1298) 2 lat temu
alpaca.sh e9a9cb0c54 examples : Improve Alpaca Default Repeat Penalty: Better Match Alpaca.cpp Experience (#1107) 2 lat temu
chat-13B.bat d9ad104440 Create chat-13B.bat (#592) 2 lat temu
chat-13B.sh 6daa09d879 examples : read chat prompts from a template file (#1196) 2 lat temu
chat-persistent.sh 1359b6aba5 chat-persistent.sh : use bracket expressions in grep (#1564) 2 lat temu
chat.sh 79b2b266db If n_predict == -1, generate forever 2 lat temu
common.cpp 35a84916fb main: add the possibility to open the prompt cache read-only (#1640) 2 lat temu
common.h 35a84916fb main: add the possibility to open the prompt cache read-only (#1640) 2 lat temu
gpt4all.sh 107980d970 examples : add -n to alpaca and gpt4all scripts (#706) 2 lat temu
reason-act.sh a6956b25a1 add example of re-act pattern (#583) 2 lat temu