slaren
|
8a052c131e
ggml-cuda : support stablelm rope (#4156)
|
пре 2 година |
Galunid
|
189d68446e
convert : fix tensors using grad in some models (#4173)
|
пре 2 година |
eastriver
|
2568a4bf54
main.swift : fix eos checking (#4197)
|
пре 2 година |
Aaryaman Vasishta
|
b35f3d0def
readme : use PATH for Windows ROCm (#4195)
|
пре 2 година |
Haohui Mai
|
55978ce09b
Fix incorrect format strings and uninitialized variables. (#4133)
|
пре 2 година |
Georgi Gerganov
|
6b0a7420d0
llama : KV cache view API + better KV cache management (#4170)
|
пре 2 година |
Georgi Gerganov
|
d103d935c0
readme : update hot topics
|
пре 2 година |
Daniel Bevenius
|
9d5949f04b
examples : fix typo in parallel example doc comment (#4181)
|
пре 2 година |
Georgi Gerganov
|
ff8238f71d
docs : add llama-star arch idea
|
пре 2 година |
Galunid
|
8e672efe63
stablelm : simplify + speedup generation (#4153)
|
пре 2 година |
Galunid
|
0b871f1a04
finetune - update readme to mention llama support only (#4148)
|
пре 2 година |
Aaryaman Vasishta
|
dfc7cd48b1
readme : update ROCm Windows instructions (#4122)
|
пре 2 година |
Seb C
|
881800d1f0
main : Add ChatML functionality to main example (#4046)
|
пре 2 година |
Galunid
|
f23c0359a3
ci : add flake8 to github actions (python linting) (#4129)
|
пре 2 година |
Branden Butler
|
40a34fe8d0
speculative : fix prompt tokenization in speculative example (#4025)
|
пре 2 година |
Georgi Gerganov
|
dae06c06e5
Revert "finetune : add --n-gpu-layers flag info to --help (#4128)"
|
пре 2 година |
Clark Saben
|
05e8301e45
finetune : add --n-gpu-layers flag info to --help (#4128)
|
пре 2 година |
SoftwareRenderer
|
936c79b227
server : relay error messages (#4131)
|
пре 2 година |
kchro3
|
262005ad9d
common : comma should be semicolon (#4137)
|
пре 2 година |
Georgi Gerganov
|
35985acffa
gitignore : tokenize
|
пре 2 година |
slaren
|
e937066420
gguf-py : export chat templates (#4125)
|
пре 2 година |
Kerfuffle
|
28a2e6e7d4
tokenize example: Respect normal add BOS token behavior (#4126)
|
пре 2 година |
Galunid
|
0b5c3b0457
scripts : Remove missed baichuan convert script (#4127)
|
пре 2 година |
Kerfuffle
|
2923f17f6f
Clean up ggml-cuda.cu warnings when compiling with clang (for ROCM) (#4124)
|
пре 2 година |
slaren
|
bbecf3f415
llama : increase max nodes (#4115)
|
пре 2 година |
Roger Meier
|
8e9361089d
build : support ppc64le build for make and CMake (#3963)
|
пре 2 година |
Georgi Gerganov
|
5ad387e994
tokenize : fix trailing whitespace
|
пре 2 година |
zakkor
|
2fa02b4b3d
examples : add tokenize (#4039)
|
пре 2 година |
Don Mahurin
|
2ab0707acb
convert : use 'model' value if it exists. This allows karpathy/tinyllamas to load (#4089)
|
пре 2 година |
John
|
11173c92d6
py : Falcon HF compatibility (#4104)
|
пре 2 година |