Valentin Mamedov
|
a0974156f3
llama : add Deepseek MoE v1 & GigaChat models (#10827)
|
1 year ago |
HimariO
|
ba1cb19cdd
llama : add Qwen2VL support + multimodal RoPE (#10361)
|
1 year ago |
Robert Collins
|
62e84d9848
llama : add 128k yarn context for Qwen (#10698)
|
1 year ago |
Djip007
|
19d8762ab6
ggml : refactor online repacking (#10446)
|
1 year ago |
JFLFY2255
|
8d0cfd554a
llama: Support MiniCPM-1B (with & w/o longrope) (#10559)
|
1 year ago |
Shane A
|
80acb7b430
Rename Olmo1124 to Olmo2 (#10500)
|
1 year ago |
Shane A
|
a88ad007de
llama : add OLMo November 2024 support (#10394)
|
1 year ago |
Brian
|
a0ec17b32e
metadata: Detailed Dataset Authorship Metadata (#8875)
|
1 year ago |
Georgi Gerganov
|
11ac9800af
llama : improve infill support and special token detection (#9798)
|
1 year ago |
compilade
|
1927378bcc
convert : refactor rope_freqs generation (#9396)
|
1 year ago |
Georgi Gerganov
|
f4d2b8846a
llama : add reranking support (#9510)
|
1 year ago |
nopperl
|
9a913110cf
llama : add support for Chameleon (#8543)
|
1 year ago |
Gabe Goodhart
|
3d6bf6919f
llama : add IBM Granite MoE architecture (#9438)
|
1 year ago |
Gabe Goodhart
|
0d2ec43833
llama : support IBM Granite architecture (#9412)
|
1 year ago |
Shane A
|
0aadac10c7
llama : support OLMoE (#9462)
|
1 year ago |
CarryFun
|
95ca85168b
llama : support MiniCPM3 (#9322)
|
1 year ago |
compilade
|
9bc6db28d0
ggml-quants : ternary packing for TriLMs and BitNet b1.58 (#8151)
|
1 year ago |
Molly Sophia
|
8f1d81a0b6
llama : support RWKV v6 models (#8980)
|
1 year ago |
Younes Belkada
|
b40eb84895
llama : support for `falcon-mamba` architecture (#9074)
|
1 year ago |
Minsoo Cheong
|
c679e0cb5c
llama : add EXAONE model support (#9025)
|
1 year ago |
Yoshi Suhara
|
2a24c8caa6
Add Nemotron/Minitron GGUF Conversion & Inference Support (#8922)
|
1 year ago |
fairydreaming
|
7c3f55c100
Add support for encoder-only T5 models (#8900)
|
1 year ago |
compilade
|
3a14e00366
gguf-py : simplify support for quant types (#8838)
|
1 year ago |
fairydreaming
|
d3f0c7166a
Stop the generation when <|eom_id|> token is encountered - needed for Llama 3.1 tool call support (#8858)
|
1 year ago |
Brian
|
672a6f1018
convert-*.py: GGUF Naming Convention Refactor and Metadata Override Refactor (#7499)
|
1 year ago |
Xuan Son Nguyen
|
97bdd26eee
Refactor lora adapter support (#8332)
|
1 year ago |
toyer
|
905942abdb
llama : support glm3 and glm4 (#8031)
|
1 year ago |
Icecream95
|
d7fd29fff1
llama : add OpenELM support (#7359)
|
1 year ago |
Faisal Zaghloul
|
968967376d
Add `JAIS` model(s) (#8118)
|
1 year ago |
Xuan Son Nguyen
|
49122a873f
gemma2: add sliding window mask (#8227)
|
1 year ago |