Tarek Dakhran
|
73d284a250
model : add LFM2-ColBert-350M (#18607)
|
3 weeks ago |
Sigbjørn Skjæret
|
eadc4184ca
llama : refactor rope_freq_base/scale_swa conversion and init (#18553)
|
3 weeks ago |
Xuan-Son Nguyen
|
4cbafad4f0
model: support MiMo-V2-Flash (#18328)
|
1 month ago |
Xuan-Son Nguyen
|
3d86c6c2b5
model: support GLM4V vision encoder (#18042)
|
1 month ago |
Johannes Gäßler
|
b1f3a6e5db
llama: automatically set parameters not set by the user in such a way that maximizes GPU utilization (#16653)
|
1 month ago |
Xuan-Son Nguyen
|
0759b09c90
graph: add f_attn_temp_offset (#18025)
|
1 month ago |
Georgi Gerganov
|
609a2d0268
models : fix YaRN regression + consolidate logic (#18006)
|
1 month ago |
Georgi Gerganov
|
7bed317f53
models : fix the attn_factor for mistral3 graphs + improve consistency (#17945)
|
1 month ago |
Xuan-Son Nguyen
|
cd3c118908
model: support Ministral3 (#17644)
|
1 month ago |
Piotr Wilkin (ilintar)
|
ff55414c42
model : Qwen3 Next (#16095)
|
2 months ago |
Sigbjørn Skjæret
|
9008027aa3
hparams : add n_embd_inp() to support extended embed (#16928)
|
2 months ago |
JJJYmmm
|
d261223d24
model: add support for qwen3vl series (#16780)
|
3 months ago |
Sigbjørn Skjæret
|
84bf3c6778
model : add BailingMoeV2 support (#16063)
|
3 months ago |
Saba Fallah
|
e08db42595
model: EmbeddingGemma Adding Support for SentenceTransformers Dense Modules (#16367)
|
3 months ago |
Piotr Wilkin (ilintar)
|
34fcc5a4ac
model : Apertus model implementation (#15852)
|
3 months ago |
Shunta Saito
|
ded67b9444
llama : parameter conversion and loading fixes for PLaMo2 variants (#16075)
|
3 months ago |
Sigbjørn Skjæret
|
835b2b915c
model : add GroveMoE support (#15510)
|
4 months ago |
Xuan-Son Nguyen
|
8f8f2274ee
convert : add Llama4ForCausalLM (#16042)
|
4 months ago |
Sigbjørn Skjæret
|
b8e09f08b9
model : add grok-2 support (#15539)
|
4 months ago |
Jie Fu (傅杰)
|
4f658855fa
llama : support T5 models with unequal number of encoder-decoder layers (#15909)
|
4 months ago |
Georgi Gerganov
|
c610b6c11b
kv-cache : fix SWA checks + disable cacheless iSWA (#15811)
|
4 months ago |
Daniel Bevenius
|
fb15d649ed
llama : add support for EmbeddingGemma 300m (#15798)
|
4 months ago |
Georgi Gerganov
|
b730706a49
kv-cache : support layer reuse (#15504)
|
5 months ago |
Georgi Gerganov
|
fd1234cb46
llama : add gpt-oss (#15091)
|
5 months ago |
Sam
|
ef0144c087
model: support GLM 4.5 family of models (#14939)
|
5 months ago |
Dongliang Wei
|
6c6e397aff
model : add support for SmallThinker series (#14898)
|
6 months ago |
Gabriel Larson
|
4762ad7316
model : make rope_yarn_log_mul optional for deepseek2 (#14896)
|
6 months ago |
Georgi Gerganov
|
225e7a1438
llama : add high-throughput mode (#14363)
|
6 months ago |
Gabriel Larson
|
4a4f426944
model : add Kimi-K2 support (#14654)
|
6 months ago |
Tarek Dakhran
|
f5e96b368f
model : support LiquidAI LFM2 hybrid family (#14620)
|
6 months ago |