Sigbjørn Skjæret
|
5936a616e4
convert : BailingMoE : fix qkv split when head_dim is 0 (#12687)
|
10 meses atrás |
Sigbjørn Skjæret
|
35782aeedb
convert : BailingMoE : avoid setting rope_dim to 0 (#12678)
|
10 meses atrás |
Sigbjørn Skjæret
|
403fbacbbc
convert : Qwerky : use lora_rank_tokenshift and lora_rank_decay if present (#12667)
|
10 meses atrás |
Sigbjørn Skjæret
|
2c3f8b850a
llama : support BailingMoE (Ling) (#12634)
|
10 meses atrás |
Juyoung Suk
|
b3de7cac73
llama : add Trillion 7B model support (#12556)
|
10 meses atrás |
Si1w
|
f125b8dccf
llama : add PLM GGUF Conversion & Inference Support (#12457)
|
10 meses atrás |
Csaba Kecskemeti
|
d5c6309d91
convert : Support Qwen2_5_VLForConditionalGeneration (#12595)
|
10 meses atrás |
Georgi Gerganov
|
df4d20cd53
convert : fix squeeze for ssm_conv tensors (#12573)
|
10 meses atrás |
Sigbjørn Skjæret
|
53af4dba42
convert: fix Mistral3/Gemma3 model hparams init (#12571)
|
10 meses atrás |
compilade
|
00d53800e0
llama-vocab : add SuperBPE pre-tokenizer (#12532)
|
10 meses atrás |
Bartowski
|
732b5fbf5e
convert : avoid calls to tokenizer.added_tokens_decoder (#12473)
|
10 meses atrás |
Sigbjørn Skjæret
|
108e53c2f1
llama : add support for GPT2, Bloom and CodeShell tied word embeddings (#12456)
|
10 meses atrás |
Xuan-Son Nguyen
|
29fff308c7
llama : support converting Mistral Small text-only (#12450)
|
10 meses atrás |
Molly Sophia
|
7dfad387e3
llama: Add support for RWKV v7 architecture (#12412)
|
10 meses atrás |
Xuan-Son Nguyen
|
7841fc723e
llama : Add Gemma 3 support (+ experimental vision capability) (#12343)
|
10 meses atrás |
Xuan-Son Nguyen
|
c43a3e7996
llama : add Phi-4-mini support (supersede #12099) (#12108)
|
11 meses atrás |
Georgi Gerganov
|
68ff663a04
repo : update links to new url (#11886)
|
11 meses atrás |
piDack
|
0cec062a63
llama : add support for GLM-Edge and GLM-Edge-V series models (#10573)
|
1 ano atrás |
Xuan Son Nguyen
|
ec7f3ac9ab
llama : add support for Deepseek-R1-Qwen distill model (#11310)
|
1 ano atrás |
RunningLeon
|
4dbc8b9cb7
llama : add internlm3 support (#11233)
|
1 ano atrás |
Daniel Bevenius
|
2739a71e4b
convert : sort print supported models [no ci] (#11179)
|
1 ano atrás |
Daniel Bevenius
|
ff3fcabc72
convert : add --print-supported-models option (#11172)
|
1 ano atrás |
Molly Sophia
|
ee7136c6d1
llama: add support for QRWKV6 model architecture (#11001)
|
1 ano atrás |
Pierrick Hymbert
|
f8feb4b01a
model: Add support for PhiMoE arch (#11003)
|
1 ano atrás |
fairydreaming
|
9394bbd484
llama : Add support for DeepSeek V3 (#11049)
|
1 ano atrás |
DAN™
|
46be942214
llama : add support for the cohere2 model architecture (#10900)
|
1 ano atrás |
ymcki
|
bc7b1f8632
convert : fix Llama-3_1-Nemotron-51B rope settings (#11008)
|
1 ano atrás |
Yun Dou
|
b92a14a841
llama : support InfiniAI Megrez 3b (#10893)
|
1 ano atrás |
ymcki
|
6f0c9e034b
llama : support for Llama-3_1-Nemotron-51B (#10669)
|
1 ano atrás |
Billel Mokeddem
|
7ae33a616f
llama : add Falcon3 support (#10883)
|
1 ano atrás |