Joan Fontanals
|
f5d7b268ec
llama : add jina v2 base code (#7596)
|
1 year ago |
zhangkaihuo
|
6f28a333c1
llama : MiniCPM support tied embeddings (#7664)
|
1 year ago |
fairydreaming
|
ee3dff6b8e
Add support for DeepseekV2ForCausalLM (#7519)
|
1 year ago |
fairydreaming
|
fbca2f27fc
Add support for ArcticForCausalLM (#7020)
|
1 year ago |
Georgi Gerganov
|
e84b71c2c6
ggml : drop support for QK_K=64 (#7473)
|
1 year ago |
liuwei-git
|
201cc11afa
llama : add phi3 128K model support (#7225)
|
1 year ago |
Georgi Gerganov
|
fabf30b4c4
llama : remove Persimmon (#7408)
|
1 year ago |
compilade
|
5a419926b0
convert-hf : support bfloat16 conversion (#7158)
|
1 year ago |
Joan Fontanals
|
b83cc3f5b3
llama : add Jina Embeddings architecture (#6826)
|
1 year ago |
compilade
|
f98eb31c51
convert-hf : save memory with lazy evaluation (#7075)
|
1 year ago |
Justine Tunney
|
3855416027
ggml : introduce bfloat16 support (#6412)
|
1 year ago |
Brian
|
a2ac89d6ef
convert.py : add python logging instead of print() (#6511)
|
1 year ago |
Georgi Gerganov
|
f4ab2a4147
llama : fix BPE pre-tokenization (#6920)
|
1 year ago |
liuwei-git
|
c8297c6af5
llama : add phi3 support (#6852)
|
1 year ago |
pmysl
|
c1386c936e
gguf-py : add IQ1_M to GGML_QUANT_SIZES (#6761)
|
1 year ago |
nopperl
|
9958c81b79
Implement the OLMo architecture (#6741)
|
1 year ago |
Sigbjørn Skjæret
|
03c0946d73
convert : support models with multiple chat templates (#6588)
|
1 year ago |
Ashish
|
dbceec87c0
llama : add StableLM2 12B (#6635)
|
1 year ago |
Shijie
|
f4dea7da18
llama : add qwen2moe (#6074)
|
1 year ago |
Daniel Bevenius
|
4fbd8098e6
gguf : add special tokens metadata for FIM/Infill (#6689)
|
1 year ago |
Pierrick Hymbert
|
4bd0f93e4a
model: support arch `DbrxForCausalLM` (#6515)
|
1 year ago |
Carolinabanana
|
5dc9dd7152
llama : add Command R Plus support (#6491)
|
1 year ago |
Brian
|
a8bd14d557
gguf.py : add licence and version to gguf writer (#6504)
|
1 year ago |
bryanSwk
|
bb43cf7e9d
llama : add SEA-LION support (#6448)
|
1 year ago |
slaren
|
08a0c02060
ggml : mul_mat_id use the same tensor for all the experts (#6387)
|
1 year ago |
hxer7963
|
069574775c
[Model] Add support for xverse (#6301)
|
1 year ago |
Kawrakow
|
55c1b2a3bb
IQ1_M: 1.75 bpw quantization (#6302)
|
1 year ago |
Julius Arkenberg
|
476b0251b2
llama : add grok-1 support (#6204)
|
1 year ago |
Andrew Canis
|
12247f4c69
llama : add Command-R support (#6033)
|
1 year ago |
Ondřej Čertík
|
7ce2c77f88
gguf : add support for I64 and F64 arrays (#6062)
|
1 year ago |