Brian
|
672a6f1018
convert-*.py: GGUF Naming Convention Refactor and Metadata Override Refactor (#7499)
|
1 year ago |
Xuan Son Nguyen
|
97bdd26eee
Refactor lora adapter support (#8332)
|
1 year ago |
toyer
|
905942abdb
llama : support glm3 and glm4 (#8031)
|
1 year ago |
Icecream95
|
d7fd29fff1
llama : add OpenELM support (#7359)
|
1 year ago |
Faisal Zaghloul
|
968967376d
Add `JAIS` model(s) (#8118)
|
1 year ago |
Xuan Son Nguyen
|
49122a873f
gemma2: add sliding window mask (#8227)
|
1 year ago |
Andrei
|
1c5eba6f8e
llama: Add attention and final logit soft-capping, update scaling factor to Gemma2 (#8197)
|
1 year ago |
pculliton
|
e57dc62057
llama: Add support for Gemma2ForCausalLM (#8156)
|
1 year ago |
Christian Zhou-Zheng
|
52fc8705a0
Option to split during conversion (#6942)
|
1 year ago |
fairydreaming
|
de0d6a68ac
gguf-py, convert-hf : model conversion support for T5 and FLAN-T5 model variants (#5763)
|
1 year ago |
Eddie-Wang
|
e112b610a1
llama : add support for BitnetForCausalLM (#7931)
|
1 year ago |
Ștefan-Gabriel Muscalu
|
a94e6ff877
update: support Qwen2-57B-A14B (#7835)
|
1 year ago |
Joan Fontanals
|
f5d7b268ec
llama : add jina v2 base code (#7596)
|
1 year ago |
zhangkaihuo
|
6f28a333c1
llama : MiniCPM support tied embeddings (#7664)
|
1 year ago |
fairydreaming
|
ee3dff6b8e
Add support for DeepseekV2ForCausalLM (#7519)
|
1 year ago |
fairydreaming
|
fbca2f27fc
Add support for ArcticForCausalLM (#7020)
|
1 year ago |
Georgi Gerganov
|
e84b71c2c6
ggml : drop support for QK_K=64 (#7473)
|
1 year ago |
liuwei-git
|
201cc11afa
llama : add phi3 128K model support (#7225)
|
1 year ago |
Georgi Gerganov
|
fabf30b4c4
llama : remove Persimmon (#7408)
|
1 year ago |
compilade
|
5a419926b0
convert-hf : support bfloat16 conversion (#7158)
|
1 year ago |
Joan Fontanals
|
b83cc3f5b3
llama : add Jina Embeddings architecture (#6826)
|
1 year ago |
compilade
|
f98eb31c51
convert-hf : save memory with lazy evaluation (#7075)
|
1 year ago |
Justine Tunney
|
3855416027
ggml : introduce bfloat16 support (#6412)
|
1 year ago |
Brian
|
a2ac89d6ef
convert.py : add python logging instead of print() (#6511)
|
1 year ago |
Georgi Gerganov
|
f4ab2a4147
llama : fix BPE pre-tokenization (#6920)
|
1 year ago |
liuwei-git
|
c8297c6af5
llama : add phi3 support (#6852)
|
1 year ago |
pmysl
|
c1386c936e
gguf-py : add IQ1_M to GGML_QUANT_SIZES (#6761)
|
1 year ago |
nopperl
|
9958c81b79
Implement the OLMo architecture (#6741)
|
1 year ago |
Sigbjørn Skjæret
|
03c0946d73
convert : support models with multiple chat templates (#6588)
|
1 year ago |
Ashish
|
dbceec87c0
llama : add StableLM2 12B (#6635)
|
1 year ago |