Douglas Hanley
|
4524290e87
Use correct type of pooling for embedding models (#5500)
|
2 years ago |
Jared Van Bortel
|
ea9c8e1143
llama : add support for Nomic Embed (#5468)
|
2 years ago |
Douglas Hanley
|
03bf161eb6
llama : support batched embeddings (#5466)
|
2 years ago |
Douglas Hanley
|
2891c8aa9a
Add support for BERT embedding models (#5423)
|
2 years ago |
runfuture
|
4aa43fab56
llama : fix MiniCPM (#5392)
|
2 years ago |
runfuture
|
316c7faf77
llama : add MiniCPM support (#5346)
|
2 years ago |
Guoteng
|
7e1ae372f3
py : fix internlm2-hf convert to gguf (#5305)
|
2 years ago |
Mirror Azure
|
2d40085c26
py : add check for '.attn.masked_bias' layers to GPT2model (#5281)
|
2 years ago |
Guoteng
|
ce32060198
llama : support InternLM2 (#5184)
|
2 years ago |
sharpHL
|
f2e69d28c0
llama : add support for Orion-14B (#5118)
|
2 years ago |
compilade
|
d6bd4d46dd
llama : support StableLM 2 1.6B (#5052)
|
2 years ago |
Jared Van Bortel
|
b43ebde3b0
convert : partially revert PR #4818 (#5041)
|
2 years ago |
Shijie
|
9b75cb2b3c
llama : support upcoming Qwen2 (#5037)
|
2 years ago |
Georgi Gerganov
|
de9a147df1
py : fix flake8 lint
|
2 years ago |
chiranko
|
2b3b999cac
llama : add CodeShell support (#5016)
|
2 years ago |
Georgi Gerganov
|
5c99960901
py : remove unnecessary hasattr (#4903)
|
2 years ago |
Georgi Gerganov
|
15ebe59210
convert : update phi-2 to latest HF repo (#4903)
|
2 years ago |
Georgi Gerganov
|
2d00741e12
py : fix lint (#4889)
|
2 years ago |
Georgi Gerganov
|
f445c0e68c
llama : fix llm_build_k_shift to use correct n_rot (#4889)
|
2 years ago |
Nam D. Tran
|
26f3071d71
py : re-enable mmap in convert hf (#4732)
|
2 years ago |
crasm
|
04ac0607e9
python : add check-requirements.sh and GitHub workflow (#4585)
|
2 years ago |
manikbhandari
|
ea5497df5d
gpt2 : Add gpt2 architecture integration (#4555)
|
2 years ago |
Nam D. Tran
|
f6793491b5
llama : add AWQ for llama, llama2, mpt, and mistral models (#4593)
|
2 years ago |
Shintarou Okada
|
753be377b6
llama : add PLaMo model (#3557)
|
2 years ago |
Ebey Abraham
|
b9e74f9bca
llama : add phi-2 + fix NeoX rope + ggml_mul_mat_set_prec (#4490)
|
2 years ago |
slaren
|
799a1cb13b
llama : add Mixtral support (#4406)
|
2 years ago |
Shijie
|
37c746d687
llama : add Qwen support (#4281)
|
2 years ago |
Galunid
|
1ddb52ec38
scripts : Use mmap in torch load (#4202)
|
2 years ago |
Galunid
|
189d68446e
convert : fix tensors using grad in some models (#4173)
|
2 years ago |
Galunid
|
f23c0359a3
ci : add flake8 to github actions (python linting) (#4129)
|
2 years ago |