Commit History

Author SHA1 Message Date
  Douglas Hanley 4524290e87 Use correct type of pooling for embedding models (#5500) 2 years ago
  Jared Van Bortel ea9c8e1143 llama : add support for Nomic Embed (#5468) 2 years ago
  Douglas Hanley 03bf161eb6 llama : support batched embeddings (#5466) 2 years ago
  Douglas Hanley 2891c8aa9a Add support for BERT embedding models (#5423) 2 years ago
  runfuture 4aa43fab56 llama : fix MiniCPM (#5392) 2 years ago
  runfuture 316c7faf77 llama : add MiniCPM support (#5346) 2 years ago
  Guoteng 7e1ae372f3 py : fix internlm2-hf convert to gguf (#5305) 2 years ago
  Mirror Azure 2d40085c26 py : add check for '.attn.masked_bias' layers to GPT2model (#5281) 2 years ago
  Guoteng ce32060198 llama : support InternLM2 (#5184) 2 years ago
  sharpHL f2e69d28c0 llama : add support for Orion-14B (#5118) 2 years ago
  compilade d6bd4d46dd llama : support StableLM 2 1.6B (#5052) 2 years ago
  Jared Van Bortel b43ebde3b0 convert : partially revert PR #4818 (#5041) 2 years ago
  Shijie 9b75cb2b3c llama : support upcoming Qwen2 (#5037) 2 years ago
  Georgi Gerganov de9a147df1 py : fix flake8 lint 2 years ago
  chiranko 2b3b999cac llama : add CodeShell support (#5016) 2 years ago
  Georgi Gerganov 5c99960901 py : remove unnecessary hasattr (#4903) 2 years ago
  Georgi Gerganov 15ebe59210 convert : update phi-2 to latest HF repo (#4903) 2 years ago
  Georgi Gerganov 2d00741e12 py : fix lint (#4889) 2 years ago
  Georgi Gerganov f445c0e68c llama : fix llm_build_k_shift to use correct n_rot (#4889) 2 years ago
  Nam D. Tran 26f3071d71 py : re-enable mmap in convert hf (#4732) 2 years ago
  crasm 04ac0607e9 python : add check-requirements.sh and GitHub workflow (#4585) 2 years ago
  manikbhandari ea5497df5d gpt2 : Add gpt2 architecture integration (#4555) 2 years ago
  Nam D. Tran f6793491b5 llama : add AWQ for llama, llama2, mpt, and mistral models (#4593) 2 years ago
  Shintarou Okada 753be377b6 llama : add PLaMo model (#3557) 2 years ago
  Ebey Abraham b9e74f9bca llama : add phi-2 + fix NeoX rope + ggml_mul_mat_set_prec (#4490) 2 years ago
  slaren 799a1cb13b llama : add Mixtral support (#4406) 2 years ago
  Shijie 37c746d687 llama : add Qwen support (#4281) 2 years ago
  Galunid 1ddb52ec38 scripts : Use mmap in torch load (#4202) 2 years ago
  Galunid 189d68446e convert : fix tensors using grad in some models (#4173) 2 years ago
  Galunid f23c0359a3 ci : add flake8 to github actions (python linting) (#4129) 2 years ago