Commit History

Author SHA1 Message Date
  Brian 672a6f1018 convert-*.py: GGUF Naming Convention Refactor and Metadata Override Refactor (#7499) 1 year ago
  Xuan Son Nguyen 97bdd26eee Refactor lora adapter support (#8332) 1 year ago
  toyer 905942abdb llama : support glm3 and glm4 (#8031) 1 year ago
  Icecream95 d7fd29fff1 llama : add OpenELM support (#7359) 1 year ago
  Faisal Zaghloul 968967376d Add `JAIS` model(s) (#8118) 1 year ago
  Xuan Son Nguyen 49122a873f gemma2: add sliding window mask (#8227) 1 year ago
  Andrei 1c5eba6f8e llama: Add attention and final logit soft-capping, update scaling factor to Gemma2 (#8197) 1 year ago
  pculliton e57dc62057 llama: Add support for Gemma2ForCausalLM (#8156) 1 year ago
  Christian Zhou-Zheng 52fc8705a0 Option to split during conversion (#6942) 1 year ago
  fairydreaming de0d6a68ac gguf-py, convert-hf : model conversion support for T5 and FLAN-T5 model variants (#5763) 1 year ago
  Eddie-Wang e112b610a1 llama : add support for BitnetForCausalLM (#7931) 1 year ago
  Ștefan-Gabriel Muscalu a94e6ff877 update: support Qwen2-57B-A14B (#7835) 1 year ago
  Joan Fontanals f5d7b268ec llama : add jina v2 base code (#7596) 1 year ago
  zhangkaihuo 6f28a333c1 llama : MiniCPM support tied embeddings (#7664) 1 year ago
  fairydreaming ee3dff6b8e Add support for DeepseekV2ForCausalLM (#7519) 1 year ago
  fairydreaming fbca2f27fc Add support for ArcticForCausalLM (#7020) 1 year ago
  Georgi Gerganov e84b71c2c6 ggml : drop support for QK_K=64 (#7473) 1 year ago
  liuwei-git 201cc11afa llama : add phi3 128K model support (#7225) 1 year ago
  Georgi Gerganov fabf30b4c4 llama : remove Persimmon (#7408) 1 year ago
  compilade 5a419926b0 convert-hf : support bfloat16 conversion (#7158) 1 year ago
  Joan Fontanals b83cc3f5b3 llama : add Jina Embeddings architecture (#6826) 1 year ago
  compilade f98eb31c51 convert-hf : save memory with lazy evaluation (#7075) 1 year ago
  Justine Tunney 3855416027 ggml : introduce bfloat16 support (#6412) 1 year ago
  Brian a2ac89d6ef convert.py : add python logging instead of print() (#6511) 1 year ago
  Georgi Gerganov f4ab2a4147 llama : fix BPE pre-tokenization (#6920) 1 year ago
  liuwei-git c8297c6af5 llama : add phi3 support (#6852) 1 year ago
  pmysl c1386c936e gguf-py : add IQ1_M to GGML_QUANT_SIZES (#6761) 1 year ago
  nopperl 9958c81b79 Implement the OLMo architecture (#6741) 1 year ago
  Sigbjørn Skjæret 03c0946d73 convert : support models with multiple chat templates (#6588) 1 year ago
  Ashish dbceec87c0 llama : add StableLM2 12B (#6635) 1 year ago