Historique des commits

Auteur SHA1 Message Date
  Georgi Gerganov 8f8c28e89c convert : auto-determine model name based on dir + scripts update il y a 2 ans
  Kerfuffle 7694adda8d Fix for main example getting stuck when -n -2 and --interactive (#2767) il y a 2 ans
  slaren fea95c682d fix convert.py for codellama, add llama 34B to the list of recognized models (#2768) il y a 2 ans
  DannyDaemonic ef955fbd23 Tag release with build number (#2732) il y a 2 ans
  Georgi Gerganov d67777c202 metal : add Q8_0 support (#2763) il y a 2 ans
  Georgi Gerganov c3e53b421a llama : escape all U+2581 in a string (#2750) il y a 2 ans
  Evan Jones 6e91a1b070 llama : fix grammar sometimes generating null char (#2756) il y a 2 ans
  Georgi Gerganov 44d5462b5c readme : fix link il y a 2 ans
  Georgi Gerganov c7868b0753 minor : fix trailing whitespace il y a 2 ans
  Georgi Gerganov 79da24b58c readme : update hot topics il y a 2 ans
  Georgi Gerganov cf658adc83 llm : add Falcon support (#2717) il y a 2 ans
  Georgi Gerganov a192860cfe minor : fix trailing whitespace il y a 2 ans
  Olivier Chafik 95385241a9 examples : restore the functionality to import llama2.c models (#2685) il y a 2 ans
  slaren 335acd2ffd fix convert-lora-to-ggml.py (#2738) il y a 2 ans
  klosax 5290c38e6e main : insert bos if no tokens (#2727) il y a 2 ans
  akawrykow cc34dbda96 gitignore : fix for windows (#2729) il y a 2 ans
  Cebtenzzre 7c2227a197 chmod : make scripts executable (#2675) il y a 2 ans
  JohnnyB f19dca04ea devops : RPM Specs (#2723) il y a 2 ans
  Kawrakow 8207214b6a Fix values shown in the quantize tool help (#2735) il y a 2 ans
  Kawrakow 62959e740e Strided perplexity (#2714) il y a 2 ans
  IgnacioFDM 7f7ddd5002 Fix ggml to gguf conversion on Windows (#2733) il y a 2 ans
  Xiao-Yong Jin b8ad1b66b2 server : allow json array in prompt or content for direct token input (#2306) il y a 2 ans
  Evan Jones f5fe98d11b docs : add grammar docs (#2701) il y a 2 ans
  Kerfuffle 777f42ba18 Improve handling of special tokens in GGML to GGUF converter (#2725) il y a 2 ans
  goerch 46ef5b5fcf llama : fix whitespace escaping in tokenizer (#2724) il y a 2 ans
  Johannes Gäßler c63bb1d16a CUDA: use mul_mat_q kernels by default (#2683) il y a 2 ans
  Alex Petenchea 3b6cfe7c92 convert.py : clarifying error message (#2718) il y a 2 ans
  Jiahao Li 800c9635b4 Fix CUDA softmax by subtracting max value before exp (#2665) il y a 2 ans
  Georgi Gerganov deb7dfca4b gguf : add ftype meta info to the model (#2710) il y a 2 ans
  Kawrakow bac66994cf Quantization imrovements for k_quants (#2707) il y a 2 ans