Historique des commits

Auteur SHA1 Message Date
  Galunid 9c4c9cc83f Move convert.py to examples/convert-legacy-llama.py (#7430) il y a 1 an
  Meng, Hengyu b864b50ce5 [SYCL] Align GEMM dispatch (#7566) il y a 1 an
  Masaya, Kato faa0e6979a ggml: aarch64: SVE kernels for q8_0_q8_0, q4_0_q8_0 vector dot (#7433) il y a 1 an
  Georgi Gerganov e84b71c2c6 ggml : drop support for QK_K=64 (#7473) il y a 1 an
  k.h.lai fcda1128bc vulkan: add workaround for iterator boundary check to fix clang-cl debug build (#7426) il y a 1 an
  junchao-loongson 65c58207ec ggml : add loongarch lsx and lasx support (#6454) il y a 1 an
  Srihari-mcw 33c8d50acc Add provisions for windows support for BF16 code including CMake provision for enabling AVX512_BF16 (#7258) il y a 1 an
  slaren d359f30921 llama : remove MPI backend (#7395) il y a 1 an
  Georgi Gerganov 059031b8c4 ci : re-enable sanitizer runs (#7358) il y a 1 an
  Engininja2 ef277de2ad cmake : fix typo in AMDGPU_TARGETS (#7356) il y a 1 an
  Gavin Zhao 82ca83db3c ROCm: use native CMake HIP support (#5966) il y a 1 an
  Max Krasnyansky 13ad16af12 Add support for properly optimized Windows ARM64 builds with LLVM and MSVC (#7191) il y a 1 an
  Radoslav Gerganov 5e31828d3e ggml : add RPC backend (#6829) il y a 1 an
  Georgi Gerganov 6f1b63606f cmake : fix version cmp (#7227) il y a 1 an
  slaren b228aba91a remove convert-lora-to-ggml.py (#7204) il y a 1 an
  Jared Van Bortel 4426e2987b cmake : fix typo (#7151) il y a 1 an
  agray3 bc4bba364f Introduction of CUDA Graphs to LLama.cpp (#6766) il y a 1 an
  William Tambellini 858f6b73f6 Add an option to build without CUDA VMM (#7067) il y a 1 an
  Georgi Gerganov dba497e0c1 cmake : restore LLAMA_LLAMAFILE_DEFAULT il y a 1 an
  Georgi Gerganov fa0b4ad252 cmake : remove obsolete ANDROID check il y a 1 an
  Justine Tunney 192090bae4 llamafile : improve sgemm.cpp (#6796) il y a 1 an
  Georgi Gerganov 3b8f1ec4b1 llamafile : tmp disable + build sgemm.o when needed (#6716) il y a 1 an
  Georgi Gerganov 666867b799 ggml : fix llamafile sgemm wdata offsets (#6710) il y a 1 an
  Justine Tunney 8cc91dc63c ggml : add llamafile sgemm (#6414) il y a 1 an
  Matt Clayton 8093987090 cmake : add explicit metal version options (#6370) il y a 1 an
  Jared Van Bortel 32c8486e1f wpm : portable unicode tolower (#6305) il y a 1 an
  Joseph Stahl e190f1fca6 nix: make `xcrun` visible in Nix sandbox for precompiling Metal shaders (#6118) il y a 1 an
  slaren 280345968d cuda : rename build flag to LLAMA_CUDA (#6299) il y a 1 an
  slaren ae1f211ce2 cuda : refactor into multiple files (#6269) il y a 1 an
  slaren 2f0e81e053 cuda : add LLAMA_CUDA_NO_PEER_COPY to workaround broken ROCm p2p copy (#6208) il y a 1 an