Commit History

Author SHA1 Message Date
  Georgi Gerganov 19514d632e cmake : do not hide GGML options + rename option (#9465) 1 year ago
  Michael Podvitskiy 2a825116b6 cmake : fix for builds without `GGML_CDEF_PUBLIC` (#9338) 1 year ago
  Borislav Stanimirov 44d28ddd5c cmake : fix use of external ggml (#8787) 1 year ago
  hipudding 1bdd8ae19f [CANN] Add Ascend NPU backend (#6035) 1 year ago
  bandoti 17eb6aa8a9 vulkan : cmake integration (#8119) 1 year ago
  Johannes Gäßler a03e8dd99d make/cmake: LLAMA_NO_CCACHE -> GGML_NO_CCACHE (#8392) 1 year ago
  Borislav Stanimirov 9925ca4087 cmake : allow external ggml (#8370) 1 year ago
  ditsuke 51d2ebadbb build: Export hf-to-gguf as snakecase 1 year ago
  ditsuke 07786a61a2 chore: Fixup requirements and build 1 year ago
  Daniel Bevenius f8c4c0738d tests : add _CRT_SECURE_NO_WARNINGS for WIN32 (#8231) 1 year ago
  slaren b851b3fba0 cmake : allow user to override default options (#8178) 1 year ago
  slaren 8172ee9da9 cmake : fix deprecated option names not working (#8171) 1 year ago
  slaren 31ec3993f6 ggml : add GGML_CUDA_USE_GRAPHS option, restore GGML_CUDA_FORCE_CUBLAS (cmake) (#8140) 1 year ago
  Georgi Gerganov f3f65429c4 llama : reorganize source code + improve CMake (#8006) 1 year ago
  Johannes Gäßler a818f3028d CUDA: use MMQ instead of cuBLAS by default (#8075) 1 year ago
  slaren 95f57bb5d5 ggml : remove ggml_task_type and GGML_PERF (#8017) 1 year ago
  luoyu-intel de391e4c80 [SYCL] Fix windows build and inference (#8003) 1 year ago
  0cc4m 7c7836d9d4 Vulkan Shader Refactor, Memory Debugging Option (#7947) 1 year ago
  Meng, Hengyu 7b2f4a7d19 [SYCL] remove global variables (#7710) 1 year ago
  slaren f578b86b21 move BLAS to a separate backend (#6210) 1 year ago
  Jared Van Bortel 864a99e7a0 cmake : fix CMake requirement for CUDA (#7821) 1 year ago
  Johannes Gäßler 7d1a378b8f CUDA: refactor mmq, dmmv, mmvq (#7716) 1 year ago
  Georgi Gerganov 554c247caf ggml : remove OpenCL (#7735) 1 year ago
  Daniele 987d743d6b Improve hipBLAS support in CMake (#7696) 1 year ago
  Masaya, Kato a5735e4426 ggml : use OpenMP as a thread pool (#7606) 1 year ago
  Andy Tai a10cda58d3 cmake : add pkg-config spec file for llama.cpp (#7702) 1 year ago
  woachk 9e405b6e2e kompute : implement op_getrows_f32 (#6403) 1 year ago
  Johannes Gäßler 9b596417af CUDA: quantized KV support for FA vec (#7527) 1 year ago
  Galunid 9c4c9cc83f Move convert.py to examples/convert-legacy-llama.py (#7430) 1 year ago
  Meng, Hengyu b864b50ce5 [SYCL] Align GEMM dispatch (#7566) 1 year ago