Jeff Bolz 4f41ee11d6 vulkan: use scalar FA rather than coopmat2 when N==1 (#13554) 8 months ago
..
ggml-blas 5931c1f233 ggml : add support for dynamic loading of backends (#10469) 1 year ago
ggml-cann 7a395f67a7 CANN: Add support for async operator submission (#12864) 9 months ago
ggml-cpu 5ab5d5fb25 arm64: optimize q6_k_q8_k kernel with i8mm (#13519) 8 months ago
ggml-cuda 4696d56749 CUDA: fix crash on large batch size for quant. MoE (#13537) 8 months ago
ggml-hip 84778e9770 CUDA/HIP: Share the same unified memory allocation logic. (#12934) 9 months ago
ggml-kompute ba1cb19cdd llama : add Qwen2VL support + multimodal RoPE (#10361) 1 year ago
ggml-metal 654a67794f metal : add FA-vec kernel for head size 64 (#13583) 8 months ago
ggml-musa b1b132efcb cuda : enable CUDA Graph on CUDA Toolkit < 12.x (#12394) 10 months ago
ggml-opencl f0d46ef157 opencl: remove unnecessary assert for `add` (#13257) 8 months ago
ggml-rpc b486ba05bf rpc : add rpc_msg_set_tensor_hash_req (#13353) 8 months ago
ggml-sycl 0a338ed013 sycl : fixed compilation warnings (#13582) 8 months ago
ggml-vulkan 4f41ee11d6 vulkan: use scalar FA rather than coopmat2 when N==1 (#13554) 8 months ago
CMakeLists.txt bba9d945c1 cmake : removed stdc++fs (whisper/3097) 8 months ago
ggml-alloc.c f057808ffa ggml: Don't assert fail when tensor data changes (#13222) 8 months ago
ggml-backend-impl.h 70680c48e5 ggml : upgrade init_tensor API to return a ggml_status (#11854) 10 months ago
ggml-backend-reg.cpp ba7654380a ggml-backend : fix backend search path (#12330) 10 months ago
ggml-backend.cpp 10d2af0eaa llama/ggml: add LLM training support (#10544) 8 months ago
ggml-common.h 492d7f1ff7 musa: fix all warnings, re-enable `-DLLAMA_FATAL_WARNINGS=ON` in ci and update doc (#12611) 9 months ago
ggml-impl.h cb79c2e7fa ggml: don't include arm_neon.h when using CUDA 12 with ARM Neon (ggml/1187) 9 months ago
ggml-opt.cpp 10d2af0eaa llama/ggml: add LLM training support (#10544) 8 months ago
ggml-quants.c 13b0a04597 whisper: remove MSVC warnings pragmas (whisper/3090) 8 months ago
ggml-quants.h ae8de6d50a ggml : build backends as libraries (#10256) 1 year ago
ggml-threading.cpp ae8de6d50a ggml : build backends as libraries (#10256) 1 year ago
ggml-threading.h cb13ef85a4 remove CMAKE_WINDOWS_EXPORT_ALL_SYMBOLS (#10797) 1 year ago
ggml.c 10d2af0eaa llama/ggml: add LLM training support (#10544) 8 months ago
gguf.cpp c6a2c9e741 gguf : use ggml log system (#13571) 8 months ago