| .. |
|
ggml-blas
|
5931c1f233
ggml : add support for dynamic loading of backends (#10469)
|
1 year ago |
|
ggml-cann
|
7a395f67a7
CANN: Add support for async operator submission (#12864)
|
9 months ago |
|
ggml-cpu
|
5ab5d5fb25
arm64: optimize q6_k_q8_k kernel with i8mm (#13519)
|
8 months ago |
|
ggml-cuda
|
4696d56749
CUDA: fix crash on large batch size for quant. MoE (#13537)
|
8 months ago |
|
ggml-hip
|
84778e9770
CUDA/HIP: Share the same unified memory allocation logic. (#12934)
|
9 months ago |
|
ggml-kompute
|
ba1cb19cdd
llama : add Qwen2VL support + multimodal RoPE (#10361)
|
1 year ago |
|
ggml-metal
|
654a67794f
metal : add FA-vec kernel for head size 64 (#13583)
|
8 months ago |
|
ggml-musa
|
b1b132efcb
cuda : enable CUDA Graph on CUDA Toolkit < 12.x (#12394)
|
10 months ago |
|
ggml-opencl
|
f0d46ef157
opencl: remove unnecessary assert for `add` (#13257)
|
8 months ago |
|
ggml-rpc
|
b486ba05bf
rpc : add rpc_msg_set_tensor_hash_req (#13353)
|
8 months ago |
|
ggml-sycl
|
0a338ed013
sycl : fixed compilation warnings (#13582)
|
8 months ago |
|
ggml-vulkan
|
4f41ee11d6
vulkan: use scalar FA rather than coopmat2 when N==1 (#13554)
|
8 months ago |
|
CMakeLists.txt
|
bba9d945c1
cmake : removed stdc++fs (whisper/3097)
|
8 months ago |
|
ggml-alloc.c
|
f057808ffa
ggml: Don't assert fail when tensor data changes (#13222)
|
8 months ago |
|
ggml-backend-impl.h
|
70680c48e5
ggml : upgrade init_tensor API to return a ggml_status (#11854)
|
10 months ago |
|
ggml-backend-reg.cpp
|
ba7654380a
ggml-backend : fix backend search path (#12330)
|
10 months ago |
|
ggml-backend.cpp
|
10d2af0eaa
llama/ggml: add LLM training support (#10544)
|
8 months ago |
|
ggml-common.h
|
492d7f1ff7
musa: fix all warnings, re-enable `-DLLAMA_FATAL_WARNINGS=ON` in ci and update doc (#12611)
|
9 months ago |
|
ggml-impl.h
|
cb79c2e7fa
ggml: don't include arm_neon.h when using CUDA 12 with ARM Neon (ggml/1187)
|
9 months ago |
|
ggml-opt.cpp
|
10d2af0eaa
llama/ggml: add LLM training support (#10544)
|
8 months ago |
|
ggml-quants.c
|
13b0a04597
whisper: remove MSVC warnings pragmas (whisper/3090)
|
8 months ago |
|
ggml-quants.h
|
ae8de6d50a
ggml : build backends as libraries (#10256)
|
1 year ago |
|
ggml-threading.cpp
|
ae8de6d50a
ggml : build backends as libraries (#10256)
|
1 year ago |
|
ggml-threading.h
|
cb13ef85a4
remove CMAKE_WINDOWS_EXPORT_ALL_SYMBOLS (#10797)
|
1 year ago |
|
ggml.c
|
10d2af0eaa
llama/ggml: add LLM training support (#10544)
|
8 months ago |
|
gguf.cpp
|
c6a2c9e741
gguf : use ggml log system (#13571)
|
8 months ago |