theo77186 622cd010ff ggml: CUDA: add head size 72 for flash-attn (#16962) 3 mesi fa
..
cmake 9a96389544 ggml: Skip backend library linking code when GGML_BACKEND_DL=ON (#15094) 6 mesi fa
include d261223d24 model: add support for qwen3vl series (#16780) 3 mesi fa
src 622cd010ff ggml: CUDA: add head size 72 for flash-attn (#16962) 3 mesi fa
.gitignore 17eb6aa8a9 vulkan : cmake integration (#8119) 1 anno fa
CMakeLists.txt 63d2fc46e1 Add experimental ggml-hexagon backend for the Hexagon NPU (#16547) 3 mesi fa