|
@@ -5014,8 +5014,8 @@ static struct ggml_tensor * llm_build_kqv(
|
|
|
ggml_mul_mat_set_prec(kq, GGML_PREC_F32);
|
|
ggml_mul_mat_set_prec(kq, GGML_PREC_F32);
|
|
|
}
|
|
}
|
|
|
|
|
|
|
|
-#if defined(GGML_USE_VULKAN) || defined(GGML_USE_KOMPUTE)
|
|
|
|
|
-#pragma message("TODO: ALiBi support in ggml_soft_max_ext is not implemented for Vulkan, and Kompute")
|
|
|
|
|
|
|
+#if defined(GGML_USE_KOMPUTE)
|
|
|
|
|
+#pragma message("TODO: ALiBi support in ggml_soft_max_ext is not implemented for Kompute")
|
|
|
#pragma message(" Falling back to ggml_alibi(). Will become an error in Mar 2024")
|
|
#pragma message(" Falling back to ggml_alibi(). Will become an error in Mar 2024")
|
|
|
#pragma message("ref: https://github.com/ggerganov/llama.cpp/pull/5488")
|
|
#pragma message("ref: https://github.com/ggerganov/llama.cpp/pull/5488")
|
|
|
if (hparams.f_max_alibi_bias > 0.0f) {
|
|
if (hparams.f_max_alibi_bias > 0.0f) {
|