|
|
@@ -193,7 +193,7 @@ llama_kv_cache_unified::llama_kv_cache_unified(
|
|
|
debug = LLAMA_KV_CACHE_DEBUG ? atoi(LLAMA_KV_CACHE_DEBUG) : 0;
|
|
|
|
|
|
const char * LLAMA_SET_ROWS = getenv("LLAMA_SET_ROWS");
|
|
|
- supports_set_rows = LLAMA_SET_ROWS ? atoi(LLAMA_SET_ROWS) != 0 : 0;
|
|
|
+ supports_set_rows = LLAMA_SET_ROWS ? atoi(LLAMA_SET_ROWS) != 0 : supports_set_rows;
|
|
|
|
|
|
if (!supports_set_rows) {
|
|
|
// ref: https://github.com/ggml-org/llama.cpp/pull/14363
|