Skip to content

Commit 580d68b

Browse files
committed
llama : enable LLAMA_SET_ROWS=1 by default
ggml-ci
1 parent 66625a5 commit 580d68b

File tree

4 files changed

+4
-4
lines changed

4 files changed

+4
-4
lines changed

src/llama-context.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -105,7 +105,7 @@ llama_context::llama_context(
105105

106106
{
107107
const char * LLAMA_SET_ROWS = getenv("LLAMA_SET_ROWS");
108-
supports_set_rows = LLAMA_SET_ROWS ? (atoi(LLAMA_SET_ROWS) != 0) : false;
108+
supports_set_rows = LLAMA_SET_ROWS ? (atoi(LLAMA_SET_ROWS) != 0) : supports_set_rows;
109109

110110
if (!supports_set_rows && !cparams.kv_unified) {
111111
LLAMA_LOG_WARN("%s: non-unified KV cache requires ggml_set_rows() - forcing unified KV cache\n", __func__);

src/llama-context.h

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -289,7 +289,7 @@ struct llama_context {
289289

290290
// env: LLAMA_SET_ROWS (temporary)
291291
// ref: https://github.com/ggml-org/llama.cpp/pull/14285
292-
bool supports_set_rows = false;
292+
bool supports_set_rows = true;
293293

294294
// perf
295295
mutable int64_t t_start_us = 0;

src/llama-kv-cache-unified.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -193,7 +193,7 @@ llama_kv_cache_unified::llama_kv_cache_unified(
193193
debug = LLAMA_KV_CACHE_DEBUG ? atoi(LLAMA_KV_CACHE_DEBUG) : 0;
194194

195195
const char * LLAMA_SET_ROWS = getenv("LLAMA_SET_ROWS");
196-
supports_set_rows = LLAMA_SET_ROWS ? atoi(LLAMA_SET_ROWS) != 0 : 0;
196+
supports_set_rows = LLAMA_SET_ROWS ? atoi(LLAMA_SET_ROWS) != 0 : supports_set_rows;
197197

198198
if (!supports_set_rows) {
199199
// ref: https://github.com/ggml-org/llama.cpp/pull/14363

src/llama-kv-cache-unified.h

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -230,7 +230,7 @@ class llama_kv_cache_unified : public llama_memory_i {
230230

231231
// env: LLAMA_SET_ROWS (temporary)
232232
// ref: https://github.com/ggml-org/llama.cpp/pull/14285
233-
bool supports_set_rows = false;
233+
bool supports_set_rows = true;
234234

235235
const llama_swa_type swa_type = LLAMA_SWA_TYPE_NONE;
236236

0 commit comments

Comments
 (0)