From 671f639c593becaaa9ad7b61028b79012047adfc Mon Sep 17 00:00:00 2001 From: Georgi Gerganov Date: Wed, 22 Nov 2023 19:30:48 +0200 Subject: [PATCH] llama : zero KV cache used upon clear ggml-ci --- llama.cpp | 1 + 1 file changed, 1 insertion(+) diff --git a/llama.cpp b/llama.cpp index 31b5ce9a4..5679c7050 100644 --- a/llama.cpp +++ b/llama.cpp @@ -1629,6 +1629,7 @@ static void llama_kv_cache_clear(struct llama_kv_cache & cache) { cache.cells[i].seq_id.clear(); } cache.head = 0; + cache.used = 0; } static void llama_kv_cache_seq_rm(