mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2025-01-31 06:03:11 +01:00
kv_cache : fix
ggml-ci
This commit is contained in:
parent
53a7c20d89
commit
422ecaf52c
@ -35,8 +35,8 @@ bool llama_kv_cache::init(
|
|||||||
size = kv_size;
|
size = kv_size;
|
||||||
used = 0;
|
used = 0;
|
||||||
|
|
||||||
type_k = type_k;
|
this->type_k = type_k;
|
||||||
type_v = type_v;
|
this->type_v = type_v;
|
||||||
|
|
||||||
cells.clear();
|
cells.clear();
|
||||||
cells.resize(kv_size);
|
cells.resize(kv_size);
|
||||||
|
Loading…
Reference in New Issue
Block a user