mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2025-01-27 04:23:06 +01:00
llama-run : fix context size (#11094)
Set `n_ctx` equal to `n_batch` in `Opt` class. Now context size is a more reasonable 2048. Signed-off-by: Eric Curtin <ecurtin@redhat.com>
This commit is contained in:
parent
ecebbd292d
commit
dc7cef9f37
@ -83,6 +83,7 @@ class Opt {
|
|||||||
}
|
}
|
||||||
|
|
||||||
ctx_params.n_batch = context_size >= 0 ? context_size : context_size_default;
|
ctx_params.n_batch = context_size >= 0 ? context_size : context_size_default;
|
||||||
|
ctx_params.n_ctx = ctx_params.n_batch;
|
||||||
model_params.n_gpu_layers = ngl >= 0 ? ngl : ngl_default;
|
model_params.n_gpu_layers = ngl >= 0 ? ngl : ngl_default;
|
||||||
temperature = temperature >= 0 ? temperature : temperature_default;
|
temperature = temperature >= 0 ? temperature : temperature_default;
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user