mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2024-11-25 01:09:22 +01:00
Fix llama.cpp loader not being random (thanks @reydeljuego12345)
This commit is contained in:
parent
49dfa0adaf
commit
c9a9f63d1b
@ -136,7 +136,7 @@ class LlamaCppModel:
|
|||||||
prompt=prompt,
|
prompt=prompt,
|
||||||
max_tokens=state['max_new_tokens'],
|
max_tokens=state['max_new_tokens'],
|
||||||
temperature=state['temperature'],
|
temperature=state['temperature'],
|
||||||
top_p=state['top_p'],
|
top_p=state['top_p'] if state['top_p'] < 1 else 0.999,
|
||||||
min_p=state['min_p'],
|
min_p=state['min_p'],
|
||||||
typical_p=state['typical_p'],
|
typical_p=state['typical_p'],
|
||||||
frequency_penalty=state['frequency_penalty'],
|
frequency_penalty=state['frequency_penalty'],
|
||||||
|
Loading…
Reference in New Issue
Block a user