mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2024-11-22 08:07:56 +01:00
Minor bug fix
This commit is contained in:
parent
2e86a1ec04
commit
b78561fba6
@ -97,12 +97,10 @@ def generate_chat_picture(picture, name1, name2):
|
|||||||
return text, visible_text
|
return text, visible_text
|
||||||
|
|
||||||
def stop_everything_event():
|
def stop_everything_event():
|
||||||
global stop_everything
|
shared.stop_everything = True
|
||||||
stop_everything = True
|
|
||||||
|
|
||||||
def chatbot_wrapper(text, tokens, do_sample, max_new_tokens, temperature, top_p, typical_p, repetition_penalty, top_k, min_length, no_repeat_ngram_size, num_beams, penalty_alpha, length_penalty, early_stopping, name1, name2, context, check, chat_prompt_size, picture=None):
|
def chatbot_wrapper(text, tokens, do_sample, max_new_tokens, temperature, top_p, typical_p, repetition_penalty, top_k, min_length, no_repeat_ngram_size, num_beams, penalty_alpha, length_penalty, early_stopping, name1, name2, context, check, chat_prompt_size, picture=None):
|
||||||
global stop_everything
|
shared.stop_everything = False
|
||||||
stop_everything = False
|
|
||||||
|
|
||||||
if 'pygmalion' in shared.model_name.lower():
|
if 'pygmalion' in shared.model_name.lower():
|
||||||
name1 = "You"
|
name1 = "You"
|
||||||
@ -126,7 +124,7 @@ def chatbot_wrapper(text, tokens, do_sample, max_new_tokens, temperature, top_p,
|
|||||||
|
|
||||||
# We need this global variable to handle the Stop event,
|
# We need this global variable to handle the Stop event,
|
||||||
# otherwise gradio gets confused
|
# otherwise gradio gets confused
|
||||||
if stop_everything:
|
if shared.stop_everything:
|
||||||
return shared.history['visible']
|
return shared.history['visible']
|
||||||
|
|
||||||
if first:
|
if first:
|
||||||
@ -215,7 +213,7 @@ def clear_chat_log(name1, name2):
|
|||||||
if shared.character != 'None':
|
if shared.character != 'None':
|
||||||
for i in range(len(shared.history['internal'])):
|
for i in range(len(shared.history['internal'])):
|
||||||
if '<|BEGIN-VISIBLE-CHAT|>' in shared.history['internal'][i][0]:
|
if '<|BEGIN-VISIBLE-CHAT|>' in shared.history['internal'][i][0]:
|
||||||
shared.history['visible'] = [['', shared.history['internal'][i][1]]]
|
shared.history['visible'] = [['', apply_extensions(shared.history['internal'][i][1], "output")]]
|
||||||
shared.history['internal'] = shared.history['internal'][:i+1]
|
shared.history['internal'] = shared.history['internal'][:i+1]
|
||||||
break
|
break
|
||||||
else:
|
else:
|
||||||
|
@ -5,11 +5,11 @@ tokenizer = None
|
|||||||
model_name = ""
|
model_name = ""
|
||||||
soft_prompt_tensor = None
|
soft_prompt_tensor = None
|
||||||
soft_prompt = False
|
soft_prompt = False
|
||||||
stop_everything = False
|
|
||||||
|
|
||||||
# Chat variables
|
# Chat variables
|
||||||
history = {'internal': [], 'visible': []}
|
history = {'internal': [], 'visible': []}
|
||||||
character = 'None'
|
character = 'None'
|
||||||
|
stop_everything = False
|
||||||
|
|
||||||
settings = {
|
settings = {
|
||||||
'max_new_tokens': 200,
|
'max_new_tokens': 200,
|
||||||
|
Loading…
Reference in New Issue
Block a user