diff --git a/modules/chat.py b/modules/chat.py index 787fb46f..ad373c28 100644 --- a/modules/chat.py +++ b/modules/chat.py @@ -94,13 +94,13 @@ def chatbot_wrapper(text, max_new_tokens, do_sample, temperature, top_p, typical # Check if any extension wants to hijack this function call visible_text = None - custom_prompt_generator = None + custom_generate_chat_prompt = None for extension, _ in extensions_module.iterator(): if hasattr(extension, 'input_hijack') and extension.input_hijack['state'] == True: extension.input_hijack['state'] = False text, visible_text = extension.input_hijack['value'] - if custom_prompt_generator is None and hasattr(extension, 'custom_prompt_generator'): - custom_prompt_generator = extension.custom_prompt_generator + if custom_generate_chat_prompt is None and hasattr(extension, 'custom_generate_chat_prompt'): + custom_generate_chat_prompt = extension.custom_generate_chat_prompt if visible_text is None: visible_text = text @@ -108,10 +108,10 @@ def chatbot_wrapper(text, max_new_tokens, do_sample, temperature, top_p, typical visible_text = visible_text.replace('\n', '
') text = apply_extensions(text, "input") - if custom_prompt_generator is None: + if custom_generate_chat_prompt is None: prompt = generate_chat_prompt(text, max_new_tokens, name1, name2, context, chat_prompt_size) else: - prompt = custom_prompt_generator(text, max_new_tokens, name1, name2, context, chat_prompt_size) + prompt = custom_generate_chat_prompt(text, max_new_tokens, name1, name2, context, chat_prompt_size) # Generate reply = ''