From a2127239debf37432489fefa41ed36bd2b6083f6 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Sun, 16 Apr 2023 01:41:37 -0300 Subject: [PATCH] Fix a bug --- modules/text_generation.py | 1 - 1 file changed, 1 deletion(-) diff --git a/modules/text_generation.py b/modules/text_generation.py index 51e2ddc0..93288316 100644 --- a/modules/text_generation.py +++ b/modules/text_generation.py @@ -193,7 +193,6 @@ def generate_reply(question, state, eos_token=None, stopping_strings=[]): # Handling the stopping strings stopping_criteria_list = transformers.StoppingCriteriaList() - print(ast.literal_eval(f"[{state['custom_stopping_strings']}]")) for st in (stopping_strings, ast.literal_eval(f"[{state['custom_stopping_strings']}]")): if type(st) is list and len(st) > 0: sentinel_token_ids = [encode(string, add_special_tokens=False) for string in st]