From b2b74c83a606c6e2c4a23d536cae4adffca8034d Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Thu, 15 Feb 2024 19:03:47 -0800 Subject: [PATCH] Fix Qwen1.5 in llamacpp_HF --- modules/models.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/modules/models.py b/modules/models.py index d8f1a9f8..ea85da84 100644 --- a/modules/models.py +++ b/modules/models.py @@ -257,7 +257,7 @@ def llamacpp_HF_loader(model_name): path = Path(f'{shared.args.model_dir}/{model_name}') # Check if a HF tokenizer is available for the model - if all((path / file).exists() for file in ['tokenizer.model', 'tokenizer_config.json']): + if all((path / file).exists() for file in ['tokenizer.json', 'tokenizer_config.json']): logger.info(f'Using tokenizer from: \"{path}\"') else: logger.error("Could not load the model because a tokenizer in Transformers format was not found.")