mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2024-11-25 09:19:23 +01:00
Fix Qwen1.5 in llamacpp_HF
This commit is contained in:
parent
080f7132c0
commit
b2b74c83a6
@ -257,7 +257,7 @@ def llamacpp_HF_loader(model_name):
|
|||||||
path = Path(f'{shared.args.model_dir}/{model_name}')
|
path = Path(f'{shared.args.model_dir}/{model_name}')
|
||||||
|
|
||||||
# Check if a HF tokenizer is available for the model
|
# Check if a HF tokenizer is available for the model
|
||||||
if all((path / file).exists() for file in ['tokenizer.model', 'tokenizer_config.json']):
|
if all((path / file).exists() for file in ['tokenizer.json', 'tokenizer_config.json']):
|
||||||
logger.info(f'Using tokenizer from: \"{path}\"')
|
logger.info(f'Using tokenizer from: \"{path}\"')
|
||||||
else:
|
else:
|
||||||
logger.error("Could not load the model because a tokenizer in Transformers format was not found.")
|
logger.error("Could not load the model because a tokenizer in Transformers format was not found.")
|
||||||
|
Loading…
Reference in New Issue
Block a user