mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2024-11-22 08:07:56 +01:00
Perplexity evaluation: add some informative error messages
This commit is contained in:
parent
59032140b5
commit
ac5a7a26ea
@ -39,8 +39,20 @@ def calculate_perplexity(models, input_dataset, stride, _max_length):
|
|||||||
https://huggingface.co/docs/transformers/perplexity#calculating-ppl-with-fixedlength-models
|
https://huggingface.co/docs/transformers/perplexity#calculating-ppl-with-fixedlength-models
|
||||||
'''
|
'''
|
||||||
|
|
||||||
|
if shared.args.loader == "llama.cpp":
|
||||||
|
logger.error("llamacpp_HF is required for perplexity evaluation with GGUF models. Please reload the model with llamacpp_HF instead of llama.cpp.")
|
||||||
|
raise ValueError
|
||||||
|
|
||||||
|
if shared.args.loader == "ExLlamav2":
|
||||||
|
logger.error("ExLlamav2_HF is required for perplexity evaluation with EXL2 models. Please reload the model with ExLlamav2_HF instead of ExLlamav2.")
|
||||||
|
raise ValueError
|
||||||
|
|
||||||
|
if shared.args.loader == "llamacpp_HF" and not shared.args.logits_all:
|
||||||
|
logger.error("--logits_all is required for perplexity evaluation with GGUF models. Please reload the model with that option set/checked.")
|
||||||
|
raise ValueError
|
||||||
|
|
||||||
if not shared.args.no_use_fast:
|
if not shared.args.no_use_fast:
|
||||||
logger.warning("--no_use_fast is not being used. If tokenizing the input dataset takes a long time, consider loading the model with that option checked.")
|
logger.warning("--no_use_fast is not set. If tokenizing the input dataset takes a long time, try reloading the model with that option set/checked.")
|
||||||
|
|
||||||
global past_evaluations
|
global past_evaluations
|
||||||
cumulative_log = ''
|
cumulative_log = ''
|
||||||
|
Loading…
Reference in New Issue
Block a user