mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2024-11-22 16:17:57 +01:00
Add a warning about ppl evaluation without --no_use_fast
This commit is contained in:
parent
f6d701624c
commit
9847809a7a
@ -7,6 +7,7 @@ from datasets import load_dataset
|
|||||||
from tqdm import tqdm
|
from tqdm import tqdm
|
||||||
|
|
||||||
from modules import shared
|
from modules import shared
|
||||||
|
from modules.logging_colors import logger
|
||||||
from modules.models import clear_torch_cache, load_model, unload_model
|
from modules.models import clear_torch_cache, load_model, unload_model
|
||||||
from modules.models_settings import get_model_metadata, update_model_parameters
|
from modules.models_settings import get_model_metadata, update_model_parameters
|
||||||
from modules.text_generation import encode
|
from modules.text_generation import encode
|
||||||
@ -38,6 +39,9 @@ def calculate_perplexity(models, input_dataset, stride, _max_length):
|
|||||||
https://huggingface.co/docs/transformers/perplexity#calculating-ppl-with-fixedlength-models
|
https://huggingface.co/docs/transformers/perplexity#calculating-ppl-with-fixedlength-models
|
||||||
'''
|
'''
|
||||||
|
|
||||||
|
if not shared.args.no_use_fast:
|
||||||
|
logger.warning("--no_use_fast is not being used. If tokenizing the input dataset takes a long time, consider loading the model with that option checked.")
|
||||||
|
|
||||||
global past_evaluations
|
global past_evaluations
|
||||||
cumulative_log = ''
|
cumulative_log = ''
|
||||||
cumulative_log += "Loading the input dataset...\n\n"
|
cumulative_log += "Loading the input dataset...\n\n"
|
||||||
|
Loading…
Reference in New Issue
Block a user