From 7ef7bba6e64da1617d0187392ae5968a392644f3 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Sun, 12 Feb 2023 09:36:27 -0300 Subject: [PATCH] Add progress bar for model loading --- server.py | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/server.py b/server.py index 6bc3673b..56645d0c 100644 --- a/server.py +++ b/server.py @@ -168,6 +168,8 @@ def load_model_wrapper(selected_model): torch.cuda.empty_cache() model, tokenizer = load_model(model_name) + return selected_model + def load_preset_values(preset_menu, return_dict=False): generate_params = { 'do_sample': True, @@ -408,7 +410,7 @@ def create_settings_menus(): min_length = gr.Slider(0, 2000, step=1, value=generate_params["min_length"] if args.no_stream else 0, label="min_length", interactive=args.no_stream) early_stopping = gr.Checkbox(value=generate_params["early_stopping"], label="early_stopping") - model_menu.change(load_model_wrapper, [model_menu], []) + model_menu.change(load_model_wrapper, [model_menu], [model_menu], show_progress=True) preset_menu.change(load_preset_values, [preset_menu], [do_sample, temperature, top_p, typical_p, repetition_penalty, top_k, min_length, no_repeat_ngram_size, num_beams, penalty_alpha, length_penalty, early_stopping]) return preset_menu, do_sample, temperature, top_p, typical_p, repetition_penalty, top_k, min_length, no_repeat_ngram_size, num_beams, penalty_alpha, length_penalty, early_stopping