From 8778b756e69144ff91180076068eeb9bcd915a60 Mon Sep 17 00:00:00 2001 From: Ayanami Rei Date: Mon, 13 Mar 2023 22:11:40 +0300 Subject: [PATCH] use updated load_quantized --- modules/models.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/modules/models.py b/modules/models.py index 9bcaca9e..46cd77ff 100644 --- a/modules/models.py +++ b/modules/models.py @@ -91,7 +91,7 @@ def load_model(model_name): elif shared.args.gptq_bits > 0: from modules.quant_loader import load_quantized - model = load_quantized(model_name, shared.args.gptq_model_type.lower()) + model = load_quantized(model_name) # Custom else: