diff --git a/modules/AutoGPTQ_loader.py b/modules/AutoGPTQ_loader.py index 5f2fd71b..8cb078af 100644 --- a/modules/AutoGPTQ_loader.py +++ b/modules/AutoGPTQ_loader.py @@ -53,13 +53,16 @@ def load_quantized(model_name): model = AutoGPTQForCausalLM.from_quantized(path_to_model, **params) # These lines fix the multimodal extension when used with AutoGPTQ - if not hasattr(model, 'dtype'): - model.dtype = model.model.dtype + if hasattr(model, 'model'): + if not hasattr(model, 'dtype'): + if hasattr(model.model, 'dtype'): + model.dtype = model.model.dtype - if not hasattr(model, 'embed_tokens'): - model.embed_tokens = model.model.model.embed_tokens + if hasattr(model.model, 'model') and hasattr(model.model.model, 'embed_tokens'): + if not hasattr(model, 'embed_tokens'): + model.embed_tokens = model.model.model.embed_tokens - if not hasattr(model.model, 'embed_tokens'): - model.model.embed_tokens = model.model.model.embed_tokens + if not hasattr(model.model, 'embed_tokens'): + model.model.embed_tokens = model.model.model.embed_tokens return model