From 991bb57e439ccfbcd5a0f154957c98d2e3d66c35 Mon Sep 17 00:00:00 2001 From: cal066 <60696996+cal066@users.noreply.github.com> Date: Mon, 14 Aug 2023 18:17:24 +0000 Subject: [PATCH] ctransformers: Fix up model_type name consistency (#3567) --- README.md | 2 +- models/config.yaml | 6 +++--- modules/loaders.py | 1 + 3 files changed, 5 insertions(+), 4 deletions(-) diff --git a/README.md b/README.md index 0cbe24cc..3a7e7eab 100644 --- a/README.md +++ b/README.md @@ -259,7 +259,7 @@ Optionally, you can use the following command-line flags: | Flag | Description | |-------------|-------------| -| `--model_type MODEL_TYPE` | Model type of pre-quantized model. Currently gpt2, gptj, gpt_neox, falcon, llama, mpt, gpt_bigcode, dolly-v2, and replit are supported. | +| `--model_type MODEL_TYPE` | Model type of pre-quantized model. Currently gpt2, gptj, gptneox, falcon, llama, mpt, starcoder (gptbigcode), dollyv2, and replit are supported. | #### AutoGPTQ diff --git a/models/config.yaml b/models/config.yaml index ba12e8bc..624840df 100644 --- a/models/config.yaml +++ b/models/config.yaml @@ -5,7 +5,7 @@ .*(gpt-j|gptj|gpt4all-j|malion-6b|pygway|pygmalion-6b|dolly-v1): model_type: 'gptj' .*(gpt-neox|koalpaca-polyglot|polyglot.*koalpaca|polyglot-ko|polyglot_ko|pythia|stablelm|incite|dolly-v2|polycoder|h2ogpt-oig|h2ogpt-oasst1|h2ogpt-gm): - model_type: 'gpt_neox' + model_type: 'gptneox' .*llama: model_type: 'llama' .*bloom: @@ -17,9 +17,9 @@ .*mpt: model_type: 'mpt' .*(starcoder|starchat): - model_type: 'gpt_bigcode' + model_type: 'starcoder' .*dolly-v2: - model_type: 'dolly-v2' + model_type: 'dollyv2' .*replit: model_type: 'replit' llama-65b-gptq-3bit: diff --git a/modules/loaders.py b/modules/loaders.py index d7bd8d48..08a11ac0 100644 --- a/modules/loaders.py +++ b/modules/loaders.py @@ -290,6 +290,7 @@ loaders_model_types = { "dollyv2" "replit", "starcoder", + "gptbigcode", "falcon" ], }