mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2024-11-22 08:07:56 +01:00
Fix NTK (alpha) and RoPE scaling for exllamav2 and exllamav2_HF (#3897)
This commit is contained in:
parent
eb9ebabec7
commit
34dc7306b8
@ -28,9 +28,9 @@ class Exllamav2Model:
|
||||
config.prepare()
|
||||
|
||||
config.max_seq_len = shared.args.max_seq_len
|
||||
config.rope_scale = shared.args.compress_pos_emb
|
||||
config.rope_alpha = shared.args.alpha_value
|
||||
|
||||
config.scale_pos_emb = shared.args.compress_pos_emb
|
||||
config.scale_alpha_value = shared.args.alpha_value
|
||||
|
||||
model = ExLlamaV2(config)
|
||||
|
||||
split = None
|
||||
|
@ -116,7 +116,7 @@ class Exllamav2HF(PreTrainedModel):
|
||||
config.prepare()
|
||||
|
||||
config.max_seq_len = shared.args.max_seq_len
|
||||
config.rope_scale = shared.args.compress_pos_emb
|
||||
config.rope_alpha = shared.args.alpha_value
|
||||
config.scale_pos_emb = shared.args.compress_pos_emb
|
||||
config.scale_alpha_value = shared.args.alpha_value
|
||||
|
||||
return Exllamav2HF(config)
|
||||
|
Loading…
Reference in New Issue
Block a user