mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-25 05:48:47 +01:00
the shapes for init model of gqa models was wrong
This commit is contained in:
parent
a847676984
commit
a03ce38455
@ -332,8 +332,8 @@ static void init_model(struct llama_model * input, struct my_llama_model * model
|
|||||||
|
|
||||||
assert_shape_1d(layer.attention_norm, hparams.n_embd);
|
assert_shape_1d(layer.attention_norm, hparams.n_embd);
|
||||||
assert_shape_2d(layer.wq, hparams.n_embd, hparams.n_embd);
|
assert_shape_2d(layer.wq, hparams.n_embd, hparams.n_embd);
|
||||||
assert_shape_2d(layer.wk, hparams.n_embd, hparams.n_embd);
|
assert_shape_2d(layer.wk, hparams.n_embd, hparams.n_embd_gqa());
|
||||||
assert_shape_2d(layer.wv, hparams.n_embd, hparams.n_embd);
|
assert_shape_2d(layer.wv, hparams.n_embd, hparams.n_embd_gqa());
|
||||||
assert_shape_2d(layer.wo, hparams.n_embd, hparams.n_embd);
|
assert_shape_2d(layer.wo, hparams.n_embd, hparams.n_embd);
|
||||||
assert_shape_1d(layer.ffn_norm, hparams.n_embd);
|
assert_shape_1d(layer.ffn_norm, hparams.n_embd);
|
||||||
assert_shape_2d(layer.w1, hparams.n_embd, hparams.n_ff);
|
assert_shape_2d(layer.w1, hparams.n_embd, hparams.n_ff);
|
||||||
|
Loading…
Reference in New Issue
Block a user