From e76c59d5245825ac1e1bf06bfc7f78ed1f6bcf16 Mon Sep 17 00:00:00 2001 From: klosax <131523366+klosax@users.noreply.github.com> Date: Fri, 11 Aug 2023 23:09:49 +0200 Subject: [PATCH] Update gptneox-main.cpp --- gptneox-main.cpp | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/gptneox-main.cpp b/gptneox-main.cpp index 1667c4d54..f2be93e4b 100644 --- a/gptneox-main.cpp +++ b/gptneox-main.cpp @@ -565,8 +565,8 @@ bool gpt_neox_model_load(const std::string & fname, gpt_neox_model & model, gpt2 std::string blocknamestart = "transformer.blocks." + std::to_string(i) + "."; - layer.ln_1_g = get_tensor_ex(ctx, blocknamestart + "attn_norm_1.weight" ); - layer.ln_1_b = get_tensor_ex(ctx, blocknamestart + "attn_norm_1.bias" ); + layer.ln_1_g = get_tensor_ex(ctx, blocknamestart + "attn_norm.weight" ); + layer.ln_1_b = get_tensor_ex(ctx, blocknamestart + "attn_norm.bias" ); layer.c_attn_attn_w = get_tensor_ex(ctx, blocknamestart + "attn_qkv.weight" ); layer.c_attn_attn_b = get_tensor_ex(ctx ,blocknamestart + "attn_qkv.bias" ); @@ -584,8 +584,8 @@ bool gpt_neox_model_load(const std::string & fname, gpt_neox_model & model, gpt2 layer.c_mlp_proj_b = get_tensor_ex(ctx, blocknamestart + "ffn_down.bias" ); // map by name - model.tensors[blocknamestart + "attn_norm_1.weight"] = layer.ln_1_g; - model.tensors[blocknamestart + "attn_norm_1.bias"] = layer.ln_1_b; + model.tensors[blocknamestart + "attn_norm.weight"] = layer.ln_1_g; + model.tensors[blocknamestart + "attn_norm.bias"] = layer.ln_1_b; model.tensors[blocknamestart + "attn_qkv.weight"] = layer.c_attn_attn_w; model.tensors[blocknamestart + "attn_qkv.bias"] = layer.c_attn_attn_b;