From 60b2e1b9c529f74f5bf881b05a6247ff6f58a71c Mon Sep 17 00:00:00 2001 From: Icecream95 Date: Sat, 18 May 2024 20:19:10 +1200 Subject: [PATCH] fixup! Initial OpenELM support (270M only so far) Fix formatting --- llama.cpp | 8 ++------ 1 file changed, 2 insertions(+), 6 deletions(-) diff --git a/llama.cpp b/llama.cpp index 88aedf66f..38285ceab 100644 --- a/llama.cpp +++ b/llama.cpp @@ -6114,9 +6114,7 @@ static bool llm_load_tensors( layer.ffn_up = ml.create_tensor(ctx_split, tn(LLM_TENSOR_FFN_UP, "weight", i), {n_embd, n_ff}); } } break; - case LLM_ARCH_OPENELM: - { - { + case LLM_ARCH_OPENELM: { std::vector num_kv_heads = {3, 3, 3, 3, 3, 4, 4, 4, 4, 4, 4, 4, 5, 5, 5, 5}; std::vector num_query_heads = {12, 12, 12, 12, 12, 16, 16, 16, 16, 16, 16, 16, 20, 20, 20, 20}; @@ -6157,9 +6155,7 @@ static bool llm_load_tensors( layer.ffn_up = ml.create_tensor(ctx_split, tn(LLM_TENSOR_FFN_UP, "weight", i), { n_embd, 2 * ffn_inter }); layer.ffn_down = ml.create_tensor(ctx_split, tn(LLM_TENSOR_FFN_DOWN, "weight", i), { ffn_inter, n_embd }); } - } - } - } break; + } break; default: throw std::runtime_error("unknown architecture"); }