mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2025-01-19 08:20:10 +01:00
llama : minor spacing changes
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
This commit is contained in:
parent
c8cdb48d10
commit
e3e33c0cbc
@ -6879,8 +6879,9 @@ static bool llm_load_tensors(
|
|||||||
|
|
||||||
const int64_t n_ff = hparams.n_ff_l(i);
|
const int64_t n_ff = hparams.n_ff_l(i);
|
||||||
|
|
||||||
ggml_context* ctx_layer = ctx_for_layer(i);
|
ggml_context * ctx_layer = ctx_for_layer(i);
|
||||||
ggml_context* ctx_split = ctx_for_layer_split(i);
|
ggml_context * ctx_split = ctx_for_layer_split(i);
|
||||||
|
|
||||||
auto & layer = model.layers[i];
|
auto & layer = model.layers[i];
|
||||||
|
|
||||||
layer.attn_norm = ml.create_tensor(ctx_layer, tn(LLM_TENSOR_ATTN_NORM, "weight", i), {n_embd});
|
layer.attn_norm = ml.create_tensor(ctx_layer, tn(LLM_TENSOR_ATTN_NORM, "weight", i), {n_embd});
|
||||||
|
Loading…
Reference in New Issue
Block a user