mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2025-01-07 11:23:56 +01:00
Update convert-llama-h5-to-gguf.py
This commit is contained in:
parent
b19c11750b
commit
4ed98bf1ab
@ -87,6 +87,7 @@ gguf_writer.write_feed_forward_length(llm_arch, hparams["intermediate_size"])
|
||||
gguf_writer.write_rope_dimension_count(llm_arch, hparams["hidden_size"] // hparams["num_attention_heads"])
|
||||
gguf_writer.write_head_count(llm_arch, hparams["num_attention_heads"])
|
||||
gguf_writer.write_float32(llm_arch + ".attention.layer_norm_rms_epsilon", hparams["rms_norm_eps"])
|
||||
gguf_writer.write_layer_norm_rms_eps(llm_arch, hparams["rms_norm_eps"])
|
||||
|
||||
|
||||
# TOKENIZATION
|
||||
|
Loading…
Reference in New Issue
Block a user