mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2025-01-08 11:46:53 +01:00
Update convert-llama-h5-to-gguf.py
This commit is contained in:
parent
b19c11750b
commit
4ed98bf1ab
@ -87,6 +87,7 @@ gguf_writer.write_feed_forward_length(llm_arch, hparams["intermediate_size"])
|
|||||||
gguf_writer.write_rope_dimension_count(llm_arch, hparams["hidden_size"] // hparams["num_attention_heads"])
|
gguf_writer.write_rope_dimension_count(llm_arch, hparams["hidden_size"] // hparams["num_attention_heads"])
|
||||||
gguf_writer.write_head_count(llm_arch, hparams["num_attention_heads"])
|
gguf_writer.write_head_count(llm_arch, hparams["num_attention_heads"])
|
||||||
gguf_writer.write_float32(llm_arch + ".attention.layer_norm_rms_epsilon", hparams["rms_norm_eps"])
|
gguf_writer.write_float32(llm_arch + ".attention.layer_norm_rms_epsilon", hparams["rms_norm_eps"])
|
||||||
|
gguf_writer.write_layer_norm_rms_eps(llm_arch, hparams["rms_norm_eps"])
|
||||||
|
|
||||||
|
|
||||||
# TOKENIZATION
|
# TOKENIZATION
|
||||||
|
Loading…
Reference in New Issue
Block a user