From 5a79863df3125824e45c09ff2052d7b622b10f1f Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Fri, 3 Mar 2023 15:54:13 -0300 Subject: [PATCH] Increase the sequence length, decrease batch size I have no idea what I am doing --- modules/LLaMA.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/modules/LLaMA.py b/modules/LLaMA.py index a98ba520..3781ccf5 100644 --- a/modules/LLaMA.py +++ b/modules/LLaMA.py @@ -70,7 +70,7 @@ class LLaMAModel: pass @classmethod - def from_pretrained(self, path, max_seq_len=512, max_batch_size=32): + def from_pretrained(self, path, max_seq_len=2048, max_batch_size=1): tokenizer_path = path / "tokenizer.model" path = os.path.abspath(path) tokenizer_path = os.path.abspath(tokenizer_path)