matiaslin faac0bae26
common : ensure llama_batch size does not exceed max size (#9668)
A crash was observed when the number of tokens added to a batch exceeds
llama_batch size. An assertion in llama_batch_add was added to protect
against llama_batch size overflow.
2024-09-29 15:25:00 +03:00
..
2024-09-28 17:42:03 +03:00
2024-09-28 17:42:03 +03:00
2024-09-15 20:46:12 +03:00
2023-11-13 14:16:23 +02:00