llama.cpp/src
fairydreaming d3f0c7166a
Stop the generation when <|eom_id|> token is encountered - needed for Llama 3.1 tool call support (#8858)
* gguf-py, llama : add constants and methods related to Llama-3.1 <|eom_id|> token

* llama : find Llama-3.1 <|eom_id|> token id during vocab loading

* llama-vocab : add Llama-3.1 <|eom_id|> token to the set of tokens stopping the generation

---------

Co-authored-by: Stanisław Szymczyk <sszymczy@gmail.com>
2024-08-05 09:38:01 +02:00
..
CMakeLists.txt llama : move vocab, grammar and sampling into separate files (#8508) 2024-07-23 13:10:17 +03:00
llama-grammar.cpp ggml : reduce hash table reset cost (#8698) 2024-07-27 04:41:55 +02:00
llama-grammar.h llama : fix build + fix fabs compile warnings (#8683) 2024-07-25 19:57:31 +03:00
llama-impl.h llama : move vocab, grammar and sampling into separate files (#8508) 2024-07-23 13:10:17 +03:00
llama-sampling.cpp llama : move vocab, grammar and sampling into separate files (#8508) 2024-07-23 13:10:17 +03:00
llama-sampling.h llama : move vocab, grammar and sampling into separate files (#8508) 2024-07-23 13:10:17 +03:00
llama-vocab.cpp Stop the generation when <|eom_id|> token is encountered - needed for Llama 3.1 tool call support (#8858) 2024-08-05 09:38:01 +02:00
llama-vocab.h Stop the generation when <|eom_id|> token is encountered - needed for Llama 3.1 tool call support (#8858) 2024-08-05 09:38:01 +02:00
llama.cpp Stop the generation when <|eom_id|> token is encountered - needed for Llama 3.1 tool call support (#8858) 2024-08-05 09:38:01 +02:00
unicode-data.cpp Removes multiple newlines at the end of files that is breaking the editorconfig step of CI. (#8258) 2024-07-02 12:18:10 -04:00
unicode-data.h llama : reorganize source code + improve CMake (#8006) 2024-06-26 18:33:02 +03:00
unicode.cpp llama : move vocab, grammar and sampling into separate files (#8508) 2024-07-23 13:10:17 +03:00
unicode.h llama : move vocab, grammar and sampling into separate files (#8508) 2024-07-23 13:10:17 +03:00