Olivier Chafik
|
6171c9d258
|
Add Jinja template support (#11016)
* Copy minja from 58f0ca6dd7
* Add --jinja and --chat-template-file flags
* Add missing <optional> include
* Avoid print in get_hf_chat_template.py
* No designated initializers yet
* Try and work around msvc++ non-macro max resolution quirk
* Update test_chat_completion.py
* Wire LLM_KV_TOKENIZER_CHAT_TEMPLATE_N in llama_model_chat_template
* Refactor test-chat-template
* Test templates w/ minja
* Fix deprecation
* Add --jinja to llama-run
* Update common_chat_format_example to use minja template wrapper
* Test chat_template in e2e test
* Update utils.py
* Update test_chat_completion.py
* Update run.cpp
* Update arg.cpp
* Refactor common_chat_* functions to accept minja template + use_jinja option
* Attempt to fix linkage of LLAMA_CHATML_TEMPLATE
* Revert LLAMA_CHATML_TEMPLATE refactor
* Normalize newlines in test-chat-templates for windows tests
* Forward decl minja::chat_template to avoid eager json dep
* Flush stdout in chat template before potential crash
* Fix copy elision warning
* Rm unused optional include
* Add missing optional include to server.cpp
* Disable jinja test that has a cryptic windows failure
* minja: fix vigogne (https://github.com/google/minja/pull/22)
* Apply suggestions from code review
Co-authored-by: Xuan Son Nguyen <thichthat@gmail.com>
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
* Finish suggested renamings
* Move chat_templates inside server_context + remove mutex
* Update --chat-template-file w/ recent change to --chat-template
* Refactor chat template validation
* Guard against missing eos/bos tokens (null token otherwise throws in llama_vocab::impl::token_get_attr)
* Warn against missing eos / bos tokens when jinja template references them
* rename: common_chat_template[s]
* reinstate assert on chat_templates.template_default
* Update minja to b8437df626
* Update minja to https://github.com/google/minja/pull/25
* Update minja from https://github.com/google/minja/pull/27
* rm unused optional header
---------
Co-authored-by: Xuan Son Nguyen <thichthat@gmail.com>
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
|
2025-01-21 13:18:51 +00:00 |
|
Georgi Gerganov
|
9f7add1cde
|
examples : fix add_special conditions (#11311)
|
2025-01-20 16:36:08 +02:00 |
|
Georgi Gerganov
|
b9daaffe02
|
simple-chat : fix BOS being added to each message (#11278)
|
2025-01-19 18:12:09 +02:00 |
|
Georgi Gerganov
|
afa8a9ec9b
|
llama : add llama_vocab , functions -> methods, naming (#11110)
* llama : functions -> methods (#11110)
* llama : add struct llama_vocab to the API (#11156)
ggml-ci
* hparams : move vocab params to llama_vocab (#11159)
ggml-ci
* vocab : more pimpl (#11165)
ggml-ci
* vocab : minor tokenization optimizations (#11160)
ggml-ci
Co-authored-by: Diego Devesa <slarengh@gmail.com>
* lora : update API names (#11167)
ggml-ci
* llama : update API names to use correct prefix (#11174)
* llama : update API names to use correct prefix
ggml-ci
* cont
ggml-ci
* cont
ggml-ci
* minor [no ci]
* vocab : llama_vocab_add_[be]os -> llama_vocab_get_add_[be]os (#11174)
ggml-ci
* vocab : llama_vocab_n_vocab -> llama_vocab_n_tokens (#11174)
ggml-ci
---------
Co-authored-by: Diego Devesa <slarengh@gmail.com>
|
2025-01-12 11:32:42 +02:00 |
|
Georgi Gerganov
|
47182dd03f
|
llama : update llama_model API names (#11063)
* llama : deprecate llama_free_model, add llama_model_free
ggml-ci
* llama : change `llama_load_model_from_file` -> `llama_model_load_from_file`
ggml-ci
|
2025-01-06 10:55:18 +02:00 |
|
Diego Devesa
|
7cc2d2c889
|
ggml : move AMX to the CPU backend (#10570)
* ggml : move AMX to the CPU backend
---------
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
|
2024-11-29 21:54:58 +01:00 |
|
Diego Devesa
|
5931c1f233
|
ggml : add support for dynamic loading of backends (#10469)
* ggml : add support for dynamic loading of backends
---------
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
|
2024-11-25 15:13:39 +01:00 |
|
Diego Devesa
|
b634f8a26f
|
simple-chat : only add bos on first prompt (#10129)
|
2024-11-02 13:08:53 +01:00 |
|
Diego Devesa
|
a6744e43e8
|
llama : add simple-chat example (#10124)
* llama : add simple-chat example
---------
Co-authored-by: Xuan Son Nguyen <thichthat@gmail.com>
|
2024-11-01 23:50:59 +01:00 |
|