llama.cpp/docs
Michał Moskal ff227703d6
sampling : support for llguidance grammars (#10224)
* initial porting of previous LLG patch

* update for new APIs

* build: integrate llguidance as an external project

* use '%llguidance' as marker to enable llg lark syntax

* add some docs

* clarify docs

* code style fixes

* remove llguidance.h from .gitignore

* fix tests when llg is enabled

* pass vocab not model to llama_sampler_init_llg()

* copy test-grammar-integration.cpp to test-llguidance.cpp

* clang fmt

* fix ref-count bug

* build and run test

* gbnf -> lark syntax

* conditionally include llguidance test based on LLAMA_LLGUIDANCE flag

* rename llguidance test file to test-grammar-llguidance.cpp

* add gh action for llg test

* align tests with LLG grammar syntax and JSON Schema spec

* llama_tokenizer() in fact requires valid utf8

* update llg

* format file

* add $LLGUIDANCE_LOG_LEVEL support

* fix whitespace

* fix warning

* include <cmath> for INFINITY

* add final newline

* fail llama_sampler_init_llg() at runtime

* Link gbnf_to_lark.py script; fix links; refer to llg docs for lexemes

* simplify #includes

* improve doc string for LLAMA_LLGUIDANCE

* typo in merge

* bump llguidance to 0.6.12
2025-02-02 09:55:32 +02:00
..
backend docs : Update readme to build targets for local docker build (#11368) 2025-01-24 14:30:13 +01:00
development model: Add support for PhiMoE arch (#11003) 2025-01-09 11:21:41 +01:00
android.md docs: fix outdated usage of llama-simple (#10565) 2024-11-28 16:03:11 +01:00
build.md docs : Update readme to build targets for local docker build (#11368) 2025-01-24 14:30:13 +01:00
cuda-fedora.md doc: add cuda guide for fedora (#11135) 2025-01-09 11:32:06 +00:00
docker.md docs : Update readme to build targets for local docker build (#11368) 2025-01-24 14:30:13 +01:00
install.md Reorganize documentation pages (#8325) 2024-07-05 18:08:32 +02:00
llguidance.md sampling : support for llguidance grammars (#10224) 2025-02-02 09:55:32 +02:00