mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-27 06:39:25 +01:00
cda0e4b648
* refactor llama_batch_get_one * adapt all examples * fix simple.cpp * fix llama_bench * fix * fix context shifting * free batch before return * use common_batch_add, reuse llama_batch in loop * null terminated seq_id list * fix save-load-state example * fix perplexity * correct token pos in llama_batch_allocr |
||
---|---|---|
.. | ||
src | ||
.gitignore | ||
build.gradle.kts | ||
consumer-rules.pro | ||
proguard-rules.pro |