llama.cpp/include
Eric Curtin 0cc63754b8
Introduce llama-run (#10291)
It's like simple-chat but it uses smart pointers to avoid manual
memory cleanups. Less memory leaks in the code now. Avoid printing
multiple dots. Split code into smaller functions. Uses no exception
handling.

Signed-off-by: Eric Curtin <ecurtin@redhat.com>
2024-11-25 22:56:24 +01:00
..
llama-cpp.h Introduce llama-run (#10291) 2024-11-25 22:56:24 +01:00
llama.h llama : accept a list of devices to use to offload a model (#10497) 2024-11-25 19:30:06 +01:00