llama.cpp/examples/parallel
2023-09-21 20:10:14 +02:00
..
CMakeLists.txt parallel : example for serving multiple users in parallel 2023-09-18 20:37:28 +03:00
parallel.cpp parallel : add disabled experimental batch chunking in powers of two 2023-09-20 20:14:05 +03:00
README.md simple : add README.md 2023-09-21 20:10:14 +02:00

llama.cpp/example/parallel

Simplified simluation for serving incoming requests in parallel