mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2025-01-11 21:10:24 +01:00
Add llmaz as another platform to run llama.cpp on Kubernetes
Signed-off-by: kerthcet <kerthcet@gmail.com>
This commit is contained in:
parent
cfac111e2b
commit
7323304092
@ -191,6 +191,7 @@ Unless otherwise noted these projects are open-source with permissive licensing:
|
||||
|
||||
**Infrastructure:**
|
||||
|
||||
- [llmaz](https://github.com/InftyAI/llmaz) - ☸️ Effortlessly serve state-of-the-art LLMs on Kubernetes, see [llama.cpp example](https://github.com/InftyAI/llmaz/tree/main/docs/examples/llamacpp) here.
|
||||
- [Paddler](https://github.com/distantmagic/paddler) - Stateful load balancer custom-tailored for llama.cpp
|
||||
- [GPUStack](https://github.com/gpustack/gpustack) - Manage GPU clusters for running LLMs
|
||||
|
||||
|
Loading…
x
Reference in New Issue
Block a user