llama.cpp/ggml
Diego Devesa a5e47592b6
cuda : optimize argmax (#10441)
* cuda : optimize argmax

* remove unused parameter

ggml-ci

* fixup : use full warps

ggml-ci

* Apply suggestions from code review

Co-authored-by: Johannes Gäßler <johannesg@5d6.de>

* fix ub

* ggml : check ne00 <= INT32_MAX in argmax and argsort

---------

Co-authored-by: Johannes Gäßler <johannesg@5d6.de>
2024-11-21 18:18:50 +01:00
..
include ggml: new optimization interface (ggml/988) 2024-11-17 08:30:29 +02:00
src cuda : optimize argmax (#10441) 2024-11-21 18:18:50 +01:00
.gitignore vulkan : cmake integration (#8119) 2024-07-13 18:12:39 +02:00
CMakeLists.txt add cmake rvv support (#10411) 2024-11-19 21:10:31 +01:00