1
0
mirror of https://github.com/ggerganov/llama.cpp.git synced 2025-01-22 17:49:10 +01:00
Commit Graph

1 Commits

Author SHA1 Message Date
Johannes Gäßler
dc685be466
CUDA: add FP32 FlashAttention vector kernel ()
* CUDA: add FP32 FlashAttention vector kernel

* fixup! CUDA: add FP32 FlashAttention vector kernel

* fixup! fixup! CUDA: add FP32 FlashAttention vector kernel

* fixup! fixup! fixup! CUDA: add FP32 FlashAttention vector kernel
2024-05-12 19:40:45 +02:00