llama.cpp/ggml
Akarshan Biswas f446c2cf6a
SYCL: Add gated linear attention kernel (#11175)
* SYCL: Add Gated Linear attention kernel

* glahpp: add a space at the end of file

* gla: Put the barrier inside the main logic loop
2025-01-15 11:20:17 +08:00
..
include llama: add support for QRWKV6 model architecture (#11001) 2025-01-10 09:58:08 +08:00
src SYCL: Add gated linear attention kernel (#11175) 2025-01-15 11:20:17 +08:00
.gitignore vulkan : cmake integration (#8119) 2024-07-13 18:12:39 +02:00
CMakeLists.txt GGUF: C++ refactor, backend support, misc fixes (#11030) 2025-01-07 18:01:58 +01:00