llama.cpp/ggml-cuda
Djip007 852aafb163
update HIP_UMA #7399 (#7414)
* update HIP_UMA #7399

add use of hipMemAdviseSetCoarseGrain when LLAMA_HIP_UMA is enable.
- get x2 on prompte eval and x1.5 on token gen with rocm6.0 on ryzen 7940HX iGPU (780M/gfx1103)

* simplify code, more consistent style

---------

Co-authored-by: slaren <slarengh@gmail.com>
2024-05-28 01:40:47 +02:00
..
acc.cu cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
acc.cuh cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
arange.cu cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
arange.cuh cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
argsort.cu ggml : mul_mat_id use the same tensor for all the experts (#6387) 2024-04-03 16:07:05 +03:00
argsort.cuh cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
binbcast.cu ggml : group all experts in a single ggml_mul_mat_id (#6505) 2024-04-18 15:18:48 +02:00
binbcast.cuh cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
clamp.cu Introduction of CUDA Graphs to LLama.cpp (#6766) 2024-05-08 22:55:49 +02:00
clamp.cuh cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
common.cuh update HIP_UMA #7399 (#7414) 2024-05-28 01:40:47 +02:00
concat.cu cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
concat.cuh cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
convert.cu ggml : drop support for QK_K=64 (#7473) 2024-05-23 10:00:21 +03:00
convert.cuh llama : add Command R Plus support (#6491) 2024-04-09 11:16:13 +03:00
cpy.cu Introduction of CUDA Graphs to LLama.cpp (#6766) 2024-05-08 22:55:49 +02:00
cpy.cuh Introduction of CUDA Graphs to LLama.cpp (#6766) 2024-05-08 22:55:49 +02:00
dequantize.cuh llama : add Command R Plus support (#6491) 2024-04-09 11:16:13 +03:00
diagmask.cu cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
diagmask.cuh cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
dmmv.cu ggml : drop support for QK_K=64 (#7473) 2024-05-23 10:00:21 +03:00
dmmv.cuh sync : ggml (#6351) 2024-03-29 17:45:46 +02:00
fattn-common.cuh CUDA: deduplicate FlashAttention code (#7352) 2024-05-18 12:36:25 +02:00
fattn-tile-f16.cu CUDA: fix FA out-of-bounds reads (#7479) 2024-05-23 00:31:20 +02:00
fattn-tile-f16.cuh CUDA: faster large batch FA without tensor cores (#7314) 2024-05-17 18:54:52 +02:00
fattn-tile-f32.cu CUDA: fix FA out-of-bounds reads (#7479) 2024-05-23 00:31:20 +02:00
fattn-tile-f32.cuh CUDA: faster large batch FA without tensor cores (#7314) 2024-05-17 18:54:52 +02:00
fattn-vec-f16.cu CUDA: fix FA out-of-bounds reads (#7479) 2024-05-23 00:31:20 +02:00
fattn-vec-f16.cuh CUDA: add FP32 FlashAttention vector kernel (#7188) 2024-05-12 19:40:45 +02:00
fattn-vec-f32.cu CUDA: fix FA out-of-bounds reads (#7479) 2024-05-23 00:31:20 +02:00
fattn-vec-f32.cuh CUDA: add FP32 FlashAttention vector kernel (#7188) 2024-05-12 19:40:45 +02:00
fattn.cu CUDA: deduplicate FlashAttention code (#7352) 2024-05-18 12:36:25 +02:00
fattn.cuh ggml : add Flash Attention (#5021) 2024-04-30 12:16:08 +03:00
getrows.cu cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
getrows.cuh cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
im2col.cu cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
im2col.cuh cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
mmq.cu ggml : drop support for QK_K=64 (#7473) 2024-05-23 10:00:21 +03:00
mmq.cuh cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
mmvq.cu Introduction of CUDA Graphs to LLama.cpp (#6766) 2024-05-08 22:55:49 +02:00
mmvq.cuh cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
norm.cu cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
norm.cuh cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
pad.cu cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
pad.cuh cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
pool2d.cu cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
pool2d.cuh cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
quantize.cu llama : add Command R Plus support (#6491) 2024-04-09 11:16:13 +03:00
quantize.cuh llama : add Command R Plus support (#6491) 2024-04-09 11:16:13 +03:00
rope.cu cuda : fix rope + add tests (#7452) 2024-05-22 11:01:35 +03:00
rope.cuh cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
scale.cu Introduction of CUDA Graphs to LLama.cpp (#6766) 2024-05-08 22:55:49 +02:00
scale.cuh cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
softmax.cu CUDA: deduplicate FlashAttention code (#7352) 2024-05-18 12:36:25 +02:00
softmax.cuh cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
sumrows.cu cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
sumrows.cuh cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
tsembd.cu cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
tsembd.cuh cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
unary.cu feat: implemented sigmoid function (ggml/806) 2024-05-11 15:38:34 +03:00
unary.cuh feat: implemented sigmoid function (ggml/806) 2024-05-11 15:38:34 +03:00
upscale.cu ggml : add ggml_upscale_ext (ggml/814) 2024-05-15 13:23:33 +03:00
upscale.cuh cuda : refactor into multiple files (#6269) 2024-03-25 13:50:23 +01:00
vecdotq.cuh ggml : drop support for QK_K=64 (#7473) 2024-05-23 10:00:21 +03:00