mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-27 06:39:25 +01:00
llama : make pos_bias contiguous for CUDA
This commit is contained in:
parent
bad0cafee9
commit
c4ded1a8fb
@ -8104,6 +8104,9 @@ struct llm_build_context {
|
|||||||
pos_bias = ggml_permute(ctx0, pos_bias, 2, 0, 1, 3);
|
pos_bias = ggml_permute(ctx0, pos_bias, 2, 0, 1, 3);
|
||||||
cb(pos_bias, "pos_bias", -1);
|
cb(pos_bias, "pos_bias", -1);
|
||||||
|
|
||||||
|
pos_bias = ggml_cont(ctx0, pos_bias);
|
||||||
|
cb(pos_bias, "pos_bias", -1);
|
||||||
|
|
||||||
return pos_bias;
|
return pos_bias;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user