mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2025-01-12 13:27:21 +01:00
cuda : print the returned error when CUDA initialization fails (#6185)
This commit is contained in:
parent
f9c7ba3447
commit
1c51f98adc
@ -294,8 +294,9 @@ static ggml_cuda_device_info ggml_cuda_init() {
|
||||
|
||||
ggml_cuda_device_info info = {};
|
||||
|
||||
if (cudaGetDeviceCount(&info.device_count) != cudaSuccess) {
|
||||
fprintf(stderr, "%s: no " GGML_CUDA_NAME " devices found, " GGML_CUDA_NAME " will be disabled\n", __func__);
|
||||
cudaError_t err = cudaGetDeviceCount(&info.device_count);
|
||||
if (err != cudaSuccess) {
|
||||
fprintf(stderr, "%s: failed to initialize " GGML_CUDA_NAME ": %s\n", __func__, cudaGetErrorString(err));
|
||||
return info;
|
||||
}
|
||||
|
||||
|
Loading…
x
Reference in New Issue
Block a user