mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2025-01-31 06:03:11 +01:00
Remove obsolete HIP workaround
This commit is contained in:
parent
b56f079e28
commit
9ad2e7dfd2
@ -119,13 +119,6 @@ static cudaError_t ggml_cuda_device_malloc(void ** ptr, size_t size, int device)
|
|||||||
}
|
}
|
||||||
|
|
||||||
static ggml_cuda_device_info ggml_cuda_init() {
|
static ggml_cuda_device_info ggml_cuda_init() {
|
||||||
#ifdef __HIP_PLATFORM_AMD__
|
|
||||||
// Workaround for a rocBLAS bug when using multiple graphics cards:
|
|
||||||
// https://github.com/ROCmSoftwarePlatform/rocBLAS/issues/1346
|
|
||||||
rocblas_initialize();
|
|
||||||
CUDA_CHECK(cudaDeviceSynchronize());
|
|
||||||
#endif
|
|
||||||
|
|
||||||
ggml_cuda_device_info info = {};
|
ggml_cuda_device_info info = {};
|
||||||
|
|
||||||
cudaError_t err = cudaGetDeviceCount(&info.device_count);
|
cudaError_t err = cudaGetDeviceCount(&info.device_count);
|
||||||
|
Loading…
Reference in New Issue
Block a user