mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-27 06:39:25 +01:00
f3f65429c4
* scripts : update sync [no ci] * files : relocate [no ci] * ci : disable kompute build [no ci] * cmake : fixes [no ci] * server : fix mingw build ggml-ci * cmake : minor [no ci] * cmake : link math library [no ci] * cmake : build normal ggml library (not object library) [no ci] * cmake : fix kompute build ggml-ci * make,cmake : fix LLAMA_CUDA + replace GGML_CDEF_PRIVATE ggml-ci * move public backend headers to the public include directory (#8122) * move public backend headers to the public include directory * nix test * spm : fix metal header --------- Co-authored-by: Georgi Gerganov <ggerganov@gmail.com> * scripts : fix sync paths [no ci] * scripts : sync ggml-blas.h [no ci] --------- Co-authored-by: slaren <slarengh@gmail.com>
35 lines
846 B
Batchfile
35 lines
846 B
Batchfile
|
|
:: MIT license
|
|
:: Copyright (C) 2024 Intel Corporation
|
|
:: SPDX-License-Identifier: MIT
|
|
|
|
|
|
IF not exist build (mkdir build)
|
|
cd build
|
|
if %errorlevel% neq 0 goto ERROR
|
|
|
|
@call "C:\Program Files (x86)\Intel\oneAPI\setvars.bat" intel64 --force
|
|
if %errorlevel% neq 0 goto ERROR
|
|
|
|
:: for FP16
|
|
:: faster for long-prompt inference
|
|
:: cmake -G "MinGW Makefiles" .. -DGGML_SYCL=ON -DCMAKE_CXX_COMPILER=icx -DBUILD_SHARED_LIBS=ON -DCMAKE_BUILD_TYPE=Release -DGGML_SYCL_F16=ON
|
|
|
|
:: for FP32
|
|
cmake -G "Ninja" .. -DGGML_SYCL=ON -DCMAKE_C_COMPILER=cl -DCMAKE_CXX_COMPILER=icx -DBUILD_SHARED_LIBS=ON -DCMAKE_BUILD_TYPE=Release
|
|
if %errorlevel% neq 0 goto ERROR
|
|
:: build example/main only
|
|
:: make main
|
|
|
|
:: build all binary
|
|
cmake --build . -j
|
|
if %errorlevel% neq 0 goto ERROR
|
|
|
|
cd ..
|
|
exit /B 0
|
|
|
|
:ERROR
|
|
echo comomand error: %errorlevel%
|
|
exit /B %errorlevel%
|
|
|