mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-11-11 13:30:35 +00:00
34 lines
845 B
Batchfile
34 lines
845 B
Batchfile
|
|
:: MIT license
|
|
:: Copyright (C) 2024 Intel Corporation
|
|
:: SPDX-License-Identifier: MIT
|
|
|
|
|
|
IF not exist build (mkdir build)
|
|
cd build
|
|
if %errorlevel% neq 0 goto ERROR
|
|
|
|
@call "C:\Program Files (x86)\Intel\oneAPI\setvars.bat" intel64 --force
|
|
if %errorlevel% neq 0 goto ERROR
|
|
|
|
:: for FP16
|
|
:: faster for long-prompt inference
|
|
:: cmake -G "MinGW Makefiles" .. -DGGML_SYCL=ON -DCMAKE_CXX_COMPILER=icx -DBUILD_SHARED_LIBS=ON -DCMAKE_BUILD_TYPE=Release -DGGML_SYCL_F16=ON
|
|
|
|
:: for FP32
|
|
cmake -G "Ninja" .. -DGGML_SYCL=ON -DCMAKE_C_COMPILER=cl -DCMAKE_CXX_COMPILER=icx -DBUILD_SHARED_LIBS=ON -DCMAKE_BUILD_TYPE=Release
|
|
if %errorlevel% neq 0 goto ERROR
|
|
:: build example/main only
|
|
:: make main
|
|
|
|
:: build all binary
|
|
cmake --build . -j
|
|
if %errorlevel% neq 0 goto ERROR
|
|
|
|
cd ..
|
|
exit /B 0
|
|
|
|
:ERROR
|
|
echo comomand error: %errorlevel%
|
|
exit /B %errorlevel%
|