mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-11-11 21:39:52 +00:00
cmake : add CUDA_ARCHITECTURES to new target ggml_static (#1917)
This commit is contained in:
parent
b2416493ab
commit
57cd69460f
@ -492,6 +492,10 @@ if (GGML_SOURCES_CUDA)
|
|||||||
message(STATUS "GGML CUDA sources found, configuring CUDA architecture")
|
message(STATUS "GGML CUDA sources found, configuring CUDA architecture")
|
||||||
set_property(TARGET ggml PROPERTY CUDA_ARCHITECTURES OFF)
|
set_property(TARGET ggml PROPERTY CUDA_ARCHITECTURES OFF)
|
||||||
set_property(TARGET ggml PROPERTY CUDA_SELECT_NVCC_ARCH_FLAGS "Auto")
|
set_property(TARGET ggml PROPERTY CUDA_SELECT_NVCC_ARCH_FLAGS "Auto")
|
||||||
|
|
||||||
|
set_property(TARGET ggml_static PROPERTY CUDA_ARCHITECTURES OFF)
|
||||||
|
set_property(TARGET ggml_static PROPERTY CUDA_SELECT_NVCC_ARCH_FLAGS "Auto")
|
||||||
|
|
||||||
set_property(TARGET llama PROPERTY CUDA_ARCHITECTURES OFF)
|
set_property(TARGET llama PROPERTY CUDA_ARCHITECTURES OFF)
|
||||||
endif()
|
endif()
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user