8000 cmake: revert CUDA arch default to 52, 61 if f16 (#1959) · Pints-AI/llama.cpp@bbca06e · GitHub
[go: up one dir, main page]

Skip to content

Commit bbca06e

Browse files
cmake: revert CUDA arch default to 52, 61 if f16 (ggml-org#1959)
1 parent fb98254 commit bbca06e

File tree

1 file changed

+9
-16
lines changed

1 file changed

+9
-16
lines changed

CMakeLists.txt

Lines changed: 9 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -250,6 +250,15 @@ if (LLAMA_CUBLAS)
250250
set(LLAMA_EXTRA_LIBS ${LLAMA_EXTRA_LIBS} CUDA::cudart CUDA::cublas CUDA::cublasLt)
251251
endif()
252252

253+
if (NOT DEFINED CMAKE_CUDA_ARCHITECTURES)
254+
if (LLAMA_CUDA_DMMV_F16)
255+
set(CMAKE_CUDA_ARCHITECTURES "61") # needed for f16 CUDA intrinsics
256+
else()
257+
set(CMAKE_CUDA_ARCHITECTURES "52") # lowest CUDA 12 standard
258+
endif()
259+
endif()
260+
message(STATUS "Using CUDA architectures: ${CMAKE_CUDA_ARCHITECTURES}")
261+
253262
else()
254263
message(WARNING "cuBLAS not found")
255264
endif()
@@ -493,22 +502,6 @@ if (BUILD_SHARED_LIBS)
493502
endif()
494503
endif()
495504

496-
if (GGML_SOURCES_CUDA)
497-
message(STATUS "GGML CUDA sources found, configuring CUDA architecture")
498-
set_property(TARGET ggml PROPERTY CUDA_ARCHITECTURES "native")
499-
set_property(TARGET ggml PROPERTY CUDA_SELECT_NVCC_ARCH_FLAGS "Auto")
500-
501-
set_property(TARGET ggml_static PROPERTY CUDA_ARCHITECTURES "native")
502-
set_property(TARGET ggml_static PROPERTY CUDA_SELECT_NVCC_ARCH_FLAGS "Auto")
503-
504-
if (BUILD_SHARED_LIBS)
505-
set_property(TARGET ggml_shared PROPERTY CUDA_ARCHITECTURES "native")
506-
set_property(TARGET ggml_shared PROPERTY CUDA_SELECT_NVCC_ARCH_FLAGS "Auto")
507-
endif()
508-
509-
set_property(TARGET llama PROPERTY CUDA_ARCHITECTURES "native")
510-
endif()
511-
512505

513506
#
514507
# programs, examples and tests

0 commit comments

Comments
 (0)
0