@@ -28,6 +28,7 @@ if (LLAMA_BUILD)
28
28
# which point to directories outside the build tree to the install RPATH
29
29
set (CMAKE_INSTALL_RPATH_USE_LINK_PATH TRUE )
30
30
set (CMAKE_INSTALL_RPATH "$ORIGIN" )
31
+ set (CMAKE_SKIP_RPATH FALSE )
31
32
32
33
# Building llama
33
34
if (APPLE AND NOT CMAKE_SYSTEM_PROCESSOR MATCHES "arm64" )
@@ -52,6 +53,10 @@ if (LLAMA_BUILD)
52
53
FRAMEWORK DESTINATION ${LLAMA_CPP_PYTHON_INSTALL_DIR}
53
54
RESOURCE DESTINATION ${LLAMA_CPP_PYTHON_INSTALL_DIR}
54
55
)
56
+ set_target_properties (llama PROPERTIES
57
+ INSTALL_RPATH "$ORIGIN"
58
+ BUILD_WITH_INSTALL_RPATH TRUE
59
+ )
55
60
install (
56
61
TARGETS ggml
57
62
LIBRARY DESTINATION ${LLAMA_CPP_PYTHON_INSTALL_DIR}
@@ -60,6 +65,10 @@ if (LLAMA_BUILD)
60
65
FRAMEWORK DESTINATION ${LLAMA_CPP_PYTHON_INSTALL_DIR}
61
66
RESOURCE DESTINATION ${LLAMA_CPP_PYTHON_INSTALL_DIR}
62
67
)
68
+ set_target_properties (ggml PROPERTIES
69
+ INSTALL_RPATH "$ORIGIN"
70
+ BUILD_WITH_INSTALL_RPATH TRUE
71
+ )
63
72
# Workaround for Windows + CUDA https://github.com/abetlen/llama-cpp-python/issues/563
64
73
if (WIN32 AND (LLAMA_CUDA OR LLAMA_CUBLAS ))
65
74
install (
@@ -93,5 +102,9 @@ if (LLAMA_BUILD)
93
102
FRAMEWORK DESTINATION ${LLAMA_CPP_PYTHON_INSTALL_DIR}
94
103
RESOURCE DESTINATION ${LLAMA_CPP_PYTHON_INSTALL_DIR}
95
104
)
105
+ set_target_properties (llava PROPERTIES
106
+ INSTALL_RPATH "$ORIGIN"
107
+ BUILD_WITH_INSTALL_RPATH TRUE
108
+ )
96
109
endif ()
97
110
endif ()
0 commit comments