@@ -135,46 +135,46 @@ if (LLAMA_BUILD)
135
135
)
136
136
endif ()
137
137
138
- if (LLAVA_BUILD )
139
- if (LLAMA_CUBLAS OR LLAMA_CUDA )
140
- add_compile_definitions (GGML_USE_CUBLAS )
141
- add_compile_definitions (GGML_USE_CUDA )
142
- endif ()
143
-
144
- if (LLAMA_METAL )
145
- add_compile_definitions (GGML_USE_METAL )
146
- endif ()
147
-
148
- # Building llava
149
- add_subdirectory (vendor/llama.cpp/tools/mtmd )
150
- set_target_properties (llava_shared PROPERTIES OUTPUT_NAME "llava" )
151
-
152
- if (WIN32 )
153
- set_target_properties (llava_shared PROPERTIES CUDA_ARCHITECTURES OFF )
154
- endif ()
155
- llama_cpp_python_install_target (llava_shared )
156
- if (WIN32 )
157
- install (
158
- FILES $< TARGET_RUNTIME_DLLS:llava_shared>
159
- DESTINATION ${CMAKE_CURRENT_SOURCE_DIR} /llama_cpp/lib
160
- )
161
- install (
162
- FILES $< TARGET_RUNTIME_DLLS:llava_shared>
163
- DESTINATION ${SKBUILD_PLATLIB_DIR} /llama_cpp/lib
164
- )
165
- endif ()
166
-
167
- # Fix for llava build: Add include directory for llama.h
168
- # Move these commands after the add_subdirectory call
169
- target_include_directories (llava PUBLIC ${CMAKE_CURRENT_SOURCE_DIR} /vendor/llama.cpp/include )
170
- target_include_directories (llava PUBLIC ${CMAKE_CURRENT_SOURCE_DIR} /vendor/llama.cpp/ggml/include )
171
-
172
- if (BUILD_SHARED_LIBS )
173
- target_include_directories (llava_shared PUBLIC ${CMAKE_CURRENT_SOURCE_DIR} /vendor/llama.cpp/include )
174
- target_include_directories (llava_shared PUBLIC ${CMAKE_CURRENT_SOURCE_DIR} /vendor/llama.cpp/ggml/include )
175
- endif ()
176
-
177
- target_include_directories (llama-llava-cli PUBLIC ${CMAKE_CURRENT_SOURCE_DIR} /vendor/llama.cpp/include )
178
- target_include_directories (llama-minicpmv-cli PUBLIC ${CMAKE_CURRENT_SOURCE_DIR} /vendor/llama.cpp/include )
179
- endif ()
138
+ # if (LLAVA_BUILD)
139
+ # if (LLAMA_CUBLAS OR LLAMA_CUDA)
140
+ # add_compile_definitions(GGML_USE_CUBLAS)
141
+ # add_compile_definitions(GGML_USE_CUDA)
142
+ # endif()
143
+ #
144
+ # if (LLAMA_METAL)
145
+ # add_compile_definitions(GGML_USE_METAL)
146
+ # endif()
147
+ #
148
+ # # Building llava
149
+ # add_subdirectory(vendor/llama.cpp/tools/mtmd)
150
+ # set_target_properties(llava_shared PROPERTIES OUTPUT_NAME "llava")
151
+ #
152
+ # if (WIN32)
153
+ # set_target_properties(llava_shared PROPERTIES CUDA_ARCHITECTURES OFF)
154
+ # endif()
155
+ # llama_cpp_python_install_target(llava_shared)
156
+ # if (WIN32)
157
+ # install(
158
+ # FILES $<TARGET_RUNTIME_DLLS:llava_shared>
159
+ # DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/llama_cpp/lib
160
+ # )
161
+ # install(
162
+ # FILES $<TARGET_RUNTIME_DLLS:llava_shared>
163
+ # DESTINATION ${SKBUILD_PLATLIB_DIR}/llama_cpp/lib
164
+ # )
165
+ # endif()
166
+ #
167
+ # # Fix for llava build: Add include directory for llama.h
168
+ # # Move these commands after the add_subdirectory call
169
+ # target_include_directories(llava PUBLIC ${CMAKE_CURRENT_SOURCE_DIR}/vendor/llama.cpp/include)
170
+ # target_include_directories(llava PUBLIC ${CMAKE_CURRENT_SOURCE_DIR}/vendor/llama.cpp/ggml/include)
171
+ #
172
+ # if (BUILD_SHARED_LIBS)
173
+ # target_include_directories(llava_shared PUBLIC ${CMAKE_CURRENT_SOURCE_DIR}/vendor/llama.cpp/include)
174
+ # target_include_directories(llava_shared PUBLIC ${CMAKE_CURRENT_SOURCE_DIR}/vendor/llama.cpp/ggml/include)
175
+ # endif()
176
+ #
177
+ # target_include_directories(llama-llava-cli PUBLIC ${CMAKE_CURRENT_SOURCE_DIR}/vendor/llama.cpp/include)
178
+ # target_include_directories(llama-minicpmv-cli PUBLIC ${CMAKE_CURRENT_SOURCE_DIR}/vendor/llama.cpp/include)
179
+ # endif()
180
180
endif ()
0 commit comments