@@ -135,46 +135,46 @@ if (LLAMA_BUILD)
135135 )
136136 endif ()
137137
138- if (LLAVA_BUILD)
139- if (LLAMA_CUBLAS OR LLAMA_CUDA)
140- add_compile_definitions (GGML_USE_CUBLAS)
141- add_compile_definitions (GGML_USE_CUDA)
142- endif ()
143-
144- if (LLAMA_METAL)
145- add_compile_definitions (GGML_USE_METAL)
146- endif ()
147-
148- # Building llava
149- add_subdirectory (vendor/llama.cpp/tools/mtmd)
150- set_target_properties (llava_shared PROPERTIES OUTPUT_NAME "llava" )
151-
152- if (WIN32 )
153- set_target_properties (llava_shared PROPERTIES CUDA_ARCHITECTURES OFF )
154- endif ()
155- llama_cpp_python_install_target(llava_shared)
156- if (WIN32 )
157- install (
158- FILES $<TARGET_RUNTIME_DLLS:llava_shared>
159- DESTINATION ${CMAKE_CURRENT_SOURCE_DIR} /llama_cpp/lib
160- )
161- install (
162- FILES $<TARGET_RUNTIME_DLLS:llava_shared>
163- DESTINATION ${SKBUILD_PLATLIB_DIR} /llama_cpp/lib
164- )
165- endif ()
166-
167- # Fix for llava build: Add include directory for llama.h
168- # Move these commands after the add_subdirectory call
169- target_include_directories (llava PUBLIC ${CMAKE_CURRENT_SOURCE_DIR} /vendor/llama.cpp/include )
170- target_include_directories (llava PUBLIC ${CMAKE_CURRENT_SOURCE_DIR} /vendor/llama.cpp/ggml/include )
171-
172- if (BUILD_SHARED_LIBS )
173- target_include_directories (llava_shared PUBLIC ${CMAKE_CURRENT_SOURCE_DIR} /vendor/llama.cpp/include )
174- target_include_directories (llava_shared PUBLIC ${CMAKE_CURRENT_SOURCE_DIR} /vendor/llama.cpp/ggml/include )
175- endif ()
176-
177- target_include_directories (llama-llava-cli PUBLIC ${CMAKE_CURRENT_SOURCE_DIR} /vendor/llama.cpp/include )
178- target_include_directories (llama-minicpmv-cli PUBLIC ${CMAKE_CURRENT_SOURCE_DIR} /vendor/llama.cpp/include )
179- endif ()
138+ # if (LLAVA_BUILD)
139+ # if (LLAMA_CUBLAS OR LLAMA_CUDA)
140+ # add_compile_definitions(GGML_USE_CUBLAS)
141+ # add_compile_definitions(GGML_USE_CUDA)
142+ # endif()
143+ #
144+ # if (LLAMA_METAL)
145+ # add_compile_definitions(GGML_USE_METAL)
146+ # endif()
147+ #
148+ # # Building llava
149+ # add_subdirectory(vendor/llama.cpp/tools/mtmd)
150+ # set_target_properties(llava_shared PROPERTIES OUTPUT_NAME "llava")
151+ #
152+ # if (WIN32)
153+ # set_target_properties(llava_shared PROPERTIES CUDA_ARCHITECTURES OFF)
154+ # endif()
155+ # llama_cpp_python_install_target(llava_shared)
156+ # if (WIN32)
157+ # install(
158+ # FILES $<TARGET_RUNTIME_DLLS:llava_shared>
159+ # DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/llama_cpp/lib
160+ # )
161+ # install(
162+ # FILES $<TARGET_RUNTIME_DLLS:llava_shared>
163+ # DESTINATION ${SKBUILD_PLATLIB_DIR}/llama_cpp/lib
164+ # )
165+ # endif()
166+ #
167+ # # Fix for llava build: Add include directory for llama.h
168+ # # Move these commands after the add_subdirectory call
169+ # target_include_directories(llava PUBLIC ${CMAKE_CURRENT_SOURCE_DIR}/vendor/llama.cpp/include)
170+ # target_include_directories(llava PUBLIC ${CMAKE_CURRENT_SOURCE_DIR}/vendor/llama.cpp/ggml/include)
171+ #
172+ # if (BUILD_SHARED_LIBS)
173+ # target_include_directories(llava_shared PUBLIC ${CMAKE_CURRENT_SOURCE_DIR}/vendor/llama.cpp/include)
174+ # target_include_directories(llava_shared PUBLIC ${CMAKE_CURRENT_SOURCE_DIR}/vendor/llama.cpp/ggml/include)
175+ # endif()
176+ #
177+ # target_include_directories(llama-llava-cli PUBLIC ${CMAKE_CURRENT_SOURCE_DIR}/vendor/llama.cpp/include)
178+ # target_include_directories(llama-minicpmv-cli PUBLIC ${CMAKE_CURRENT_SOURCE_DIR}/vendor/llama.cpp/include)
179+ # endif()
180180endif ()
0 commit comments