forked from jiuyuan/InfiniTensor
293 lines
11 KiB
CMake
293 lines
11 KiB
CMake
cmake_minimum_required(VERSION 3.17) # FindCUDAToolkit
|
|
include(CMakeDependentOption)
|
|
project(InfiniTensor C CXX)
|
|
|
|
# Do not change these options in this file. Use cmake.config, cmake -DOPTION=VALUE, or ccmake to specify them.
|
|
option(USE_CUDA "Support CUDA GPU" OFF)
|
|
option(USE_BANG "Support BANG MLU" OFF)
|
|
option(USE_INTELCPU "Support INTELCPU" OFF)
|
|
option(USE_BACKTRACE "Print backtrace on exception and segmentation fault" ON)
|
|
option(USE_PROTOBUF "Serialize and deserialize tensors" OFF)
|
|
option(BUILD_DIST "Build project for distributed running" OFF)
|
|
option(BUILD_TEST "Build tests" OFF)
|
|
|
|
cmake_dependent_option(BUILD_TEST_CORE "Build tests for core components" ON BUILD_TEST OFF)
|
|
cmake_dependent_option(BUILD_TEST_PET "Build tests for PET" OFF BUILD_TEST OFF)
|
|
cmake_dependent_option(BUILD_TEST_EINNET "Build tests for EINNET" OFF BUILD_TEST OFF)
|
|
|
|
set(DEFAULT_BUILD_TYPE "RelWithDebInfo")
|
|
# Build Type
|
|
if(CMAKE_BUILD_TYPE STREQUAL "Debug")
|
|
message("Configuring for Debug build.")
|
|
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -g -O0")
|
|
add_compile_definitions(DEBUG_MODE)
|
|
elseif(CMAKE_BUILD_TYPE STREQUAL "Release")
|
|
message("Configuring for Release build.")
|
|
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -O2")
|
|
add_compile_definitions(NDEBUG)
|
|
elseif(CMAKE_BUILD_TYPE STREQUAL "RelWithDebInfo")
|
|
message("Configuring for RelWithDebInfo build.")
|
|
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -g -O2")
|
|
else()
|
|
message("Build type not specified. Configuring for RelWithDebInfo build.")
|
|
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -g -O2")
|
|
endif()
|
|
|
|
|
|
if(EXISTS ${CMAKE_CURRENT_BINARY_DIR}/config.cmake)
|
|
message(STATUS "Using config.cmake in CMAKE_CURRENT_BINARY_DIR directory")
|
|
include(${CMAKE_CURRENT_BINARY_DIR}/config.cmake)
|
|
else()
|
|
if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/config.cmake)
|
|
message(STATUS "Using config.cmake in CMAKE_CURRENT_SOURCE_DIR directory")
|
|
include(${CMAKE_CURRENT_SOURCE_DIR}/config.cmake)
|
|
endif()
|
|
endif()
|
|
|
|
set(CMAKE_CXX_STANDARD 17)
|
|
set(CMAKE_CXX_EXTENSIONS OFF) # -std=gnu++11 when on, -std=c++11 when off
|
|
|
|
find_package(
|
|
Python
|
|
COMPONENTS Interpreter Development
|
|
REQUIRED)
|
|
# OpenMP
|
|
find_package(OpenMP)
|
|
if(OpenMP_C_FOUND)
|
|
set(CMAKE_C_FLAGS "${CMAKE_C_FLAGS} ${OpenMP_C_FLAGS}")
|
|
endif()
|
|
if(OpenMP_CXX_FOUND)
|
|
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} ${OpenMP_CXX_FLAGS}")
|
|
endif()
|
|
|
|
#Protobuf
|
|
if(USE_PROTOBUF)
|
|
add_definitions(-D TENSOR_PROTOBUF)
|
|
find_package(Protobuf REQUIRED)
|
|
message(STATUS "protobuf include: " ${PROTOBUF_INCLUDE_DIRS})
|
|
message(STATUS "protobuf libraries: " ${PROTOBUF_LIBRARIES})
|
|
message(STATUS "protoc executable: " ${PROTOBUF_PROTOC_EXECUTABLE})
|
|
include_directories(${PROTOBUF_INCLUDE_DIR})
|
|
include_directories(${CMAKE_CURRENT_BINARY_DIR})
|
|
set(PROTO_PATH "${CMAKE_CURRENT_SOURCE_DIR}/proto")
|
|
file(GLOB PROTO_FILES "${PROTO_PATH}/data.proto")
|
|
protobuf_generate_cpp(PROTO_SRCS PROTO_HDRS ${PROTO_FILES})
|
|
set_source_files_properties (${PROTO_SRCS} PROPERTIES COMPILE_FLAGS -Wno-unused-variable)
|
|
add_library(tensor_proto SHARED ${PROTO_SRCS} ${PROTO_HDRS})
|
|
target_link_libraries(tensor_proto PUBLIC ${PROTOBUF_LIBRARIES})
|
|
endif()
|
|
|
|
include_directories(include)
|
|
# Pybind11
|
|
add_subdirectory(3rd-party/pybind11)
|
|
include_directories(3rd-party/pybind11/include)
|
|
|
|
# nlohmann_json
|
|
add_subdirectory(3rd-party/nlohmann_json_cmake_fetchcontent)
|
|
include_directories(3rd-party/nlohmann_json_cmake_fetchcontent/single_include)
|
|
|
|
# TVM backend
|
|
if(BUILD_TEST_EINNET)
|
|
if (NOT TVM_INCLUDE_DIR OR NOT DMLC_INCLUDE_DIR OR NOT DLPACK_INCLUDE_DIR OR NOT DLPACK_INCLUDE_DIR)
|
|
message(FATAL_ERROR "TVM_INCLUDE_DIR, DMLC_INCLUDE_DIR, and DLPACK_INCLUDE_DIR must be set when BUILD_TEST_EINNET is ON")
|
|
endif()
|
|
# TVM and DMLC for invoking TVM packed functions
|
|
include_directories(${TVM_INCLUDE_DIR})
|
|
include_directories(${DMLC_INCLUDE_DIR})
|
|
include_directories(${DLPACK_INCLUDE_DIR})
|
|
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -DDMLC_USE_LOGGING_LIBRARY=\\\<${TVM_INCLUDE_DIR}/tvm/runtime/logging.h\\\> ")
|
|
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -DINFINI_USE_TVM=1") # Enable TVM codegen kernels
|
|
endif()
|
|
|
|
if(BUILD_TEST)
|
|
set(BUILD_GMOCK
|
|
OFF
|
|
CACHE BOOL "Do not build gmock" FORCE)
|
|
set(INSTALL_GTEST
|
|
OFF
|
|
CACHE BOOL "Do not install gtest" FORCE)
|
|
add_subdirectory(3rd-party/googletest)
|
|
include_directories(3rd-party/googletest/googletest/include)
|
|
endif()
|
|
|
|
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -g -Wall -Werror -Wno-error=deprecated-declarations")
|
|
set(CMAKE_CXX_FLAGS_RELEASE "${CMAKE_CXX_FLAGS_RELEASE} -UNDEBUG") # Enable assertion
|
|
set(CMAKE_CXX_FLAGS_RELWITHDEBINFO "${CMAKE_CXX_FLAGS_RELWITHDEBINFO} -UNDEBUG") # Enable assertion
|
|
|
|
|
|
# Source files
|
|
file(GLOB_RECURSE SRC src/ffi/*.cc src/core/*.cc src/kernels/cpu/*.cc src/nnet/*.cc src/operators/*.cc src/utils/*.cc)
|
|
|
|
if(USE_CUDA)
|
|
file(GLOB_RECURSE SRC_CUDA src/cuda/*.cc src/cuda/*.cu src/kernels/cuda/*.cc src/kernels/cuda/*.cu)
|
|
list (APPEND SRC ${SRC_CUDA})
|
|
endif()
|
|
|
|
if(USE_BANG)
|
|
file(GLOB_RECURSE SRC_BANG src/bang/*.cc src/kernels/bang/*.cc )
|
|
list (APPEND SRC ${SRC_BANG})
|
|
endif()
|
|
|
|
if(USE_INTELCPU)
|
|
file(GLOB_RECURSE SRC_INTELCPU src/intelcpu/*.cc src/kernels/intelcpu/*.cc )
|
|
list (APPEND SRC ${SRC_INTELCPU})
|
|
endif()
|
|
|
|
# Libraries
|
|
add_library(InfiniTensor SHARED ${SRC})
|
|
if(USE_PROTOBUF)
|
|
target_link_libraries(InfiniTensor tensor_proto)
|
|
endif()
|
|
|
|
target_link_libraries(InfiniTensor pybind11::embed)
|
|
|
|
# TVM backend
|
|
if(BUILD_TEST_EINNET)
|
|
target_link_libraries(InfiniTensor ${TVM_LIB_DIR}/libtvm.so)
|
|
endif()
|
|
|
|
# Python bindings
|
|
file(GLOB_RECURSE FFIS src/ffi/ffi_infinitensor.cc)
|
|
pybind11_add_module(backend MODULE ${FFIS})
|
|
target_link_libraries(backend PRIVATE InfiniTensor)
|
|
|
|
if(USE_BACKTRACE)
|
|
add_definitions(-D BACKWARD_TRACE)
|
|
add_subdirectory(3rd-party/backward-cpp)
|
|
include_directories(3rd-party/backward-cpp)
|
|
add_backward(InfiniTensor)
|
|
target_link_libraries(InfiniTensor dw)
|
|
endif()
|
|
|
|
if(USE_INTELCPU)
|
|
add_compile_definitions(USE_INTELCPU=1)
|
|
find_package(MKL CONFIG REQUIRED)
|
|
|
|
# Refer to https://www.intel.com/content/www/us/en/developer/tools/oneapi/onemkl-link-line-advisor.html
|
|
target_link_libraries(InfiniTensor sycl OpenCL)
|
|
|
|
set(DNNL_CONFIGURATION "cpu_gomp")
|
|
find_package(dnnl CONFIG REQUIRED)
|
|
if(dnnl_FOUND)
|
|
add_compile_definitions(USE_MKL=1)
|
|
include_directories(BEFORE ${dnnl_DIR}/../../../cpu_gomp/include/)
|
|
link_directories(${dnnl_DIR}/../../../cpu_gomp/lib)
|
|
target_link_libraries(InfiniTensor dnnl)
|
|
else()
|
|
message(FATAL_ERROR "dnnl library not found")
|
|
endif()
|
|
set(WNO_ERRORS "-Wno-error=unused-parameter -Wno-error=unused-function -Wno-error=unused-private-field -Wno-error=ignored-attributes -Wno-error=unused-const-variable -Wno-error=inconsistent-missing-override -Wno-error=unused-variable -Wno-error=tautological-constant-compare")
|
|
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -DMKL_ILP64 -qmkl=parallel -Werror ${WNO_ERRORS}")
|
|
set(CMAKE_CXX_FLAGS_RELEASE "${CMAKE_CXX_FLAGS_RELEASE} -DMKL_ILP64 -qmkl=parallel ${WNO_ERRORS}") # Enable assertion
|
|
set(CMAKE_CXX_FLAGS_RELWITHDEBINFO "${CMAKE_CXX_FLAGS_RELWITHDEBINFO} -DMKL_ILP64 -qmkl=parallel ${WNO_ERRORS}") # Enable assertion
|
|
|
|
find_package(IntelDPCPP REQUIRED)
|
|
endif()
|
|
|
|
if(USE_CUDA)
|
|
add_compile_definitions(USE_CUDA=1)
|
|
# Since enable_language only executes once, rerun cmake is required if CMAKE_CUDA_HOST_COMPILER is wrong
|
|
set(CMAKE_CUDA_HOST_COMPILER
|
|
${CMAKE_CXX_COMPILER}
|
|
CACHE STRING "Set cuda host compiler path")
|
|
# CMP0104 requires CUDA_ARCHITECTURES
|
|
set_target_properties(InfiniTensor PROPERTIES CUDA_ARCHITECTURES "70;80")
|
|
enable_language(CUDA)
|
|
find_package(CUDAToolkit) # For nvrtc and cuda driver
|
|
target_link_libraries(InfiniTensor cudnn CUDA::curand CUDA::cublas CUDA::nvrtc CUDA::cudart CUDA::cuda_driver)
|
|
if (BUILD_DIST)
|
|
message(STATUS "Add BUILD_DIST, use NCCL with CUDA")
|
|
list(APPEND CMAKE_MODULE_PATH ${PROJECT_SOURCE_DIR}/cmake)
|
|
find_package(NCCL REQUIRED)
|
|
add_compile_definitions(INFINI_USE_NCCL=1)
|
|
target_link_libraries(InfiniTensor nccl)
|
|
endif()
|
|
endif()
|
|
|
|
if(USE_BANG)
|
|
add_compile_definitions(USE_BANG=1)
|
|
include_directories(src/kernels/mlu/include)
|
|
################################################################################
|
|
# Neuware Evironment
|
|
################################################################################
|
|
# cnrt cndrv cnnl
|
|
if ((NOT DEFINED NEUWARE_HOME) AND (NOT DEFINED ENV{NEUWARE_HOME}))
|
|
message(FATAL_ERROR "NEUWARE_HOME is not defined from cmake or env")
|
|
elseif (DEFINED NEUWARE_HOME)
|
|
set(NEUWARE_HOME ${NEUWARE_HOME} CACHE STRING "NEUWARE_HOME directory for Cambricon Neuware development")
|
|
else()
|
|
set(NEUWARE_HOME $ENV{NEUWARE_HOME} CACHE STRING "NEUWARE_HOME directory for Cambricon Neuware development")
|
|
endif()
|
|
message(STATUS "NEUWARE_HOME: ${NEUWARE_HOME}")
|
|
|
|
include_directories("${NEUWARE_HOME}/include")
|
|
find_library(CAMBRICON_CNNL libcnnl.so "${NEUWARE_HOME}/lib64")
|
|
find_library(CAMBRICON_CNRT libcnrt.so "${NEUWARE_HOME}/lib64")
|
|
find_library(CAMBRICON_CNDRV libcndrv.so "${NEUWARE_HOME}/lib64")
|
|
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -lstdc++ -Wall -Werror")
|
|
|
|
if ((NOT DEFINED TARGET_CPU_ARCH) AND (NOT DEFINED ENV{TARGET_CPU_ARCH}))
|
|
execute_process(COMMAND uname -m OUTPUT_VARIABLE _uname_m OUTPUT_STRIP_TRAILING_WHITESPACE)
|
|
set(TARGET_CPU_ARCH "${_uname_m}" CACHE STRING "Target CPU ARCH")
|
|
elseif(DEFINED TARGET_CPU_ARCH)
|
|
set(TARGET_CPU_ARCH ${TARGET_CPU_ARCH} CACHE STRING "Target CPU ARCH")
|
|
else()
|
|
set(TARGET_CPU_ARCH $ENV{TARGET_CPU_ARCH} CACHE STRING "Target CPU ARCH")
|
|
endif()
|
|
message(STATUS "TARGET_CPU_ARCH: ${TARGET_CPU_ARCH}")
|
|
|
|
################################################################################
|
|
# BangC Kernels
|
|
################################################################################
|
|
|
|
target_link_libraries(InfiniTensor ${CAMBRICON_CNNL} ${CAMBRICON_CNRT} ${CAMBRICON_CNDRV} stdc++)
|
|
endif()
|
|
|
|
# # Python bindings
|
|
# pybind11_add_module(infini MODULE ${FFI})
|
|
# target_link_libraries(infini PRIVATE infini_cpp)
|
|
|
|
function(build_test files)
|
|
# Non-recursive glob for skip failed tests
|
|
file(GLOB TEST_SOURCES ${files})
|
|
foreach(testsourcefile ${TEST_SOURCES})
|
|
get_filename_component(testname ${testsourcefile} NAME_WE)
|
|
add_executable(${testname} ${testsourcefile})
|
|
target_link_libraries(${testname} InfiniTensor GTest::gtest_main)
|
|
add_test(NAME ${testname} COMMAND ${testname})
|
|
endforeach(testsourcefile ${TEST_SOURCES})
|
|
endfunction()
|
|
|
|
if(BUILD_TEST)
|
|
add_compile_definitions(BUILD_TEST=1)
|
|
enable_testing()
|
|
if(USE_TRACE)
|
|
build_test(test/trace/*.cc)
|
|
endif()
|
|
if(BUILD_TEST_CORE)
|
|
build_test(test/core/*.cc)
|
|
build_test(test/operators/*.cc)
|
|
build_test(test/kernels/nativecpu/*.cc)
|
|
if (USE_CUDA)
|
|
build_test(test/kernels/cuda/*.cc)
|
|
build_test(test/cuda/*.cc)
|
|
endif()
|
|
if (USE_BANG)
|
|
build_test(test/kernels/bang/*.cc)
|
|
endif()
|
|
if (USE_INTELCPU)
|
|
build_test(test/kernels/intelcpu/*.cc)
|
|
endif()
|
|
endif()
|
|
if(BUILD_TEST_PET)
|
|
build_test(test/pet/*.cc)
|
|
endif()
|
|
if(BUILD_TEST_EINNET)
|
|
build_test(test/nnet/test_*.cc)
|
|
|
|
# Build expression reader
|
|
add_executable(nnet_reader test/nnet/readlog.cc)
|
|
target_link_libraries(nnet_reader InfiniTensor)
|
|
endif()
|
|
endif()
|