GitOrigin-RevId: 51af7508de
tags/v1.7.2.m1
| @@ -798,10 +798,15 @@ endif() | |||
| if(MGE_WITH_CAMBRICON) | |||
| include_directories("$ENV{NEUWARE_HOME}/include") | |||
| link_directories("$ENV{NEUWARE_HOME}/lib64") | |||
| include(cmake/cnrt.cmake) | |||
| include(cmake/cndev.cmake) | |||
| include(cmake/cnml.cmake) | |||
| list(APPEND MGE_CAMBRICON_LIBS libcnrt libcndev libcnml) | |||
| if (BANG_ARCH LESS 300) | |||
| include(cmake/cnml.cmake) | |||
| list(APPEND MGE_CAMBRICON_LIBS libcnml) | |||
| else() | |||
| include(cmake/cnnl.cmake) | |||
| include(cmake/cnlight.cmake) | |||
| include(cmake/magicmind.cmake) | |||
| list(APPEND MGE_CAMBRICON_LIBS libcnnl libcnnl_extra libcnlight libmagicmind libmagicmid_runtime) | |||
| endif() | |||
| set(MGE_CAMBRICON_LIBS "${MGE_CAMBRICON_LIBS}") | |||
| endif() | |||
| @@ -0,0 +1,40 @@ | |||
| find_library(CNLIGHT_LIBRARY | |||
| NAMES libcnlight.so | |||
| PATHS ${ALTER_LD_LIBRARY_PATHS} "$ENV{NEUWARE_HOME}/lib64" ${CMAKE_INSTALL_PREFIX} | |||
| HINTS ${ALTER_LIBRARY_PATHS} | |||
| PATH_SUFFIXES lib lib64 | |||
| DOC "CNLIGHT library." ) | |||
| if(CNLIGHT_LIBRARY STREQUAL "CNLIGHT_LIBRARY-NOTFOUND") | |||
| message(FATAL_ERROR "Can not find CNLIGHT Library") | |||
| endif() | |||
| get_filename_component(__found_cnlight_root "${CNLIGHT_LIBRARY}/../include" REALPATH) | |||
| find_path(CNLIGHT_INCLUDE_DIR | |||
| NAMES cnlight.h | |||
| HINTS "$ENV{NEUWARE_HOME}/include" ${__found_cnlight_root} | |||
| PATH_SUFFIXES include | |||
| DOC "Path to CNLIGHT include directory." ) | |||
| if(CNLIGHT_INCLUDE_DIR STREQUAL "CNLIGHT_INCLUDE_DIR-NOTFOUND") | |||
| message(FATAL_ERROR "Can not find CNLIGHT Library") | |||
| endif() | |||
| file(STRINGS "${CNLIGHT_INCLUDE_DIR}/cnlight.h" CNLIGHT_MAJOR REGEX "^#define CNLIGHT_MAJOR_VERSION [0-9]+.*$") | |||
| file(STRINGS "${CNLIGHT_INCLUDE_DIR}/cnlight.h" CNLIGHT_MINOR REGEX "^#define CNLIGHT_MINOR_VERSION [0-9]+.*$") | |||
| file(STRINGS "${CNLIGHT_INCLUDE_DIR}/cnlight.h" CNLIGHT_PATCH REGEX "^#define CNLIGHT_PATCH_VERSION [0-9]+.*$") | |||
| string(REGEX REPLACE "^#define CNLIGHT_MAJOR_VERSION ([0-9]+).*$" "\\1" CNLIGHT_VERSION_MAJOR "${CNLIGHT_MAJOR}") | |||
| string(REGEX REPLACE "^#define CNLIGHT_MINOR_VERSION ([0-9]+).*$" "\\1" CNLIGHT_VERSION_MINOR "${CNLIGHT_MINOR}") | |||
| string(REGEX REPLACE "^#define CNLIGHT_PATCH_VERSION ([0-9]+).*$" "\\1" CNLIGHT_VERSION_PATCH "${CNLIGHT_PATCH}") | |||
| set(CNLIGHT_VERSION_STRING "${CNLIGHT_VERSION_MAJOR}.${CNLIGHT_VERSION_MINOR}.${CNLIGHT_VERSION_PATCH}") | |||
| add_library(libcnlight SHARED IMPORTED) | |||
| set_target_properties(libcnlight PROPERTIES | |||
| IMPORTED_LOCATION ${CNLIGHT_LIBRARY} | |||
| INTERFACE_INCLUDE_DIRECTORIES ${CNLIGHT_INCLUDE_DIR} | |||
| ) | |||
| message(STATUS "Found CNLIGHT: ${__found_cnlight_root} (found version: ${CNLIGHT_VERSION_STRING})") | |||
| @@ -0,0 +1,80 @@ | |||
| find_library(CNNL_LIBRARY | |||
| NAMES libcnnl.so | |||
| PATHS ${ALTER_LD_LIBRARY_PATHS} "$ENV{NEUWARE_HOME}/lib64" ${CMAKE_INSTALL_PREFIX} | |||
| HINTS ${ALTER_LIBRARY_PATHS} | |||
| PATH_SUFFIXES lib lib64 | |||
| DOC "CNNL library." ) | |||
| if(CNNL_LIBRARY STREQUAL "CNNL_LIBRARY-NOTFOUND") | |||
| message(FATAL_ERROR "Can not find CNNL Library") | |||
| endif() | |||
| get_filename_component(__found_cnnl_root "${CNNL_LIBRARY}/../include" REALPATH) | |||
| find_path(CNNL_INCLUDE_DIR | |||
| NAMES cnnl.h | |||
| HINTS "$ENV{NEUWARE_HOME}/include" ${__found_cnnl_root} | |||
| PATH_SUFFIXES include | |||
| DOC "Path to CNNL include directory." ) | |||
| if(CNNL_INCLUDE_DIR STREQUAL "CNNL_INCLUDE_DIR-NOTFOUND") | |||
| message(FATAL_ERROR "Can not find CNNL Library") | |||
| endif() | |||
| file(STRINGS "${CNNL_INCLUDE_DIR}/cnnl.h" CNNL_MAJOR REGEX "^#define CNNL_MAJOR [0-9]+.*$") | |||
| file(STRINGS "${CNNL_INCLUDE_DIR}/cnnl.h" CNNL_MINOR REGEX "^#define CNNL_MINOR [0-9]+.*$") | |||
| file(STRINGS "${CNNL_INCLUDE_DIR}/cnnl.h" CNNL_PATCH REGEX "^#define CNNL_PATCHLEVEL [0-9]+.*$") | |||
| string(REGEX REPLACE "^#define CNNL_MAJOR ([0-9]+).*$" "\\1" CNNL_VERSION_MAJOR "${CNNL_MAJOR}") | |||
| string(REGEX REPLACE "^#define CNNL_MINOR ([0-9]+).*$" "\\1" CNNL_VERSION_MINOR "${CNNL_MINOR}") | |||
| string(REGEX REPLACE "^#define CNNL_PATCHLEVEL ([0-9]+).*$" "\\1" CNNL_VERSION_PATCH "${CNNL_PATCH}") | |||
| set(CNNL_VERSION_STRING "${CNNL_VERSION_MAJOR}.${CNNL_VERSION_MINOR}.${CNNL_VERSION_PATCH}") | |||
| add_library(libcnnl SHARED IMPORTED) | |||
| set_target_properties(libcnnl PROPERTIES | |||
| IMPORTED_LOCATION ${CNNL_LIBRARY} | |||
| INTERFACE_INCLUDE_DIRECTORIES ${CNNL_INCLUDE_DIR} | |||
| ) | |||
| message(STATUS "Found CNNL: ${__found_cnnl_root} (found version: ${CNNL_VERSION_STRING})") | |||
| find_library(CNNL_EXTRA_LIBRARY | |||
| NAMES libcnnl_extra.so | |||
| PATHS ${ALTER_LD_LIBRARY_PATHS} "$ENV{NEUWARE_HOME}/lib64" ${CMAKE_INSTALL_PREFIX} | |||
| HINTS ${ALTER_LIBRARY_PATHS} | |||
| PATH_SUFFIXES lib lib64 | |||
| DOC "CNNL_EXTRA library." ) | |||
| if(CNNL_EXTRA_LIBRARY STREQUAL "CNNL_EXTRA_LIBRARY-NOTFOUND") | |||
| message(FATAL_ERROR "Can not find CNNL_EXTRA Library") | |||
| endif() | |||
| get_filename_component(__found_cnnl_extra_root "${CNNL_EXTRA_LIBRARY}/../include" REALPATH) | |||
| find_path(CNNL_EXTRA_INCLUDE_DIR | |||
| NAMES cnnl_extra.h | |||
| HINTS "$ENV{NEUWARE_HOME}/include" ${__found_cnnl_extra_root} | |||
| PATH_SUFFIXES include | |||
| DOC "Path to CNNL_EXTRA include directory." ) | |||
| if(CNNL_EXTRA_INCLUDE_DIR STREQUAL "CNNL_EXTRA_INCLUDE_DIR-NOTFOUND") | |||
| message(FATAL_ERROR "Can not find CNNL_EXTRA Library") | |||
| endif() | |||
| file(STRINGS "${CNNL_EXTRA_INCLUDE_DIR}/cnnl_extra.h" CNNL_EXTRA_MAJOR REGEX "^#define CNNL_EXTRA_MAJOR [0-9]+.*$") | |||
| file(STRINGS "${CNNL_EXTRA_INCLUDE_DIR}/cnnl_extra.h" CNNL_EXTRA_MINOR REGEX "^#define CNNL_EXTRA_MINOR [0-9]+.*$") | |||
| file(STRINGS "${CNNL_EXTRA_INCLUDE_DIR}/cnnl_extra.h" CNNL_EXTRA_PATCH REGEX "^#define CNNL_EXTRA_PATCHLEVEL [0-9]+.*$") | |||
| string(REGEX REPLACE "^#define CNNL_EXTRA_MAJOR ([0-9]+).*$" "\\1" CNNL_EXTRA_VERSION_MAJOR "${CNNL_EXTRA_MAJOR}") | |||
| string(REGEX REPLACE "^#define CNNL_EXTRA_MINOR ([0-9]+).*$" "\\1" CNNL_EXTRA_VERSION_MINOR "${CNNL_EXTRA_MINOR}") | |||
| string(REGEX REPLACE "^#define CNNL_EXTRA_PATCHLEVEL ([0-9]+).*$" "\\1" CNNL_EXTRA_VERSION_PATCH "${CNNL_EXTRA_PATCH}") | |||
| set(CNNL_EXTRA_VERSION_STRING "${CNNL_EXTRA_VERSION_MAJOR}.${CNNL_EXTRA_VERSION_MINOR}.${CNNL_EXTRA_VERSION_PATCH}") | |||
| add_library(libcnnl_extra SHARED IMPORTED) | |||
| set_target_properties(libcnnl_extra PROPERTIES | |||
| IMPORTED_LOCATION ${CNNL_EXTRA_LIBRARY} | |||
| INTERFACE_INCLUDE_DIRECTORIES ${CNNL_EXTRA_INCLUDE_DIR} | |||
| ) | |||
| message(STATUS "Found CNNL_EXTRA: ${__found_cnnl_extra_root} (found version: ${CNNL_EXTRA_VERSION_STRING})") | |||
| @@ -0,0 +1,54 @@ | |||
| find_library(MAGICMIND_LIBRARY | |||
| NAMES libmagicmind.so | |||
| PATHS ${ALTER_LD_LIBRARY_PATHS} "$ENV{NEUWARE_HOME}/lib64" ${CMAKE_INSTALL_PREFIX} | |||
| HINTS ${ALTER_LIBRARY_PATHS} | |||
| PATH_SUFFIXES lib lib64 | |||
| DOC "MAGICMIND library." ) | |||
| if(MAGICMIND_LIBRARY STREQUAL "MAGICMIND_LIBRARY-NOTFOUND") | |||
| message(FATAL_ERROR "Can not find MAGICMIND Library") | |||
| endif() | |||
| get_filename_component(__found_magicmind_root "${MAGICMIND_LIBRARY}/../include" REALPATH) | |||
| find_path(MAGICMIND_INCLUDE_DIR | |||
| NAMES common.h | |||
| HINTS "$ENV{NEUWARE_HOME}/include" ${__found_magicmind_root} | |||
| PATH_SUFFIXES include | |||
| DOC "Path to MAGICMIND include directory." ) | |||
| if(MAGICMIND_INCLUDE_DIR STREQUAL "MAGICMIND_INCLUDE_DIR-NOTFOUND") | |||
| message(FATAL_ERROR "Can not find MAGICMIND Library") | |||
| endif() | |||
| file(STRINGS "${MAGICMIND_INCLUDE_DIR}/common.h" MAGICMIND_MAJOR REGEX "^#define MM_MAJOR_VERSION [0-9]+.*$") | |||
| file(STRINGS "${MAGICMIND_INCLUDE_DIR}/common.h" MAGICMIND_MINOR REGEX "^#define MM_MINOR_VERSION [0-9]+.*$") | |||
| file(STRINGS "${MAGICMIND_INCLUDE_DIR}/common.h" MAGICMIND_PATCH REGEX "^#define MM_PATCH_VERSION [0-9]+.*$") | |||
| string(REGEX REPLACE "^#define MM_MAJOR_VERSION ([0-9]+).*$" "\\1" MAGICMIND_VERSION_MAJOR "${MAGICMIND_MAJOR}") | |||
| string(REGEX REPLACE "^#define MM_MINOR_VERSION ([0-9]+).*$" "\\1" MAGICMIND_VERSION_MINOR "${MAGICMIND_MINOR}") | |||
| string(REGEX REPLACE "^#define MM_PATCH_VERSION ([0-9]+).*$" "\\1" MAGICMIND_VERSION_PATCH "${MAGICMIND_PATCH}") | |||
| set(MAGICMIND_VERSION_STRING "${MAGICMIND_VERSION_MAJOR}.${MAGICMIND_VERSION_MINOR}.${MAGICMIND_VERSION_PATCH}") | |||
| add_library(libmagicmind SHARED IMPORTED) | |||
| set_target_properties(libmagicmind PROPERTIES | |||
| IMPORTED_LOCATION ${MAGICMIND_LIBRARY} | |||
| INTERFACE_INCLUDE_DIRECTORIES ${MAGICMIND_INCLUDE_DIR} | |||
| ) | |||
| message(STATUS "Found MAGICMIND: ${__found_magicmind_root} (found version: ${MAGICMIND_VERSION_STRING})") | |||
| find_library(MAGICMIND_RUNTIME_LIBRARY | |||
| NAMES libmagicmind_runtime.so | |||
| PATHS "${__found_magicmind_root}/../lib64" | |||
| ) | |||
| if(MAGICMIND_RUNTIME_LIBRARY STREQUAL "MAGICMIND_RUNTIME_LIBRARY-NOTFOUND") | |||
| message(FATAL_ERROR "Can not find MAGICMIND_RUNTIME Library") | |||
| else() | |||
| message(STATUS "Found MAGICMIND_RUNTIME: ${MAGICMIND_RUNTIME_LIBRARY}") | |||
| endif() | |||
| add_library(libmagicmind_runtime SHARED IMPORTED) | |||
| set_target_properties(libmagicmind_runtime PROPERTIES | |||
| IMPORTED_LOCATION ${MAGICMIND_RUNTIME_LIBRARY} | |||
| ) | |||
| @@ -0,0 +1,397 @@ | |||
| /** | |||
| * \file src/cambricon/impl/magicmind_runtime_opr.cpp | |||
| * MegEngine is Licensed under the Apache License, Version 2.0 (the "License") | |||
| * | |||
| * Copyright (c) 2014-2021 Megvii Inc. All rights reserved. | |||
| * | |||
| * Unless required by applicable law or agreed to in writing, | |||
| * software distributed under the License is distributed on an | |||
| * "AS IS" BASIS, WITHOUT ARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
| */ | |||
| #include "megbrain/cambricon/magicmind_runtime_opr.h" | |||
| #include <sstream> | |||
| #include "megbrain/common.h" | |||
| #include "megbrain/comp_node_env.h" | |||
| #if MGB_CAMBRICON | |||
| using namespace mgb; | |||
| using namespace opr; | |||
| using namespace magicmind; | |||
| #define MM_CHECK(stmt) \ | |||
| do { \ | |||
| auto ret = (stmt); \ | |||
| if (ret != magicmind::Status::OK()) { \ | |||
| std::ostringstream msg; \ | |||
| msg << ret; \ | |||
| mgb_throw(MegBrainError, "mm failure(extra msg:%s)", msg.str().c_str()); \ | |||
| } \ | |||
| } while (0) | |||
| namespace { | |||
| Dims mgb_shape_to_mm_dims(TensorShape mgb_shp) { | |||
| size_t ndim = mgb_shp.ndim; | |||
| std::vector<int64_t> dimensions(ndim); | |||
| for (size_t i = 0; i < ndim; ++i) { | |||
| dimensions[i] = mgb_shp[i]; | |||
| } | |||
| return Dims{dimensions}; | |||
| } | |||
| TensorShape mm_dims_to_mgb_shape(const Dims& dims) { | |||
| TensorShape ret; | |||
| ret.ndim = dims.GetDimsNum(); | |||
| auto&& dimensions = dims.GetDims(); | |||
| for (size_t i = 0; i < ret.ndim; ++i) { | |||
| ret[i] = dimensions[i]; | |||
| } | |||
| return ret; | |||
| } | |||
| DType mm_dtype_to_mgb_dtype(DataType data_type) { | |||
| switch (data_type) { | |||
| case DataType::FLOAT16: | |||
| #if !MEGDNN_DISABLE_FLOAT16 | |||
| return dtype::Float16(); | |||
| #else | |||
| mgb_throw(MegBrainError, "Float16 support is disabled at compile time."); | |||
| #endif | |||
| case DataType::FLOAT32: | |||
| return dtype::Float32(); | |||
| case DataType::INT8: | |||
| return dtype::QuantizedS8(1.f); | |||
| case DataType::INT16: | |||
| return dtype::Int16(); | |||
| case DataType::INT32: | |||
| return dtype::Int32(); | |||
| case DataType::UINT8: | |||
| return dtype::Uint8(); | |||
| //! TODO: check scale | |||
| case DataType::QINT8: | |||
| return dtype::QuantizedS8(1.f); | |||
| case DataType::INT4: | |||
| return dtype::QuantizedS4(1.f); | |||
| case DataType::UINT4: | |||
| return dtype::Quantized4Asymm(1.f, static_cast<uint8_t>(8)); | |||
| default: | |||
| mgb_throw( | |||
| MegBrainError, "DataType %u is not supported by MegEngine.", | |||
| static_cast<uint32_t>(data_type)); | |||
| } | |||
| } | |||
| DataType mgb_dtype_to_mm_dtype(DType data_type) { | |||
| switch (data_type.enumv()) { | |||
| #if !MEGDNN_DISABLE_FLOAT16 | |||
| case DTypeEnum::Float16: | |||
| return DataType::FLOAT16; | |||
| #endif | |||
| case DTypeEnum::Float32: | |||
| return DataType::FLOAT32; | |||
| case DTypeEnum::QuantizedS8: | |||
| return DataType::QINT8; | |||
| case DTypeEnum::Int8: | |||
| return DataType::INT8; | |||
| case DTypeEnum::Int32: | |||
| return DataType::INT32; | |||
| case DTypeEnum::Uint8: | |||
| return DataType::UINT8; | |||
| case DTypeEnum::QuantizedS4: | |||
| return DataType::INT4; | |||
| case DTypeEnum::Quantized4Asymm: | |||
| return DataType::UINT4; | |||
| default: | |||
| mgb_throw( | |||
| MegBrainError, | |||
| "megengine data type %s is not supported by magicmind.", | |||
| data_type.name()); | |||
| } | |||
| } | |||
| }; // namespace | |||
| /* =========== MagicMindRuntimeOpr::CambriconAllocator =========== */ | |||
| class MagicMindRuntimeOpr::CambriconAllocator final : public IAllocator { | |||
| CompNode m_cn; | |||
| std::mutex m_ptr2size_mtx; | |||
| ThinHashMap<void*, size_t> m_ptr2size; | |||
| public: | |||
| explicit CambriconAllocator(CompNode cn); | |||
| ~CambriconAllocator() noexcept; | |||
| void* AllocateRaw(size_t size, size_t alignment) override; | |||
| void DeallocateRaw(void* ptr) override; | |||
| CompNode comp_node() const { return m_cn; } | |||
| }; | |||
| MagicMindRuntimeOpr::CambriconAllocator::CambriconAllocator(CompNode cn) : m_cn{cn} { | |||
| mgb_assert( | |||
| cn.device_type() == CompNode::DeviceType::CAMBRICON, | |||
| "invalid comp node %s for CambriconAllocator", cn.to_string().c_str()); | |||
| } | |||
| MagicMindRuntimeOpr::CambriconAllocator::~CambriconAllocator() noexcept { | |||
| MGB_LOCK_GUARD(m_ptr2size_mtx); | |||
| if (!m_ptr2size.empty()) { | |||
| std::string msg{"there are unreleased magicmind mem buffers:\n"}; | |||
| for (auto&& i : m_ptr2size) { | |||
| msg.append(ssprintf(" %p: %zu\n", i.first, i.second)); | |||
| } | |||
| mgb_log_error("%sabort now", msg.c_str()); | |||
| mgb_trap(); | |||
| } | |||
| } | |||
| void* MagicMindRuntimeOpr::CambriconAllocator::AllocateRaw( | |||
| size_t size, size_t alignment) { | |||
| static bool enable_log = getenv("MGE_LOG_MAGICMIND_MEM_ALLOC"); | |||
| mgb_assert(!(alignment & (alignment - 1)), "invalid alignment(%zu)", alignment); | |||
| auto ret = m_cn.alloc_device(size); | |||
| mgb_assert( | |||
| !(reinterpret_cast<uintptr_t>(ret) & (alignment - 1)), | |||
| "alignment not required(ptr:%p,alignment:%zu)", ret, alignment); | |||
| if (enable_log) { | |||
| mgb_log("magicmind mem alloc on %s: size=%zu, align=%zu, ptr=%p", | |||
| m_cn.to_string().c_str(), size, alignment, ret); | |||
| } | |||
| { | |||
| MGB_LOCK_GUARD(m_ptr2size_mtx); | |||
| m_ptr2size[ret] = size; | |||
| } | |||
| return ret; | |||
| } | |||
| void MagicMindRuntimeOpr::CambriconAllocator::DeallocateRaw(void* ptr) { | |||
| { | |||
| auto iter = m_ptr2size.find(ptr); | |||
| mgb_assert(iter != m_ptr2size.end(), "ptr %p not found", ptr); | |||
| m_ptr2size.erase(iter); | |||
| } | |||
| m_cn.free_device(ptr); | |||
| } | |||
| /* ====================== MagicMindRuntimeOpr ==================== */ | |||
| MGB_DYN_TYPE_OBJ_FINAL_IMPL(MagicMindRuntimeOpr); | |||
| MagicMindRuntimeOpr::MagicMindRuntimeOpr( | |||
| IModelPtr model, CambriconAllocatorPtr allocator, const VarNodeArray& inputs, | |||
| const OperatorNodeConfig& config) | |||
| : Super(inputs[0]->owner_graph(), config, "magic_runtime", inputs), | |||
| m_allocator{std::move(allocator)}, | |||
| m_context{nullptr}, | |||
| m_engine{nullptr}, | |||
| m_model{std::move(model)} { | |||
| mgb_assert( | |||
| inputs[0]->comp_node().device_type() == CompNode::DeviceType::CAMBRICON, | |||
| "MagicMindRuntimeOpr can only be used on cambricon comp node; " | |||
| "got %s", | |||
| inputs[0]->comp_node().to_string().c_str()); | |||
| size_t nr_inputs = m_model->GetInputNum(); | |||
| mgb_assert( | |||
| nr_inputs == inputs.size(), "input number mismatch(got:%zu,expected:%zu)", | |||
| inputs.size(), nr_inputs); | |||
| for (auto i : inputs) { | |||
| add_input({i}); | |||
| } | |||
| size_t nr_outputs = m_model->GetOutputNum(); | |||
| for (size_t i = 0; i < nr_outputs; ++i) { | |||
| add_output(m_model->GetOutputName(i)); | |||
| } | |||
| IModel::EngineConfig engine_config; | |||
| engine_config.device_type = "MLU"; | |||
| engine_config.allocator = m_allocator.get(); | |||
| auto&& cnrt_env = CompNodeEnv::from_comp_node(m_allocator->comp_node()).cnrt_env(); | |||
| cnrt_env.activate(); | |||
| m_engine = { | |||
| m_model->CreateIEngine(engine_config), | |||
| magicmind_intl::MagicMindDeleter<IEngine>()}; | |||
| mgb_assert( | |||
| m_engine != nullptr, | |||
| "create IEngine failed, corresponding MagicMindRuntimeOpr(%s)", cname()); | |||
| cg::add_workspace_output(this); | |||
| add_equivalence_component<mgb::ScalarHash<void*>>(m_model.get()); | |||
| }; | |||
| void MagicMindRuntimeOpr::scn_do_execute() { | |||
| mgb_assert(m_engine != nullptr); | |||
| mgb_assert(m_context != nullptr); | |||
| auto&& cnrt_env = CompNodeEnv::from_comp_node(input(0)->comp_node()).cnrt_env(); | |||
| cnrt_env.activate(); | |||
| std::vector<IRTTensor*> inputs, outputs; | |||
| MM_CHECK(CreateInputTensors(m_context.get(), &inputs)); | |||
| MM_CHECK(CreateInputTensors(m_context.get(), &outputs)); | |||
| size_t nr_inputs = input().size(); | |||
| mgb_assert(nr_inputs == inputs.size()); | |||
| for (size_t i = 0; i < nr_inputs; ++i) { | |||
| auto&& iname = m_model->GetInputName(i); | |||
| auto tensor = FindIRTTensorByName(inputs, iname); | |||
| mgb_assert( | |||
| tensor != nullptr, "failed to find input tensor(name:%s)", | |||
| iname.c_str()); | |||
| MM_CHECK(tensor->SetDimensions(mgb_shape_to_mm_dims(input(i)->shape()))); | |||
| MM_CHECK(tensor->SetData(input(i)->dev_tensor().raw_ptr())); | |||
| } | |||
| size_t nr_outputs = output().size(); | |||
| mgb_assert(nr_outputs == outputs.size() + 1); | |||
| for (size_t i = 0; i < nr_outputs - 1; ++i) { | |||
| auto&& oname = m_model->GetOutputName(i); | |||
| auto tensor = FindIRTTensorByName(outputs, oname); | |||
| mgb_assert( | |||
| tensor != nullptr, "failed to find output tensor(name:%s)", | |||
| oname.c_str()); | |||
| MM_CHECK(tensor->SetDimensions(mgb_shape_to_mm_dims(output(i)->shape()))); | |||
| MM_CHECK(tensor->SetData(output(i)->dev_tensor().raw_ptr())); | |||
| } | |||
| auto size = output().back()->dev_tensor().layout().span().dist_byte(); | |||
| MM_CHECK(m_context->SetWorkspace(output().back()->dev_tensor().raw_ptr(), size)); | |||
| MM_CHECK(m_context->Enqueue(inputs, outputs, cnrt_env.queue)); | |||
| for (auto&& i : inputs) { | |||
| i->SetData(nullptr); | |||
| i->Destroy(); | |||
| } | |||
| for (auto&& o : outputs) { | |||
| o->SetData(nullptr); | |||
| o->Destroy(); | |||
| } | |||
| } | |||
| void MagicMindRuntimeOpr::get_output_var_shape( | |||
| const TensorShapeArray& inp_shape, TensorShapeArray& out_shape) const { | |||
| mgb_assert(m_engine != nullptr); | |||
| mgb_assert(input().size() == inp_shape.size()); | |||
| auto&& cnrt_env = CompNodeEnv::from_comp_node(input(0)->comp_node()).cnrt_env(); | |||
| cnrt_env.activate(); | |||
| if (m_context == nullptr) { | |||
| m_context = { | |||
| m_engine->CreateIContext(), | |||
| magicmind_intl::MagicMindDeleter<IContext>()}; | |||
| mgb_assert( | |||
| m_context != nullptr, | |||
| "failed to create IContext, corresponding MagicMindRuntimeOpr(%s)", | |||
| cname()); | |||
| } | |||
| std::vector<IRTTensor*> inputs, outputs; | |||
| MM_CHECK(CreateInputTensors(m_context.get(), &inputs)); | |||
| MM_CHECK(CreateInputTensors(m_context.get(), &outputs)); | |||
| size_t nr_inputs = input().size(); | |||
| mgb_assert(nr_inputs == inputs.size()); | |||
| for (size_t i = 0; i < nr_inputs; ++i) { | |||
| auto&& iname = m_model->GetInputName(i); | |||
| auto tensor = FindIRTTensorByName(inputs, iname); | |||
| mgb_assert( | |||
| tensor != nullptr, "failed to find input tensor(name:%s)", | |||
| iname.c_str()); | |||
| MM_CHECK(tensor->SetDimensions(mgb_shape_to_mm_dims(input(i)->shape()))); | |||
| } | |||
| if (Status::OK() == m_context->InferOutputShape(inputs, outputs)) { | |||
| size_t nr_outputs = output().size(); | |||
| mgb_assert(nr_outputs == outputs.size() + 1); | |||
| for (size_t i = 0; i < nr_outputs - 1; ++i) { | |||
| auto&& oname = m_model->GetOutputName(i); | |||
| auto tensor = FindIRTTensorByName(outputs, oname); | |||
| mgb_assert( | |||
| tensor != nullptr, "failed to find output tensor(name:%s)", | |||
| oname.c_str()); | |||
| auto&& dims = tensor->GetDimensions(); | |||
| out_shape[i] = mm_dims_to_mgb_shape(dims); | |||
| } | |||
| std::vector<Dims> shape(inp_shape.size()); | |||
| for (size_t i = 0; i < nr_inputs; ++i) { | |||
| shape[i] = mgb_shape_to_mm_dims(input(i)->shape()); | |||
| } | |||
| size_t wk_size = 0; | |||
| MM_CHECK(m_engine->QueryContextMaxWorkspaceSize(shape, &wk_size)); | |||
| out_shape.back() = {wk_size}; | |||
| } else { | |||
| mgb_assert( | |||
| false, "static shape infer for MagicMindRuntimeOpr(%s) failed", | |||
| cname()); | |||
| } | |||
| for (auto&& i : inputs) { | |||
| i->SetData(nullptr); | |||
| i->Destroy(); | |||
| } | |||
| for (auto&& o : outputs) { | |||
| o->SetData(nullptr); | |||
| o->Destroy(); | |||
| } | |||
| } | |||
| void MagicMindRuntimeOpr::add_input_layout_constraint() { | |||
| //! default contiguous | |||
| for (auto i : input()) { | |||
| i->add_layout_constraint_contiguous(); | |||
| } | |||
| } | |||
| void MagicMindRuntimeOpr::init_output_dtype() { | |||
| std::vector<DataType> inp_dtypes = m_model->GetInputDataTypes(); | |||
| mgb_assert( | |||
| inp_dtypes.size() == input().size(), | |||
| "input size mismatch(got:%zu,expected:%zu)", inp_dtypes.size(), | |||
| input().size()); | |||
| size_t nr_inputs = input().size(); | |||
| for (size_t i = 0; i < nr_inputs; ++i) { | |||
| auto dt_mm = mm_dtype_to_mgb_dtype(inp_dtypes[i]); | |||
| auto dt_inp = input(i)->dtype(); | |||
| MGB_MARK_USED_VAR(dt_mm); | |||
| MGB_MARK_USED_VAR(dt_inp); | |||
| mgb_assert( | |||
| dt_mm.valid() && dt_inp.valid() && dt_mm.enumv() == dt_inp.enumv(), | |||
| "input %zu's data type mismatch with that in " | |||
| "IModel: expected %s, got %s", | |||
| i, dt_mm.name(), dt_inp.name()); | |||
| } | |||
| std::vector<DataType> out_dtypes = m_model->GetOutputDataTypes(); | |||
| mgb_assert( | |||
| out_dtypes.size() == output().size(), | |||
| "output size mismatch(got:%zu,expected:%zu)", out_dtypes.size(), | |||
| output().size()); | |||
| size_t nr_outputs = output().size(); | |||
| for (size_t i = 0; i < nr_outputs; ++i) { | |||
| auto dt_mm = mm_dtype_to_mgb_dtype(out_dtypes[i]); | |||
| mgb_assert( | |||
| dt_mm.valid(), "output dtype checking failed: invalid dtype returned."); | |||
| if (dt_mm.enumv() == DTypeEnum::QuantizedS8) { | |||
| mgb_assert( | |||
| output(i)->dtype().valid(), | |||
| "user should specify scale of output tensor of " | |||
| "MagicMindRuntimeOpr."); | |||
| } | |||
| if (!output(i)->dtype().valid()) | |||
| output(i)->dtype(dt_mm); | |||
| } | |||
| } | |||
| SymbolVarArray MagicMindRuntimeOpr::make( | |||
| IModelPtr model, CambriconAllocatorPtr allocator, const SymbolVarArray& src, | |||
| const OperatorNodeConfig& config) { | |||
| VarNodeArray var_node_array = cg::to_var_node_array(src); | |||
| auto magicmind_runtime_opr = std::make_unique<MagicMindRuntimeOpr>( | |||
| std::move(model), std::move(allocator), var_node_array, config); | |||
| auto ret = cg::to_symbol_var_array( | |||
| src[0].node() | |||
| ->owner_graph() | |||
| ->insert_opr(std::move(magicmind_runtime_opr)) | |||
| ->output()); | |||
| ret.pop_back(); // remove workspace | |||
| return ret; | |||
| } | |||
| SymbolVarArray MagicMindRuntimeOpr::make( | |||
| const void* buf, size_t size, const SymbolVarArray& src, | |||
| const OperatorNodeConfig& config) { | |||
| mgb_throw_if( | |||
| !CompNode::get_device_count(CompNode::DeviceType::CAMBRICON), SystemError, | |||
| "can not create MagicMindRuntimeOpr when MagicMind is not " | |||
| "available"); | |||
| auto cambricon_allocator = | |||
| std::make_shared<CambriconAllocator>(src[0].node()->comp_node()); | |||
| IModelPtr model = make_model_ptr(CreateIModel()); | |||
| model->DeserializeFromMemory(const_cast<void*>(buf), size); | |||
| return make(std::move(model), std::move(cambricon_allocator), src, config); | |||
| } | |||
| #undef MM_CHECK | |||
| #endif // MGB_CAMBRICON | |||
| // vim: syntax=cpp.doxygen foldmethod=marker foldmarker=f{{{,f}}} | |||
| @@ -0,0 +1,77 @@ | |||
| /** | |||
| * \file src/cambricon/impl/magicmind_runtime_opr.sereg.h | |||
| * MegEngine is Licensed under the Apache License, Version 2.0 (the "License") | |||
| * | |||
| * Copyright (c) 2014-2021 Megvii Inc. All rights reserved. | |||
| * | |||
| * Unless required by applicable law or agreed to in writing, | |||
| * software distributed under the License is distributed on an | |||
| * "AS IS" BASIS, WITHOUT ARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
| */ | |||
| #include <sstream> | |||
| #include "megbrain/cambricon/magicmind_runtime_opr.h" | |||
| #include "megbrain/serialization/sereg.h" | |||
| namespace mgb { | |||
| namespace serialization { | |||
| #define MM_CHECK(stmt) \ | |||
| do { \ | |||
| auto ret = (stmt); \ | |||
| if (ret != magicmind::Status::OK()) { \ | |||
| std::ostringstream msg; \ | |||
| msg << ret; \ | |||
| mgb_throw(MegBrainError, "mm failure(extra msg:%s)", msg.str().c_str()); \ | |||
| } \ | |||
| } while (0) | |||
| template <> | |||
| struct OprLoadDumpImpl<opr::MagicMindRuntimeOpr, 0> { | |||
| static void dump(OprDumpContext& ctx, const cg::OperatorNodeBase& opr_) { | |||
| auto&& opr = opr_.cast_final_safe<opr::MagicMindRuntimeOpr>(); | |||
| auto&& model = opr.inference_model(); | |||
| size_t size = 0; | |||
| MM_CHECK(model->GetSerializedModelSize(&size)); | |||
| std::string buf; | |||
| buf.resize(size); | |||
| MM_CHECK(model->SerializeToMemory( | |||
| reinterpret_cast<void*>(buf.data()), buf.size())); | |||
| ctx.dump_buf_with_len(buf.data(), buf.size()); | |||
| } | |||
| static cg::OperatorNodeBase* load( | |||
| OprLoadContext& ctx, const cg::VarNodeArray& inputs, | |||
| const OperatorNodeConfig& config) { | |||
| auto buf = ctx.load_shared_buf_with_len(); | |||
| return opr::MagicMindRuntimeOpr::make( | |||
| reinterpret_cast<const void*>(buf.data()), buf.size(), | |||
| cg::to_symbol_var_array(inputs), config) | |||
| .at(0) | |||
| .node() | |||
| ->owner_opr(); | |||
| } | |||
| }; | |||
| } // namespace serialization | |||
| namespace opr { | |||
| cg::OperatorNodeBase* opr_shallow_copy_magicmind_runtime_opr( | |||
| const serialization::OprShallowCopyContext& ctx, | |||
| const cg::OperatorNodeBase& opr_, const VarNodeArray& inputs, | |||
| const OperatorNodeConfig& config) { | |||
| auto&& opr = opr_.cast_final_safe<MagicMindRuntimeOpr>(); | |||
| return MagicMindRuntimeOpr::make( | |||
| opr.inference_model(), opr.cambricon_allocator(), | |||
| cg::to_symbol_var_array(inputs), config) | |||
| .at(0) | |||
| .node() | |||
| ->owner_opr(); | |||
| } | |||
| MGB_SEREG_OPR(MagicMindRuntimeOpr, 0); | |||
| MGB_REG_OPR_SHALLOW_COPY(MagicMindRuntimeOpr, opr_shallow_copy_magicmind_runtime_opr); | |||
| #undef MM_CHECK | |||
| } // namespace opr | |||
| } // namespace mgb | |||
| // vim: syntax=cpp.doxygen foldmethod=marker foldmarker=f{{{,f}}} | |||
| @@ -0,0 +1,92 @@ | |||
| /** | |||
| * \file src/cambricon/include/megbrain/cambricon/magicmind_runtime_opr.h | |||
| * MegEngine is Licensed under the Apache License, Version 2.0 (the "License") | |||
| * | |||
| * Copyright (c) 2014-2021 Megvii Inc. All rights reserved. | |||
| * | |||
| * Unless required by applicable law or agreed to in writing, | |||
| * software distributed under the License is distributed on an | |||
| * "AS IS" BASIS, WITHOUT ARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
| */ | |||
| #pragma once | |||
| #include "megbrain/graph.h" | |||
| #include "megbrain/serialization/file.h" | |||
| #if MGB_CAMBRICON | |||
| #include "interface_runtime.h" | |||
| namespace mgb { | |||
| namespace opr { | |||
| namespace magicmind_intl { | |||
| template <typename T> | |||
| struct MagicMindDeleter { | |||
| void operator()(T* p) { | |||
| if (p != nullptr) | |||
| p->Destroy(); | |||
| } | |||
| }; | |||
| template <typename T> | |||
| using MagicMindUniquePtr = std::unique_ptr<T, MagicMindDeleter<T>>; | |||
| } // namespace magicmind_intl | |||
| MGB_DEFINE_OPR_CLASS( | |||
| MagicMindRuntimeOpr, cg::SingleCNOutshapePureByInshapeOprBase) // { | |||
| void scn_do_execute() override; | |||
| void get_output_var_shape( | |||
| const TensorShapeArray& inp_shape, | |||
| TensorShapeArray& out_shape) const override; | |||
| void add_input_layout_constraint() override; | |||
| void init_output_dtype() override; | |||
| public: | |||
| template <typename T> | |||
| using MagicMindUniquePtr = magicmind_intl::MagicMindUniquePtr<T>; | |||
| //! Due to the requirement of shallow copy, the IModel should be shared among | |||
| //! instances of magicmind operators. | |||
| using IModelPtr = std::shared_ptr<magicmind::IModel>; | |||
| using IContextPtr = MagicMindUniquePtr<magicmind::IContext>; | |||
| using IEnginePtr = MagicMindUniquePtr<magicmind::IEngine>; | |||
| class CambriconAllocator; | |||
| using CambriconAllocatorPtr = std::shared_ptr<CambriconAllocator>; | |||
| MagicMindRuntimeOpr( | |||
| IModelPtr model, CambriconAllocatorPtr allocator, | |||
| const VarNodeArray& inputs, const OperatorNodeConfig& config); | |||
| //! get underlying inference model | |||
| const IModelPtr& inference_model() const { return m_model; } | |||
| //! get underlying cambricon allocator | |||
| const CambriconAllocatorPtr& cambricon_allocator() const { return m_allocator; } | |||
| //! | |||
| static SymbolVarArray make( | |||
| IModelPtr model, CambriconAllocatorPtr allocator, const SymbolVarArray& src, | |||
| const OperatorNodeConfig& config); | |||
| //! creator a magicmind runtime operator from a serialized memory buffer | |||
| static SymbolVarArray make( | |||
| const void* buf, size_t buf_size, const SymbolVarArray& src, | |||
| const OperatorNodeConfig& config = {}); | |||
| static IModelPtr make_model_ptr(magicmind::IModel* model) { | |||
| return {model, magicmind_intl::MagicMindDeleter<magicmind::IModel>()}; | |||
| } | |||
| private: | |||
| CambriconAllocatorPtr m_allocator; | |||
| mutable IContextPtr m_context; | |||
| IEnginePtr m_engine; | |||
| IModelPtr m_model; | |||
| }; | |||
| } // namespace opr | |||
| } // namespace mgb | |||
| #endif // MGB_CAMBRICON | |||
| // vim: syntax=cpp.doxygen foldmethod=marker foldmarker=f{{{,f}}} | |||
| @@ -22,6 +22,7 @@ | |||
| using namespace mgb; | |||
| #if CNRT_MAJOR_VERSION < 5 | |||
| namespace { | |||
| class CnmlModelContext { | |||
| public: | |||
| @@ -620,6 +621,7 @@ TEST(TestCambriconRuntimeOpr, CrossCNCopy) { | |||
| MGB_ASSERT_TENSOR_NEAR(out_cnml, out_mgb, 1e-4); | |||
| } | |||
| #endif | |||
| #endif | |||
| // vim: syntax=cpp.doxygen foldmethod=marker foldmarker=f{{{,f}}} | |||
| @@ -193,11 +193,19 @@ public: | |||
| std::pair<size_t, size_t> get_mem_status_bytes() override { | |||
| m_env.cnrt_env().activate(); | |||
| cndevMemoryInfo_t mem_info; | |||
| #if CNRT_MAJOR_VERSION >= 5 | |||
| mem_info.version = CNDEV_VERSION_5; | |||
| #endif | |||
| MGB_CNDEV_CHECK(cndevGetMemoryUsage(&mem_info, m_env.cnrt_env().device)); | |||
| size_t tot, used, free; | |||
| constexpr size_t mb2size = 1024 * 1024; | |||
| #if CNRT_MAJOR_VERSION >= 5 | |||
| tot = static_cast<size_t>(mem_info.physicalMemoryTotal) * mb2size; | |||
| used = static_cast<size_t>(mem_info.physicalMemoryUsed) * mb2size; | |||
| #else | |||
| tot = static_cast<size_t>(mem_info.PhysicalMemoryTotal) * mb2size; | |||
| used = static_cast<size_t>(mem_info.PhysicalMemoryUsed) * mb2size; | |||
| #endif | |||
| free = tot - used + m_mem_alloc->get_free_memory_dev().tot; | |||
| return {tot, free}; | |||
| } | |||
| @@ -417,10 +425,18 @@ size_t CambriconCompNodeImpl::DeviceInfo::get_mem_reserve_size() { | |||
| } | |||
| size_t tot, free; | |||
| cndevMemoryInfo_t mem_info; | |||
| #if CNRT_MAJOR_VERSION >= 5 | |||
| mem_info.version = CNDEV_VERSION_5; | |||
| #endif | |||
| MGB_CNDEV_CHECK(cndevGetMemoryUsage(&mem_info, dev_num)); | |||
| constexpr size_t mb2size = 1024 * 1024; | |||
| #if CNRT_MAJOR_VERSION >= 5 | |||
| tot = static_cast<size_t>(mem_info.physicalMemoryTotal) * mb2size; | |||
| size_t used = static_cast<size_t>(mem_info.physicalMemoryUsed) * mb2size; | |||
| #else | |||
| tot = static_cast<size_t>(mem_info.PhysicalMemoryTotal) * mb2size; | |||
| size_t used = static_cast<size_t>(mem_info.PhysicalMemoryUsed) * mb2size; | |||
| #endif | |||
| free = tot - used; | |||
| return free - get_min_system_memory(free); | |||
| } else { | |||
| @@ -701,11 +717,19 @@ void mgb::mem_alloc::CambriconRawAlloctor::get_mem_info(size_t& free, size_t& to | |||
| } | |||
| mgb_assert(device >= 0, "current device has not been initialized in static data"); | |||
| cndevMemoryInfo_t mem_info; | |||
| #if CNRT_MAJOR_VERSION >= 5 | |||
| mem_info.version = CNDEV_VERSION_5; | |||
| #endif | |||
| auto ret = cndevGetMemoryUsage(&mem_info, device); | |||
| if (ret == CNDEV_SUCCESS) { | |||
| constexpr size_t mb2size = 1024 * 1024; | |||
| #if CNRT_MAJOR_VERSION >= 5 | |||
| tot = static_cast<size_t>(mem_info.physicalMemoryTotal) * mb2size; | |||
| size_t used = static_cast<size_t>(mem_info.physicalMemoryUsed) * mb2size; | |||
| #else | |||
| tot = static_cast<size_t>(mem_info.PhysicalMemoryTotal) * mb2size; | |||
| size_t used = static_cast<size_t>(mem_info.PhysicalMemoryUsed) * mb2size; | |||
| #endif | |||
| free = tot - used; | |||
| return; | |||
| } | |||
| @@ -318,6 +318,7 @@ void CompNodeEnv::init_rocm_async( | |||
| #endif | |||
| #if MGB_CAMBRICON | |||
| #if CNRT_MAJOR_VERSION < 5 | |||
| const char* mgb::cnml_get_error_string(cnmlStatus_t err) { | |||
| switch (err) { | |||
| #define cb(_err) \ | |||
| @@ -341,6 +342,7 @@ const char* mgb::cnml_get_error_string(cnmlStatus_t err) { | |||
| } | |||
| return "Unknown CNML error"; | |||
| } | |||
| #endif | |||
| void mgb::_on_cnrt_error( | |||
| const char* expr, cnrtRet_t err, const char* file, const char* func, int line) { | |||
| @@ -357,6 +359,7 @@ void mgb::_on_cndev_error( | |||
| cndevGetErrorString(err), expr, file, func, line); | |||
| } | |||
| #if CNRT_MAJOR_VERSION < 5 | |||
| void mgb::_on_cnml_error( | |||
| const char* expr, cnmlStatus_t err, const char* file, const char* func, | |||
| int line) { | |||
| @@ -365,6 +368,7 @@ void mgb::_on_cnml_error( | |||
| cnml_get_error_string(err), expr, file, func, line); | |||
| } | |||
| #endif | |||
| #endif | |||
| void CompNodeEnv::init_cpu(const CpuEnv& env, CompNode comp_node) { | |||
| m_comp_node = comp_node; | |||
| @@ -116,8 +116,10 @@ | |||
| #if MGB_CAMBRICON | |||
| #include <cndev.h> | |||
| #include <cnml.h> | |||
| #include <cnrt.h> | |||
| #if CNRT_MAJOR_VERSION < 5 | |||
| #include <cnml.h> | |||
| #endif | |||
| #if MGB_ENABLE_LOGGING | |||
| #define MGB_CNRT_CHECK(expr) \ | |||
| @@ -204,15 +206,19 @@ namespace mgb { | |||
| #endif | |||
| #if MGB_CAMBRICON | |||
| #if CNRT_MAJOR_VERSION < 5 | |||
| const char* cnml_get_error_string(cnmlStatus_t err); | |||
| #endif | |||
| [[noreturn]] void _on_cnrt_error( | |||
| const char* expr, cnrtRet_t err, const char* file, const char* func, int line); | |||
| [[noreturn]] void _on_cndev_error( | |||
| const char* expr, cndevRet_t err, const char* file, const char* func, int line); | |||
| #if CNRT_MAJOR_VERSION < 5 | |||
| [[noreturn]] void _on_cnml_error( | |||
| const char* expr, cnmlStatus_t err, const char* file, const char* func, | |||
| int line); | |||
| #endif | |||
| #endif | |||
| class CPUDispatcher : public MegcoreCPUDispatcher { | |||
| public: | |||
| @@ -456,6 +462,13 @@ public: | |||
| initialized = cnrt_err == CNRT_RET_SUCCESS; | |||
| auto cndev_err = cndevInit(0); | |||
| initialized &= cndev_err == CNDEV_SUCCESS; | |||
| #if CNRT_MAJOR_VERSION >= 5 | |||
| mgb_throw_if( | |||
| !initialized, CnrtError, | |||
| "cnrt/cndev initialize failed: (cnrt:%d, " | |||
| "cndev:%d)", | |||
| static_cast<int>(cnrt_err), static_cast<int>(cndev_err)); | |||
| #else | |||
| auto cnml_err = cnmlInit(0); | |||
| initialized &= cnml_err == CNML_STATUS_SUCCESS; | |||
| mgb_throw_if( | |||
| @@ -464,11 +477,14 @@ public: | |||
| "cndev:%d, cnml: %d)", | |||
| static_cast<int>(cnrt_err), static_cast<int>(cndev_err), | |||
| static_cast<int>(cnml_err)); | |||
| #endif | |||
| } | |||
| } | |||
| ~InitStatus() { | |||
| if (initialized) { | |||
| #if CNRT_MAJOR_VERSION < 5 | |||
| MGB_CNML_CHECK(cnmlExit()); | |||
| #endif | |||
| MGB_CNDEV_CHECK(cndevRelease()); | |||
| cnrtDestroy(); | |||
| initialized = false; | |||
| @@ -674,6 +674,7 @@ TEST(TestCompNodeCambricon, D2DCopy) { | |||
| REQUIRE_CAMBRICON_DEVICE(1); | |||
| auto cn = CompNode::load("cambricon0"); | |||
| run(cn); | |||
| REQUIRE_CAMBRICON_DEVICE(2); | |||
| cn = CompNode::load("cambricon1"); | |||
| run(cn); | |||
| } | |||
| @@ -43,6 +43,7 @@ void call_sereg() {} | |||
| #endif | |||
| #if MGB_CAMBRICON | |||
| #include "../../cambricon/impl/cambricon_runtime_opr.sereg.h" | |||
| #include "../../cambricon/impl/magicmind_runtime_opr.sereg.h" | |||
| #endif | |||
| #if MGB_CUSTOM_OP | |||