From cdd3e50e986fd85375bd3cb89eabfa60ecc31896 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E9=99=88=E5=BF=A0=E9=91=AB?= <554473243@qq.com> Date: Thu, 5 Oct 2023 22:27:15 +0800 Subject: [PATCH] mindir --- .../backend/common/optimizer/common_backend_optimization.h | 2 +- mindspore/ccsrc/backend/common/session/ascend_session.h | 2 +- mindspore/ccsrc/backend/common/session/cpu_session.h | 2 +- mindspore/ccsrc/backend/common/session/session_basic.h | 2 +- mindspore/lite/tools/converter/import/mindspore_importer.cc | 2 +- .../micro/coder/generator/component/common_component.cc | 2 +- mindspore/python/mindspore/train/_utils.py | 4 ++-- 7 files changed, 8 insertions(+), 8 deletions(-) diff --git a/mindspore/ccsrc/backend/common/optimizer/common_backend_optimization.h b/mindspore/ccsrc/backend/common/optimizer/common_backend_optimization.h index 762fc6a273..ddc3b5c820 100644 --- a/mindspore/ccsrc/backend/common/optimizer/common_backend_optimization.h +++ b/mindspore/ccsrc/backend/common/optimizer/common_backend_optimization.h @@ -21,7 +21,7 @@ namespace mindspore { namespace opt { void BackendCommonOptimization(const std::shared_ptr &kernel_graph); void CommonFinalOptimization(const std::shared_ptr &kernel_graph); -void CommonUnifyMindIR(const std::shared_ptr &kernel_graph); +void CommonUnifyMindIR(const std::shared_ptr &kernel_graph); // for debug void AddDynamicShapeAttrPass(const std::shared_ptr &kernel_graph); void EliminateIllegalDataTypePass(const std::shared_ptr &kernel_graph); void DynamicShapeConvertPass(const std::shared_ptr &kernel_graph); diff --git a/mindspore/ccsrc/backend/common/session/ascend_session.h b/mindspore/ccsrc/backend/common/session/ascend_session.h index 9efb511b57..cf8d00add0 100644 --- a/mindspore/ccsrc/backend/common/session/ascend_session.h +++ b/mindspore/ccsrc/backend/common/session/ascend_session.h @@ -47,7 +47,7 @@ class AscendSession : public SessionBasic { static void BatchBuildKernel(const std::vector> &build_tasks); - protected: + protected: // load graph to device related void UnifyMindIR(const KernelGraphPtr &graph) override; GraphId CompileGraphImpl(const AnfNodePtrList &lst, const AnfNodePtrList &outputs) override; GraphId CompileGraphImpl(NotNull func_graph) override; diff --git a/mindspore/ccsrc/backend/common/session/cpu_session.h b/mindspore/ccsrc/backend/common/session/cpu_session.h index 7bc583b6ad..5274143e71 100644 --- a/mindspore/ccsrc/backend/common/session/cpu_session.h +++ b/mindspore/ccsrc/backend/common/session/cpu_session.h @@ -31,7 +31,7 @@ class CPUSession : public SessionBasic { ~CPUSession() override = default; void Init(uint32_t device_id) override; - protected: + protected: // load graph void UnifyMindIR(const KernelGraphPtr &graph) override { SessionBasic::UnifyMindIR(graph); } void CreateOutputTensors(const GraphId &graph_id, const std::vector &input_tensors, VectorRef *, std::map *tensor_to_node, diff --git a/mindspore/ccsrc/backend/common/session/session_basic.h b/mindspore/ccsrc/backend/common/session/session_basic.h index ab292a573a..d173931a0d 100644 --- a/mindspore/ccsrc/backend/common/session/session_basic.h +++ b/mindspore/ccsrc/backend/common/session/session_basic.h @@ -228,7 +228,7 @@ class BACKEND_EXPORT SessionBasic : public std::enable_shared_from_this &tensor_to_node, std::map *); virtual void UnifyMindIR(const KernelGraphPtr &graph); diff --git a/mindspore/lite/tools/converter/import/mindspore_importer.cc b/mindspore/lite/tools/converter/import/mindspore_importer.cc index d411f48ffe..d6a5af7899 100644 --- a/mindspore/lite/tools/converter/import/mindspore_importer.cc +++ b/mindspore/lite/tools/converter/import/mindspore_importer.cc @@ -42,7 +42,7 @@ constexpr size_t kDependInputNum = 3; constexpr size_t kDependFirstInputIdx = 1; constexpr size_t kTupleGetItemFirstInputIdx = 1; } // namespace -STATUS MindsporeImporter::Mindir2AnfAdjust(const FuncGraphPtr &func_graph, const converter::Flags &flag) { +STATUS MindsporeImporter::Mindir2AnfAdjust(const FuncGraphPtr &func_graph, const converter::Flags &flag) { // fmk: Ms MS_ASSERT(func_graph != nullptr); auto primitive_adjust_pass = std::make_shared(); MS_CHECK_TRUE_MSG(primitive_adjust_pass != nullptr, RET_NULL_PTR, "primitive_adjust_pass is nullptr."); diff --git a/mindspore/lite/tools/converter/micro/coder/generator/component/common_component.cc b/mindspore/lite/tools/converter/micro/coder/generator/component/common_component.cc index f34fbec310..2326d62608 100644 --- a/mindspore/lite/tools/converter/micro/coder/generator/component/common_component.cc +++ b/mindspore/lite/tools/converter/micro/coder/generator/component/common_component.cc @@ -77,7 +77,7 @@ void CodeMSModelBuild(std::ofstream &ofs, const Configurator *config) { " return kMSStatusLiteNotSupport;\n" " }\n"; ofs << " int ret = RET_OK;\n"; - if (config->target() != kARM32M) { + if (config->target() != kARM32M) { // only support ARM32M ofs << " ret = Init((void*)model_data, data_size);\n"; } if (config->support_parallel()) { diff --git a/mindspore/python/mindspore/train/_utils.py b/mindspore/python/mindspore/train/_utils.py index 951ba053fe..0e3467c2d2 100644 --- a/mindspore/python/mindspore/train/_utils.py +++ b/mindspore/python/mindspore/train/_utils.py @@ -23,7 +23,7 @@ from mindspore.common.dtype import dtype_to_nptype, pytype_to_dtype from mindspore.common import dtype as mstype from mindspore import log as logger from mindspore.common.api import _cell_graph_executor -from mindspore.train.mind_ir_pb2 import ModelProto as mindir_model +from mindspore.train.mind_ir_pb2 import ModelProto as mindir_model from mindspore.train.checkpoint_pb2 import Checkpoint from mindspore.train.node_strategy_pb2 import ParallelStrategyMap as ckpt_strategy @@ -42,7 +42,7 @@ def _convert_type(types): Returns: list, list of element in dataset. """ - ms_types = [] + ms_types = [] for np_type in types: ms_type = pytype_to_dtype(np_type) ms_types.append(ms_type)