diff --git a/mindspore/ccsrc/runtime/framework/actor/control_flow/exit_actor.cc b/mindspore/ccsrc/runtime/framework/actor/control_flow/exit_actor.cc index 011c5090cb..323dd26c62 100644 --- a/mindspore/ccsrc/runtime/framework/actor/control_flow/exit_actor.cc +++ b/mindspore/ccsrc/runtime/framework/actor/control_flow/exit_actor.cc @@ -87,7 +87,7 @@ void ExitActor::SendOutput(OpContext *const context) { " current:" + std::to_string(input_partials_.size()) + " for actor:" + GetAID().Name(); SET_OPCONTEXT_FAIL_RET_WITH_ERROR((*context), error_info); } - auto output_partial = input_partials_[arrow->from_output_index_]; + auto output_partial = input_partials_[IntToSize(arrow->from_output_index_)]; MS_EXCEPTION_IF_NULL(output_partial->func_graph_); ActorDispatcher::Send(arrow->to_op_id_, &ControlActor::RunOpPartial, output_partial, IntToSize(arrow->to_input_index_), context); diff --git a/mindspore/ccsrc/runtime/framework/actor/control_flow/stack_actor.cc b/mindspore/ccsrc/runtime/framework/actor/control_flow/stack_actor.cc index f843362270..2b1ad2c5ae 100644 --- a/mindspore/ccsrc/runtime/framework/actor/control_flow/stack_actor.cc +++ b/mindspore/ccsrc/runtime/framework/actor/control_flow/stack_actor.cc @@ -45,7 +45,7 @@ void StackActor::Init() { } // Fetch the total number of input partial. - int total_partials_num = 0; + size_t total_partials_num = 0; for (const auto &formal_parameter : formal_parameters_) { const auto &abstract = formal_parameter.first->abstract(); MS_EXCEPTION_IF_NULL(abstract); @@ -117,8 +117,8 @@ void StackActor::RunOpPartial(const OpPartialPtr &partial, size_t position, OpCo auto self_partial = std::make_shared(); *self_partial = *partial; // The parameters from the inside of the subgraph need to be put into the stack. - if (IntToSize(position) < input_stack_data_num_ + device_tensor_store_keys_.size() + input_stack_partials_num_ + - local_device_tensors_.size()) { + if (position < input_stack_data_num_ + device_tensor_store_keys_.size() + input_stack_partials_num_ + + local_device_tensors_.size()) { input_stack_partials_[context->sequential_num_][position].push(self_partial); } else { (void)input_op_partials_[context->sequential_num_].emplace_back(position, self_partial); diff --git a/mindspore/ccsrc/runtime/framework/actor/control_flow/switch_actor.cc b/mindspore/ccsrc/runtime/framework/actor/control_flow/switch_actor.cc index 1ce78ec47f..07c83a6559 100644 --- a/mindspore/ccsrc/runtime/framework/actor/control_flow/switch_actor.cc +++ b/mindspore/ccsrc/runtime/framework/actor/control_flow/switch_actor.cc @@ -39,7 +39,7 @@ void SwitchActor::Init() { } auto data = std::make_unique>(data_arrow->to_op_id_, nullptr, data_arrow->to_input_index_); MS_EXCEPTION_IF_NULL(data); - (void)output_data_by_output_index_[data_arrow->from_output_index_].emplace_back(data.get()); + (void)output_data_by_output_index_[IntToSize(data_arrow->from_output_index_)].emplace_back(data.get()); (void)output_data_.emplace_back(std::move(data)); } } @@ -57,7 +57,7 @@ void SwitchActor::FetchInput(OpContext *const context) { input_partials_[0] = input_partials_[index + kSwitchCondPos]; } - for (auto &output_data : output_data_by_output_index_[kSwitchDefaultOutputNum - 1]) { + for (auto &output_data : output_data_by_output_index_[0]) { MS_EXCEPTION_IF_NULL(output_data); MS_EXCEPTION_IF_NULL(input_device_tensors_[index + kSwitchCondPos]); output_data->data_ = input_device_tensors_[index + kSwitchCondPos]; diff --git a/mindspore/ccsrc/runtime/framework/actor/data_prepare_actor.cc b/mindspore/ccsrc/runtime/framework/actor/data_prepare_actor.cc index af69de752f..782991f7ba 100644 --- a/mindspore/ccsrc/runtime/framework/actor/data_prepare_actor.cc +++ b/mindspore/ccsrc/runtime/framework/actor/data_prepare_actor.cc @@ -103,11 +103,11 @@ void ValueTupleToValue(const ValuePtr &value, std::vector *const value if (element->isa()) { ValueTupleToValue(element, values); } else { - values->emplace_back(element); + (void)values->emplace_back(element); } } } else { - values->emplace_back(value); + (void)values->emplace_back(value); } } diff --git a/mindspore/ccsrc/runtime/framework/control_node_parser.cc b/mindspore/ccsrc/runtime/framework/control_node_parser.cc index d04604882b..4ec164ed07 100644 --- a/mindspore/ccsrc/runtime/framework/control_node_parser.cc +++ b/mindspore/ccsrc/runtime/framework/control_node_parser.cc @@ -24,15 +24,6 @@ namespace mindspore { namespace runtime { namespace { -// Check whether node has a partial structure, a node is a partial structure whicih: -// 1. a partial cnode. -// 2. a funcgraph value node. -bool IsPartial(const AnfNodePtr &node) { - MS_EXCEPTION_IF_NULL(node); - return (node->isa() && IsValueNode(node)) || - AnfAlgo::CheckPrimitiveType(node, prim::kPrimPartial); -} - // Check if node is a value node need to create a device tensor. bool IsFrontValueNode(const KernelWithIndex &node_with_index) { const auto &node = node_with_index.first; @@ -58,135 +49,6 @@ bool IsFrontValueNode(const KernelWithIndex &node_with_index) { return !sub_abstracts[index]->isa(); } -// Get funcgraph in partial structure. -// Depth represents the number of layers of the call. When the first input of the call node is a call node, -// the funcgraph in the return value of the inner call needs to be returned. -FuncGraphPtr GetFuncGraphFromPartial(const AnfNodePtr &node, std::stack *output_indexs) { - MS_EXCEPTION_IF_NULL(node); - MS_EXCEPTION_IF_NULL(output_indexs); - - if (output_indexs->empty()) { - if (node->isa() && IsValueNode(node)) { - // Value node of funcgraph. - return GetValueNode(node); - } else if (AnfAlgo::CheckPrimitiveType(node, prim::kPrimPartial)) { - // Partial cnode. - const auto &partial_inputs = node->cast()->inputs(); - return GetValueNode(partial_inputs[kPartialFuncGraphPos]); - } else { - MS_LOG(EXCEPTION) << "Invalid partial construct node:" << node->DebugString(); - } - return nullptr; - } - - // Get funcgraph in the output of inner call. - FuncGraphPtr func_graph; - if (node->isa() && IsValueNode(node)) { - func_graph = GetValueNode(node); - } else if (AnfAlgo::CheckPrimitiveType(node, prim::kPrimPartial)) { - const auto &partial_inputs = node->cast()->inputs(); - const auto &func_graph_node = partial_inputs[kPartialFuncGraphPos]; - if (!func_graph_node->isa() || !IsValueNode(func_graph_node)) { - MS_LOG(EXCEPTION) << "Invalid partial node:" << node->DebugString(); - } - func_graph = GetValueNode(func_graph_node); - } else { - MS_LOG(EXCEPTION) << "Invalid partial node:" << node->DebugString(); - } - MS_EXCEPTION_IF_NULL(func_graph); - - size_t index = output_indexs->top(); - output_indexs->pop(); - const auto &output = func_graph->output(); - MS_EXCEPTION_IF_NULL(output); - KernelWithIndex real_output_with_index = AnfAlgo::VisitKernelWithReturnType(output, 0); - auto real_output = real_output_with_index.first; - MS_EXCEPTION_IF_NULL(real_output); - if (AnfAlgo::CheckPrimitiveType(real_output, prim::kPrimMakeTuple)) { - const auto &cnode = real_output->cast(); - const auto &inputs = cnode->inputs(); - if (inputs.size() <= index + kMakeTupleInputStartPos) { - MS_LOG(EXCEPTION) << "Invalid output index:" << index << " for node:" << real_output->DebugString(); - } - real_output = inputs[index + kMakeTupleInputStartPos]; - } - MS_EXCEPTION_IF_NULL(real_output); - return GetFuncGraphFromPartial(real_output, output_indexs); -} - -// Find all funcgraphs that the call node will call. -// The output index represents the index of the funcgraph in the graph output. When the funcgraph is passed through -// the function return value, the index in the return value needs to be placed on the stack so that the index can be -// obtained from the stack when the graph output is found. -std::set GetFuncGraphbyCallNode(const AnfNodePtr &node, std::stack *output_indexs, - AnfNodePtr make_tuple = nullptr) { - MS_EXCEPTION_IF_NULL(node); - std::set func_graphs; - if (!node->isa()) { - return func_graphs; - } - - const auto &cnode = node->cast(); - MS_EXCEPTION_IF_NULL(cnode); - const auto &call_input0 = cnode->input(0); - MS_EXCEPTION_IF_NULL(call_input0); - const auto &call_input_with_index = AnfAlgo::VisitKernelWithReturnType(call_input0, 0); - const auto &real_input = call_input_with_index.first; - if (AnfAlgo::IsCallNode(real_input)) { - output_indexs->push(call_input_with_index.second); - return GetFuncGraphbyCallNode(real_input, output_indexs); - } - - if (AnfAlgo::CheckPrimitiveType(real_input, prim::kPrimSwitch)) { - // First input node of call is switch node. - const auto &input_cnode = real_input->cast(); - MS_EXCEPTION_IF_NULL(input_cnode); - const auto &switch_inputs = input_cnode->inputs(); - for (size_t i = kSwitchTrueBranchPos; i < switch_inputs.size(); ++i) { - MS_EXCEPTION_IF_NULL(switch_inputs[i]); - std::stack tmp_output_indexs = *output_indexs; - (void)func_graphs.emplace(GetFuncGraphFromPartial(switch_inputs[i], &tmp_output_indexs)); - } - } else if (AnfAlgo::CheckPrimitiveType(real_input, prim::kPrimSwitchLayer)) { - // First input node of call is switch layer node. - const auto &input_cnode = real_input->cast(); - MS_EXCEPTION_IF_NULL(input_cnode); - const auto &tuple_node = input_cnode->input(kSwitchLayerBranchPos); - if (AnfAlgo::CheckPrimitiveType(tuple_node, prim::kPrimMakeTuple)) { - const auto &tuple_inputs = tuple_node->cast()->inputs(); - for (size_t i = kMakeTupleInputStartPos; i < tuple_inputs.size(); ++i) { - MS_EXCEPTION_IF_NULL(tuple_inputs[i]); - std::stack tmp_output_indexs = *output_indexs; - (void)func_graphs.emplace(GetFuncGraphFromPartial(tuple_inputs[i], &tmp_output_indexs)); - } - } else if (tuple_node->isa()) { - const auto &abstract = tuple_node->abstract(); - MS_EXCEPTION_IF_NULL(abstract); - if (!abstract->isa()) { - MS_LOG(EXCEPTION) << "Invalid abstract:" << abstract->ToString() << " in node:" << tuple_node->DebugString() - << " for switch layer node:" << real_input->DebugString(); - } - MS_EXCEPTION_IF_NULL(make_tuple); - auto partials = make_tuple->cast()->inputs(); - for (const auto partial : partials) { - if (IsPartial(partial)) { - std::stack tmp_output_indexs; - (void)func_graphs.emplace(GetFuncGraphFromPartial(partial, &tmp_output_indexs)); - } - } - } else { - MS_LOG(EXCEPTION) << "Invalid input tuple node:" << tuple_node->DebugString() - << " for switch layer node:" << cnode->DebugString(); - } - } else if (IsPartial(real_input)) { - // First input node of call is partial node or value node of funcgraph. - (void)func_graphs.emplace(GetFuncGraphFromPartial(real_input, output_indexs)); - } else { - MS_LOG(EXCEPTION) << "Unable to identify call node" << real_input->DebugString(); - } - return func_graphs; -} - // Fetch real input node in maketuple. KernelWithIndex FetchRealInputNode(const KernelWithIndex &node_with_index) { const auto &node = node_with_index.first; @@ -456,7 +318,7 @@ std::vector FetchAllOutputWithIndex(const AnfNodePtr &node) { const auto &cnode = node_with_index.first->cast(); const auto &inputs = cnode->inputs(); const auto &tmp_list = FetchAllOutputWithIndex(inputs[i + kMakeTupleInputStartPos]); - result.insert(result.end(), tmp_list.begin(), tmp_list.end()); + (void)result.insert(result.end(), tmp_list.begin(), tmp_list.end()); } else if (AnfAlgo::CheckPrimitiveType(node_with_index.first, prim::kPrimSwitch) || AnfAlgo::CheckPrimitiveType(node_with_index.first, prim::kPrimSwitchLayer)) { } else if (AnfAlgo::IsCallNode(node_with_index.first)) { @@ -517,7 +379,6 @@ void CreateDeviceTensorForValueNode(const KernelWithIndex &front_node_with_index void CreateDeviceTensorForFrontNode(const KernelWithIndex &front_node_with_index, const DeviceContext *device_context) { MS_EXCEPTION_IF_NULL(device_context); const auto &node = front_node_with_index.first; - MS_EXCEPTION_IF_NULL(device_context); TypeId type_id = AnfAlgo::GetOutputInferDataType(node, 0); if (node->kernel_info() == nullptr) { @@ -579,7 +440,7 @@ std::vector FetchInputNodeByNode(const AnfNodePtr &node) { const auto &inputs = cnode->inputs(); for (size_t i = kMakeTupleInputStartPos; i < inputs.size(); ++i) { const auto &sub_results = FetchInputNodeByNode(inputs[i]); - results.insert(results.end(), sub_results.begin(), sub_results.end()); + (void)results.insert(results.end(), sub_results.begin(), sub_results.end()); } return results; } @@ -608,8 +469,9 @@ std::vector FetchInputNodeByNode(const AnfNodePtr &node) { if (csr_tensor_inputs.size() <= kMakeCSRTensorInputNum) { MS_LOG(EXCEPTION) << "Invalid make csr tensor node:" << cnode->DebugString(); } - const auto &sub_results = FetchInputNodeByNode(csr_tensor_inputs[iter->second + kMakeCSRTensorInputStartPos]); - results.insert(results.end(), sub_results.begin(), sub_results.end()); + const auto &sub_results = + FetchInputNodeByNode(csr_tensor_inputs[LongToSize(iter->second) + kMakeCSRTensorInputStartPos]); + (void)results.insert(results.end(), sub_results.begin(), sub_results.end()); } else { // Csr node from parameter or call node. auto abstract = src_node->abstract(); @@ -637,7 +499,7 @@ std::vector FetchInputNodeByNode(const AnfNodePtr &node) { MS_EXCEPTION_IF_NULL(get_item_src_node); if (index_stack.empty()) { const auto &sub_results = FetchInputNodeByNode(get_item_src_node); - results.insert(results.end(), sub_results.begin(), sub_results.end()); + (void)results.insert(results.end(), sub_results.begin(), sub_results.end()); return results; } auto get_item_src_abstract = get_item_src_node->abstract(); @@ -698,7 +560,7 @@ bool IsFirstControlNode(const AnfNodePtr &node, std::set *checked_no if (!node->isa() || checked_nodes->find(node) != checked_nodes->end()) { return true; } - checked_nodes->emplace(node); + (void)checked_nodes->emplace(node); const auto &cnode = node->cast(); MS_EXCEPTION_IF_NULL(cnode); @@ -722,7 +584,7 @@ size_t ParseControlNodeLevel(const AnfNodePtr &node, std::set *check if (!node->isa() || checked_nodes->find(node) != checked_nodes->end()) { return 0; } - checked_nodes->emplace(node); + (void)checked_nodes->emplace(node); auto iter = node_to_level.find(node); if (iter != node_to_level.end()) { @@ -832,7 +694,7 @@ std::vector FetchInputNodeByCNode(const AnfNodePtr &node) { continue; } const auto &sub_results = FetchInputNodeByNode(inputs[i]); - results.insert(results.end(), sub_results.begin(), sub_results.end()); + (void)results.insert(results.end(), sub_results.begin(), sub_results.end()); } return results; } @@ -1259,8 +1121,9 @@ void ControlNodeParser::ParseDeviceContextForReturnNode(const DeviceContext *def if (func_graph_iter == func_graph_to_device_contexts_.end()) { MS_LOG(EXCEPTION) << "Cannot find device context for funcgraph:" << func_graph->ToString(); } - MS_EXCEPTION_IF_NULL(func_graph_iter->second[iter - func_graph->parameters().begin()]); - (void)return_device_contexts.emplace_back(func_graph_iter->second[iter - func_graph->parameters().begin()]); + size_t index = LongToSize(iter - func_graph->parameters().begin()); + MS_EXCEPTION_IF_NULL(func_graph_iter->second[index]); + (void)return_device_contexts.emplace_back(func_graph_iter->second[index]); } else if (output_node.first->isa()) { // If the output is parameter, used the default context type. MS_EXCEPTION_IF_NULL(default_context); @@ -1449,13 +1312,14 @@ void ControlNodeParser::ParseFormalToRealParameter(const std::vector const auto &func_graphs = FetchFuncGraphbyCallNode(node); for (const auto func_graph : func_graphs) { const auto ¶meters = func_graph->parameters(); - for (size_t i = inputs.size() - 1, j = parameters.size() - 1; i >= kCallInputStartPos && j >= 0; --i, --j) { - MS_EXCEPTION_IF_NULL(inputs[i]); - MS_EXCEPTION_IF_NULL(parameters[j]); - if (HasAbstractMonad(inputs[i])) { + for (int i = inputs.size() - 1, j = parameters.size() - 1; i >= 1 && j >= 0; --i, --j) { + MS_EXCEPTION_IF_NULL(inputs[IntToSize(i)]); + MS_EXCEPTION_IF_NULL(parameters[IntToSize(j)]); + if (HasAbstractMonad(inputs[IntToSize(i)])) { continue; } - AddFormalToRealParameter(parameters[j], inputs[i], call_node_to_func_graphs_, &formal_to_real_parameters); + AddFormalToRealParameter(parameters[IntToSize(j)], inputs[IntToSize(i)], call_node_to_func_graphs_, + &formal_to_real_parameters); } } } else if (AnfAlgo::CheckPrimitiveType(node, prim::kPrimPartial)) { @@ -1968,7 +1832,7 @@ void ControlNodeParser::ParseNodeLevel(const std::vector &control_no kernel_graph_group_info->need_stack_ = true; kernel_graph_group_info->level_ = max_level; for (const auto &kernel_graph : kernel_graph_group_info->graphs_) { - call_input_kernel_graphs_.emplace(kernel_graph.get()); + (void)call_input_kernel_graphs_.emplace(kernel_graph.get()); } } } diff --git a/mindspore/ccsrc/runtime/framework/control_node_scheduler.cc b/mindspore/ccsrc/runtime/framework/control_node_scheduler.cc index 428d66cfb0..123c7a7411 100644 --- a/mindspore/ccsrc/runtime/framework/control_node_scheduler.cc +++ b/mindspore/ccsrc/runtime/framework/control_node_scheduler.cc @@ -223,8 +223,9 @@ std::vector ControlNodeScheduler::BuildEntranceActor(const Gra : context_iter->second.size()); } entrance_actor->device_contexts_.clear(); - entrance_actor->device_contexts_.insert(entrance_actor->device_contexts_.begin(), context_iter->second.begin(), - context_iter->second.begin() + formal_parameters.size()); + (void)entrance_actor->device_contexts_.insert( + entrance_actor->device_contexts_.begin(), context_iter->second.begin(), + context_iter->second.begin() + SizeToLong(formal_parameters.size())); (void)entrance_actors.emplace_back(entrance_actor); InsertActor(entrance_actor.get()); } @@ -329,8 +330,8 @@ std::vector ControlNodeScheduler::BuildStackActor(const GraphComp (void)formal_parameters.emplace_back(node_with_context.first); (void)device_contexts.emplace_back(node_with_context.second); } else { - formal_parameters.insert(formal_parameters.begin(), node_with_context.first); - device_contexts.insert(device_contexts.begin(), node_with_context.second); + (void)formal_parameters.insert(formal_parameters.begin(), node_with_context.first); + (void)device_contexts.insert(device_contexts.begin(), node_with_context.second); input_parameter_data_num++; } } @@ -410,8 +411,8 @@ void ControlNodeScheduler::BuildStackActorForControlNode(const GraphCompilerInfo (void)formal_parameters.emplace_back(parameter); (void)device_contexts.emplace_back(device_context); } else { - formal_parameters.insert(formal_parameters.begin(), parameter); - device_contexts.insert(device_contexts.begin(), device_context); + (void)formal_parameters.insert(formal_parameters.begin(), parameter); + (void)device_contexts.insert(device_contexts.begin(), device_context); const auto &abstract = parameter.first->abstract(); MS_EXCEPTION_IF_NULL(abstract); @@ -725,7 +726,7 @@ void ControlNodeScheduler::LinkArrowByCallNode(const AnfNodePtr &call_node, Cont auto actor = FetchActor(actor_name); MS_EXCEPTION_IF_NULL(actor); auto exit_actor = dynamic_cast(actor); - size_t branch_id = parser->FetchBranchIDByCallNode(from_node); + auto branch_id = parser->FetchBranchIDByCallNode(from_node); if (real_abstract->isa()) { LinkPartialArrowForExitActor(exit_actor, to_actor, from_node_with_index.second, to_node_with_index.second, branch_id); @@ -1367,7 +1368,7 @@ void ControlNodeScheduler::LinkDataWithBranchIDArrow(GatherActor *const gather_a const FuncGraphPtr &func_graph) { MS_EXCEPTION_IF_NULL(gather_actor); MS_EXCEPTION_IF_NULL(entrance_actor); - gather_actor->output_data_with_branch_id_arrows_[func_graph.get()].emplace_back(entrance_actor->GetAID()); + (void)gather_actor->output_data_with_branch_id_arrows_[func_graph.get()].emplace_back(entrance_actor->GetAID()); } void ControlNodeScheduler::LinkPartialArrow(ControlActor *const from_actor, ControlActor *const to_actor, @@ -1410,7 +1411,7 @@ bool ControlNodeScheduler::CheckActorValid(const ActorSet *actor_set) const { return true; } -bool ControlNodeScheduler::IsNoInputActor(const ControlActor *control_actor) { +bool ControlNodeScheduler::IsNoInputActor(const ControlActor *control_actor) const { return (control_actor->input_datas_num_ == 0 && control_actor->input_controls_num_ == 0 && control_actor->input_partials_num_ == 0 && control_actor->input_branch_ids_num_ == 0); } diff --git a/mindspore/ccsrc/runtime/framework/control_node_scheduler.h b/mindspore/ccsrc/runtime/framework/control_node_scheduler.h index aca62081ec..c6550fb548 100644 --- a/mindspore/ccsrc/runtime/framework/control_node_scheduler.h +++ b/mindspore/ccsrc/runtime/framework/control_node_scheduler.h @@ -106,7 +106,7 @@ class ControlNodeScheduler { size_t to_index, int branch_id); void LinkPartialArrowForExitActor(ExitActor *const exit_actor, ControlActor *const to_actor, size_t from_index, size_t to_index, int branch_id); - bool IsNoInputActor(const ControlActor *control_actor); + bool IsNoInputActor(const ControlActor *control_actor) const; // Fill the device tensors of backend input nodes corresponding to ref formal parameters. void AddFormalParameterDeviceTensor(ControlActor *const from_actor, size_t from_index, const AnfNodePtr &input_node); diff --git a/mindspore/ccsrc/vm/backend.cc b/mindspore/ccsrc/vm/backend.cc index 585d9d4816..a315b107ac 100644 --- a/mindspore/ccsrc/vm/backend.cc +++ b/mindspore/ccsrc/vm/backend.cc @@ -565,9 +565,9 @@ void MindRTBackend::CompileGraph(const GraphSegmentPtr &segment, bool contain_mu const auto &func_graph = segment->nodes_[0]->func_graph(); MS_EXCEPTION_IF_NULL(func_graph); if (func_graph_to_kernel_graph_ids_.find(func_graph) == func_graph_to_kernel_graph_ids_.end()) { - func_graph_to_kernel_graph_ids_[func_graph].emplace_back(std::vector{graph_id}); + (void)func_graph_to_kernel_graph_ids_[func_graph].emplace_back(std::vector{graph_id}); } else { - func_graph_to_kernel_graph_ids_[func_graph].back().emplace_back(graph_id); + (void)func_graph_to_kernel_graph_ids_[func_graph].back().emplace_back(graph_id); } } else { // Compile the cut node. @@ -579,7 +579,7 @@ void MindRTBackend::CompileGraph(const GraphSegmentPtr &segment, bool contain_mu AnfAlgo::CheckPrimitiveType(cut_node, prim::kPrimSwitchLayer)) { const auto &func_graph = cut_node->func_graph(); MS_EXCEPTION_IF_NULL(func_graph); - func_graph_to_kernel_graph_ids_[func_graph].emplace_back(std::vector()); + (void)func_graph_to_kernel_graph_ids_[func_graph].emplace_back(std::vector()); } } } @@ -799,7 +799,7 @@ void PushTupleTensor(const VectorRef &args, const std::vector ¶m if (position >= args.size()) { MS_LOG(INFO) << "Position out of args range, position value is " << position << " and args size is " << args.size() << "."; - input_tensor->push_back(nullptr); + (void)input_tensor->emplace_back(nullptr); return; } auto value_tuple = utils::cast(args[position]); @@ -1125,9 +1125,9 @@ std::unique_ptr MindRTBackend::ConstructGraphCompilerInfo(con for (const auto &graph_id : sub_kernel_graphs_ids) { const auto &kernel_graph = graph_compiler_->Fetch(graph_id); MS_EXCEPTION_IF_NULL(kernel_graph); - kernel_graphs.emplace_back(kernel_graph); + (void)kernel_graphs.emplace_back(kernel_graph); } - func_graph_to_kernel_graphs[func_graph].emplace_back(kernel_graphs); + (void)func_graph_to_kernel_graphs[func_graph].emplace_back(kernel_graphs); } }