/** * This is the C++ adaptation and derivative work of Myia (https://github.com/mila-iqia/myia/). * * Copyright 2019 Huawei Technologies Co., Ltd * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include "pipeline/jit/parse/function_block.h" #include #include #include "pipeline/jit/parse/resolve.h" #include "pipeline/jit/parse/parse.h" #include "frontend/operator/ops.h" #include "utils/info.h" #include "debug/trace.h" #include "pybind11/pybind11.h" namespace mindspore { namespace py = pybind11; namespace parse { FunctionBlock::FunctionBlock(const Parser &parser) : parser_(parser) { func_graph_ = std::make_shared(); matured_ = false; } void FunctionBlock::AddPrevBlock(const FunctionBlockPtr &block) { prev_blocks_.push_back(block.get()); } // write variable records the variable name to corresponding node void FunctionBlock::WriteVariable(const std::string &var_name, const AnfNodePtr &node) { MS_LOG(DEBUG) << func_graph_->ToString() << " write var " << var_name << " with node " << node->DebugString(); vars_[var_name] = node; } // read variable from predecessors AnfNodePtr FunctionBlock::ReadVariable(const std::string &var) { // get var node if it is found if (vars_.count(var)) { AnfNodePtr node = vars_[var]; MS_EXCEPTION_IF_NULL(node); if (node->isa()) { return NewValueNode(GetValueNode(node)); } else { return node; } } // get var from predecessor block ,if can't get the make a resolve node to it if (matured_) { // If only one predecessor block, read the definition of var from it. if (prev_blocks_.size() == 1) { auto block = prev_blocks_[0]; MS_EXCEPTION_IF_NULL(block); return block->ReadVariable(var); } else if (prev_blocks_.empty()) { // get namespace and make Reslove return MakeResolveSymbol(var); } } // If have more than one predecessor blocks then build a phi node. auto debug_info = std::make_shared(); debug_info->set_name(var); TraceManager::DebugTrace(std::make_shared(debug_info)); ParameterPtr phi_param = std::make_shared(func_graph()); TraceManager::EndTrace(); MS_LOG(DEBUG) << func_graph_->ToString() << " generate phi node " << phi_param->ToString() << " for " << var; func_graph()->add_parameter(phi_param); phi_nodes_[phi_param] = var; WriteVariable(var, phi_param); if (matured_) { SetPhiArgument(phi_param); } return phi_param; } // Resolve Ast operator node AnfNodePtr FunctionBlock::MakeResolveAstOp(const py::object &op) { auto ast = parser_.ast(); MS_EXCEPTION_IF_NULL(ast); TraceGuard trace_guard(parser_.GetLocation(op)); py::tuple namespace_var = ast->CallParserObjMethod(PYTHON_PARSE_GET_AST_NAMESPACE_SYMBOL, op); if (namespace_var.size() != 2) { MS_LOG(EXCEPTION) << "Resolve ast op failed, get namespace tuple size=" << namespace_var.size(); } NameSpacePtr name_space = std::make_shared(RESOLVE_NAMESPACE_NAME_AST, namespace_var[0]); SymbolPtr symbol = std::make_shared(namespace_var[1].cast()); return MakeResolve(name_space, symbol); } // Resolve class member, two possible: method, member variable AnfNodePtr FunctionBlock::MakeResolveClassMember(const std::string &attr) { py::object namespace_var = parser_.ast()->CallParseModFunction(PYTHON_MOD_GET_MEMBER_NAMESPACE_SYMBOL, parser_.ast()->obj()); NameSpacePtr name_space = std::make_shared(RESOLVE_NAMESPACE_NAME_CLASS_MEMBER, namespace_var); SymbolPtr symbol = std::make_shared(attr); return MakeResolve(name_space, symbol); } // Make a resolve node for symbol string AnfNodePtr FunctionBlock::MakeResolveSymbol(const std::string &value) { if (value.compare(0, strlen("self"), "self") == 0) { auto start = value.find_first_of('.') + 1; if (start >= value.size()) { MS_LOG(ERROR) << "Find invalid resolve symbol str: " << value; return nullptr; } auto bits_str = value.substr(start); return MakeResolveClassMember(bits_str); } py::tuple namespace_var = parser_.ast()->CallParserObjMethod(PYTHON_PARSE_GET_NAMESPACE_SYMBOL, value); NameSpacePtr name_space = std::make_shared(RESOLVE_NAMESPACE_NAME_SYMBOL_STR, namespace_var[0]); SymbolPtr symbol = std::make_shared(namespace_var[1].cast()); return MakeResolve(name_space, symbol); } AnfNodePtr FunctionBlock::MakeResolveOperation(const std::string &value) { py::tuple namespace_var = parser_.ast()->CallParserObjMethod(PYTHON_PARSE_GET_OPERATION_NAMESPACE_SYMBOL, value); NameSpacePtr name_space = std::make_shared(RESOLVE_NAMESPACE_NAME_COMMON_OPS, namespace_var[0]); SymbolPtr symbol = std::make_shared(namespace_var[1].cast()); return MakeResolve(name_space, symbol); } AnfNodePtr FunctionBlock::MakeResolve(const NameSpacePtr &name_space, const SymbolPtr &resolve_symbol) { MS_LOG(DEBUG) << "MakeResolve for " << ((std::string)py::str(name_space->obj())) << " , " << ((std::string)resolve_symbol->symbol()); ValueNodePtr module_node = NewValueNode(name_space); ValueNodePtr symbol_node = NewValueNode(resolve_symbol); auto node = func_graph()->NewCNode({NewValueNode(prim::kPrimResolve), module_node, symbol_node}); return node; } // add input for the block's phi parameter void FunctionBlock::SetPhiArgument(const ParameterPtr &phi) { std::string var = phi_nodes_[phi]; MS_LOG(DEBUG) << "graph " << func_graph_->ToString() << " set phi " << phi->ToString() << " for var " << var; auto removable = CollectRemovablePhi(phi); // If the phi node is not necessary, not need to add to jumps_ of the prev blocks. if (removable) { MS_LOG(DEBUG) << "remove the phi when call graph " << func_graph_->ToString() << " var " << var; return; } for (auto &pred : prev_blocks_) { MS_EXCEPTION_IF_NULL(pred); MS_LOG(DEBUG) << "graph " << func_graph_->ToString() << " pred_blocks_ " << pred->func_graph_->ToString(); AnfNodePtr arg_node = pred->ReadVariable(var); CNodePtr jump = pred->jumps_[this]; jump->add_input(arg_node); } } AnfNodePtr FunctionBlock::SearchReplaceNode(const std::string &var, const ParameterPtr &phi) { AnfNodePtr arg_node = nullptr; for (auto &prev : prev_blocks_) { MS_EXCEPTION_IF_NULL(prev); AnfNodePtr temp_node = prev->ReadVariable(var); MS_LOG(DEBUG) << "graph " << prev->func_graph_->ToString() << " phi " << phi->ToString() << " for var " << var << " is " << temp_node->DebugString(); if (temp_node != phi) { if (arg_node == nullptr) { arg_node = temp_node; MS_LOG(DEBUG) << "graph " << prev->func_graph_->ToString() << " phi " << phi->ToString() << " may be replaced by node " << arg_node->DebugString(); } else if (temp_node == arg_node) { MS_LOG(DEBUG) << "graph " << prev->func_graph_->ToString() << " phi " << phi->ToString() << " is same as node " << arg_node->DebugString(); } else { MS_LOG(DEBUG) << "phi " << phi->ToString() << " cannot be removed as it assigns to different node. node1: " << arg_node->DebugString() << ", node2: " << temp_node->DebugString(); return nullptr; } } } return arg_node; } // Check if there is removable unnecessary phi node in this graph. // as per the FIRM TR 3.2, a phi node can be remove if: // // If all arguments of a φ-function are the same value s or the φfunction itself, // then we remove the φ-function and let all users directly uses. We call such a // φ-function obviously unnecessary. // When we removed a φ-function p, then we recursively try to apply this simplification // rule with all (former) users of p, because they may have become obviously unnecessary // due to the removal of p // // phi node in graph will be removed after the whole function is parsed in a DFS visit // of that graph.The reason is : // 1. when this function is called, not all usage of this phi node had bound to the // graph of this function block, some may stay in vars_ in other blocks. // 2. it's costly to iterate the graph to replace the phi for each phi. // Args : // phi : This parameter node is functioning as a phi node. bool FunctionBlock::CollectRemovablePhi(const ParameterPtr &phi) { MS_EXCEPTION_IF_NULL(phi); std::string var = phi_nodes_[phi]; MS_LOG(DEBUG) << "check phi " << phi->DebugString() << " for " << var; if (prev_blocks_.size() == 0) { MS_LOG(DEBUG) << "no phi " << phi->DebugString() << " for var " << var; return false; } AnfNodePtr arg_node = SearchReplaceNode(var, phi); if (arg_node != nullptr) { MS_LOG(DEBUG) << "graph " << func_graph_->ToString() << " phi " << phi->ToString() << " can be replaced with " << arg_node->DebugString(); // replace var with new one. This equal to statement in TR "v0 is immediately replaced by v1." WriteVariable(var, arg_node); removable_phis_[phi] = arg_node; // The following equal to statement "The φ-function defining v1, which now reads φ(v2, v1), is optimized // recursively". check if phi1 is assigned with this phi before, then phi1 can be replaced with arg_node. for (auto &prev : prev_blocks_) { MS_EXCEPTION_IF_NULL(prev); if (!prev->matured_) { continue; } for (auto &phi_iter : prev->removable_phis_) { MS_EXCEPTION_IF_NULL(phi_iter.second); if (phi_iter.second->isa()) { const auto ¶m = phi_iter.second->cast(); if (param == phi) { MS_LOG(DEBUG) << "graph " << prev->func_graph_->ToString() << " var " << phi_iter.first->DebugString() << " can be replaced from " << param->DebugString() << " with " << arg_node->DebugString() << " in graph " << arg_node->func_graph()->ToString(); prev->removable_phis_[phi_iter.first] = arg_node; } } } } return true; } return false; } // A block should be marked matured if its predecessor blocks have been processed void FunctionBlock::Mature() { const auto &graphParamVec = func_graph_->parameters(); for (auto ¶mItr : graphParamVec) { MS_EXCEPTION_IF_NULL(paramItr); ParameterPtr param = paramItr->cast(); if (phi_nodes_.find(param) != phi_nodes_.cend()) { SetPhiArgument(param); } } matured_ = true; } // Force the conditIon node to bool using bool operation CNodePtr FunctionBlock::ForceToBoolNode(const AnfNodePtr &cond) { TraceManager::DebugTrace(std::make_shared(cond->debug_info())); CNodePtr op_apply_node = func_graph()->NewCNode({MakeResolveOperation(NAMED_PRIMITIVE_BOOL), cond}); TraceManager::EndTrace(); return op_apply_node; } CNodePtr FunctionBlock::ForceToWhileCond(const AnfNodePtr &cond) { TraceManager::DebugTrace(std::make_shared(cond->debug_info())); CNodePtr op_apply_node = func_graph()->NewCNode({MakeResolveOperation("while_cond"), cond}); TraceManager::EndTrace(); return op_apply_node; } // Perform a jump from this block to target block void FunctionBlock::Jump(const FunctionBlockPtr &target_block, AnfNodePtr node) { if (func_graph()->get_return() != nullptr) { MS_LOG(EXCEPTION) << "Failure: have return node! NodeInfo: " << trace::GetDebugInfo(func_graph()->get_return()->debug_info()); } std::vector input_nodes; input_nodes.emplace_back(NewValueNode(target_block->func_graph())); if (node != nullptr) { input_nodes.emplace_back(node); } CNodePtr jump = func_graph()->NewCNode(input_nodes); jumps_[target_block.get()] = jump; target_block->AddPrevBlock(shared_from_this()); func_graph()->set_output(jump); InsertDependItemsBeforeReturn(); } // Perform a conditional jump using switch operation. // The first CNode select graph with condition, and than execute this graph void FunctionBlock::ConditionalJump(AnfNodePtr condNode, const FunctionBlockPtr &true_block, const FunctionBlockPtr &false_block, bool unroll_loop) { if (func_graph()->get_return() != nullptr) { MS_LOG(EXCEPTION) << "Failure: have return node! NodeInfo: " << trace::GetDebugInfo(func_graph()->get_return()->debug_info()); } CNodePtr switch_app = func_graph()->NewCNode({NewValueNode(prim::kPrimSwitch), condNode, NewValueNode(true_block->func_graph()), NewValueNode(false_block->func_graph())}); CNodePtr switch_app_new = func_graph()->NewCNode({switch_app}); func_graph()->set_output(switch_app_new); InsertDependItemsBeforeReturn(); } void FunctionBlock::SetStateAssgin(const AnfNodePtr &target, const std::string &readid) { const std::string primitive_name("assign"); const std::string module_name("mindspore.ops.functional"); ValueNodePtr assign_op = NewValueNode(prim::GetPythonOps(primitive_name, module_name, true)); auto source = ReadVariable(readid); auto assign = func_graph()->NewCNode({assign_op, target, source}); WriteVariable(readid, assign); MS_LOG(INFO) << "SetState read " << target->DebugString() << ", " << readid; AddAutoDepend(assign); } void FunctionBlock::AddAutoDepend(const AnfNodePtr &target) { auto_depends_.push_back(target); } void FunctionBlock::InsertDependItemsBeforeReturn() { if (!prev_blocks_.empty()) { for (auto &prev_block : prev_blocks_) { MS_LOG(DEBUG) << "Has prev_block " << prev_block->func_graph()->debug_info().get(); } } ValueNodePtr make_tuple_op = NewValueNode(prim::kPrimMakeTuple); ValueNodePtr depend_op = NewValueNode(prim::kPrimDepend); ValueNodePtr stop_gradient_op = NewValueNode(prim::kPrimStopGradient); if (auto_depends_.size() == 0) { return; } AnfNodePtr state = nullptr; std::vector vec_states; vec_states.emplace_back(make_tuple_op); for (auto &item : auto_depends_) { MS_LOG(DEBUG) << "auto_depends " << item->ToString(); vec_states.emplace_back(item); } // if there are only make_tuple_op and another node in vec_states(the vec_states size is 2) // do not need to make_tuple, just use the node. if (vec_states.size() == 2) { state = vec_states[1]; } else { state = func_graph()->NewCNode(vec_states); } AnfNodePtr old_ret = nullptr; auto return_node = func_graph()->get_return(); if (return_node) { if (return_node->inputs().size() < 1) { MS_LOG(EXCEPTION) << "Length of inputs of output node is less than 2"; } old_ret = return_node->input(1); } else { old_ret = NewValueNode(kNone); } AnfNodePtr stopped = func_graph()->NewCNode({stop_gradient_op, state}); AnfNodePtr ret = func_graph()->NewCNode({depend_op, old_ret, stopped}); func_graph()->set_output(ret, true); } } // namespace parse } // namespace mindspore