| @@ -1312,7 +1312,7 @@ class IrParser { | |||||
| return nullptr; | return nullptr; | ||||
| } | } | ||||
| PrimitivePtr prim = std::make_shared<Primitive>("return"); | |||||
| PrimitivePtr prim = std::make_shared<Primitive>("Return"); | |||||
| ValueNodePtr input0 = std::make_shared<ValueNode>(prim); | ValueNodePtr input0 = std::make_shared<ValueNode>(prim); | ||||
| std::vector<AnfNodePtr> inputs; | std::vector<AnfNodePtr> inputs; | ||||
| inputs.push_back(input0); | inputs.push_back(input0); | ||||
| @@ -345,7 +345,7 @@ constexpr char REDUCED_SHAPE[] = "reduced_shape"; | |||||
| constexpr char TUPLE_DIV[] = "tuple_div"; | constexpr char TUPLE_DIV[] = "tuple_div"; | ||||
| constexpr char TUPLE_TO_ARRAY[] = "tuple_to_array"; | constexpr char TUPLE_TO_ARRAY[] = "tuple_to_array"; | ||||
| constexpr char VIRTUALLOSS[] = "VirtualLoss"; | constexpr char VIRTUALLOSS[] = "VirtualLoss"; | ||||
| constexpr char RETURN[] = "return"; | |||||
| constexpr char RETURN[] = "Return"; | |||||
| constexpr char ENV_GETITEM[] = "env_getitem"; | constexpr char ENV_GETITEM[] = "env_getitem"; | ||||
| constexpr char IDENTITY[] = "identity"; | constexpr char IDENTITY[] = "identity"; | ||||
| constexpr char PARTIAL[] = "partial"; | constexpr char PARTIAL[] = "partial"; | ||||
| @@ -181,7 +181,7 @@ constexpr auto kBNInferGradOpName = "BNInferGrad"; | |||||
| constexpr auto kCallOpName = "call"; | constexpr auto kCallOpName = "call"; | ||||
| constexpr auto kPartialOpName = "partial"; | constexpr auto kPartialOpName = "partial"; | ||||
| constexpr auto kSwitchOpName = "switch"; | constexpr auto kSwitchOpName = "switch"; | ||||
| constexpr auto kReturnOpName = "return"; | |||||
| constexpr auto kReturnOpName = "Return"; | |||||
| constexpr auto kLarsV2OpName = "LarsV2"; | constexpr auto kLarsV2OpName = "LarsV2"; | ||||
| constexpr auto kLarsV2UpdateOpName = "LarsV2Update"; | constexpr auto kLarsV2UpdateOpName = "LarsV2Update"; | ||||
| constexpr auto kSquareSumAllOpName = "SquareSumAll"; | constexpr auto kSquareSumAllOpName = "SquareSumAll"; | ||||
| @@ -63,7 +63,7 @@ using InstType = std::pair<Instruction, VectorRef>; | |||||
| using InstSet = std::vector<InstType>; | using InstSet = std::vector<InstType>; | ||||
| using InstFunctionMap = std::map<Instruction, std::function<void(const VectorRef &)>>; | using InstFunctionMap = std::map<Instruction, std::function<void(const VectorRef &)>>; | ||||
| const std::vector<std::string> inst_str{"call", "tail_call", "return", "partial", "switch", | |||||
| const std::vector<std::string> inst_str{"call", "tail_call", "Return", "partial", "switch", | |||||
| "switch_return", "tuple", "input", "external", "push", | "switch_return", "tuple", "input", "external", "push", | ||||
| "primitive", "graph", "pad_stack", "switch_layer"}; | "primitive", "graph", "pad_stack", "switch_layer"}; | ||||
| class StructPartial : public Base { | class StructPartial : public Base { | ||||
| @@ -402,7 +402,7 @@ inline const PrimitivePtr kPrimFloorMod = std::make_shared<Primitive>("FloorMod" | |||||
| inline const PrimitivePtr kPrimWhere = std::make_shared<Primitive>("Where"); | inline const PrimitivePtr kPrimWhere = std::make_shared<Primitive>("Where"); | ||||
| // Statements | // Statements | ||||
| inline const PrimitivePtr kPrimReturn = std::make_shared<Primitive>("return"); | |||||
| inline const PrimitivePtr kPrimReturn = std::make_shared<Primitive>("Return"); | |||||
| inline const PrimitivePtr kPrimSwitch = std::make_shared<Primitive>("switch"); | inline const PrimitivePtr kPrimSwitch = std::make_shared<Primitive>("switch"); | ||||
| inline const PrimitivePtr kPrimSwitchLayer = std::make_shared<Primitive>("switch_layer"); | inline const PrimitivePtr kPrimSwitchLayer = std::make_shared<Primitive>("switch_layer"); | ||||
| inline const PrimitivePtr kPrimAssign = std::make_shared<Primitive>("Assign"); | inline const PrimitivePtr kPrimAssign = std::make_shared<Primitive>("Assign"); | ||||
| @@ -26,7 +26,7 @@ namespace mindspore { | |||||
| static const std::set<std::string> PARALLEL_BLACK_LIST_ = {prim::kTupleGetItem, "J", "list_getitem", | static const std::set<std::string> PARALLEL_BLACK_LIST_ = {prim::kTupleGetItem, "J", "list_getitem", | ||||
| "array_getitem", "tuple_setitem", "Depend", "list_setitem", "array_setitem", "dict_getitem", | "array_getitem", "tuple_setitem", "Depend", "list_setitem", "array_setitem", "dict_getitem", | ||||
| "list_append", "list_map", "list_reduce", "tuple_reversed", "tile_shape", "tuple_div", "tuple_to_array", | "list_append", "list_map", "list_reduce", "tuple_reversed", "tile_shape", "tuple_div", "tuple_to_array", | ||||
| "make_dict", "make_slice", "make_record", "string_equal", "VirtualLoss", "return", "env_getitem", | |||||
| "make_dict", "make_slice", "make_record", "string_equal", "VirtualLoss", "Return", "env_getitem", | |||||
| "identity", "partial", "env_setitem", "env_getitem", "env_add", "MakeRefKey", "make_ref", "get_ref_key", | "identity", "partial", "env_setitem", "env_getitem", "env_add", "MakeRefKey", "make_ref", "get_ref_key", | ||||
| "get_ref_value", "get_ref_origin", "dot", "im2col", "col2im", "im2col_v1", "state_setitem", "ScalarSummary", | "get_ref_value", "get_ref_origin", "dot", "im2col", "col2im", "im2col_v1", "state_setitem", "ScalarSummary", | ||||
| "ImageSummary", "TensorSummary", "Debug", "HistogramSummary", "col2im_v1", "resolve", "BroadcastGradientArgs", | "ImageSummary", "TensorSummary", "Debug", "HistogramSummary", "col2im_v1", "resolve", "BroadcastGradientArgs", | ||||
| @@ -249,7 +249,7 @@ int AnfImporterFromMetaGraphT::AddReturnCNode() { | |||||
| op_inputs.emplace_back(make_tuple_cnode); | op_inputs.emplace_back(make_tuple_cnode); | ||||
| auto cnode = func_graph_->NewCNode(op_inputs); | auto cnode = func_graph_->NewCNode(op_inputs); | ||||
| MS_ASSERT(nullptr != cnode); | MS_ASSERT(nullptr != cnode); | ||||
| cnode->set_fullname_with_scope("return"); | |||||
| cnode->set_fullname_with_scope("Return"); | |||||
| func_graph_->set_return(cnode); | func_graph_->set_return(cnode); | ||||
| } else { | } else { | ||||
| auto return_prim_ptr = GetReturnPrim(); | auto return_prim_ptr = GetReturnPrim(); | ||||
| @@ -270,7 +270,7 @@ int AnfImporterFromMetaGraphT::AddReturnCNode() { | |||||
| MS_LOG(ERROR) << "NewCNode is nullptr"; | MS_LOG(ERROR) << "NewCNode is nullptr"; | ||||
| return RET_NULL_PTR; | return RET_NULL_PTR; | ||||
| } | } | ||||
| return_cnode->set_fullname_with_scope("return"); | |||||
| return_cnode->set_fullname_with_scope("Return"); | |||||
| func_graph_->set_return(return_cnode); | func_graph_->set_return(return_cnode); | ||||
| } | } | ||||
| return RET_OK; | return RET_OK; | ||||
| @@ -272,7 +272,7 @@ STATUS CaffeModelParser::ConvertGraphOutputs() { | |||||
| op_inputs.emplace_back(value_node); | op_inputs.emplace_back(value_node); | ||||
| op_inputs.emplace_back(make_tuple_cnode); | op_inputs.emplace_back(make_tuple_cnode); | ||||
| auto cnode = func_graph_ptr_->NewCNode(op_inputs); | auto cnode = func_graph_ptr_->NewCNode(op_inputs); | ||||
| cnode->set_fullname_with_scope("return"); | |||||
| cnode->set_fullname_with_scope("Return"); | |||||
| func_graph_ptr_->set_return(cnode); | func_graph_ptr_->set_return(cnode); | ||||
| } else { | } else { | ||||
| auto returnPrim = GetReturnPrim(); | auto returnPrim = GetReturnPrim(); | ||||
| @@ -293,7 +293,7 @@ STATUS CaffeModelParser::ConvertGraphOutputs() { | |||||
| } | } | ||||
| opInputs.emplace_back(cnode); | opInputs.emplace_back(cnode); | ||||
| auto returnCnode = func_graph_ptr_->NewCNode(opInputs); | auto returnCnode = func_graph_ptr_->NewCNode(opInputs); | ||||
| returnCnode->set_fullname_with_scope("return"); | |||||
| returnCnode->set_fullname_with_scope("Return"); | |||||
| func_graph_ptr_->set_return(returnCnode); | func_graph_ptr_->set_return(returnCnode); | ||||
| } | } | ||||
| return RET_OK; | return RET_OK; | ||||
| @@ -392,7 +392,7 @@ STATUS OnnxModelParser::BuildReturnNode(const FuncGraphPtr &anf_graph, const std | |||||
| MS_LOG(ERROR) << "new cnode error"; | MS_LOG(ERROR) << "new cnode error"; | ||||
| return RET_ERROR; | return RET_ERROR; | ||||
| } | } | ||||
| returnCnode->set_fullname_with_scope("return"); | |||||
| returnCnode->set_fullname_with_scope("Return"); | |||||
| anf_graph->set_return(returnCnode); | anf_graph->set_return(returnCnode); | ||||
| return RET_OK; | return RET_OK; | ||||
| } | } | ||||
| @@ -908,7 +908,7 @@ STATUS TFModelParser::MakeAnfGraphOutputs(std::vector<AnfNodePtr> *output_nodes, | |||||
| auto value_node = NewValueNode(return_prim_ptr); | auto value_node = NewValueNode(return_prim_ptr); | ||||
| std::vector<AnfNodePtr> op_inputs = {value_node, make_tuple_cnode}; | std::vector<AnfNodePtr> op_inputs = {value_node, make_tuple_cnode}; | ||||
| auto cnode = anf_graph->NewCNode(op_inputs); | auto cnode = anf_graph->NewCNode(op_inputs); | ||||
| cnode->set_fullname_with_scope("return"); | |||||
| cnode->set_fullname_with_scope("Return"); | |||||
| anf_graph->set_return(cnode); | anf_graph->set_return(cnode); | ||||
| } else { | } else { | ||||
| auto return_prim_ptr = GetReturnPrim(); | auto return_prim_ptr = GetReturnPrim(); | ||||
| @@ -919,7 +919,7 @@ STATUS TFModelParser::MakeAnfGraphOutputs(std::vector<AnfNodePtr> *output_nodes, | |||||
| auto value_node = NewValueNode(return_prim_ptr); | auto value_node = NewValueNode(return_prim_ptr); | ||||
| std::vector<AnfNodePtr> op_inputs{value_node, output_nodes->front()}; | std::vector<AnfNodePtr> op_inputs{value_node, output_nodes->front()}; | ||||
| auto return_cnode = anf_graph->NewCNode(op_inputs); | auto return_cnode = anf_graph->NewCNode(op_inputs); | ||||
| return_cnode->set_fullname_with_scope("return"); | |||||
| return_cnode->set_fullname_with_scope("Return"); | |||||
| anf_graph->set_return(return_cnode); | anf_graph->set_return(return_cnode); | ||||
| } | } | ||||
| return RET_OK; | return RET_OK; | ||||
| @@ -319,7 +319,7 @@ STATUS TfliteModelParser::ConvertGraphOutputs() { | |||||
| op_inputs.emplace_back(value_node); | op_inputs.emplace_back(value_node); | ||||
| op_inputs.emplace_back(make_tuple_cnode); | op_inputs.emplace_back(make_tuple_cnode); | ||||
| auto cnode = func_graph_->NewCNode(op_inputs); | auto cnode = func_graph_->NewCNode(op_inputs); | ||||
| cnode->set_fullname_with_scope("return"); | |||||
| cnode->set_fullname_with_scope("Return"); | |||||
| func_graph_->set_return(cnode); | func_graph_->set_return(cnode); | ||||
| } else { | } else { | ||||
| auto returnPrim = GetReturnPrim(); | auto returnPrim = GetReturnPrim(); | ||||
| @@ -339,7 +339,7 @@ STATUS TfliteModelParser::ConvertGraphOutputs() { | |||||
| } | } | ||||
| op_inputs.emplace_back(cnode); | op_inputs.emplace_back(cnode); | ||||
| auto returnCnode = func_graph_->NewCNode(op_inputs); | auto returnCnode = func_graph_->NewCNode(op_inputs); | ||||
| returnCnode->set_fullname_with_scope("return"); | |||||
| returnCnode->set_fullname_with_scope("Return"); | |||||
| func_graph_->set_return(returnCnode); | func_graph_->set_return(returnCnode); | ||||
| } | } | ||||
| return RET_OK; | return RET_OK; | ||||
| @@ -90,7 +90,7 @@ TEST_F(TestCloner, test_clone_simple) { | |||||
| Cloner cl2(gs); | Cloner cl2(gs); | ||||
| auto g3 = cl2[g]; | auto g3 = cl2[g]; | ||||
| std::vector<Primitive> results = {Primitive(prim::kScalarAdd), Primitive(prim::kScalarMul), Primitive("return")}; | |||||
| std::vector<Primitive> results = {Primitive(prim::kScalarAdd), Primitive(prim::kScalarMul), Primitive("Return")}; | |||||
| AnfNodeSet d3 = AnfNodeSet(DeepScopedGraphSearch(g3->get_return())); | AnfNodeSet d3 = AnfNodeSet(DeepScopedGraphSearch(g3->get_return())); | ||||
| common = d1 & d3; | common = d1 & d3; | ||||
| for (auto& x : common) { | for (auto& x : common) { | ||||
| @@ -315,7 +315,7 @@ TEST_F(TestOps, SwitchTest) { | |||||
| } | } | ||||
| TEST_F(TestOps, ReturnTest) { | TEST_F(TestOps, ReturnTest) { | ||||
| auto prim = std::make_shared<Primitive>("return"); | |||||
| auto prim = std::make_shared<Primitive>("Return"); | |||||
| ASSERT_EQ(prim->name(), kPrimReturn->name()); | ASSERT_EQ(prim->name(), kPrimReturn->name()); | ||||
| } | } | ||||
| @@ -348,17 +348,17 @@ TEST_F(TestStepParallel, CreatOpInstance1) { | |||||
| } | } | ||||
| TEST_F(TestStepParallel, OperatorInstance) { | TEST_F(TestStepParallel, OperatorInstance) { | ||||
| // creat attrs and prim | |||||
| // create attrs and prim | |||||
| PrimitivePtr prim = NewValueNode(prim::kPrimMatMul)->value()->cast<PrimitivePtr>(); | PrimitivePtr prim = NewValueNode(prim::kPrimMatMul)->value()->cast<PrimitivePtr>(); | ||||
| ValuePtr transpose_a = MakeValue(false); | ValuePtr transpose_a = MakeValue(false); | ||||
| ValuePtr transpose_b = MakeValue(false); | ValuePtr transpose_b = MakeValue(false); | ||||
| prim->set_attr("transpose_a", transpose_a); | prim->set_attr("transpose_a", transpose_a); | ||||
| prim->set_attr("transpose_b", transpose_b); | prim->set_attr("transpose_b", transpose_b); | ||||
| auto attrs = prim->attrs(); | auto attrs = prim->attrs(); | ||||
| // creat strategy | |||||
| // create strategy | |||||
| Strategys strategy = {{2, 2}, {2, 4}}; | Strategys strategy = {{2, 2}, {2, 4}}; | ||||
| StrategyPtr strategyPtr = parallel::NewStrategy(0, strategy); | StrategyPtr strategyPtr = parallel::NewStrategy(0, strategy); | ||||
| // creat shape | |||||
| // create shape | |||||
| Shapes inputs_shape = std::vector<Shape>{{64, 32}, {32, 64}}; | Shapes inputs_shape = std::vector<Shape>{{64, 32}, {32, 64}}; | ||||
| Shapes outputs_shape = std::vector<Shape>{{64, 64}}; | Shapes outputs_shape = std::vector<Shape>{{64, 64}}; | ||||
| std::vector<Shapes> shape = {inputs_shape, outputs_shape}; | std::vector<Shapes> shape = {inputs_shape, outputs_shape}; | ||||
| @@ -433,7 +433,7 @@ TEST_F(TestStepParallel, ForwardCommunication1) { | |||||
| } | } | ||||
| auto &inputs = node->cast<CNodePtr>()->inputs(); | auto &inputs = node->cast<CNodePtr>()->inputs(); | ||||
| PrimitivePtr prim = inputs[0]->cast<ValueNodePtr>()->value()->cast<PrimitivePtr>(); | PrimitivePtr prim = inputs[0]->cast<ValueNodePtr>()->value()->cast<PrimitivePtr>(); | ||||
| if (prim->name() == "return" || prim->name() == "MatMul") { | |||||
| if (prim->name() == "Return" || prim->name() == "MatMul") { | |||||
| if (!inputs[1]->isa<Parameter>()) { | if (!inputs[1]->isa<Parameter>()) { | ||||
| CNodePtr pre_node = inputs[1]->cast<CNodePtr>(); | CNodePtr pre_node = inputs[1]->cast<CNodePtr>(); | ||||
| PrimitivePtr pre_prim = pre_node->input(0)->cast<ValueNodePtr>()->value()->cast<PrimitivePtr>(); | PrimitivePtr pre_prim = pre_node->input(0)->cast<ValueNodePtr>()->value()->cast<PrimitivePtr>(); | ||||
| @@ -497,7 +497,7 @@ TEST_F(TestStepParallel, ForwardCommunication3) { | |||||
| } | } | ||||
| TEST_F(TestStepParallel, GetTensorInLayout) { | TEST_F(TestStepParallel, GetTensorInLayout) { | ||||
| // creat attrs and prim | |||||
| // create attrs and prim | |||||
| FuncGraphPtr func_graph = std::make_shared<FuncGraph>(); | FuncGraphPtr func_graph = std::make_shared<FuncGraph>(); | ||||
| Shape inputs_x_dims = {64, 32}; | Shape inputs_x_dims = {64, 32}; | ||||
| Shape inputs_y_dims = {32, 64}; | Shape inputs_y_dims = {32, 64}; | ||||
| @@ -511,10 +511,10 @@ TEST_F(TestStepParallel, GetTensorInLayout) { | |||||
| prim->set_attr("transpose_a", transpose_a); | prim->set_attr("transpose_a", transpose_a); | ||||
| prim->set_attr("transpose_b", transpose_b); | prim->set_attr("transpose_b", transpose_b); | ||||
| auto attrs = prim->attrs(); | auto attrs = prim->attrs(); | ||||
| // creat strategy | |||||
| // create strategy | |||||
| Strategys strategy = {{2, 2}, {2, 4}}; | Strategys strategy = {{2, 2}, {2, 4}}; | ||||
| StrategyPtr strategyPtr = parallel::NewStrategy(0, strategy); | StrategyPtr strategyPtr = parallel::NewStrategy(0, strategy); | ||||
| // creat shape | |||||
| // create shape | |||||
| Shapes inputs_shape = std::vector<Shape>{{64, 32}, {32, 64}}; | Shapes inputs_shape = std::vector<Shape>{{64, 32}, {32, 64}}; | ||||
| Shapes outputs_shape = std::vector<Shape>{{64, 64}}; | Shapes outputs_shape = std::vector<Shape>{{64, 64}}; | ||||
| std::vector<Shapes> shape = {inputs_shape, outputs_shape}; | std::vector<Shapes> shape = {inputs_shape, outputs_shape}; | ||||
| @@ -58,7 +58,7 @@ void TestSpecializeGraph::SetUp() { | |||||
| */ | */ | ||||
| graph_g_ = std::make_shared<FuncGraph>(); | graph_g_ = std::make_shared<FuncGraph>(); | ||||
| ParameterPtr y = graph_g_->add_parameter(); | ParameterPtr y = graph_g_->add_parameter(); | ||||
| auto prim_return = std::make_shared<Primitive>("return"); | |||||
| auto prim_return = std::make_shared<Primitive>("Return"); | |||||
| std::vector<AnfNodePtr> inputs; | std::vector<AnfNodePtr> inputs; | ||||
| inputs.push_back(NewValueNode(prim_return)); | inputs.push_back(NewValueNode(prim_return)); | ||||
| inputs.push_back(y); | inputs.push_back(y); | ||||
| @@ -101,7 +101,7 @@ void TestSpecializeGraph::SetUp() { | |||||
| inputs.push_back(y); | inputs.push_back(y); | ||||
| CNodePtr cnode_add = graph_beta_->NewCNode(inputs); | CNodePtr cnode_add = graph_beta_->NewCNode(inputs); | ||||
| inputs.clear(); | inputs.clear(); | ||||
| inputs.push_back(NewValueNode(std::make_shared<Primitive>("return"))); | |||||
| inputs.push_back(NewValueNode(std::make_shared<Primitive>("Return"))); | |||||
| inputs.push_back(cnode_add); | inputs.push_back(cnode_add); | ||||
| CNodePtr cnode_return = graph_beta_->NewCNode(inputs); | CNodePtr cnode_return = graph_beta_->NewCNode(inputs); | ||||
| graph_beta_->set_return(cnode_return); | graph_beta_->set_return(cnode_return); | ||||
| @@ -173,7 +173,7 @@ class MetaScalarAdd : public MetaFuncGraph { | |||||
| inputs.push_back(x); | inputs.push_back(x); | ||||
| inputs.push_back(y); | inputs.push_back(y); | ||||
| CNodePtr cnode_add = graph_g->NewCNode(inputs); | CNodePtr cnode_add = graph_g->NewCNode(inputs); | ||||
| auto prim_return = std::make_shared<Primitive>("return"); | |||||
| auto prim_return = std::make_shared<Primitive>("Return"); | |||||
| inputs.clear(); | inputs.clear(); | ||||
| inputs.push_back(NewValueNode(prim_return)); | inputs.push_back(NewValueNode(prim_return)); | ||||
| inputs.push_back(cnode_add); | inputs.push_back(cnode_add); | ||||
| @@ -202,7 +202,7 @@ void TestSpecializeMetaFuncGraph::SetUp() { | |||||
| inputs.push_back(x); | inputs.push_back(x); | ||||
| inputs.push_back(y); | inputs.push_back(y); | ||||
| CNodePtr cnode_add = graph_->NewCNode(inputs); | CNodePtr cnode_add = graph_->NewCNode(inputs); | ||||
| auto prim_return = std::make_shared<Primitive>("return"); | |||||
| auto prim_return = std::make_shared<Primitive>("Return"); | |||||
| inputs.clear(); | inputs.clear(); | ||||
| inputs.push_back(NewValueNode(prim_return)); | inputs.push_back(NewValueNode(prim_return)); | ||||
| inputs.push_back(cnode_add); | inputs.push_back(cnode_add); | ||||
| @@ -133,7 +133,7 @@ static KernelGraphPtr CreateKernelGraph() { | |||||
| next_cnode_ptr = kernelptr_floor; | next_cnode_ptr = kernelptr_floor; | ||||
| // return res | // return res | ||||
| auto p_return = std::make_shared<Primitive>("return"); | |||||
| auto p_return = std::make_shared<Primitive>("Return"); | |||||
| inputs.clear(); | inputs.clear(); | ||||
| inputs.push_back(NewValueNode(p_return)); | inputs.push_back(NewValueNode(p_return)); | ||||
| inputs.push_back(next_cnode_ptr); | inputs.push_back(next_cnode_ptr); | ||||
| @@ -171,7 +171,7 @@ TEST_F(TestConvert, TestConvertBatchNorm) { | |||||
| CNodePtr cnode_relu = anf_graph->NewCNode(inputs); | CNodePtr cnode_relu = anf_graph->NewCNode(inputs); | ||||
| inputs.clear(); | inputs.clear(); | ||||
| inputs.push_back(NewValueNode(std::make_shared<Primitive>("return"))); | |||||
| inputs.push_back(NewValueNode(std::make_shared<Primitive>("Return"))); | |||||
| inputs.push_back(cnode_relu); | inputs.push_back(cnode_relu); | ||||
| CNodePtr cnode_return = anf_graph->NewCNode(inputs); | CNodePtr cnode_return = anf_graph->NewCNode(inputs); | ||||
| anf_graph->set_return(cnode_return); | anf_graph->set_return(cnode_return); | ||||
| @@ -811,7 +811,7 @@ TEST_F(TestConvert, TestConvertMakeTuple) { | |||||
| } | } | ||||
| CNodePtr cnode_prim = func_graph->NewCNode(inputs); | CNodePtr cnode_prim = func_graph->NewCNode(inputs); | ||||
| inputs.clear(); | inputs.clear(); | ||||
| inputs.push_back(NewValueNode(std::make_shared<Primitive>("return"))); | |||||
| inputs.push_back(NewValueNode(std::make_shared<Primitive>("Return"))); | |||||
| inputs.push_back(cnode_prim); | inputs.push_back(cnode_prim); | ||||
| CNodePtr cnode_return = func_graph->NewCNode(inputs); | CNodePtr cnode_return = func_graph->NewCNode(inputs); | ||||
| func_graph->set_return(cnode_return); | func_graph->set_return(cnode_return); | ||||
| @@ -90,7 +90,7 @@ FuncGraphPtr MakeFuncGraph(const PrimitivePtr prim, unsigned int nparam) { | |||||
| } | } | ||||
| CNodePtr cnode_prim = func_graph->NewCNode(inputs); | CNodePtr cnode_prim = func_graph->NewCNode(inputs); | ||||
| inputs.clear(); | inputs.clear(); | ||||
| inputs.push_back(NewValueNode(std::make_shared<Primitive>("return"))); | |||||
| inputs.push_back(NewValueNode(std::make_shared<Primitive>("Return"))); | |||||
| inputs.push_back(cnode_prim); | inputs.push_back(cnode_prim); | ||||
| CNodePtr cnode_return = func_graph->NewCNode(inputs); | CNodePtr cnode_return = func_graph->NewCNode(inputs); | ||||
| func_graph->set_return(cnode_return); | func_graph->set_return(cnode_return); | ||||