You can not select more than 25 topics Topics must start with a chinese character,a letter or number, can include dashes ('-') and can be up to 35 characters long.

convert.cc 66 kB

6 years ago
6 years ago
5 years ago
5 years ago
123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371137213731374137513761377137813791380138113821383138413851386138713881389139013911392139313941395139613971398139914001401140214031404140514061407140814091410141114121413141414151416141714181419142014211422142314241425142614271428142914301431143214331434143514361437143814391440144114421443144414451446144714481449145014511452145314541455145614571458145914601461146214631464146514661467146814691470147114721473147414751476147714781479148014811482148314841485148614871488148914901491149214931494149514961497149814991500150115021503150415051506150715081509151015111512151315141515151615171518151915201521152215231524152515261527152815291530153115321533153415351536153715381539154015411542154315441545154615471548154915501551155215531554155515561557155815591560156115621563156415651566156715681569157015711572157315741575157615771578157915801581158215831584158515861587158815891590159115921593159415951596159715981599160016011602160316041605160616071608160916101611161216131614161516161617161816191620162116221623162416251626162716281629163016311632163316341635163616371638163916401641164216431644164516461647164816491650165116521653165416551656165716581659166016611662166316641665166616671668166916701671167216731674167516761677167816791680168116821683168416851686168716881689169016911692169316941695169616971698169917001701170217031704170517061707170817091710171117121713171417151716171717181719172017211722172317241725172617271728172917301731173217331734173517361737173817391740174117421743174417451746174717481749175017511752175317541755175617571758175917601761176217631764176517661767176817691770177117721773177417751776177717781779178017811782178317841785178617871788178917901791179217931794179517961797179817991800180118021803180418051806180718081809181018111812181318141815181618171818181918201821182218231824182518261827182818291830
  1. /**
  2. * Copyright 2019-2021 Huawei Technologies Co., Ltd
  3. *
  4. * Licensed under the Apache License, Version 2.0 (the "License");
  5. * you may not use this file except in compliance with the License.
  6. * You may obtain a copy of the License at
  7. *
  8. * http://www.apache.org/licenses/LICENSE-2.0
  9. *
  10. * Unless required by applicable law or agreed to in writing, software
  11. * distributed under the License is distributed on an "AS IS" BASIS,
  12. * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  13. * See the License for the specific language governing permissions and
  14. * limitations under the License.
  15. */
  16. #include "transform/graph_ir/convert.h"
  17. #include <inttypes.h>
  18. #include <algorithm>
  19. #include <stack>
  20. #include "utils/utils.h"
  21. #include "base/core_ops.h"
  22. #include "frontend/operator/ops.h"
  23. #include "utils/log_adapter.h"
  24. #include "ir/graph_utils.h"
  25. #include "utils/symbolic.h"
  26. #include "utils/config_manager.h"
  27. #include "utils/convert_utils.h"
  28. #include "utils/ms_context.h"
  29. #include "transform/graph_ir/op_adapter_map.h"
  30. #include "ops/state_ops.h"
  31. #include "ops/array_ops.h"
  32. #include "ops/elewise_calculation_ops.h"
  33. #include "ops/math_ops.h"
  34. #ifdef ENABLE_GE
  35. #include "ops/save_ops.h"
  36. #endif
  37. namespace mindspore {
  38. namespace transform {
  39. using std::endl;
  40. using ge::Operator;
  41. using mindspore::kAnyValue;
  42. using std::make_shared;
  43. using std::shared_ptr;
  44. using std::string;
  45. using std::vector;
  46. using Variable = ge::op::Variable;
  47. using Constant = ge::op::Constant;
  48. using Assign = ge::op::Assign;
  49. using Data = ge::op::Data;
  50. namespace {
  51. std::vector<AnfNodePtr> GetOrderedCNodes(const FuncGraphPtr fg) {
  52. auto BelongSameGraph = std::bind(IncludeBelongGraph, fg, std::placeholders::_1);
  53. auto succ_include_fv = [&fg](const AnfNodePtr &node) -> std::vector<AnfNodePtr> {
  54. std::vector<AnfNodePtr> vecs;
  55. if (node == nullptr) {
  56. return vecs;
  57. }
  58. if (node->isa<CNode>()) {
  59. auto cnode = node->cast<CNodePtr>();
  60. auto &inputs = cnode->inputs();
  61. // Check if free variables used.
  62. for (const auto &input : inputs) {
  63. auto input_fg = GetValueNode<FuncGraphPtr>(input);
  64. if (input_fg) {
  65. for (auto &fv : input_fg->free_variables_nodes()) {
  66. if (fv->func_graph() == fg && fg->nodes().contains(fv)) {
  67. vecs.push_back(fv);
  68. }
  69. }
  70. }
  71. }
  72. (void)vecs.insert(vecs.end(), inputs.begin(), inputs.end());
  73. }
  74. return vecs;
  75. };
  76. return TopoSort(fg->get_return(), succ_include_fv, BelongSameGraph);
  77. }
  78. } // namespace
  79. // ---------------implement of DfGraphConvertor-------------
  80. PrimType GetCNodeFuncType(const CNodePtr cnode) {
  81. if (cnode->inputs().empty()) {
  82. return kPrimTypeUnknown;
  83. }
  84. AnfNodePtr valuenode = cnode->input(0);
  85. if (IsValueNode<Primitive>(valuenode)) {
  86. // check whether the valuenode is primitive
  87. return GetValueNode<PrimitivePtr>(valuenode)->prim_type();
  88. }
  89. return kPrimTypeUnknown;
  90. }
  91. bool IsCaseNode(const CNodePtr node) {
  92. if (!node->inputs().empty() && node->input(0)->isa<CNode>() &&
  93. GetCNodeFuncName(node->input(0)->cast<CNodePtr>()) == "switch_layer") {
  94. return true;
  95. }
  96. return false;
  97. }
  98. std::string GetCNodeTargetFuncName(const CNodePtr cnode) {
  99. if (IsCaseNode(cnode)) {
  100. return string(kNameCase);
  101. }
  102. auto name = GetCNodeFuncName(cnode);
  103. if (name == "switch_layer") {
  104. name = "";
  105. }
  106. return name;
  107. }
  108. OpAdapterPtr DfGraphConvertor::FindAdapter(const AnfNodePtr node, bool train) {
  109. if (node->isa<CNode>()) {
  110. auto cnode = node->cast<CNodePtr>();
  111. std::string name = kNameCustomOp;
  112. if (!IsCustomCNode(cnode)) {
  113. name = GetCNodeTargetFuncName(cnode);
  114. }
  115. auto it_adpt = OpAdapterMap::get().find(name);
  116. if (it_adpt != OpAdapterMap::get().end()) {
  117. return it_adpt->second->Get(train);
  118. }
  119. MS_LOG(EXCEPTION) << "Can't find OpAdapter for " << name;
  120. }
  121. if (node->isa<ValueNode>()) {
  122. return OpAdapterMap::get()[kNameConst]->Get(train);
  123. }
  124. if (node->isa<Parameter>()) {
  125. return OpAdapterMap::get()[kNameParam]->Get(train);
  126. }
  127. return OpAdapterPtr(nullptr);
  128. }
  129. void DfGraphConvertor::InitLoopVar(std::vector<ge::Operator> *init_input) {
  130. if (this->training_) {
  131. GeTensorDesc desc(GeShape(), ge::FORMAT_NCHW, ge::DT_INT64);
  132. auto var_iter_num = std::make_shared<Variable>("npu_runconfig/iterations_per_loop");
  133. auto var_loop_cond = std::make_shared<Variable>("npu_runconfig/loop_cond");
  134. auto var_one = std::make_shared<Variable>("npu_runconfig/one");
  135. auto var_zero = std::make_shared<Variable>("npu_runconfig/zero");
  136. (void)var_iter_num->update_output_desc_y(desc);
  137. (void)var_loop_cond->update_output_desc_y(desc);
  138. (void)var_one->update_output_desc_y(desc);
  139. (void)var_zero->update_output_desc_y(desc);
  140. vars_["npu_runconfig/iterations_per_loop"] = var_iter_num;
  141. vars_["npu_runconfig/loop_cond"] = var_loop_cond;
  142. vars_["npu_runconfig/one"] = var_one;
  143. vars_["npu_runconfig/zero"] = var_zero;
  144. int64_t value = 0;
  145. auto const_iter_num = std::make_shared<Constant>("const/npu_runconfig/iterations_per_loop");
  146. if (ConfigManager::GetInstance().dataset_mode() == DS_SINK_MODE) {
  147. value = ConfigManager::GetInstance().iter_num();
  148. } else {
  149. MS_LOG(INFO) << "Run with normal(non-sink) mode, the iterator number will always be 1";
  150. value = 1;
  151. ConfigManager::GetInstance().set_iter_num(value);
  152. }
  153. value -= 1; // iteration start from 0, the max iteration number for n loop should be n-1
  154. (void)const_iter_num->set_attr_value(GeTensor(desc, reinterpret_cast<uint8_t *>(&value), sizeof(int64_t)));
  155. auto const_loop_cond = std::make_shared<Constant>("const/npu_runconfig/loop_cond");
  156. value = 0;
  157. (void)const_loop_cond->set_attr_value(GeTensor(desc, reinterpret_cast<uint8_t *>(&value), sizeof(int64_t)));
  158. auto const_one = std::make_shared<Constant>("const/npu_runconfig/one");
  159. value = 1;
  160. (void)const_one->set_attr_value(GeTensor(desc, reinterpret_cast<uint8_t *>(&value), sizeof(int64_t)));
  161. auto const_zero = std::make_shared<Constant>("const/npu_runconfig/zero");
  162. value = 0;
  163. (void)const_zero->set_attr_value(GeTensor(desc, reinterpret_cast<uint8_t *>(&value), sizeof(int64_t)));
  164. (void)const_iter_num->update_output_desc_y(desc);
  165. (void)const_loop_cond->update_output_desc_y(desc);
  166. (void)const_one->update_output_desc_y(desc);
  167. (void)const_zero->update_output_desc_y(desc);
  168. auto assign_iter_num = std::make_shared<Assign>("assign/npu_runconfig/iterations_per_loop");
  169. (void)assign_iter_num->set_input_ref(*var_iter_num).set_input_value(*const_iter_num);
  170. auto assign_loop_cond = std::make_shared<Assign>("assign/npu_runconfig/loop_cond");
  171. (void)assign_loop_cond->set_input_ref(*var_loop_cond).set_input_value(*const_loop_cond);
  172. auto assign_one = std::make_shared<Assign>("assign/npu_runconfig/one");
  173. (void)assign_one->set_input_ref(*var_one).set_input_value(*const_one);
  174. auto assign_zero = std::make_shared<Assign>("assign/npu_runconfig/zero");
  175. (void)assign_zero->set_input_ref(*var_zero).set_input_value(*const_zero);
  176. init_input->push_back(*var_iter_num);
  177. init_input->push_back(*var_loop_cond);
  178. init_input->push_back(*var_one);
  179. init_input->push_back(*var_zero);
  180. init_ops_.push_back(var_iter_num);
  181. init_ops_.push_back(var_loop_cond);
  182. init_ops_.push_back(var_one);
  183. init_ops_.push_back(var_zero);
  184. init_ops_.push_back(const_iter_num);
  185. init_ops_.push_back(const_loop_cond);
  186. init_ops_.push_back(const_one);
  187. init_ops_.push_back(const_zero);
  188. init_ops_.push_back(assign_iter_num);
  189. init_ops_.push_back(assign_loop_cond);
  190. init_ops_.push_back(assign_one);
  191. init_ops_.push_back(assign_zero);
  192. }
  193. }
  194. OpAdapterPtr DfGraphConvertor::FindAdapter(const std::string &name, bool train) {
  195. auto it = OpAdapterMap::get().find(name);
  196. if (it != OpAdapterMap::get().end()) {
  197. return it->second->Get(train);
  198. }
  199. MS_LOG(EXCEPTION) << "Can't find OpAdapter for " << name;
  200. }
  201. void DfGraphConvertor::DrawParamInitSubGraph(const std::string &name, const AnfNodePtr &it) {
  202. // draw init subgraph
  203. init_sout_ << "op_assign" << it.get() << "[label=<";
  204. init_sout_ << "<table border='1' cellborder='1'>" << endl;
  205. init_sout_ << "<tr>";
  206. init_sout_ << "<td port='1'>resource</td>";
  207. init_sout_ << "<td port='2'>value</td>";
  208. init_sout_ << "</tr>" << endl;
  209. init_sout_ << "<tr><td colspan=\"2\">"
  210. << "\"assign_" << name << "\"</td></tr>" << endl;
  211. init_sout_ << "</table>> shape=plaintext]" << endl;
  212. init_sout_ << "param" << it.get() << "[shape=octagon, label=\"" << name << "\"]" << endl;
  213. init_sout_ << "const" << it.get() << "[label= \"" << name << "_const"
  214. << "\" shape=ellipse]" << endl;
  215. init_sout_ << "param" << it.get() << "->"
  216. << "op_assign" << it.get() << ":1" << endl;
  217. init_sout_ << "const" << it.get() << "->"
  218. << "op_assign" << it.get() << ":2" << endl;
  219. }
  220. void DfGraphConvertor::SetupParamInitSubGraph(const TensorOrderMap &tensors, std::vector<ge::Operator> *init_input) {
  221. DfGraphPtr init_graph = std::make_shared<DfGraph>("init");
  222. std::vector<AnfNodePtr> nodes = GetOrderedCNodes(anf_graph_);
  223. for (auto &it : nodes) {
  224. if (it->isa<ValueNode>()) {
  225. if (IsValueNode<SymbolicKeyInstance>(it)) {
  226. auto symbolic = GetValueNode<SymbolicKeyInstancePtr>(it);
  227. auto name = std::static_pointer_cast<Parameter>(symbolic->node())->name();
  228. auto iter = vars_.find(name); // get corresponding variable op
  229. if (iter != vars_.end()) {
  230. op_cache_[it.get()] = iter->second;
  231. // #ifdef DRAW_GE_GRAPH
  232. compute_sout_ << op_draw_name_[params_[name].get()] << " -> " << op_draw_name_[it.get()]
  233. << "[style=\"dotted\"]" << endl;
  234. // #endif
  235. }
  236. } else if (IsValueNode<RefKey>(it)) {
  237. auto refkey = GetValueNode<RefKeyPtr>(it);
  238. auto name = refkey->tag();
  239. auto iter = vars_.find(name); // get corresponding variable op
  240. if (iter != vars_.end()) {
  241. op_cache_[it.get()] = iter->second;
  242. compute_sout_ << op_draw_name_[params_[name].get()] << " -> " << op_draw_name_[it.get()]
  243. << "[style=\"dotted\"]" << endl;
  244. }
  245. }
  246. }
  247. }
  248. for (auto &it : tensors) {
  249. if (vars_.find(it.first) == vars_.end()) {
  250. MS_LOG(WARNING) << "Init parameter " << it.first << " didn't appear in graph.";
  251. vars_[it.first] = nullptr;
  252. }
  253. }
  254. // set up init sub graph
  255. if (init_input->size()) {
  256. // init sub graph needs no input
  257. MS_LOG(INFO) << "Build data init subgraph.";
  258. (void)init_graph->SetInputs(*init_input);
  259. this->init_graph_ = init_graph;
  260. } else {
  261. this->init_graph_ = nullptr;
  262. }
  263. }
  264. void DfGraphConvertor::MakeDatasetHandler(const std::string &name, const size_t &input_idx, const AnfNodePtr &it) {
  265. MS_LOG(INFO) << "The " << name << " is the " << input_idx << "(st/nd/th) input";
  266. if (ConfigManager::GetInstance().dataset_mode() == DS_SINK_MODE) {
  267. auto getnext_idx = static_cast<int64_t>(input_idx);
  268. DatasetGraphParam param = ConfigManager::GetInstance().dataset_param();
  269. if (!param.input_indexes().empty() && input_idx <= param.input_indexes().size()) {
  270. getnext_idx = param.input_indexes()[input_idx] - 1; // input_idx start from 0.
  271. MS_LOG(INFO) << "remap input_index:" << input_idx << " to getnext_index:" << getnext_idx << ".";
  272. }
  273. // use iterator_getnext op with output_name instead of data op in BuildGraph.
  274. out_handle_cache_[it.get()] = OutHandler(dataset_iter_getnext_, "y" + std::to_string(getnext_idx));
  275. }
  276. }
  277. void DfGraphConvertor::SetupBroadcast(const std::shared_ptr<HcomBroadcast> &broadcast,
  278. const std::vector<GeTensorDesc> &broadcast_desc,
  279. const DfGraphPtr &broadcast_graph, std::vector<ge::Operator> broadcast_input) {
  280. MS_LOG(INFO) << "build broadcast subgraph";
  281. if (broadcast_desc.size() != broadcast_input.size()) {
  282. MS_LOG(EXCEPTION) << "Desc number of BroadCast is not equal to number of Input";
  283. }
  284. (void)broadcast->create_dynamic_input_x(static_cast<unsigned int>(broadcast_input.size()));
  285. (void)broadcast->create_dynamic_output_y(static_cast<unsigned int>(broadcast_desc.size()));
  286. for (unsigned int i = 0; i < broadcast_input.size(); i++) {
  287. (void)broadcast->set_dynamic_input_x(i, broadcast_input[i]);
  288. (void)broadcast->update_dynamic_output_desc_y(i, broadcast_desc[i]);
  289. }
  290. (void)broadcast_graph->SetInputs(broadcast_input);
  291. this->broadcast_graph_ = broadcast_graph;
  292. }
  293. void DfGraphConvertor::InitParamWithData(const TensorOrderMap &tensors) {
  294. int index = 0;
  295. std::vector<Operator> init_input;
  296. for (auto it : tensors) {
  297. std::string name = it.first;
  298. auto node_itor = params_.find(name);
  299. // if name not in params_, create a node in graph
  300. if (node_itor == params_.end()) {
  301. MS_LOG(WARNING) << name << " is not in params, and create a new node.";
  302. ParameterPtr param = std::make_shared<Parameter>(nullptr);
  303. name = name + "_temp";
  304. param->set_name(name);
  305. (void)ConvertParameter(param);
  306. node_itor = params_.find(name);
  307. }
  308. auto node = node_itor->second;
  309. auto op_itor = op_cache_.find(node.get());
  310. if (op_itor == op_cache_.end()) {
  311. MS_LOG(EXCEPTION) << "Can not find op for node " << node->ToString() << ".";
  312. }
  313. auto adpt = FindAdapter(kNameParam, training_);
  314. if (adpt == nullptr) continue;
  315. auto param_op = adpt->generate(name + "_data");
  316. MS_LOG(INFO) << "Add parameter " << name << " as input, index " << index << ".";
  317. if (!training_) {
  318. auto adpt_const = FindAdapter(kNameConst, training_);
  319. if (adpt_const == nullptr) continue;
  320. auto const_op = adpt_const->generate(name + "_const");
  321. (void)adpt_const->setAttr(const_op, "value", it.second);
  322. auto const_op_desc = TransformUtil::GetGeTensorDesc(it.second->shape_c(), it.second->data_type(), kOpFormat_NCHW);
  323. if (const_op_desc == nullptr) {
  324. MS_LOG(ERROR) << "Create variable " << name << " output descriptor failed!";
  325. continue;
  326. }
  327. (void)std::static_pointer_cast<Constant>(const_op)->update_output_desc_y(*const_op_desc);
  328. vars_[name] = const_op;
  329. op_itor->second = const_op;
  330. continue;
  331. }
  332. // create tensor descriptor for output descriptor
  333. auto desc = TransformUtil::GetGeTensorDesc(it.second->shape_c(), it.second->data_type(), kOpFormat_NCHW);
  334. if (desc == nullptr) {
  335. MS_LOG(ERROR) << "Create variable " << name << " output descriptor failed!";
  336. continue;
  337. }
  338. // we need three variable ops for each graph with same name
  339. // build init subgraph
  340. if (it.second->is_init() == 0) {
  341. (void)std::static_pointer_cast<Data>(param_op)->set_attr_index(index++);
  342. auto init_var = std::make_shared<Variable>(name);
  343. auto assign_op = std::make_shared<Assign>("assign_" + name);
  344. (void)init_var->update_output_desc_y(*desc);
  345. (void)assign_op->set_input_ref(*init_var).set_input_value(*param_op);
  346. init_input.push_back(*init_var);
  347. init_ops_.push_back(param_op);
  348. init_ops_.push_back(assign_op);
  349. init_ops_.push_back(init_var);
  350. }
  351. auto variable = std::make_shared<Variable>(name);
  352. (void)variable->update_output_desc_y(*desc);
  353. // do not use read variable while variable sink
  354. MS_LOG(DEBUG) << "InitParam, op_name = " << name << ", var = " << variable->GetName() << ".";
  355. op_itor->second = variable; // replace parameter with variable
  356. vars_[name] = variable; // prevent the variable operator from being freed
  357. DrawParamInitSubGraph(name, node);
  358. }
  359. InitLoopVar(&init_input);
  360. SetupParamInitSubGraph(tensors, &init_input);
  361. }
  362. // convert all parameter need initialize to variable
  363. DfGraphConvertor &DfGraphConvertor::InitParam(const TensorOrderMap &tensors) {
  364. size_t input_idx = 0;
  365. if (error_ != 0) {
  366. return *this;
  367. }
  368. if (anf_graph_ == nullptr || anf_graph_->output() == nullptr) {
  369. error_ = INVALID_ARGUMENT;
  370. MS_LOG(ERROR) << "Invalid AnfGraph in InitParam.";
  371. return *this;
  372. }
  373. // Processing input with MakeDatasetHandler
  374. for (auto &it : anf_graph_->parameters()) {
  375. auto op_itor = op_cache_.find(it.get()); // converted node
  376. if (it->isa<Parameter>() && op_itor != op_cache_.end()) {
  377. string name = std::static_pointer_cast<Parameter>(it)->name();
  378. auto tensor_itor = tensors.find(name); // in init value map
  379. if (tensor_itor == tensors.end()) {
  380. DfGraphConvertor::MakeDatasetHandler(name, input_idx, it);
  381. input_idx++;
  382. }
  383. }
  384. }
  385. InitParamWithData(tensors);
  386. init_sout_ << "}" << endl;
  387. return *this;
  388. }
  389. #if (defined ENABLE_GE)
  390. void DfGraphConvertor::BuildSaveCheckpointGraph() {
  391. std::vector<Operator> graph_inputs;
  392. ge::op::Save save_op("save_parms");
  393. int save_op_is_active = 0;
  394. size_t index = 0;
  395. string name;
  396. int32_t count_size = std::count_if(vars_.begin(), vars_.end(), [](const std::pair<std::string, OperatorPtr> &it) {
  397. return (it.second == nullptr || it.first.find("/") != std::string::npos);
  398. });
  399. (void)save_op.create_dynamic_input_tensors(vars_.size() - static_cast<size_t>(count_size));
  400. // for each "parameter" in anf graph excluding "input"
  401. for (const auto &it : vars_) {
  402. name = it.first;
  403. if (it.second == nullptr || name.find("/") != std::string::npos) continue;
  404. Variable variable(name);
  405. (void)variable.update_output_desc_y(it.second->GetOutputDesc(0));
  406. (void)save_op.set_dynamic_input_tensors(index++, variable);
  407. graph_inputs.push_back(variable);
  408. if (save_op_is_active == 0) {
  409. checkpoint_sout_ << "op_save" << &save_op << "[label=<";
  410. checkpoint_sout_ << "<table border='1' cellborder='1'>" << endl;
  411. checkpoint_sout_ << "<tr><td port='1'>tensor</td></tr>" << endl;
  412. checkpoint_sout_ << "<tr><td colspan=\"1\">"
  413. << "\"saveop"
  414. << "\"</td></tr>" << endl;
  415. checkpoint_sout_ << "</table>> shape=plaintext]" << endl;
  416. }
  417. checkpoint_sout_ << "param" << it.second << "[shape=octagon, label=\"" << name << "\"]" << endl;
  418. checkpoint_sout_ << "param" << it.second << "->"
  419. << "op_save" << &save_op << ":1" << endl;
  420. save_op_is_active = 1;
  421. }
  422. if (save_op_is_active) {
  423. std::vector<Operator> graph_output;
  424. graph_output.emplace_back(save_op);
  425. DfGraphPtr checkpoint_graph = std::make_shared<DfGraph>("checkpoint");
  426. (void)checkpoint_graph->SetInputs(graph_inputs);
  427. (void)checkpoint_graph->SetOutputs(graph_output);
  428. this->save_ckp_graph_ = checkpoint_graph;
  429. } else {
  430. this->save_ckp_graph_ = nullptr;
  431. }
  432. checkpoint_sout_ << "}" << endl;
  433. return;
  434. }
  435. #endif
  436. DfGraphConvertor &DfGraphConvertor::GenerateBroadcastGraph(const TensorOrderMap &tensors) {
  437. if (error_ != 0) {
  438. return *this;
  439. }
  440. if (anf_graph_ == nullptr || anf_graph_->output() == nullptr) {
  441. error_ = INVALID_ARGUMENT;
  442. MS_LOG(ERROR) << "Invalid AnfGraph in generate broadcast graph";
  443. return *this;
  444. }
  445. DfGraphPtr broadcast_graph = std::make_shared<DfGraph>("broadcast");
  446. // collect the operators create for broadcast sub graph, in order to avoid auto release
  447. std::vector<Operator> broadcast_input;
  448. std::vector<GeTensorDesc> broadcast_desc;
  449. auto broadcast = std::make_shared<HcomBroadcast>("broadcast_parameter");
  450. (void)broadcast->set_attr_root_rank(0);
  451. (void)broadcast->set_attr_group("hccl_world_group");
  452. broadcast_ops_.push_back(broadcast);
  453. // find every parameter, build broadcast subgraph (or initialize the parameter with constant)
  454. for (auto &it : anf_graph_->parameters()) {
  455. auto op_itor = op_cache_.find(it.get()); // converted node
  456. if (it->isa<Parameter>() && op_itor != op_cache_.end()) {
  457. string name = std::static_pointer_cast<Parameter>(it)->name();
  458. auto tensor_itor = tensors.find(name); // in init tensor map
  459. if (tensor_itor != tensors.end()) {
  460. auto tensor = tensor_itor->second;
  461. auto shape_ge = tensor->shape_c();
  462. // create tensor descriptor for output descriptor
  463. auto desc = TransformUtil::GetGeTensorDesc(shape_ge, tensor->data_type(), kOpFormat_NCHW);
  464. if (desc == nullptr) {
  465. MS_LOG(ERROR) << "Create variable " << name << " output descriptor failed!";
  466. continue;
  467. }
  468. // build broadcast subgraph
  469. if (distribute_) {
  470. auto broadcast_var = std::make_shared<Variable>(name);
  471. (void)broadcast_var->update_output_desc_y(*desc);
  472. broadcast_input.push_back(*broadcast_var);
  473. broadcast_desc.push_back(*desc);
  474. broadcast_ops_.push_back(broadcast_var);
  475. }
  476. }
  477. }
  478. }
  479. // set up broadcast sub graph
  480. if (!broadcast_input.empty()) {
  481. DfGraphConvertor::SetupBroadcast(broadcast, broadcast_desc, broadcast_graph, broadcast_input);
  482. } else {
  483. this->broadcast_graph_ = nullptr;
  484. }
  485. return *this;
  486. }
  487. DfGraphConvertor &DfGraphConvertor::GenerateCheckpointGraph() {
  488. if (error_ != 0) {
  489. MS_LOG(ERROR) << "Generate checkpoint graph failed, found error code " << error_ << ".";
  490. return *this;
  491. }
  492. if (anf_graph_ == nullptr || anf_graph_->output() == nullptr) {
  493. error_ = INVALID_ARGUMENT;
  494. MS_LOG(ERROR) << "Invalid AnfGraph in GenerateCheckpointGraph";
  495. return *this;
  496. }
  497. #if (defined ENABLE_GE)
  498. BuildSaveCheckpointGraph();
  499. // Restoring from checkpoint file is done by pyfront, not in graph now.
  500. #endif
  501. return *this;
  502. }
  503. DfGraphConvertor &DfGraphConvertor::ConvertAllNode() {
  504. if (error_ != 0) {
  505. return *this;
  506. }
  507. if (anf_graph_ == nullptr || anf_graph_->output() == nullptr) {
  508. MS_LOG(ERROR) << "Invalid AnfGraph";
  509. error_ = FAILED;
  510. return *this;
  511. }
  512. compute_sout_.clear();
  513. compute_sout_ << "digraph {" << endl;
  514. init_sout_.clear();
  515. init_sout_ << "digraph {" << endl;
  516. #if (defined ENABLE_GE)
  517. checkpoint_sout_.clear();
  518. checkpoint_sout_ << "digraph {" << endl;
  519. #endif
  520. restore_checkpoint_sout_.clear();
  521. restore_checkpoint_sout_ << "digraph {" << endl;
  522. // Convert all anf node to Operator
  523. MS_LOG(DEBUG) << "convert all node";
  524. std::vector<AnfNodePtr> nodes = GetOrderedCNodes(anf_graph_);
  525. for (auto &it : nodes) {
  526. (void)Convert(it);
  527. if (this->error_ != 0) {
  528. MS_LOG(ERROR) << "failed to convert node: " << it->DebugString() << ".";
  529. }
  530. }
  531. // Create dataset iterator and iterator_getnext node
  532. if (ConfigManager::GetInstance().dataset_mode() == DS_SINK_MODE) {
  533. DatasetGraphParam param = ConfigManager::GetInstance().dataset_param();
  534. MS_LOG(INFO) << "Dataset param is " << param.ToString() << ".";
  535. // GetNext
  536. auto iter_getnext_op = make_shared<ge::op::GetNext>("get_next_tmp");
  537. (void)iter_getnext_op->set_attr_output_types(param.ge_types());
  538. (void)iter_getnext_op->set_attr_output_shapes(param.shapes());
  539. (void)iter_getnext_op->set_attr_channel_name(param.queue_name());
  540. // save iter_getnext_op for later use
  541. dataset_iter_getnext_ = iter_getnext_op;
  542. }
  543. // return the data flow graph
  544. return *this;
  545. }
  546. void DfGraphConvertor::TraceOutputFromTupleGetItem(const AnfNodePtr &anf_out) {
  547. auto it = out_handle_cache_.find(anf_out.get());
  548. if (it != out_handle_cache_.end()) {
  549. OutHandler handle = it->second;
  550. auto op = handle.op;
  551. if (op != nullptr) {
  552. MS_LOG(INFO) << "op name: " << op->GetName() << ", op type: " << op->GetOpType() << ", out_name: " << handle.out;
  553. graph_outputs_.emplace_back(std::make_pair(*op, handle.out));
  554. } else {
  555. MS_LOG(EXCEPTION) << "tuple_getitem: " << anf_out->fullname_with_scope() << " is not converted";
  556. }
  557. } else {
  558. // invalid tuple_getitem e.g. tuple_getitem(tuple_getitem())/tuple_getitem(depend())/tuple_getitem(make_tuple())
  559. MS_LOG(WARNING) << "Invalid tuple_getitem: " << anf_out->fullname_with_scope();
  560. }
  561. }
  562. void DfGraphConvertor::TraceOutput(const AnfNodePtr node) {
  563. AnfNodePtr anf_out = node;
  564. AnfNodePtr pre_node = nullptr;
  565. // Trace value node
  566. if (node->isa<ValueNode>()) {
  567. auto op = Convert(anf_out);
  568. if (op != nullptr) {
  569. graph_outputs_.emplace_back(std::make_pair(*op, ""));
  570. AddGraphConstInput(op);
  571. }
  572. return;
  573. }
  574. // Trace Parameter node
  575. TraceOutputFromParameter(anf_out);
  576. // Then trace cnode
  577. if (!node->isa<CNode>()) {
  578. return;
  579. }
  580. // trace tuple_getitem
  581. while (anf_out->isa<CNode>() && IsPrimitiveCNode(anf_out, prim::kPrimTupleGetItem)) {
  582. pre_node = anf_out;
  583. anf_out = anf_out->cast<CNodePtr>()->input(1);
  584. }
  585. // trace every element of make_tuple
  586. auto c = anf_out->cast<CNodePtr>();
  587. std::string name = "";
  588. if (anf_out->isa<CNode>()) {
  589. name = GetCNodeTargetFuncName(c);
  590. }
  591. if (name == "MakeTuple") {
  592. for (unsigned int i = 1; i < c->inputs().size(); i++) {
  593. TraceOutput(c->input(i));
  594. }
  595. } else if (name == prim::kPrimDepend->name()) {
  596. if (c->inputs().size() < 3) { // "Depend" primitive have 3 inputs
  597. MS_LOG(EXCEPTION) << "length of inputs is " << c->inputs().size() << ", which is less than 3";
  598. }
  599. TraceOutput(c->input(1));
  600. } else if (name == prim::kTupleGetItem) {
  601. TraceOutputFromTupleGetItem(anf_out);
  602. } else {
  603. // add outputs;
  604. auto op = Convert(anf_out);
  605. std::string index;
  606. if (op != nullptr) {
  607. if ((pre_node != nullptr) && IsPrimitiveCNode(pre_node, prim::kPrimTupleGetItem)) {
  608. auto item = out_handle_cache_.find(pre_node.get());
  609. if (item != out_handle_cache_.end()) {
  610. index = item->second.out;
  611. } else {
  612. MS_LOG(WARNING) << "Can't get operator: " << anf_out->fullname_with_scope() << " 's output item";
  613. }
  614. }
  615. MS_LOG(INFO) << "Add graph output: " << anf_out->fullname_with_scope() << ":" << index;
  616. graph_outputs_.emplace_back(make_pair(*op, index));
  617. }
  618. }
  619. }
  620. void DfGraphConvertor::TraceOutputFromParameter(const AnfNodePtr &anf_out) {
  621. if (anf_out->isa<Parameter>()) {
  622. MS_LOG(INFO) << "Add graph output: " << anf_out->fullname_with_scope();
  623. auto it = out_handle_cache_.find(anf_out.get());
  624. if (it != out_handle_cache_.end()) {
  625. // For dataset graph mode, input parameter is converted to a "iterator_get_next:yn" OutHandler.
  626. OutHandler handle = it->second;
  627. auto op = handle.op;
  628. MS_LOG(INFO) << "op name: " << op->GetName() << ", op type: " << op->GetOpType() << ", out_name: " << handle.out;
  629. graph_outputs_.emplace_back(make_pair(*op, handle.out));
  630. } else {
  631. // common parameter case
  632. auto op = Convert(anf_out);
  633. if (op != nullptr) {
  634. MS_LOG(INFO) << "op name: " << op->GetName() << ", op type: " << op->GetOpType();
  635. graph_outputs_.emplace_back(std::make_pair(*op, ""));
  636. }
  637. }
  638. }
  639. }
  640. void SetupDatasetIterGetNextNode(const OperatorPtr &op) {
  641. if (ConfigManager::GetInstance().dataset_mode() == DS_SINK_MODE) {
  642. DatasetGraphParam param = ConfigManager::GetInstance().dataset_param();
  643. size_t output_num = param.ge_types().size();
  644. MS_LOG(INFO) << "Set iterator_getnext op's output num = " << output_num << ".";
  645. // set iterator_getnext op's output num
  646. shared_ptr<ge::op::GetNext> iter_getnext = std::static_pointer_cast<ge::op::GetNext>(op);
  647. (void)iter_getnext->create_dynamic_output_y(static_cast<unsigned int>(output_num));
  648. for (uint32_t i = 0; i < output_num; i++) {
  649. ge::TensorDesc desc(GeShape(param.shapes()[i]), ge::FORMAT_NCHW, (ge::DataType)param.ge_types()[i]);
  650. // we don't SetRealDimCnt here since GE do not use this output's real-dim
  651. (void)iter_getnext->update_dynamic_output_desc_y((i), desc);
  652. }
  653. }
  654. return;
  655. }
  656. void DfGraphConvertor::SetSubgraph(AnfNodePtr node) {
  657. if (!node->isa<CNode>()) {
  658. return;
  659. }
  660. auto cnode = node->cast<CNodePtr>();
  661. if (!IsCaseNode(cnode)) {
  662. return;
  663. }
  664. std::vector<AnfNodePtr> case_inputs;
  665. for (size_t i = 1; i < cnode->inputs().size(); i++) {
  666. case_inputs.emplace_back(cnode->input(i));
  667. }
  668. std::shared_ptr<std::vector<DfGraph>> branches = std::make_shared<std::vector<DfGraph>>();
  669. auto bnode = cnode->input(0)->cast<CNodePtr>()->input(2)->cast<CNodePtr>();
  670. for (size_t i = 1; i < bnode->inputs().size(); i++) {
  671. auto branch_node = bnode->input(i)->cast<CNodePtr>();
  672. for (size_t j = 2; j < branch_node->inputs().size(); j++) {
  673. if (std::find(case_inputs.begin(), case_inputs.end(), branch_node->input(j)) == case_inputs.end()) {
  674. case_inputs.emplace_back(branch_node->input(j));
  675. }
  676. }
  677. }
  678. for (size_t i = 1; i < bnode->inputs().size(); i++) {
  679. ProcessSubgraph(bnode->input(i), case_inputs);
  680. }
  681. for (size_t i = 1; i < bnode->inputs().size(); i++) {
  682. branches->emplace_back(branches_map_[bnode->input(i).get()]);
  683. }
  684. if (op_cache_.find(node.get()) == op_cache_.end()) {
  685. return;
  686. }
  687. OpAdapterPtr adpt = FindAdapter(node, training_);
  688. if (nullptr == adpt) {
  689. MS_LOG(DEBUG) << "Not found adapter";
  690. return;
  691. }
  692. OperatorPtr op = Convert(node);
  693. adpt->setSubgraph(op, 0, branches);
  694. return;
  695. }
  696. void DfGraphConvertor::GetCaseNodeInput(const CNodePtr node, const CNodePtr input_node) {
  697. std::vector<AnfNodePtr> case_inputs;
  698. for (size_t i = 1; i < node->inputs().size(); i++) {
  699. case_inputs.emplace_back(node->input(i));
  700. }
  701. auto bnode = input_node->input(2)->cast<CNodePtr>();
  702. for (size_t i = 1; i < bnode->inputs().size(); i++) {
  703. auto branch_node = bnode->input(i)->cast<CNodePtr>();
  704. for (size_t j = 2; j < branch_node->inputs().size(); j++) {
  705. if (std::find(case_inputs.begin(), case_inputs.end(), branch_node->input(j)) == case_inputs.end()) {
  706. case_inputs.emplace_back(branch_node->input(j));
  707. }
  708. }
  709. }
  710. const size_t case_index = 1;
  711. const size_t make_tuple_index = 2;
  712. AnfNodePtr case_index_iter = input_node->input(case_index);
  713. AnfNodePtr make_tuple_iter = input_node->input(make_tuple_index);
  714. auto make_tuple_node = make_tuple_iter->cast<CNodePtr>();
  715. std::shared_ptr<std::vector<OutHandler>> tuple_items = std::make_shared<std::vector<OutHandler>>();
  716. for (size_t i = 0; i < case_inputs.size(); i++) {
  717. auto item = case_inputs[i];
  718. auto op = Convert(item);
  719. if (op != nullptr) {
  720. tuple_items->emplace_back(OutHandler(op, "", item));
  721. } else if (out_handle_cache_.find(item.get()) != out_handle_cache_.end()) {
  722. tuple_items->push_back(out_handle_cache_[item.get()]);
  723. } else {
  724. MS_LOG(DEBUG) << "Add an empty out handler: " << item->ToString();
  725. tuple_items->push_back(OutHandler());
  726. }
  727. }
  728. tuple_out_handle_cache_[make_tuple_node.get()] = tuple_items;
  729. std::shared_ptr<std::vector<AnfNodePtr>> case_input_items = std::make_shared<std::vector<AnfNodePtr>>();
  730. case_input_items->emplace_back(case_index_iter);
  731. case_input_items->emplace_back(make_tuple_iter);
  732. case_input_handle_cache_[node.get()] = case_input_items;
  733. }
  734. void DfGraphConvertor::UpdateTupleOutCache() {
  735. for (auto &it : tuple_out_handle_cache_) {
  736. std::size_t len = it.second->size();
  737. for (std::size_t i = 0; i < len; i++) {
  738. OutHandler handle = (*it.second)[i];
  739. if (handle.op == nullptr) {
  740. continue;
  741. }
  742. string name = handle.op->GetName();
  743. if (vars_.count(name) && (vars_[name] != nullptr)) {
  744. (*it.second)[i] = OutHandler(vars_[name], handle.out, handle.node);
  745. MS_LOG(INFO) << "update tuple_out_handle_cache_ " << name;
  746. }
  747. }
  748. }
  749. }
  750. DfGraphConvertor &DfGraphConvertor::BuildGraph() {
  751. SetupDatasetIterGetNextNode(dataset_iter_getnext_);
  752. if (error_ != 0) {
  753. return *this;
  754. }
  755. // Case node set input.
  756. std::vector<AnfNodePtr> nodes = GetOrderedCNodes(anf_graph_);
  757. for (auto &it : nodes) {
  758. if (it->isa<CNode>() && IsCaseNode(it->cast<CNodePtr>())) {
  759. auto node = it->cast<CNodePtr>();
  760. auto input_node = node->input(0)->cast<CNodePtr>();
  761. GetCaseNodeInput(node, input_node);
  762. }
  763. }
  764. // update tuple_out_handle_cache_
  765. UpdateTupleOutCache();
  766. // set up dependencies
  767. MS_LOG(DEBUG) << "set up dependencies";
  768. nodes = GetOrderedCNodes(anf_graph_);
  769. for (auto &it : nodes) {
  770. SetNodeInput(it);
  771. SetOpControlInput(it);
  772. SetSubgraph(it);
  773. UpdateOpDesc(it);
  774. }
  775. if (error_ == 0) {
  776. df_graph_ = make_shared<DfGraph>(anf_graph_->ToString());
  777. } else {
  778. return *this;
  779. }
  780. // set graph input according to the order from anf graph
  781. std::vector<Operator> inputs;
  782. if (ConfigManager::GetInstance().dataset_mode() == DS_SINK_MODE) {
  783. inputs.push_back(*dataset_iter_getnext_);
  784. } else {
  785. auto params = anf_graph_->parameters();
  786. if (use_inputs_) {
  787. params = inputs_;
  788. auto anf_params = anf_graph_->parameters();
  789. for (size_t i = 0; i < params.size(); i++) {
  790. for (size_t j = 0; j < anf_params.size(); j++) {
  791. if (params[i]->ToString() == anf_params[j]->ToString()) {
  792. params[i] = anf_params[j];
  793. }
  794. }
  795. }
  796. }
  797. int index = 0;
  798. for (auto &it : params) {
  799. auto name = std::static_pointer_cast<Parameter>(it)->name();
  800. // the parameters which has not been converted to var
  801. if (vars_.find(name) == vars_.end()) {
  802. auto op = Convert(it);
  803. MS_EXCEPTION_IF_NULL(op);
  804. MS_LOG(INFO) << "add not var input " << it->ToString() << ", index " << index;
  805. if (op == nullptr) {
  806. MS_LOG(ERROR) << "Convert graph failed!";
  807. return *this;
  808. }
  809. UpdateDataOpDesc(it, op);
  810. MS_LOG(INFO) << "add input " << it->ToString() << ", index " << index;
  811. (void)std::static_pointer_cast<Data>(op)->set_attr_index(index++);
  812. inputs.push_back(*op);
  813. } else if (vars_[name] != nullptr) {
  814. MS_LOG(INFO) << "add var input " << it->ToString();
  815. auto op = Convert(it);
  816. MS_EXCEPTION_IF_NULL(op);
  817. inputs.push_back(*op);
  818. }
  819. }
  820. }
  821. MS_LOG(DEBUG) << "trace output";
  822. graph_outputs_.clear();
  823. TraceOutput(anf_graph_->get_return()->input(1));
  824. // Add const nodes as graph input for some operator work with constant
  825. MS_LOG(INFO) << "graph const input size: " << graph_const_inputs_.size();
  826. std::transform(graph_const_inputs_.begin(), graph_const_inputs_.end(), std::back_inserter(inputs),
  827. [](OperatorPtr x) { return *x; });
  828. MS_LOG(INFO) << "set graph input num: " << inputs.size();
  829. (void)df_graph_->SetInputs(inputs);
  830. // set graph output
  831. // set the value of finale return apply node as the output of dataflow graph
  832. MS_LOG(DEBUG) << "set output";
  833. MS_LOG(INFO) << "set graph output num: " << graph_outputs_.size();
  834. (void)df_graph_->SetOutputs(graph_outputs_);
  835. compute_sout_ << "}" << endl;
  836. // For the graph(e.g. eval_subgraph) whose IterNum is 1, donot set NeedIteration flag.
  837. if (ConfigManager::GetInstance().iter_num() > 1) {
  838. df_graph_->SetNeedIteration(true);
  839. }
  840. return *this;
  841. }
  842. void DfGraphConvertor::UpdateDataOpDesc(const AnfNodePtr &it, const OperatorPtr &op) const {
  843. auto node = std::static_pointer_cast<AnfNode>(it);
  844. if (node == nullptr) {
  845. MS_LOG(ERROR) << "Update data op descriptor failed! Invalid node.";
  846. return;
  847. }
  848. auto normal_shape_ptr = dyn_cast<abstract::Shape>(node->Shape());
  849. std::vector<int64_t> shape;
  850. if (normal_shape_ptr == nullptr) {
  851. MS_LOG(INFO) << "Invalid shape to update data op descriptor.";
  852. return;
  853. }
  854. shape = normal_shape_ptr->shape();
  855. if (node->Type() == nullptr) {
  856. MS_LOG(INFO) << "Invalid type to update data op descriptor.";
  857. return;
  858. }
  859. TypeId me_type = node->Type()->type_id();
  860. if (kObjectTypeTensorType == me_type) {
  861. me_type = dyn_cast<TensorType>(node->Type())->element()->type_id();
  862. }
  863. std::ostringstream buf;
  864. buf << "[" << shape << "]";
  865. MS_LOG(INFO) << "input shape is " << buf.str() << ", type is " << me_type;
  866. auto desc = TransformUtil::GetGeTensorDesc(shape, me_type, "NCHW");
  867. if (desc == nullptr) {
  868. MS_LOG(ERROR) << "Update data op descriptor failed! TensorDesc is null.";
  869. } else {
  870. (void)std::static_pointer_cast<Data>(op)->update_input_desc_x(*desc);
  871. (void)std::static_pointer_cast<Data>(op)->update_output_desc_y(*desc);
  872. }
  873. }
  874. DfGraphPtr DfGraphConvertor::GetComputeGraph() { return df_graph_; }
  875. DfGraphPtr DfGraphConvertor::GetInitGraph() { return init_graph_; }
  876. DfGraphPtr DfGraphConvertor::GetSaveCheckpointGraph() { return save_ckp_graph_; }
  877. DfGraphPtr DfGraphConvertor::GetBroadcastGraph() { return broadcast_graph_; }
  878. bool DfGraphConvertor::IsSourceEdgeNode(const AnfNodePtr &node) {
  879. if (!node->isa<CNode>()) {
  880. return false;
  881. }
  882. auto cnode = node->cast<CNodePtr>();
  883. if (!IsCustomCNode(cnode)) {
  884. std::string name = GetCNodeTargetFuncName(cnode);
  885. if (name.empty()) {
  886. return false;
  887. }
  888. // Ignore apply node Depend, UpdateState, make_tuple. make_tuple in ge pipeline.
  889. if ((name == prim::kPrimDepend->name()) || (name == prim::kPrimUpdateState->name()) ||
  890. (name == prim::kPrimReturn->name()) || (name == prim::kPrimMakeTuple->name())) {
  891. return false;
  892. }
  893. }
  894. // Load and other normal primitives which contain monad node.
  895. auto has_monad = std::any_of(cnode->inputs().begin(), cnode->inputs().end(),
  896. [](const AnfNodePtr &node) -> bool { return HasAbstractMonad(node); });
  897. if (has_monad) {
  898. return true;
  899. }
  900. // primitive with make_tuple as input
  901. for (auto &input : cnode->inputs()) {
  902. if (IsPrimitiveCNode(input, prim::kPrimMakeTuple)) {
  903. auto tuple = input->cast<CNodePtr>();
  904. auto ret = std::any_of(tuple->inputs().begin(), tuple->inputs().end(),
  905. [](const AnfNodePtr &node) -> bool { return HasAbstractMonad(node); });
  906. if (ret) {
  907. return true;
  908. }
  909. }
  910. }
  911. return false;
  912. }
  913. bool DfGraphConvertor::IsControlEdgeNode(const AnfNodePtr &node) {
  914. if (!node->isa<CNode>()) {
  915. return false;
  916. }
  917. auto cnode = node->cast<CNodePtr>();
  918. if (!IsCustomCNode(cnode)) {
  919. std::string name = GetCNodeTargetFuncName(cnode);
  920. if (name.empty()) {
  921. return false;
  922. }
  923. // Ignore apply node of Load, Depend, UpdateState, make_tuple, return
  924. if ((name == prim::kPrimLoad->name()) || (name == prim::kPrimDepend->name()) ||
  925. (name == prim::kPrimUpdateState->name()) || (name == prim::kPrimMakeTuple->name()) ||
  926. (name == prim::kPrimReturn->name())) {
  927. return false;
  928. }
  929. }
  930. return true;
  931. }
  932. OperatorPtr DfGraphConvertor::ToOperatorPtr(const AnfNodePtr &node) {
  933. auto op = Convert(GetRealOpNode(node));
  934. if (op == nullptr) {
  935. MS_LOG(ERROR) << "Convert real op node to operator failed, " << node->ToString();
  936. error_ = FAILED;
  937. return nullptr;
  938. }
  939. return op;
  940. }
  941. void DfGraphConvertor::AddEdgeToCache(const AnfNodePtr &src, const AnfNodePtr &dest) {
  942. auto item = monad_control_edge_cache_.find(src);
  943. if (item == monad_control_edge_cache_.end()) {
  944. monad_control_edge_cache_[src] = std::set<AnfNodePtr>{dest};
  945. } else {
  946. item->second.insert(dest);
  947. }
  948. }
  949. void DfGraphConvertor::AddEdgeForLoad(const AnfNodePtr &node) {
  950. auto func_graph = node->func_graph();
  951. MS_EXCEPTION_IF_NULL(func_graph);
  952. auto mng = func_graph->manager();
  953. if (mng == nullptr) {
  954. mng = Manage(func_graph, true);
  955. func_graph->set_manager(mng);
  956. }
  957. auto manager = func_graph->manager();
  958. MS_EXCEPTION_IF_NULL(manager);
  959. auto &users = manager->node_users()[node];
  960. std::shared_ptr<std::vector<AnfNodePtr>> src_node_list = std::make_shared<std::vector<AnfNodePtr>>();
  961. std::shared_ptr<std::vector<AnfNodePtr>> dst_node_list = std::make_shared<std::vector<AnfNodePtr>>();
  962. for (const auto &iter : users) {
  963. auto user_node = iter.first;
  964. auto name = GetCNodeTargetFuncName(user_node->cast<CNodePtr>());
  965. if (name == prim::kPrimUpdateState->name()) {
  966. FindDestOps(user_node, dst_node_list, false);
  967. continue;
  968. }
  969. if (IsControlEdgeNode(user_node)) {
  970. src_node_list->push_back(user_node);
  971. continue;
  972. }
  973. FindDestOps(user_node, src_node_list, false);
  974. }
  975. // add to cache
  976. for (auto &dest : *dst_node_list) {
  977. for (auto &src : *src_node_list) {
  978. AddEdgeToCache(src, dest);
  979. }
  980. }
  981. }
  982. void DfGraphConvertor::FindDestOps(const AnfNodePtr &node, const std::shared_ptr<std::vector<AnfNodePtr>> &node_list,
  983. bool top) {
  984. auto func_graph = node->func_graph();
  985. MS_EXCEPTION_IF_NULL(func_graph);
  986. auto mng = func_graph->manager();
  987. if (mng == nullptr) {
  988. mng = Manage(func_graph, true);
  989. func_graph->set_manager(mng);
  990. }
  991. auto manager = func_graph->manager();
  992. MS_EXCEPTION_IF_NULL(manager);
  993. auto users = manager->node_users()[node];
  994. for (const auto &iter : users) {
  995. auto user_node = iter.first;
  996. if (IsControlEdgeNode(user_node)) {
  997. if (!top) {
  998. node_list->push_back(user_node);
  999. }
  1000. } else {
  1001. FindDestOps(user_node, node_list, false);
  1002. }
  1003. }
  1004. }
  1005. void DfGraphConvertor::AutoMonadCollectInput(const AnfNodePtr &node) {
  1006. if (!IsSourceEdgeNode(node)) {
  1007. return;
  1008. }
  1009. // Add control edge if contain monad input.
  1010. std::string name = GetCNodeTargetFuncName(node->cast<CNodePtr>());
  1011. if (name == prim::kPrimLoad->name()) {
  1012. AddEdgeForLoad(node);
  1013. } else {
  1014. auto src_ops = ToOperatorPtr(node);
  1015. if (src_ops != nullptr) {
  1016. // Find dest ops list
  1017. std::shared_ptr<std::vector<AnfNodePtr>> dst_node_list = std::make_shared<std::vector<AnfNodePtr>>();
  1018. FindDestOps(node, dst_node_list, true);
  1019. for (auto &dest : *dst_node_list) {
  1020. AddEdgeToCache(node, dest);
  1021. }
  1022. }
  1023. }
  1024. }
  1025. void DfGraphConvertor::AutoMonadSetInput(const AnfNodePtr &node) {
  1026. if (monad_control_edge_cache_.find(node) == monad_control_edge_cache_.end()) {
  1027. return;
  1028. }
  1029. auto src_ops = ToOperatorPtr(node);
  1030. if (src_ops != nullptr) {
  1031. for (auto &dest : monad_control_edge_cache_[node]) {
  1032. auto dest_ops = ToOperatorPtr(dest);
  1033. if (dest_ops == nullptr) {
  1034. continue;
  1035. }
  1036. (void)dest_ops->AddControlInput(*src_ops);
  1037. #ifdef DRAW_GE_GRAPH
  1038. compute_sout_ << op_draw_name_[node.get()] << " -> " << op_draw_name_[dest.get()] << "[style=\"dotted\"]" << endl;
  1039. #endif
  1040. }
  1041. }
  1042. }
  1043. void DfGraphConvertor::AutoMonadSetControlInput(const AnfNodePtr &node) {
  1044. AutoMonadCollectInput(node);
  1045. AutoMonadSetInput(node);
  1046. }
  1047. void DfGraphConvertor::SetOpControlInput(const AnfNodePtr &node) {
  1048. AutoMonadSetControlInput(node);
  1049. if (control_edge_cache_.find(node.get()) == control_edge_cache_.end()) {
  1050. return;
  1051. }
  1052. std::vector<ControlEdge> control_edges = control_edge_cache_[node.get()];
  1053. if ((control_edges.empty())) {
  1054. MS_LOG(ERROR) << "Get control edge node's src or dest operator failed";
  1055. return;
  1056. }
  1057. for (auto &item : control_edges) {
  1058. (void)item.dest_op->AddControlInput(*item.src_op);
  1059. }
  1060. }
  1061. const std::vector<std::string> trans_var_list = {string(kNameAssign), string(kNameAssignAdd), string(kNameAssignSub)};
  1062. AnfNodePtr DfGraphConvertor::ParseLoadInput(const CNodePtr &cnode) {
  1063. if (cnode->inputs().size() < 3) {
  1064. MS_LOG(EXCEPTION) << "input size error, " << cnode->ToString();
  1065. }
  1066. const size_t para_index = 1;
  1067. return cnode->input(para_index);
  1068. }
  1069. void DfGraphConvertor::SetTupleOpInput(const OpAdapterPtr &adpt, const CNodePtr &node, const AnfNodePtr &pred,
  1070. const OperatorPtr &src, int index) {
  1071. std::shared_ptr<std::vector<OutHandler>> handler_vec = tuple_out_handle_cache_[pred.get()];
  1072. std::shared_ptr<std::vector<OutHandler>> handler_vec_without_monad = std::make_shared<std::vector<OutHandler>>();
  1073. bool with_monad = false;
  1074. for (auto &handler : *handler_vec) {
  1075. // when tuple with monad type element, the handler operator is nullptr, should be ignored.
  1076. if (handler.op == nullptr) {
  1077. if ((handler.node != nullptr) && !HasAbstractMonad(handler.node)) {
  1078. MS_LOG(WARNING) << "Unsupported node in tuple : " << node->ToString();
  1079. }
  1080. continue;
  1081. }
  1082. with_monad = true;
  1083. handler_vec_without_monad->push_back(handler);
  1084. }
  1085. int ret = adpt->setInput(src, index, handler_vec_without_monad);
  1086. if ((ret == 0) && pred->isa<CNode>() && (pred->cast<CNodePtr>()->inputs().size() == handler_vec->size() + 1)) {
  1087. for (unsigned int j = 0; j < handler_vec_without_monad->size(); j++) {
  1088. AnfNodePtr input_node = pred->cast<CNodePtr>()->input(j + 1);
  1089. if (with_monad) {
  1090. input_node = handler_vec_without_monad->at(j).node;
  1091. }
  1092. compute_sout_ << op_draw_name_[input_node.get()] << " -> " << op_draw_name_[node.get()] << ":" << index << endl;
  1093. AddGraphConstInput(handler_vec_without_monad->at(j).op);
  1094. }
  1095. return;
  1096. }
  1097. MS_LOG(WARNING) << "This anf node is not supported as a tuple item : " << node->ToString();
  1098. }
  1099. AnfNodePtr DfGraphConvertor::GetRealInputNode(const CNodePtr &node, const AnfNodePtr &input) {
  1100. if (input == nullptr || node == nullptr) {
  1101. return nullptr;
  1102. }
  1103. AnfNodePtr pred = input;
  1104. while (pred->isa<CNode>() && GetCNodeTargetFuncName(pred->cast<CNodePtr>()) == prim::kPrimDepend->name()) {
  1105. pred = pred->cast<CNodePtr>()->input(1);
  1106. }
  1107. // skip input of UMonad, IOMonad
  1108. if (IsValueNode<UMonad>(pred) || IsValueNode<IOMonad>(pred)) {
  1109. return nullptr;
  1110. }
  1111. // skip input of the None, UpdateState
  1112. if (IsValueNode<None>(pred) || IsPrimitiveCNode(pred, prim::kPrimUpdateState)) {
  1113. return nullptr;
  1114. }
  1115. if (IsPrimitiveCNode(pred, prim::kPrimLoad)) {
  1116. pred = ParseLoadInput(pred->cast<CNodePtr>());
  1117. }
  1118. // transform "Const" op to "Variable" op when the next node is "Assign" op.
  1119. std::string c_name = GetCNodeTargetFuncName(node);
  1120. auto pos = std::find(trans_var_list.begin(), trans_var_list.end(), c_name);
  1121. if (!training_ && pos != trans_var_list.end() && pred->isa<Parameter>()) {
  1122. std::string name = std::static_pointer_cast<Parameter>(pred)->name();
  1123. auto op_itor = op_cache_.find(pred.get());
  1124. if (op_itor == op_cache_.end()) {
  1125. MS_LOG(EXCEPTION) << "Can not find op for node " << pred->ToString() << ".";
  1126. }
  1127. if (op_itor->second != nullptr &&
  1128. (op_itor->second->GetOpType() == "Constant" || op_itor->second->GetOpType() == "Const") &&
  1129. vars_.find(name) != vars_.end()) {
  1130. auto variable = std::make_shared<Variable>(name);
  1131. auto desc = vars_[name]->GetOutputDesc("y");
  1132. (void)variable->update_output_desc_y(desc);
  1133. MS_LOG(DEBUG) << "Trans to variable, var = " << variable->GetName() << ".";
  1134. op_itor->second = variable; // replace parameter with variable
  1135. vars_[name] = variable;
  1136. }
  1137. }
  1138. return pred;
  1139. }
  1140. void DfGraphConvertor::SetOpInput(const OpAdapterPtr &adpt, const CNodePtr &node) {
  1141. OperatorPtr src = Convert(node);
  1142. int case_flag = 0;
  1143. auto &inputs = node->inputs();
  1144. size_t input_size = inputs.size();
  1145. if (case_input_handle_cache_.find(node.get()) != case_input_handle_cache_.end()) {
  1146. case_flag = 1;
  1147. input_size = case_input_handle_cache_[node.get()]->size() + 1;
  1148. }
  1149. for (size_t i = 1; i < input_size; i++) {
  1150. AnfNodePtr pred = nullptr;
  1151. if (case_flag != 0) {
  1152. pred = case_input_handle_cache_[node.get()]->at(i - 1);
  1153. } else {
  1154. pred = inputs[i];
  1155. }
  1156. pred = GetRealInputNode(node, pred);
  1157. if (pred == nullptr) {
  1158. continue;
  1159. }
  1160. int index = SizeToInt(i);
  1161. // find in out_hadnle_cache_ first
  1162. auto it = out_handle_cache_.find(pred.get());
  1163. if (it != out_handle_cache_.end()) {
  1164. int ret = adpt->setInput(src, index, it->second);
  1165. if (ret == 0) {
  1166. if (pred->isa<CNode>() && GetCNodeTargetFuncName(pred->cast<CNodePtr>()) == prim::kTupleGetItem) {
  1167. compute_sout_ << op_draw_name_[pred->cast<CNodePtr>()->input(1).get()] << " -> " << op_draw_name_[node.get()]
  1168. << ":" << i << endl;
  1169. } else if (pred->isa<Parameter>()) {
  1170. compute_sout_ << op_draw_name_[pred.get()] << " -> " << op_draw_name_[node.get()] << ":" << i << endl;
  1171. } else {
  1172. // don't draw anything.
  1173. MS_LOG(INFO) << "DRAW_GE_GRAPH: Shouldn't have this case.";
  1174. }
  1175. AddGraphConstInput(it->second.op);
  1176. }
  1177. } else if (tuple_out_handle_cache_.find(pred.get()) != tuple_out_handle_cache_.end()) {
  1178. SetTupleOpInput(adpt, node, pred, src, index);
  1179. } else {
  1180. auto op = Convert(pred);
  1181. int ret = adpt->setInput(src, index, op);
  1182. if (ret == 0) {
  1183. compute_sout_ << op_draw_name_[pred.get()] << " -> " << op_draw_name_[node.get()] << ":" << i << endl;
  1184. AddGraphConstInput(op);
  1185. }
  1186. }
  1187. }
  1188. }
  1189. void DfGraphConvertor::AddGraphConstInput(const OperatorPtr &op) {
  1190. if (op->GetOpType() == "Constant" || op->GetOpType() == "Const") {
  1191. graph_const_inputs_.push_back(op);
  1192. }
  1193. }
  1194. void DfGraphConvertor::SetNodeInput(const AnfNodePtr node) {
  1195. if (!node->isa<CNode>()) {
  1196. return;
  1197. }
  1198. if (op_cache_.find(node.get()) == op_cache_.end()) {
  1199. return;
  1200. }
  1201. auto cnode = node->cast<CNodePtr>();
  1202. OpAdapterPtr adpt = FindAdapter(cnode, training_);
  1203. if (adpt == nullptr) {
  1204. error_ = NOT_FOUND;
  1205. return;
  1206. }
  1207. // get Operator from op_cache_, use adapter to set Inputs
  1208. DfGraphConvertor::SetOpInput(adpt, cnode);
  1209. }
  1210. void DfGraphConvertor::ProcessSubgraph(AnfNodePtr node, const std::vector<AnfNodePtr> &inputs) {
  1211. if (!node->isa<CNode>() || GetCNodeFuncName(node->cast<CNodePtr>()) != "Partial") {
  1212. return;
  1213. }
  1214. auto graph_node = node->cast<CNodePtr>()->input(1)->cast<ValueNodePtr>();
  1215. FuncGraphPtr anf_graph = graph_node->value()->cast<FuncGraphPtr>();
  1216. DfGraphConvertor converter(anf_graph);
  1217. converter.use_inputs_ = true;
  1218. converter.inputs_ = inputs;
  1219. (void)converter.ConvertAllNode().BuildGraph();
  1220. std::string name = graph_node->ToString() + "_ge_graph.dot";
  1221. if (MsContext::GetInstance()->get_param<bool>(MS_CTX_SAVE_GRAPHS_FLAG)) {
  1222. converter.DrawComputeGraph(name);
  1223. }
  1224. branches_map_[node.get()] = *(converter.df_graph_);
  1225. }
  1226. // Update GE op's shape and type info
  1227. void DfGraphConvertor::UpdateOpDesc(const AnfNodePtr node) {
  1228. if (nullptr == node || !node->isa<CNode>()) {
  1229. return;
  1230. }
  1231. if (op_cache_.find(node.get()) == op_cache_.end()) {
  1232. return;
  1233. }
  1234. OpAdapterPtr adpt = FindAdapter(node, training_);
  1235. if (adpt == nullptr) {
  1236. error_ = NOT_FOUND;
  1237. return;
  1238. }
  1239. // get Operator from op_cache_
  1240. OperatorPtr op = Convert(node);
  1241. adpt->updateOutputDesc(op, node->Shape(), node->Type(), node);
  1242. }
  1243. OperatorPtr DfGraphConvertor::Convert(const AnfNodePtr node) {
  1244. if (node == nullptr) {
  1245. MS_LOG(ERROR) << "node is nullptr";
  1246. error_ = NOT_FOUND;
  1247. return nullptr;
  1248. }
  1249. // find in cache
  1250. if (op_cache_.count(node.get())) {
  1251. return op_cache_[node.get()];
  1252. }
  1253. // do not convert primitive node, Load, UpdateState
  1254. if (IsValueNode<Primitive>(node) || IsPrimitiveCNode(node, prim::kPrimLoad) ||
  1255. IsPrimitiveCNode(node, prim::kPrimUpdateState)) {
  1256. return nullptr;
  1257. }
  1258. // convert a new one
  1259. if (node->isa<CNode>()) {
  1260. return ConvertCNode(node->cast<CNodePtr>());
  1261. }
  1262. if (node->isa<Parameter>()) {
  1263. return ConvertParameter(node);
  1264. }
  1265. if (node->isa<ValueNode>()) {
  1266. if (IsValueNode<Monad>(node)) {
  1267. return nullptr;
  1268. }
  1269. return ConvertValueNode(node->cast<ValueNodePtr>());
  1270. }
  1271. MS_LOG(ERROR) << "Invalid AnfNode";
  1272. error_ = INVALID_ARGUMENT;
  1273. return nullptr;
  1274. }
  1275. void DfGraphConvertor::ConvertMakeTuple(const CNodePtr node) {
  1276. std::shared_ptr<std::vector<OutHandler>> tuple_items = std::make_shared<std::vector<OutHandler>>();
  1277. // convert each tuple item to a OutHandler
  1278. for (size_t i = 1; i < node->inputs().size(); i++) {
  1279. AnfNodePtr item = node->input(i);
  1280. if (IsPrimitiveCNode(item, prim::kPrimLoad)) {
  1281. item = ParseLoadInput(item->cast<CNodePtr>());
  1282. }
  1283. OperatorPtr op = Convert(item);
  1284. if (op != nullptr) {
  1285. tuple_items->emplace_back(OutHandler(op, "", item));
  1286. } else if (out_handle_cache_.find(item.get()) != out_handle_cache_.end()) {
  1287. tuple_items->push_back(out_handle_cache_[item.get()]);
  1288. } else {
  1289. tuple_items->push_back(OutHandler(nullptr, "", item));
  1290. }
  1291. }
  1292. MS_LOG(DEBUG) << "ConvertMakeTuple: " << node.get() << " " << tuple_items->size();
  1293. tuple_out_handle_cache_[node.get()] = tuple_items;
  1294. }
  1295. void DfGraphConvertor::ConvertTopK(const CNodePtr node) {
  1296. MS_LOG(INFO) << "Convert TopK second input's type from int64 to int32.";
  1297. auto value_ptr = node->input(2)->cast<ValueNodePtr>();
  1298. std::ostringstream ss;
  1299. ss << "op" << value_ptr.get();
  1300. op_draw_name_[value_ptr.get()] = ss.str();
  1301. compute_sout_ << ss.str() << "[label= \"" << value_ptr->value()->ToString() << "\" shape=ellipse]" << endl;
  1302. auto int64_value = value_ptr->value()->cast<Int64ImmPtr>()->value();
  1303. OpAdapterPtr adpt = FindAdapter(value_ptr, training_);
  1304. auto op = adpt->generate(value_ptr);
  1305. adpt->setAttr(op, "value", static_cast<int32_t>(int64_value));
  1306. op_cache_[value_ptr.get()] = op;
  1307. }
  1308. AnfNodePtr DfGraphConvertor::TraceTupleGetItem(const CNodePtr &node, uint64_t *index) {
  1309. const int TUPLE_GET_ITEM_INDEX = 2;
  1310. if (node->inputs().size() < 3) { // "tuple_getitem" primitive must have 3 inputs
  1311. MS_LOG(EXCEPTION) << "length of inputs of TupleGetItem is less than 3";
  1312. }
  1313. auto index_node = node->inputs()[TUPLE_GET_ITEM_INDEX];
  1314. if (!index_node->isa<ValueNode>()) {
  1315. error_ = INVALID_ARGUMENT;
  1316. MS_LOG(EXCEPTION) << "can't convert get item with non-constant index";
  1317. }
  1318. *index = LongToUlong(GetValue<int64_t>(GetValueNode(index_node)));
  1319. return node->inputs()[1];
  1320. }
  1321. AnfNodePtr DfGraphConvertor::TraceDepend(const CNodePtr &node) {
  1322. auto cnode = node->cast<CNodePtr>();
  1323. if (cnode->inputs().size() < 3) { // "Depend" primitive have 3 inputs
  1324. MS_LOG(EXCEPTION) << "length of inputs of depend is less than 3";
  1325. }
  1326. return cnode->inputs()[1];
  1327. }
  1328. AnfNodePtr DfGraphConvertor::TraceMakeTuple(const CNodePtr &node, uint64_t index) {
  1329. if (index + 1 >= node->inputs().size()) {
  1330. MS_LOG(EXCEPTION) << "length of make_tuple is less than index: " << index;
  1331. }
  1332. return node->inputs()[index + 1];
  1333. }
  1334. OutHandler DfGraphConvertor::GetHandler(const AnfNodePtr &node, const std::stack<uint64_t> &index_stack,
  1335. AnfNode *const draw_index) {
  1336. if (node == nullptr) {
  1337. MS_LOG(ERROR) << "Get nullptr while trace real op";
  1338. return OutHandler(nullptr, "");
  1339. }
  1340. std::ostringstream ss;
  1341. ss << "op" << node.get();
  1342. if (index_stack.empty()) {
  1343. op_draw_name_[draw_index] = ss.str();
  1344. return OutHandler(Convert(node), "");
  1345. } else {
  1346. OpAdapterPtr adpt = FindAdapter(node, training_);
  1347. if (nullptr == adpt) {
  1348. MS_LOG(ERROR) << "Can not get node output as adpt is nullptr!";
  1349. error_ = NOT_FOUND;
  1350. return OutHandler(nullptr, "");
  1351. }
  1352. OperatorPtr op = Convert(node);
  1353. if (op == nullptr) {
  1354. error_ = NOT_FOUND;
  1355. MS_LOG(ERROR) << "Can not convert node for trace real op";
  1356. return OutHandler(nullptr, "");
  1357. }
  1358. op_draw_name_[draw_index] = ss.str();
  1359. return adpt->getOutput(Convert(node), UintToInt(index_stack.top()));
  1360. }
  1361. }
  1362. // get the real operator through maketuple tuple_getitem depend
  1363. OutHandler DfGraphConvertor::TraceRealOp(AnfNodePtr node) {
  1364. bool flag = IsPrimitiveCNode(node, prim::kPrimTupleGetItem) || IsPrimitiveCNode(node, prim::kPrimMakeTuple) ||
  1365. IsPrimitiveCNode(node, prim::kPrimDepend);
  1366. std::stack<uint64_t> index_stack;
  1367. auto draw_index = node.get();
  1368. while (flag) {
  1369. flag = false;
  1370. if (IsPrimitiveCNode(node, prim::kPrimTupleGetItem)) {
  1371. uint64_t index;
  1372. node = TraceTupleGetItem(node->cast<CNodePtr>(), &index);
  1373. index_stack.push(index);
  1374. flag = true;
  1375. } else if (IsPrimitiveCNode(node, prim::kPrimMakeTuple)) {
  1376. if (index_stack.empty()) {
  1377. MS_LOG(ERROR) << "TraceRealOp find a make_tuple node";
  1378. return OutHandler(nullptr, "");
  1379. } else {
  1380. node = TraceMakeTuple(node->cast<CNodePtr>(), index_stack.top());
  1381. index_stack.pop();
  1382. flag = true;
  1383. }
  1384. } else if (IsPrimitiveCNode(node, prim::kPrimDepend)) {
  1385. node = TraceDepend(node->cast<CNodePtr>());
  1386. flag = true;
  1387. }
  1388. }
  1389. return GetHandler(node, index_stack, draw_index);
  1390. }
  1391. void DfGraphConvertor::ConvertTupleGetItem(const CNodePtr node) {
  1392. auto handle = TraceRealOp(node);
  1393. if (handle.op == nullptr) {
  1394. MS_LOG(ERROR) << "Failed to trace tuple get item";
  1395. return;
  1396. }
  1397. out_handle_cache_[node.get()] = handle;
  1398. }
  1399. // Get the real op for tuple_getitem through make tuple, or depend
  1400. AnfNodePtr DfGraphConvertor::GetRealOpNode(AnfNodePtr node) {
  1401. const int TUPLE_GET_ITEM_INDEX = 2;
  1402. if (IsPrimitiveCNode(node, prim::kPrimTupleGetItem)) {
  1403. auto node_inputs = node->cast<CNodePtr>()->inputs();
  1404. if (node_inputs.size() != 3) { // "tuple_getitem" primitive must have 3 inputs
  1405. MS_LOG(ERROR) << "tuple get item node not correct!";
  1406. error_ = FAILED;
  1407. return node;
  1408. }
  1409. MS_EXCEPTION_IF_NULL(node_inputs[TUPLE_GET_ITEM_INDEX]);
  1410. if (!node_inputs[TUPLE_GET_ITEM_INDEX]->isa<ValueNode>()) {
  1411. error_ = INVALID_ARGUMENT;
  1412. MS_LOG(EXCEPTION) << "can't convert get item with non-constant index";
  1413. }
  1414. auto value_ptr = GetValueNode(node_inputs[TUPLE_GET_ITEM_INDEX])->cast<Int32ImmPtr>();
  1415. if (value_ptr == nullptr) {
  1416. MS_LOG(ERROR) << "Can not convert get item as value is nullptr!";
  1417. error_ = FAILED;
  1418. return node;
  1419. }
  1420. int64_t index = value_ptr->value();
  1421. // make_tuple apply inputs:make_tuple, [tuple_items,]
  1422. if (IsPrimitiveCNode(node_inputs[1], prim::kPrimMakeTuple)) {
  1423. auto tuple_inputs = node->cast<CNodePtr>()->inputs();
  1424. if (tuple_inputs.size() < IntToSize(index + 1)) {
  1425. MS_LOG(ERROR) << "make tuple input items node not correct! size:" << tuple_inputs.size()
  1426. << ", item index:" << index;
  1427. error_ = FAILED;
  1428. return node;
  1429. }
  1430. return GetRealOpNode(tuple_inputs[IntToSize(index + 1)]);
  1431. }
  1432. return GetRealOpNode(node_inputs[1]);
  1433. }
  1434. // depend apply inputs: depend,output,depended_node
  1435. if (IsPrimitiveCNode(node, prim::kPrimDepend)) {
  1436. auto depend_inputs = node->cast<CNodePtr>()->inputs();
  1437. if (depend_inputs.size() != 3) { // "Depend" primitive have 3 inputs
  1438. MS_LOG(ERROR) << "depend input items not correct";
  1439. error_ = FAILED;
  1440. return node;
  1441. }
  1442. return GetRealOpNode(depend_inputs[1]);
  1443. }
  1444. return node;
  1445. }
  1446. // convert the anf node to corresponding operator list
  1447. std::vector<OperatorPtr> DfGraphConvertor::ConvertDependNode(const AnfNodePtr node) {
  1448. if (IsPrimitiveCNode(node, prim::kPrimMakeTuple)) {
  1449. std::vector<OperatorPtr> op_lists;
  1450. auto node_inputs = node->cast<CNodePtr>()->inputs();
  1451. for (size_t index = 1; index < node_inputs.size(); index++) {
  1452. auto op = Convert(GetRealOpNode(node_inputs[index]));
  1453. if (op == nullptr) {
  1454. MS_LOG(ERROR) << "Convert real op node to operator failed";
  1455. error_ = FAILED;
  1456. return std::vector<OperatorPtr>({});
  1457. }
  1458. op_lists.push_back(op);
  1459. }
  1460. return op_lists;
  1461. }
  1462. auto op = Convert(GetRealOpNode(node));
  1463. if (op == nullptr) {
  1464. MS_LOG(ERROR) << "Convert real op node to operator failed";
  1465. error_ = FAILED;
  1466. return std::vector<OperatorPtr>({});
  1467. }
  1468. return std::vector<OperatorPtr>({op});
  1469. }
  1470. bool DfGraphConvertor::CheckCNode(const std::string &name, const CNodePtr node) {
  1471. // ignore apply node of return
  1472. if (name == "" || name == prim::kPrimReturn->name() || name == prim::kPrimDepend->name() ||
  1473. name == prim::kPrimSwitchLayer->name() || name == prim::kPrimPartial->name()) {
  1474. return false;
  1475. }
  1476. // Convert TopK second input from int64 to int32.
  1477. if (name == prim::kPrimTopK->name()) {
  1478. ConvertTopK(node);
  1479. return true;
  1480. }
  1481. // make_tuple is used for a dynamic_input, convert it to a vector of OutHandlers
  1482. if (name == prim::kPrimMakeTuple->name()) {
  1483. ConvertMakeTuple(node);
  1484. return false;
  1485. }
  1486. // As for nodes with multi outputs, convert tuple_getitem to OutHandle
  1487. if (name == prim::kPrimTupleGetItem->name()) {
  1488. ConvertTupleGetItem(node);
  1489. return false;
  1490. }
  1491. return true;
  1492. }
  1493. OperatorPtr DfGraphConvertor::ConvertCNode(const CNodePtr node) {
  1494. std::string name = GetCNodeTargetFuncName(node);
  1495. if (!CheckCNode(name, node)) {
  1496. return nullptr;
  1497. }
  1498. // get corresponding OpAdapter
  1499. OpAdapterPtr adpt = FindAdapter(node, training_);
  1500. if (adpt == nullptr) {
  1501. error_ = NOT_FOUND;
  1502. return nullptr;
  1503. }
  1504. // get operator
  1505. OperatorPtr op = nullptr;
  1506. auto it_op = op_cache_.find(node.get());
  1507. if (it_op != op_cache_.end()) {
  1508. op = it_op->second;
  1509. } else {
  1510. op = adpt->generate(node);
  1511. }
  1512. // set attribute for primitive
  1513. (void)adpt->setAttr(op, node);
  1514. // add into cache
  1515. (void)op_cache_.insert(std::make_pair(node.get(), op));
  1516. DrawCNode(node, adpt);
  1517. return op_cache_[node.get()];
  1518. }
  1519. OperatorPtr DfGraphConvertor::ConvertParameter(const AnfNodePtr node) {
  1520. // convert Parameter in ANF to variable in DataFlow
  1521. auto op = FindAdapter(node, training_)->generate(node);
  1522. op_cache_[node.get()] = op;
  1523. // build index for parameter using name
  1524. std::string name = std::static_pointer_cast<Parameter>(node)->name();
  1525. params_[name] = node;
  1526. std::ostringstream ss;
  1527. ss << "op" << node.get();
  1528. op_draw_name_[node.get()] = ss.str();
  1529. compute_sout_ << ss.str() << "[shape=octagon, label=\"" << name << "\"]" << endl;
  1530. return op_cache_[node.get()];
  1531. }
  1532. Status DfGraphConvertor::TryConvertValueNodeToMultiConst(const ValueNodePtr node) {
  1533. MS_EXCEPTION_IF_NULL(node);
  1534. ValuePtr value = node->value();
  1535. MS_EXCEPTION_IF_NULL(value);
  1536. if (!value->isa<ValueList>() && !value->isa<ValueTuple>()) {
  1537. return FAILED;
  1538. }
  1539. auto vec = value->isa<ValueTuple>() ? value->cast<ValueTuplePtr>()->value() : value->cast<ValueListPtr>()->value();
  1540. if (vec.empty()) {
  1541. return FAILED;
  1542. }
  1543. std::shared_ptr<std::vector<OutHandler>> tuple_items = std::make_shared<std::vector<OutHandler>>();
  1544. for (size_t i = 0; i < vec.size(); i++) {
  1545. MS_EXCEPTION_IF_NULL(vec[i]);
  1546. if (vec[i]->isa<MeTensor>()) {
  1547. GeTensorPtr ge_tensor = transform::TransformUtil::ConvertTensor(vec[i]->cast<MeTensorPtr>(), kOpFormat_NCHW);
  1548. auto const_op = std::make_shared<Constant>(node->fullname_with_scope() + "/const/inputs/" + std::to_string(i));
  1549. (void)const_op->set_attr_value(*ge_tensor);
  1550. (void)const_op->update_output_desc_y(ge_tensor->GetTensorDesc());
  1551. tuple_items->emplace_back(OutHandler(const_op, ""));
  1552. } else {
  1553. return FAILED;
  1554. }
  1555. }
  1556. if (tuple_items->empty()) {
  1557. return FAILED;
  1558. }
  1559. tuple_out_handle_cache_[node.get()] = tuple_items;
  1560. return SUCCESS;
  1561. }
  1562. OperatorPtr DfGraphConvertor::ConvertValueNode(const ValueNodePtr node) {
  1563. // convert valuenode in ANF to Const in DataFlow
  1564. // find paramerte referenced by SymbolicKeyInstance of valuenode
  1565. std::ostringstream ss;
  1566. ss << "op" << node.get();
  1567. op_draw_name_[node.get()] = ss.str();
  1568. compute_sout_ << ss.str() << "[label= \"" << node->value()->ToString() << "\" shape=ellipse]" << endl;
  1569. if (TryConvertValueNodeToMultiConst(node) == SUCCESS) {
  1570. MS_LOG(INFO) << "Convert value node to multi Constant OP success";
  1571. return nullptr;
  1572. }
  1573. OpAdapterPtr adpt = FindAdapter(node, training_);
  1574. if (adpt == nullptr) {
  1575. error_ = NOT_FOUND;
  1576. return nullptr;
  1577. }
  1578. auto op = adpt->generate(node);
  1579. // set const's attrs
  1580. if (adpt->setAttr(op, "value", node->value()) != 0) {
  1581. MS_LOG(WARNING) << "set attr value for const failed";
  1582. }
  1583. #if (defined ENABLE_GE)
  1584. auto const_op = std::static_pointer_cast<Constant>(op);
  1585. if (const_op == nullptr) {
  1586. MS_LOG(ERROR) << "Get Constant operator failed";
  1587. return nullptr;
  1588. }
  1589. auto ge_tensor = const_op->get_attr_value();
  1590. auto ge_desc = ge_tensor.GetTensorDesc();
  1591. (void)const_op->update_output_desc_y(ge_desc);
  1592. #endif
  1593. op_cache_[node.get()] = op;
  1594. return op_cache_[node.get()];
  1595. }
  1596. void DfGraphConvertor::DrawCNode(const CNodePtr node, const OpAdapterPtr adpt) {
  1597. if (nullptr == adpt || nullptr == node) {
  1598. MS_LOG(ERROR) << "Failed to draw apply node as adpt or node is nullptr!";
  1599. return;
  1600. }
  1601. std::ostringstream ss;
  1602. ss << "op" << node.get();
  1603. op_draw_name_[node.get()] = ss.str();
  1604. compute_sout_ << ss.str() << "[label=<";
  1605. compute_sout_ << "<table border='1' cellborder='1'>" << endl;
  1606. auto input_map = adpt->getInputMap();
  1607. auto dyn_input_map = adpt->getDynInputMap();
  1608. if (input_map.size() + dyn_input_map.size() > 0) {
  1609. compute_sout_ << "<tr>";
  1610. for (auto &it : input_map) {
  1611. compute_sout_ << "<td port='" << it.first << "'>" << it.second.name << "</td>";
  1612. }
  1613. for (auto &it : dyn_input_map) {
  1614. compute_sout_ << "<td port='" << it.first << "'>" << it.second.name << "</td>";
  1615. }
  1616. compute_sout_ << "</tr>" << endl;
  1617. }
  1618. compute_sout_ << "<tr><td colspan=\"" << (input_map.size() + dyn_input_map.size()) << "\">\"" << node->ToString()
  1619. << ":" << GetCNodeTargetFuncName(node) << "\"</td></tr>" << endl;
  1620. // print attrs' values
  1621. auto atts = adpt->GetAttrsFromDrawGraph();
  1622. for (auto &it : atts) {
  1623. compute_sout_ << "<tr><td colspan=\"" << (input_map.size() + dyn_input_map.size()) << "\">\"" << it
  1624. << "\"</td></tr>";
  1625. }
  1626. adpt->clearAttrVect();
  1627. compute_sout_ << "</table>> shape=plaintext]" << endl;
  1628. }
  1629. void DfGraphConvertor::RegisterAdapter(const std::string &name, OpAdapterPtr adpt) {
  1630. OpAdapterMap::get()[name] = std::make_shared<OpAdapterDesc>(adpt);
  1631. }
  1632. void DfGraphConvertor::RegisterAdapter(const std::string &name, OpAdapterPtr train_adpt, OpAdapterPtr infer_adpt) {
  1633. OpAdapterMap::get()[name] = std::make_shared<OpAdapterDesc>(train_adpt, infer_adpt);
  1634. }
  1635. } // namespace transform
  1636. } // namespace mindspore