You can not select more than 25 topics Topics must start with a chinese character,a letter or number, can include dashes ('-') and can be up to 35 characters long.

pynative_execute_ge.cc 12 kB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312
  1. /**
  2. * Copyright 2020 Huawei Technologies Co., Ltd
  3. *
  4. * Licensed under the Apache License, Version 2.0 (the "License");
  5. * you may not use this file except in compliance with the License.
  6. * You may obtain a copy of the License at
  7. *
  8. * http://www.apache.org/licenses/LICENSE-2.0
  9. *
  10. * Unless required by applicable law or agreed to in writing, software
  11. * distributed under the License is distributed on an "AS IS" BASIS,
  12. * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  13. * See the License for the specific language governing permissions and
  14. * limitations under the License.
  15. */
  16. #include "pynative/pynative_execute_ge.h"
  17. #include <typeinfo>
  18. #include <map>
  19. #include <set>
  20. #include <unordered_set>
  21. #include "utils/any.h"
  22. #include "utils/utils.h"
  23. #include "utils/context/ms_context.h"
  24. #include "operator/ops.h"
  25. #include "pipeline/parse/data_converter.h"
  26. #include "pipeline/static_analysis/prim.h"
  27. #include "session/session_factory.h"
  28. #include "ir/tensor_py.h"
  29. const char SINGLE_OP_GRAPH[] = "single_op_graph";
  30. using mindspore::tensor::TensorPy;
  31. namespace mindspore {
  32. namespace pynative {
  33. using MeTensor = mindspore::tensor::Tensor;
  34. using MeTensorPtr = mindspore::tensor::TensorPtr;
  35. using GeOperator = ge::Operator;
  36. using GeOperatorPtr = std::shared_ptr<GeOperator>;
  37. using transform::GraphRunner;
  38. using transform::GraphRunnerOptions;
  39. using transform::OperatorPtr;
  40. static std::shared_ptr<session::SessionBasic> session = nullptr;
  41. inline ValuePtr PyAttrValue(const py::object &obj) {
  42. ValuePtr converted_ret = nullptr;
  43. bool converted = parse::ConvertData(obj, &converted_ret);
  44. if (!converted) {
  45. MS_LOG(EXCEPTION) << "Attribute convert error with type:" << std::string(py::str(obj));
  46. }
  47. return converted_ret;
  48. }
  49. MeTensorPtr ConvertPyObjToTensor(const py::object &obj) {
  50. MeTensorPtr me_tensor_ptr = nullptr;
  51. if (py::isinstance<MeTensor>(obj)) {
  52. me_tensor_ptr = py::cast<MeTensorPtr>(obj);
  53. } else if (py::isinstance<py::tuple>(obj)) {
  54. me_tensor_ptr = TensorPy::MakeTensor(py::array(py::cast<py::tuple>(obj)), nullptr);
  55. } else if (py::isinstance<py::float_>(obj)) {
  56. me_tensor_ptr = TensorPy::MakeTensor(py::array(py::cast<py::float_>(obj)), nullptr);
  57. } else if (py::isinstance<py::int_>(obj)) {
  58. me_tensor_ptr = TensorPy::MakeTensor(py::array(py::cast<py::int_>(obj)), nullptr);
  59. } else if (py::isinstance<py::list>(obj)) {
  60. me_tensor_ptr = TensorPy::MakeTensor(py::array(py::cast<py::list>(obj)), nullptr);
  61. } else if (py::isinstance<py::array>(obj)) {
  62. me_tensor_ptr = TensorPy::MakeTensor(py::cast<py::array>(obj), nullptr);
  63. } else {
  64. MS_LOG(EXCEPTION) << "Run op inputs type is invalid!";
  65. }
  66. return me_tensor_ptr;
  67. }
  68. bool SetInputsForSingleOpGraph(const OpExecInfoPtr &op_exec_info, const std::vector<GeTensorPtr> &inputs,
  69. const OperatorPtr &op, std::vector<GeOperator> *graph_input_nodes) {
  70. MS_EXCEPTION_IF_NULL(op_exec_info);
  71. MS_EXCEPTION_IF_NULL(graph_input_nodes);
  72. auto op_inputs = op_exec_info->op_inputs;
  73. std::string op_name = op_exec_info->op_name;
  74. transform::OpAdapterPtr adapter = transform::DfGraphConvertor::FindAdapter(op_name, true);
  75. if (adapter == nullptr) {
  76. return false;
  77. }
  78. int op_input_idx = 1;
  79. size_t size = inputs.size();
  80. for (size_t i = 0; i < size; i++) {
  81. if (inputs[i] == nullptr) {
  82. continue;
  83. }
  84. auto const_op = std::make_shared<transform::Constant>();
  85. MS_EXCEPTION_IF_NULL(const_op);
  86. (void)const_op->set_attr_value(*inputs[i]);
  87. MeTensorPtr me_tensor_ptr = ConvertPyObjToTensor(op_inputs[i]);
  88. MS_EXCEPTION_IF_NULL(me_tensor_ptr);
  89. auto const_op_desc =
  90. transform::TransformUtil::GetGeTensorDesc(me_tensor_ptr->shape_c(), me_tensor_ptr->data_type(), kOpFormat_NCHW);
  91. if (const_op_desc == nullptr) {
  92. MS_LOG(ERROR) << "Create variable " << op_name << " output descriptor failed!";
  93. return false;
  94. }
  95. auto pointer_cast_const_op = std::static_pointer_cast<transform::Constant>(const_op);
  96. MS_EXCEPTION_IF_NULL(pointer_cast_const_op);
  97. (void)pointer_cast_const_op->update_output_desc_y(*const_op_desc);
  98. auto &input_map = adapter->getInputMap();
  99. if (input_map.find(op_input_idx) == input_map.end()) {
  100. continue;
  101. }
  102. if (adapter->setInput(op, op_input_idx++, const_op)) {
  103. MS_LOG(ERROR) << "Failed to set params, index is " << op_input_idx;
  104. return false;
  105. }
  106. graph_input_nodes->push_back(*const_op);
  107. }
  108. return true;
  109. }
  110. bool BuildSingleOpGraph(const OpExecInfoPtr &op_exec_info, const std::vector<GeTensorPtr> &inputs,
  111. const std::unordered_map<std::string, ValuePtr> &attrs, const GeGraphPtr &graph) {
  112. MS_EXCEPTION_IF_NULL(op_exec_info);
  113. std::string op_name = op_exec_info->op_name;
  114. auto op_inputs = op_exec_info->op_inputs;
  115. transform::OpAdapterPtr adapter = transform::DfGraphConvertor::FindAdapter(op_name, true);
  116. if (adapter == nullptr) {
  117. MS_LOG(ERROR) << "Unable to find Adapter for " << ((std::string)py::str(op_name));
  118. return false;
  119. }
  120. OperatorPtr op = adapter->generate(op_name);
  121. MS_EXCEPTION_IF_NULL(op);
  122. std::vector<GeOperator> graph_input_nodes;
  123. // hold param nodes after setting input and output for the graph
  124. // set input
  125. if (!SetInputsForSingleOpGraph(op_exec_info, inputs, op, &graph_input_nodes)) {
  126. return false;
  127. }
  128. // set attributes
  129. for (auto attr : attrs) {
  130. (void)adapter->setAttr(op, attr.first, attr.second);
  131. }
  132. // set default attributes
  133. auto extra_attrs = adapter->GetExtraAttr();
  134. for (auto attr : extra_attrs) {
  135. (void)adapter->setAttr(op, attr.first, attr.second);
  136. }
  137. // set input attributes
  138. auto &input_attr_map = adapter->getInputAttrMap();
  139. for (auto &it : input_attr_map) {
  140. if (op_inputs.size() < it.first) {
  141. continue;
  142. }
  143. auto const_value = PyAttrValue(op_inputs[it.first - 1]);
  144. if (const_value->isa<None>()) {
  145. continue;
  146. }
  147. it.second.set_attr(op, const_value);
  148. }
  149. // construct output data nodes
  150. std::vector<GeOperator> graph_outputs{*op};
  151. // set input and output nodes for the graph
  152. MS_EXCEPTION_IF_NULL(graph);
  153. (void)graph->SetInputs(graph_input_nodes).SetOutputs(graph_outputs);
  154. MS_LOG(INFO) << "BuildSingleOpGraph done";
  155. return true;
  156. }
  157. void ToTensorPtr(const OpExecInfoPtr op_exec_info, std::vector<GeTensorPtr> *const inputs) {
  158. MS_EXCEPTION_IF_NULL(inputs);
  159. MS_EXCEPTION_IF_NULL(op_exec_info);
  160. auto op_inputs = op_exec_info->op_inputs;
  161. size_t size = op_inputs.size();
  162. for (size_t i = 0; i < size; i++) {
  163. if (py::isinstance<py::none>(op_inputs[i])) {
  164. inputs->emplace_back(nullptr);
  165. continue;
  166. }
  167. MeTensorPtr me_tensor_ptr = ConvertPyObjToTensor(op_inputs[i]);
  168. auto ge_tensor_ptr = transform::TransformUtil::ConvertTensor(me_tensor_ptr, kOpFormat_NCHW);
  169. if (ge_tensor_ptr == nullptr) {
  170. MS_LOG(EXCEPTION) << "Convert inputs to GE tensor failed in op " << op_exec_info->op_name << ".";
  171. }
  172. // set inputs for operator to build single node graph
  173. inputs->push_back(ge_tensor_ptr);
  174. }
  175. }
  176. PynativeStatusCode ConvertAttributes(const OpExecInfoPtr &op_exec_info, const std::vector<GeTensorPtr> &inputs) {
  177. MS_EXCEPTION_IF_NULL(op_exec_info);
  178. auto op_attrs = op_exec_info->op_attrs;
  179. std::unordered_map<std::string, ValuePtr> attrs{};
  180. for (auto &item : op_attrs) {
  181. if (!py::isinstance<py::str>(item.first)) {
  182. MS_LOG(ERROR) << "Type error in py dict convert";
  183. return PYNATIVE_OP_ATTRS_ERR;
  184. }
  185. std::string name = py::cast<std::string>(item.first);
  186. auto attr_value = PyAttrValue(py::cast<py::object>(item.second));
  187. (void)attrs.emplace(name, attr_value);
  188. }
  189. // build graph
  190. GeGraphPtr graph = std::make_shared<GeGraph>(op_exec_info->op_name);
  191. if (BuildSingleOpGraph(op_exec_info, inputs, attrs, graph) == false) {
  192. MS_LOG(ERROR) << "Failed to BuildSingleOpGraph";
  193. return PYNATIVE_GRAPH_GE_BUILD_ERR;
  194. }
  195. // add the single op graph into the graph manager, which will be iterated by session.
  196. transform::Status ret =
  197. transform::DfGraphManager::GetInstance().AddGraph(SINGLE_OP_GRAPH, std::shared_ptr<transform::DfGraph>(graph));
  198. if (ret != transform::SUCCESS) {
  199. MS_LOG(ERROR) << "Failed to AddGraph into graph manager";
  200. return PYNATIVE_GRAPH_MANAGER_ERR;
  201. }
  202. return PYNATIVE_SUCCESS;
  203. }
  204. std::vector<MeTensorPtr> ConvertOutputTensors(const OpExecInfoPtr &op_exec_info,
  205. const std::vector<GeTensorPtr> &ge_tensors) {
  206. std::vector<MeTensorPtr> outputs;
  207. AbstractBasePtr abs_base = op_exec_info->abstract;
  208. std::vector<std::vector<int>> shapes;
  209. if (abs_base != nullptr && abs_base->isa<abstract::AbstractTensor>()) {
  210. auto arg_tensor = dyn_cast<abstract::AbstractTensor>(abs_base);
  211. shapes.emplace_back(arg_tensor->shape()->shape());
  212. outputs = transform::TransformUtil::ConvertGeTensors(ge_tensors, shapes);
  213. return outputs;
  214. }
  215. if (abs_base != nullptr && abs_base->isa<abstract::AbstractTuple>()) {
  216. auto arg_tuple = dyn_cast<abstract::AbstractTuple>(abs_base);
  217. size_t len = arg_tuple->size();
  218. for (size_t i = 0; i < len; i++) {
  219. if (arg_tuple->elements()[i]->isa<abstract::AbstractTensor>()) {
  220. auto arg_tensor = dyn_cast<abstract::AbstractTensor>(arg_tuple->elements()[i]);
  221. shapes.emplace_back(arg_tensor->shape()->shape());
  222. }
  223. }
  224. outputs = transform::TransformUtil::ConvertGeTensors(ge_tensors, shapes);
  225. return outputs;
  226. }
  227. for (auto &it : ge_tensors) {
  228. auto tensor = transform::TransformUtil::ConvertGeTensor(it);
  229. if (tensor != nullptr) {
  230. outputs.emplace_back(tensor);
  231. }
  232. }
  233. return outputs;
  234. }
  235. py::object RunOpInGE(const OpExecInfoPtr &op_exec_info, PynativeStatusCode *status) {
  236. MS_LOG(INFO) << "RunOpInGe start";
  237. MS_EXCEPTION_IF_NULL(op_exec_info);
  238. MS_EXCEPTION_IF_NULL(status);
  239. // returns a null py::tuple on error
  240. py::tuple err_ret(0);
  241. auto op_name = op_exec_info->op_name;
  242. transform::OpAdapterPtr adapter = transform::DfGraphConvertor::FindAdapter(op_name, true);
  243. if (adapter == nullptr) {
  244. MS_LOG(ERROR) << "Unable to find GE Adapter for " << ((std::string)py::str(op_name));
  245. *status = PYNATIVE_OP_NOT_IMPLEMENTED_ERR;
  246. return std::move(err_ret);
  247. }
  248. std::vector<GeTensorPtr> inputs{};
  249. ToTensorPtr(op_exec_info, &inputs);
  250. // convert me attr to ge AttrValue
  251. PynativeStatusCode ret = ConvertAttributes(op_exec_info, inputs);
  252. if (ret != PYNATIVE_SUCCESS) {
  253. *status = ret;
  254. return std::move(err_ret);
  255. }
  256. // run graph
  257. transform::RunOptions run_options;
  258. run_options.name = SINGLE_OP_GRAPH;
  259. std::vector<GeTensorPtr> ge_inputs;
  260. std::vector<GeTensorPtr> ge_outputs;
  261. transform::GraphRunnerOptions graph_runner_options;
  262. graph_runner_options.options["ge.trainFlag"] = "1";
  263. auto graph_runner = std::make_shared<transform::GraphRunner>(graph_runner_options);
  264. transform::Status run_ret;
  265. {
  266. // Release GIL before calling into (potentially long-running) C++ code
  267. py::gil_scoped_release release;
  268. run_ret = graph_runner->RunGraph(run_options, ge_inputs, &ge_outputs);
  269. }
  270. if (run_ret != transform::Status::SUCCESS) {
  271. MS_LOG(ERROR) << "GraphRunner fails to run graph";
  272. *status = PYNATIVE_GRAPH_GE_RUN_ERR;
  273. return std::move(err_ret);
  274. }
  275. std::vector<MeTensorPtr> graph_outputs = ConvertOutputTensors(op_exec_info, ge_outputs);
  276. size_t output_size = graph_outputs.size();
  277. py::tuple result(output_size);
  278. for (size_t i = 0; i < output_size; i++) {
  279. MS_EXCEPTION_IF_NULL(graph_outputs[i]);
  280. result[i] = *graph_outputs[i];
  281. }
  282. *status = PYNATIVE_SUCCESS;
  283. MS_LOG(INFO) << "RunOpInGe end";
  284. return std::move(result);
  285. }
  286. } // namespace pynative
  287. } // namespace mindspore