Merge pull request !146 from leonwanghui/typo-fixtags/v0.2.0-alpha
| @@ -122,7 +122,7 @@ bool CastKernel(const TypeIdArgs &args, void *dst, const size_t data_size, const | |||||
| TransDataSrc2Dst<uint16_t, int32_t>(args, dst, data_size); | TransDataSrc2Dst<uint16_t, int32_t>(args, dst, data_size); | ||||
| break; | break; | ||||
| default: | default: | ||||
| MS_LOG(ERROR) << "unsupported datatype trans"; | |||||
| MS_LOG(ERROR) << "Unsupported datatype trans"; | |||||
| return false; | return false; | ||||
| } | } | ||||
| return true; | return true; | ||||
| @@ -132,7 +132,7 @@ size_t CubeSizeByType(const TypeId data_type) { | |||||
| const size_t default_error = 0; | const size_t default_error = 0; | ||||
| auto dt_size = TypeIdSize(data_type); | auto dt_size = TypeIdSize(data_type); | ||||
| if (dt_size < 1) { | if (dt_size < 1) { | ||||
| MS_LOG(ERROR) << "illegal dtype."; | |||||
| MS_LOG(ERROR) << "Illegal dtype."; | |||||
| return default_error; | return default_error; | ||||
| } else if (dt_size == 1) { | } else if (dt_size == 1) { | ||||
| return kCubeSize * 2; | return kCubeSize * 2; | ||||
| @@ -146,12 +146,12 @@ size_t ShapeSize(const std::vector<size_t> &shape) { | |||||
| } | } | ||||
| size_t TypeIdSize(const TypeId data_type) { | size_t TypeIdSize(const TypeId data_type) { | ||||
| const size_t unsupport_type_error = 0; | |||||
| const size_t unsupported_type_error = 0; | |||||
| auto iter = type_map.find(data_type); | auto iter = type_map.find(data_type); | ||||
| if (iter != type_map.end()) { | if (iter != type_map.end()) { | ||||
| return iter->second; | return iter->second; | ||||
| } | } | ||||
| return unsupport_type_error; | |||||
| return unsupported_type_error; | |||||
| } | } | ||||
| std::vector<size_t> TransShapeTo4d(const std::vector<size_t> &shape) { | std::vector<size_t> TransShapeTo4d(const std::vector<size_t> &shape) { | ||||
| @@ -174,7 +174,7 @@ std::vector<size_t> TransShapeTo4d(const std::vector<size_t> &shape) { | |||||
| } | } | ||||
| break; | break; | ||||
| default: | default: | ||||
| MS_LOG(EXCEPTION) << "Unexpeted shape size = " << shape.size(); | |||||
| MS_LOG(EXCEPTION) << "Unexpected shape size = " << shape.size(); | |||||
| } | } | ||||
| return shape_4d; | return shape_4d; | ||||
| } | } | ||||
| @@ -183,7 +183,7 @@ std::vector<size_t> TransShapeToDevice(const std::vector<size_t> &shape, const s | |||||
| std::vector<size_t> device_shape; | std::vector<size_t> device_shape; | ||||
| if (format == kOpFormat_FRAC_NZ) { | if (format == kOpFormat_FRAC_NZ) { | ||||
| if (shape.size() < 2) { | if (shape.size() < 2) { | ||||
| MS_EXCEPTION(NotSupportError) << "format " << format << " is not support shape " << shape.size(); | |||||
| MS_EXCEPTION(NotSupportError) << "Format " << format << " is not support shape " << shape.size(); | |||||
| } | } | ||||
| if (shape.size() > 2) { | if (shape.size() > 2) { | ||||
| (void)std::copy(shape.begin(), shape.end() - 2, std::back_inserter(device_shape)); | (void)std::copy(shape.begin(), shape.end() - 2, std::back_inserter(device_shape)); | ||||
| @@ -231,37 +231,37 @@ std::vector<size_t> TransShapeToDevice(const std::vector<size_t> &shape, const s | |||||
| } | } | ||||
| bool TransDataType(const TypeIdArgs &args, void *result) { | bool TransDataType(const TypeIdArgs &args, void *result) { | ||||
| MS_LOG(DEBUG) << "begin trans datatype from " << TypeIdLabel(args.host_data_type) << " to " | |||||
| MS_LOG(DEBUG) << "Begin trans datatype from " << TypeIdLabel(args.host_data_type) << " to " | |||||
| << TypeIdLabel(args.device_data_type); | << TypeIdLabel(args.device_data_type); | ||||
| MS_EXCEPTION_IF_NULL(result); | MS_EXCEPTION_IF_NULL(result); | ||||
| std::pair<TypeId, TypeId> type_info(args.host_data_type, args.device_data_type); | std::pair<TypeId, TypeId> type_info(args.host_data_type, args.device_data_type); | ||||
| auto iter = mode_map.find(type_info); | auto iter = mode_map.find(type_info); | ||||
| if (iter == mode_map.end()) { | if (iter == mode_map.end()) { | ||||
| MS_LOG(ERROR) << "unsupported datatype trans. src_type :" << TypeIdLabel(args.host_data_type) | |||||
| MS_LOG(ERROR) << "Unsupported datatype trans. src_type :" << TypeIdLabel(args.host_data_type) | |||||
| << ", dst_type:" << TypeIdLabel(args.device_data_type); | << ", dst_type:" << TypeIdLabel(args.device_data_type); | ||||
| return false; | return false; | ||||
| } | } | ||||
| auto trans_mode = iter->second; | auto trans_mode = iter->second; | ||||
| auto type_size = TypeIdSize(args.device_data_type); | auto type_size = TypeIdSize(args.device_data_type); | ||||
| if (type_size < 1) { | if (type_size < 1) { | ||||
| MS_LOG(ERROR) << "invalid host data type."; | |||||
| MS_LOG(ERROR) << "Invalid host data type."; | |||||
| return false; | return false; | ||||
| } | } | ||||
| if (args.host_shape_size < 1) { | if (args.host_shape_size < 1) { | ||||
| MS_LOG(ERROR) << "invalid host data size."; | |||||
| MS_LOG(ERROR) << "Invalid host data size."; | |||||
| return false; | return false; | ||||
| } | } | ||||
| if (!CastKernel(args, result, args.host_shape_size, trans_mode)) { | if (!CastKernel(args, result, args.host_shape_size, trans_mode)) { | ||||
| MS_LOG(ERROR) << "failed to trans datatype.."; | |||||
| MS_LOG(ERROR) << "Failed to trans datatype.."; | |||||
| return false; | return false; | ||||
| } | } | ||||
| return true; | return true; | ||||
| } | } | ||||
| bool TransFormat(const FormatArgs &args, void *result) { | bool TransFormat(const FormatArgs &args, void *result) { | ||||
| MS_LOG(DEBUG) << "start trans format."; | |||||
| MS_LOG(DEBUG) << "Start trans format."; | |||||
| if (TypeIdSize(args.src_data_type) < 1) { | if (TypeIdSize(args.src_data_type) < 1) { | ||||
| MS_LOG(ERROR) << "invalid datatype.."; | |||||
| MS_LOG(ERROR) << "Invalid datatype.."; | |||||
| return false; | return false; | ||||
| } | } | ||||
| if ((args.host_format == kOpFormat_NCHW || args.host_format == kOpFormat_ND) && | if ((args.host_format == kOpFormat_NCHW || args.host_format == kOpFormat_ND) && | ||||
| @@ -276,9 +276,9 @@ bool TransFormat(const FormatArgs &args, void *result) { | |||||
| } | } | ||||
| bool TransFormatFromDeviceToHost(const FormatArgs &args, void *result) { | bool TransFormatFromDeviceToHost(const FormatArgs &args, void *result) { | ||||
| MS_LOG(DEBUG) << "start trans format."; | |||||
| MS_LOG(DEBUG) << "Start trans format."; | |||||
| if (TypeIdSize(args.src_data_type) < 1) { | if (TypeIdSize(args.src_data_type) < 1) { | ||||
| MS_LOG(ERROR) << "invalid datatype.."; | |||||
| MS_LOG(ERROR) << "Invalid datatype.."; | |||||
| return false; | return false; | ||||
| } | } | ||||
| if ((args.host_format == kOpFormat_NCHW || args.host_format == kOpFormat_ND) && | if ((args.host_format == kOpFormat_NCHW || args.host_format == kOpFormat_ND) && | ||||
| @@ -293,15 +293,15 @@ bool TransFormatFromDeviceToHost(const FormatArgs &args, void *result) { | |||||
| } | } | ||||
| bool NchwToFracZ(const FormatArgs &args, void *result) { | bool NchwToFracZ(const FormatArgs &args, void *result) { | ||||
| MS_LOG(DEBUG) << "trans format from nchw to frac_z"; | |||||
| MS_LOG(DEBUG) << "Trans format from nchw to frac_z"; | |||||
| MS_EXCEPTION_IF_NULL(result); | MS_EXCEPTION_IF_NULL(result); | ||||
| if (args.host_shape.size() != kNchwDims) { | if (args.host_shape.size() != kNchwDims) { | ||||
| MS_LOG(ERROR) << "invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims; | |||||
| MS_LOG(ERROR) << "Invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims; | |||||
| return false; | return false; | ||||
| } | } | ||||
| size_t size = TypeIdSize(args.src_data_type); | size_t size = TypeIdSize(args.src_data_type); | ||||
| if (size < 1) { | if (size < 1) { | ||||
| MS_LOG(ERROR) << "illegal dtype."; | |||||
| MS_LOG(ERROR) << "Illegal dtype."; | |||||
| return false; | return false; | ||||
| } | } | ||||
| auto n = args.host_shape[0]; | auto n = args.host_shape[0]; | ||||
| @@ -311,7 +311,7 @@ bool NchwToFracZ(const FormatArgs &args, void *result) { | |||||
| size_t c0 = CubeSizeByType(args.src_data_type); | size_t c0 = CubeSizeByType(args.src_data_type); | ||||
| if (c0 < 1) { | if (c0 < 1) { | ||||
| MS_LOG(ERROR) << "illegal dtype."; | |||||
| MS_LOG(ERROR) << "Illegal dtype."; | |||||
| return false; | return false; | ||||
| } | } | ||||
| size_t c1 = Ceil(c, c0); | size_t c1 = Ceil(c, c0); | ||||
| @@ -327,7 +327,7 @@ bool NchwToFracZ(const FormatArgs &args, void *result) { | |||||
| size_t dst_size = total_ele_cnt * size; | size_t dst_size = total_ele_cnt * size; | ||||
| if (dst_size != args.device_size) { | if (dst_size != args.device_size) { | ||||
| MS_LOG(ERROR) << "illegal total data size." | |||||
| MS_LOG(ERROR) << "Illegal total data size." | |||||
| << "dst size is :" << dst_size << "device size is :" << args.device_size; | << "dst size is :" << dst_size << "device size is :" << args.device_size; | ||||
| return false; | return false; | ||||
| } | } | ||||
| @@ -369,20 +369,20 @@ bool NchwToFracZ(const FormatArgs &args, void *result) { | |||||
| } | } | ||||
| bool FracZToNchw(const FormatArgs &args, void *result) { | bool FracZToNchw(const FormatArgs &args, void *result) { | ||||
| MS_LOG(DEBUG) << "trans format from frac_z to nchw"; | |||||
| MS_LOG(DEBUG) << "Trans format from frac_z to nchw"; | |||||
| MS_EXCEPTION_IF_NULL(result); | MS_EXCEPTION_IF_NULL(result); | ||||
| if (args.host_shape.size() != kNchwDims) { | if (args.host_shape.size() != kNchwDims) { | ||||
| MS_LOG(ERROR) << "invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims; | |||||
| MS_LOG(ERROR) << "Invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims; | |||||
| return false; | return false; | ||||
| } | } | ||||
| size_t size = TypeIdSize(args.src_data_type); | size_t size = TypeIdSize(args.src_data_type); | ||||
| if (size < 1) { | if (size < 1) { | ||||
| MS_LOG(ERROR) << "illegal dtype."; | |||||
| MS_LOG(ERROR) << "Illegal dtype."; | |||||
| return false; | return false; | ||||
| } | } | ||||
| size_t total_size = ShapeSize(args.device_shape) * size; | size_t total_size = ShapeSize(args.device_shape) * size; | ||||
| if (total_size != args.device_size) { | if (total_size != args.device_size) { | ||||
| MS_LOG(ERROR) << "illegal total data size, total_size:" << total_size << ", device_size:" << args.device_size; | |||||
| MS_LOG(ERROR) << "Illegal total data size, total_size:" << total_size << ", device_size:" << args.device_size; | |||||
| return false; | return false; | ||||
| } | } | ||||
| @@ -435,7 +435,7 @@ bool FracZToNchw(const FormatArgs &args, void *result) { | |||||
| bool TransShapeToNz(const std::vector<size_t> &host_shape, std::vector<size_t> *hw_shape) { | bool TransShapeToNz(const std::vector<size_t> &host_shape, std::vector<size_t> *hw_shape) { | ||||
| MS_EXCEPTION_IF_NULL(hw_shape); | MS_EXCEPTION_IF_NULL(hw_shape); | ||||
| if (host_shape.empty()) { | if (host_shape.empty()) { | ||||
| MS_LOG(ERROR) << "size of vector is 0."; | |||||
| MS_LOG(ERROR) << "Size of vector is 0."; | |||||
| return false; | return false; | ||||
| } | } | ||||
| switch (host_shape.size()) { | switch (host_shape.size()) { | ||||
| @@ -447,7 +447,7 @@ bool TransShapeToNz(const std::vector<size_t> &host_shape, std::vector<size_t> * | |||||
| default: | default: | ||||
| auto size = host_shape.size(); | auto size = host_shape.size(); | ||||
| if (size < 2) { | if (size < 2) { | ||||
| MS_LOG(ERROR) << "illegal size."; | |||||
| MS_LOG(ERROR) << "Illegal size."; | |||||
| return false; | return false; | ||||
| } | } | ||||
| size_t times = 1; | size_t times = 1; | ||||
| @@ -462,26 +462,26 @@ bool TransShapeToNz(const std::vector<size_t> &host_shape, std::vector<size_t> * | |||||
| } | } | ||||
| bool NchwToFracNz(const FormatArgs &args, void *result) { | bool NchwToFracNz(const FormatArgs &args, void *result) { | ||||
| MS_LOG(DEBUG) << "trans format from nchw to frac_nz."; | |||||
| MS_LOG(DEBUG) << "Trans format from nchw to frac_nz."; | |||||
| MS_EXCEPTION_IF_NULL(result); | MS_EXCEPTION_IF_NULL(result); | ||||
| std::vector<size_t> hw_shape; | std::vector<size_t> hw_shape; | ||||
| if (!TransShapeToNz(args.host_shape, &hw_shape)) { | if (!TransShapeToNz(args.host_shape, &hw_shape)) { | ||||
| MS_LOG(ERROR) << "trans shape failed.."; | |||||
| MS_LOG(ERROR) << "Trans shape failed.."; | |||||
| return false; | return false; | ||||
| } | } | ||||
| if (hw_shape.size() < 3 || args.device_shape.size() < 4) { | if (hw_shape.size() < 3 || args.device_shape.size() < 4) { | ||||
| MS_LOG(ERROR) << "invalid shape size."; | |||||
| MS_LOG(ERROR) << "Invalid shape size."; | |||||
| return false; | return false; | ||||
| } | } | ||||
| auto size = TypeIdSize(args.src_data_type); | auto size = TypeIdSize(args.src_data_type); | ||||
| if (size < 1) { | if (size < 1) { | ||||
| MS_LOG(ERROR) << "illegal dtype"; | |||||
| MS_LOG(ERROR) << "Illegal dtype"; | |||||
| return false; | return false; | ||||
| } | } | ||||
| auto dst_size = ShapeSize(args.device_shape) * size; | auto dst_size = ShapeSize(args.device_shape) * size; | ||||
| if (dst_size != args.device_size) { | if (dst_size != args.device_size) { | ||||
| MS_LOG(ERROR) << "illegal total data size, total_size:" << dst_size << ", device_size:" << args.device_size; | |||||
| MS_LOG(ERROR) << "Illegal total data size, total_size:" << dst_size << ", device_size:" << args.device_size; | |||||
| return false; | return false; | ||||
| } | } | ||||
| auto times = hw_shape.at(0); | auto times = hw_shape.at(0); | ||||
| @@ -538,26 +538,26 @@ bool NchwToFracNz(const FormatArgs &args, void *result) { | |||||
| } | } | ||||
| bool FracNzToNchw(const FormatArgs &args, void *result) { | bool FracNzToNchw(const FormatArgs &args, void *result) { | ||||
| MS_LOG(DEBUG) << "trans format from frac_nz to nchw"; | |||||
| MS_LOG(DEBUG) << "Trans format from frac_nz to nchw"; | |||||
| MS_EXCEPTION_IF_NULL(result); | MS_EXCEPTION_IF_NULL(result); | ||||
| std::vector<size_t> hw_shape; | std::vector<size_t> hw_shape; | ||||
| if (!TransShapeToNz(args.host_shape, &hw_shape)) { | if (!TransShapeToNz(args.host_shape, &hw_shape)) { | ||||
| MS_LOG(ERROR) << "trans shape failed.."; | |||||
| MS_LOG(ERROR) << "Trans shape failed.."; | |||||
| return false; | return false; | ||||
| } | } | ||||
| if (hw_shape.size() < 3 || args.device_shape.size() < 4) { | if (hw_shape.size() < 3 || args.device_shape.size() < 4) { | ||||
| MS_LOG(ERROR) << "invalid shape size."; | |||||
| MS_LOG(ERROR) << "Invalid shape size."; | |||||
| return false; | return false; | ||||
| } | } | ||||
| auto size = TypeIdSize(args.src_data_type); | auto size = TypeIdSize(args.src_data_type); | ||||
| if (size < 1) { | if (size < 1) { | ||||
| MS_LOG(ERROR) << "illegal dtype"; | |||||
| MS_LOG(ERROR) << "Illegal dtype"; | |||||
| return false; | return false; | ||||
| } | } | ||||
| auto dst_size = ShapeSize(args.device_shape) * size; | auto dst_size = ShapeSize(args.device_shape) * size; | ||||
| if (dst_size != args.device_size) { | if (dst_size != args.device_size) { | ||||
| MS_LOG(ERROR) << "illegal total data size, total_size:" << dst_size << ", device_size:" << args.device_size; | |||||
| MS_LOG(ERROR) << "Illegal total data size, total_size:" << dst_size << ", device_size:" << args.device_size; | |||||
| return false; | return false; | ||||
| } | } | ||||
| auto times = hw_shape.at(0); | auto times = hw_shape.at(0); | ||||
| @@ -614,20 +614,20 @@ bool FracNzToNchw(const FormatArgs &args, void *result) { | |||||
| } | } | ||||
| bool NchwToNc1hwc0(const FormatArgs &args, void *result) { | bool NchwToNc1hwc0(const FormatArgs &args, void *result) { | ||||
| MS_LOG(DEBUG) << "trans format from nchw to Nc1h1wc0"; | |||||
| MS_LOG(DEBUG) << "Trans format from nchw to Nc1h1wc0"; | |||||
| MS_EXCEPTION_IF_NULL(result); | MS_EXCEPTION_IF_NULL(result); | ||||
| if (args.host_shape.size() != kNchwDims) { | if (args.host_shape.size() != kNchwDims) { | ||||
| MS_LOG(ERROR) << "invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims; | |||||
| MS_LOG(ERROR) << "Invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims; | |||||
| return false; | return false; | ||||
| } | } | ||||
| size_t size = TypeIdSize(args.src_data_type); | size_t size = TypeIdSize(args.src_data_type); | ||||
| if (size < 1) { | if (size < 1) { | ||||
| MS_LOG(ERROR) << "illegal dtype."; | |||||
| MS_LOG(ERROR) << "Illegal dtype."; | |||||
| return false; | return false; | ||||
| } | } | ||||
| auto total_size = ShapeSize(args.device_shape) * size; | auto total_size = ShapeSize(args.device_shape) * size; | ||||
| if (total_size != args.device_size) { | if (total_size != args.device_size) { | ||||
| MS_LOG(ERROR) << "illegal total data size, total_size:" << total_size << ", device_size:" << args.device_size; | |||||
| MS_LOG(ERROR) << "Illegal total data size, total_size:" << total_size << ", device_size:" << args.device_size; | |||||
| return false; | return false; | ||||
| } | } | ||||
| @@ -637,7 +637,7 @@ bool NchwToNc1hwc0(const FormatArgs &args, void *result) { | |||||
| auto w = args.host_shape[3]; | auto w = args.host_shape[3]; | ||||
| size_t c0 = CubeSizeByType(args.src_data_type); | size_t c0 = CubeSizeByType(args.src_data_type); | ||||
| if (c0 < 1) { | if (c0 < 1) { | ||||
| MS_LOG(ERROR) << "illegal dtype."; | |||||
| MS_LOG(ERROR) << "Illegal dtype."; | |||||
| return false; | return false; | ||||
| } | } | ||||
| size_t c1 = Ceil(c, c0); | size_t c1 = Ceil(c, c0); | ||||
| @@ -687,20 +687,20 @@ bool NchwToNc1hwc0(const FormatArgs &args, void *result) { | |||||
| } | } | ||||
| bool Nc1hwc0ToNchw(const FormatArgs &args, void *result) { | bool Nc1hwc0ToNchw(const FormatArgs &args, void *result) { | ||||
| MS_LOG(DEBUG) << "trans format from nc1h1wc0 to nchw"; | |||||
| MS_LOG(DEBUG) << "Trans format from nc1h1wc0 to nchw"; | |||||
| MS_EXCEPTION_IF_NULL(result); | MS_EXCEPTION_IF_NULL(result); | ||||
| if (args.host_shape.size() != kNchwDims) { | if (args.host_shape.size() != kNchwDims) { | ||||
| MS_LOG(ERROR) << "invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims; | |||||
| MS_LOG(ERROR) << "Invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims; | |||||
| return false; | return false; | ||||
| } | } | ||||
| size_t size = TypeIdSize(args.src_data_type); | size_t size = TypeIdSize(args.src_data_type); | ||||
| if (size < 1) { | if (size < 1) { | ||||
| MS_LOG(ERROR) << "illegal dtype."; | |||||
| MS_LOG(ERROR) << "Illegal dtype."; | |||||
| return false; | return false; | ||||
| } | } | ||||
| size_t total_size = ShapeSize(args.device_shape) * size; | size_t total_size = ShapeSize(args.device_shape) * size; | ||||
| if (total_size != args.device_size) { | if (total_size != args.device_size) { | ||||
| MS_LOG(ERROR) << "illegal total data size, total_size:" << total_size << ", device_size:" << args.device_size; | |||||
| MS_LOG(ERROR) << "Illegal total data size, total_size:" << total_size << ", device_size:" << args.device_size; | |||||
| return false; | return false; | ||||
| } | } | ||||
| @@ -141,7 +141,7 @@ void DumpKernelInfo(const CNodePtr &node, const std::shared_ptr<SubGraphIRInfo> | |||||
| void DumpParams(const FuncGraphPtr &graph, std::ostringstream &buffer, OrderedMap<AnfNodePtr, int32_t> *para_map) { | void DumpParams(const FuncGraphPtr &graph, std::ostringstream &buffer, OrderedMap<AnfNodePtr, int32_t> *para_map) { | ||||
| if (graph == nullptr) { | if (graph == nullptr) { | ||||
| MS_LOG(INFO) << "param graph is nullptr."; | |||||
| MS_LOG(INFO) << "Param graph is nullptr."; | |||||
| return; | return; | ||||
| } | } | ||||
| std::vector<AnfNodePtr> parameters = graph->parameters(); | std::vector<AnfNodePtr> parameters = graph->parameters(); | ||||
| @@ -175,17 +175,17 @@ void DumpParams(const FuncGraphPtr &graph, std::ostringstream &buffer, OrderedMa | |||||
| if (para_map != nullptr) { | if (para_map != nullptr) { | ||||
| (*para_map)[p] = para++; | (*para_map)[p] = para++; | ||||
| } | } | ||||
| MS_LOG(DEBUG) << "record param: " << p->ToString() << " graph belong : " << p->func_graph()->ToString(); | |||||
| MS_LOG(DEBUG) << "Record param: " << p->ToString() << " graph belong : " << p->func_graph()->ToString(); | |||||
| } | } | ||||
| } | } | ||||
| void DumpOperator(const AnfNodePtr &op, const std::shared_ptr<SubGraphIRInfo> &gsub) { | void DumpOperator(const AnfNodePtr &op, const std::shared_ptr<SubGraphIRInfo> &gsub) { | ||||
| if (op == nullptr) { | if (op == nullptr) { | ||||
| MS_LOG(INFO) << "param op is nullptr"; | |||||
| MS_LOG(INFO) << "Param op is nullptr"; | |||||
| return; | return; | ||||
| } | } | ||||
| if (gsub == nullptr) { | if (gsub == nullptr) { | ||||
| MS_LOG(INFO) << "param gsub is nullptr"; | |||||
| MS_LOG(INFO) << "Param gsub is nullptr"; | |||||
| return; | return; | ||||
| } | } | ||||
| @@ -338,7 +338,7 @@ void DumpCNode(const CNodePtr &nd, const FuncGraphPtr &sub_graph, OrderedMap<Anf | |||||
| } | } | ||||
| if (nd->inputs().empty()) { | if (nd->inputs().empty()) { | ||||
| MS_LOG(EXCEPTION) << "input of apply node is empty"; | |||||
| MS_LOG(EXCEPTION) << "Input of apply node is empty"; | |||||
| } | } | ||||
| // print operator | // print operator | ||||
| @@ -376,7 +376,7 @@ void DumpIRInSubgraph(const std::vector<AnfNodePtr> &nodes, OrderedMap<AnfNodePt | |||||
| MS_EXCEPTION_IF_NULL(nd); | MS_EXCEPTION_IF_NULL(nd); | ||||
| FuncGraphPtr sub_graph = nd->func_graph(); | FuncGraphPtr sub_graph = nd->func_graph(); | ||||
| if (sub_graph == nullptr) { | if (sub_graph == nullptr) { | ||||
| MS_LOG(DEBUG) << "node[" << nd->ToString() << "] belongs to no graph!"; | |||||
| MS_LOG(DEBUG) << "Node[" << nd->ToString() << "] belongs to no graph!"; | |||||
| continue; | continue; | ||||
| } | } | ||||
| std::shared_ptr<SubGraphIRInfo> gsub = (*sub_graphs)[sub_graph]; | std::shared_ptr<SubGraphIRInfo> gsub = (*sub_graphs)[sub_graph]; | ||||
| @@ -430,12 +430,12 @@ void DumpIR(const std::string &filename, const FuncGraphPtr &graph, bool dump_fu | |||||
| return; | return; | ||||
| } | } | ||||
| if (filename.size() > PATH_MAX) { | if (filename.size() > PATH_MAX) { | ||||
| MS_LOG(ERROR) << "file path " << filename << " is too long."; | |||||
| MS_LOG(ERROR) << "File path " << filename << " is too long."; | |||||
| return; | return; | ||||
| } | } | ||||
| char real_path[PATH_MAX] = {0}; | char real_path[PATH_MAX] = {0}; | ||||
| if (nullptr == realpath(filename.c_str(), real_path)) { | if (nullptr == realpath(filename.c_str(), real_path)) { | ||||
| MS_LOG(DEBUG) << "dir " << filename << " does not exit."; | |||||
| MS_LOG(DEBUG) << "Dir " << filename << " does not exit."; | |||||
| } | } | ||||
| OrderedMap<AnfNodePtr, int32_t> para_map; | OrderedMap<AnfNodePtr, int32_t> para_map; | ||||
| @@ -49,7 +49,7 @@ std::string GetMsIrPath(void) { | |||||
| path = path_ptr; | path = path_ptr; | ||||
| char real_path[PATH_MAX] = {0}; | char real_path[PATH_MAX] = {0}; | ||||
| if (path.size() > PATH_MAX || nullptr == realpath(path.c_str(), real_path)) { | if (path.size() > PATH_MAX || nullptr == realpath(path.c_str(), real_path)) { | ||||
| MS_LOG(EXCEPTION) << "MS IR Path error, " << path_ptr; | |||||
| MS_LOG(EXCEPTION) << "MS IR path error, " << path_ptr; | |||||
| } | } | ||||
| path = real_path; | path = real_path; | ||||
| } | } | ||||
| @@ -144,8 +144,8 @@ std::string AnfExporter::GetValueNodeText(const FuncGraphPtr& fg, const ValueNod | |||||
| } | } | ||||
| std::string AnfExporter::GetMultitypeFuncGraphText(const prim::MultitypeFuncGraphPtr& mt_func_graph) { | std::string AnfExporter::GetMultitypeFuncGraphText(const prim::MultitypeFuncGraphPtr& mt_func_graph) { | ||||
| auto py_funs = mt_func_graph->GetPyFunctions(); | |||||
| if (py_funs.empty()) { | |||||
| auto py_funcs = mt_func_graph->GetPyFunctions(); | |||||
| if (py_funcs.empty()) { | |||||
| return ""; | return ""; | ||||
| } | } | ||||
| @@ -153,7 +153,7 @@ std::string AnfExporter::GetMultitypeFuncGraphText(const prim::MultitypeFuncGrap | |||||
| oss << "{"; | oss << "{"; | ||||
| bool is_first = true; | bool is_first = true; | ||||
| for (const auto& py_func : py_funs) { | |||||
| for (const auto& py_func : py_funcs) { | |||||
| if (is_first) { | if (is_first) { | ||||
| is_first = false; | is_first = false; | ||||
| } else { | } else { | ||||
| @@ -626,7 +626,7 @@ void AnfExporter::ExportFuncGraph(const std::string& filename, const FuncGraphPt | |||||
| ofs << "\n\n"; | ofs << "\n\n"; | ||||
| (void)func_graph_set.erase(fg); | (void)func_graph_set.erase(fg); | ||||
| } | } | ||||
| ofs << "# num of total funcgraphs: " << exported.size(); | |||||
| ofs << "# num of total function graphs: " << exported.size(); | |||||
| ofs.close(); | ofs.close(); | ||||
| } | } | ||||
| @@ -651,7 +651,7 @@ void AnfExporter::ExportFuncGraph(const std::string& filename, const std::vector | |||||
| ofs << "\n\n"; | ofs << "\n\n"; | ||||
| } | } | ||||
| ofs << "# num of total funcgraphs: " << graphs.size(); | |||||
| ofs << "# num of total function graphs: " << graphs.size(); | |||||
| ofs.close(); | ofs.close(); | ||||
| } | } | ||||
| @@ -763,7 +763,7 @@ class Lexer { | |||||
| fin.close(); | fin.close(); | ||||
| } | } | ||||
| } catch (const std::exception& e) { | } catch (const std::exception& e) { | ||||
| MS_LOG(ERROR) << "exception when closing file"; | |||||
| MS_LOG(ERROR) << "Exception when closing file"; | |||||
| } catch (...) { | } catch (...) { | ||||
| std::string exName(abi::__cxa_current_exception_type()->name()); | std::string exName(abi::__cxa_current_exception_type()->name()); | ||||
| MS_LOG(ERROR) << "Error occurred when closing file. Exception name: " << exName; | MS_LOG(ERROR) << "Error occurred when closing file. Exception name: " << exName; | ||||
| @@ -802,7 +802,7 @@ class Lexer { | |||||
| Token token = GetNextTokenInner(); | Token token = GetNextTokenInner(); | ||||
| const char* str = token_text[token]; | const char* str = token_text[token]; | ||||
| std::string text = (str == nullptr ? GetTokenText() : str); | std::string text = (str == nullptr ? GetTokenText() : str); | ||||
| MS_LOG(DEBUG) << "------parse token] " << text; | |||||
| MS_LOG(DEBUG) << "------Parse token] " << text; | |||||
| return token; | return token; | ||||
| } | } | ||||
| @@ -1642,7 +1642,7 @@ class IrParser { | |||||
| MS_LOG(EXCEPTION) << "Expect @file at line " << lexer_.GetLineNo(); | MS_LOG(EXCEPTION) << "Expect @file at line " << lexer_.GetLineNo(); | ||||
| } | } | ||||
| // load prameter default value from serialized file | |||||
| // load parameter default value from serialized file | |||||
| py::object default_obj = LoadObject(lexer_.GetTokenText()); | py::object default_obj = LoadObject(lexer_.GetTokenText()); | ||||
| param->set_default_param(default_obj); | param->set_default_param(default_obj); | ||||
| @@ -1950,7 +1950,7 @@ class IrParser { | |||||
| return TOK_ERROR; | return TOK_ERROR; | ||||
| } | } | ||||
| // restore python funciton of PrimitivePy from serialized file | |||||
| // restore python function of PrimitivePy from serialized file | |||||
| py::object py_obj = LoadObject(lexer_.GetTokenText()); | py::object py_obj = LoadObject(lexer_.GetTokenText()); | ||||
| PrimitivePyPtr ptr = nullptr; | PrimitivePyPtr ptr = nullptr; | ||||
| if (py::hasattr(py_obj, "__setattr_flag__") && py::hasattr(py_obj, "_clone")) { | if (py::hasattr(py_obj, "__setattr_flag__") && py::hasattr(py_obj, "_clone")) { | ||||
| @@ -1958,7 +1958,7 @@ class IrParser { | |||||
| py::object new_obj = clone_fn(); | py::object new_obj = clone_fn(); | ||||
| ptr = new_obj.cast<PrimitivePyPtr>(); | ptr = new_obj.cast<PrimitivePyPtr>(); | ||||
| if (ptr == nullptr) { | if (ptr == nullptr) { | ||||
| MS_LOG(EXCEPTION) << "cast to type 'PrimitivePyPtr' error"; | |||||
| MS_LOG(EXCEPTION) << "Cast to type 'PrimitivePyPtr' error"; | |||||
| } | } | ||||
| } else { | } else { | ||||
| ptr = std::make_shared<PrimitivePy>(id.substr(strlen("PrimitivePy::")), py_obj); | ptr = std::make_shared<PrimitivePy>(id.substr(strlen("PrimitivePy::")), py_obj); | ||||
| @@ -2221,15 +2221,15 @@ class IrParser { | |||||
| }; | }; | ||||
| std::vector<FuncGraphPtr> ImportIR(const std::string& filename) { | std::vector<FuncGraphPtr> ImportIR(const std::string& filename) { | ||||
| IrParser paser(filename.c_str()); | |||||
| paser.ParseFile(); | |||||
| return paser.GetFuncGraphs(); | |||||
| IrParser parser(filename.c_str()); | |||||
| parser.ParseFile(); | |||||
| return parser.GetFuncGraphs(); | |||||
| } | } | ||||
| #ifdef ENABLE_DUMP_IR | #ifdef ENABLE_DUMP_IR | ||||
| void DumpIRProto(const FuncGraphPtr& func_graph, const std::string& suffix) { | void DumpIRProto(const FuncGraphPtr& func_graph, const std::string& suffix) { | ||||
| if (func_graph == nullptr) { | if (func_graph == nullptr) { | ||||
| MS_LOG(ERROR) << "func graph is nullptr"; | |||||
| MS_LOG(ERROR) << "Func graph is nullptr"; | |||||
| return; | return; | ||||
| } | } | ||||
| auto ms_context = MsContext::GetInstance(); | auto ms_context = MsContext::GetInstance(); | ||||
| @@ -2243,16 +2243,16 @@ void DumpIRProto(const FuncGraphPtr& func_graph, const std::string& suffix) { | |||||
| } | } | ||||
| std::string file_path = save_graphs_path + "/" + "ms_output_" + suffix + ".pb"; | std::string file_path = save_graphs_path + "/" + "ms_output_" + suffix + ".pb"; | ||||
| if (file_path.size() > PATH_MAX) { | if (file_path.size() > PATH_MAX) { | ||||
| MS_LOG(ERROR) << "file path " << file_path << " is too long."; | |||||
| MS_LOG(ERROR) << "File path " << file_path << " is too long."; | |||||
| return; | return; | ||||
| } | } | ||||
| char real_path[PATH_MAX] = {0}; | char real_path[PATH_MAX] = {0}; | ||||
| if (nullptr == realpath(file_path.c_str(), real_path)) { | if (nullptr == realpath(file_path.c_str(), real_path)) { | ||||
| MS_LOG(DEBUG) << "dir " << file_path << " does not exit."; | |||||
| MS_LOG(DEBUG) << "Dir " << file_path << " does not exit."; | |||||
| } else { | } else { | ||||
| std::string path_string = real_path; | std::string path_string = real_path; | ||||
| if (chmod(common::SafeCStr(path_string), S_IRUSR | S_IWUSR) == -1) { | if (chmod(common::SafeCStr(path_string), S_IRUSR | S_IWUSR) == -1) { | ||||
| MS_LOG(ERROR) << "modify file:" << real_path << " to rw fail."; | |||||
| MS_LOG(ERROR) << "Modify file:" << real_path << " to rw fail."; | |||||
| return; | return; | ||||
| } | } | ||||
| } | } | ||||
| @@ -362,7 +362,7 @@ Digraph::~Digraph() { | |||||
| fout_.close(); | fout_.close(); | ||||
| } | } | ||||
| } catch (const std::exception& e) { | } catch (const std::exception& e) { | ||||
| MS_LOG(ERROR) << "exception when closing file " << filename_; | |||||
| MS_LOG(ERROR) << "Exception when closing file " << filename_; | |||||
| } | } | ||||
| } | } | ||||
| @@ -208,7 +208,7 @@ void ProtoExporter::SetValueToProto(const ValuePtr& val, irpb::ValueProto* value | |||||
| TypePtr elem_type = dyn_cast<TensorType>(val)->element(); | TypePtr elem_type = dyn_cast<TensorType>(val)->element(); | ||||
| type_proto->mutable_tensor_type()->set_elem_type(GetNumberDataType(elem_type)); | type_proto->mutable_tensor_type()->set_elem_type(GetNumberDataType(elem_type)); | ||||
| } else { | } else { | ||||
| MS_LOG(WARNING) << "Not supported type " << val->type_name(); | |||||
| MS_LOG(WARNING) << "Unsupported type " << val->type_name(); | |||||
| } | } | ||||
| } | } | ||||
| @@ -101,7 +101,7 @@ bool Dump::IsConfigValid(const nlohmann::json& dumpSettings) { | |||||
| auto kernels = dumpSettings.at("kernels"); | auto kernels = dumpSettings.at("kernels"); | ||||
| if (!(enable.is_boolean() && trans_flag.is_boolean() && mode.is_number() && path.is_string() && | if (!(enable.is_boolean() && trans_flag.is_boolean() && mode.is_number() && path.is_string() && | ||||
| net_name.is_string() && iteration.is_number() && kernels.is_array())) { | net_name.is_string() && iteration.is_number() && kernels.is_array())) { | ||||
| MS_LOG(ERROR) << "element's type in Dump config json is invalid."; | |||||
| MS_LOG(ERROR) << "Element's type in Dump config json is invalid."; | |||||
| dump_enable_ = false; | dump_enable_ = false; | ||||
| return false; | return false; | ||||
| } | } | ||||
| @@ -121,7 +121,7 @@ bool Dump::IsConfigValid(const nlohmann::json& dumpSettings) { | |||||
| bool Dump::SetDumpConfFromJsonFile() { | bool Dump::SetDumpConfFromJsonFile() { | ||||
| const char* config_path_str = std::getenv("MINDSPORE_CONFIG_PATH"); | const char* config_path_str = std::getenv("MINDSPORE_CONFIG_PATH"); | ||||
| if (config_path_str != nullptr) { | if (config_path_str != nullptr) { | ||||
| MS_LOG(INFO) << "getenv MINDSPORE_CONFIG_PATH :" << config_path_str; | |||||
| MS_LOG(INFO) << "Getenv MINDSPORE_CONFIG_PATH :" << config_path_str; | |||||
| } else { | } else { | ||||
| MS_LOG(INFO) << "No need E2E Dump. please export MINDSPORE_CONFIG_PATH eg: MINDSPORE_CONFIG_PATH=/etc"; | MS_LOG(INFO) << "No need E2E Dump. please export MINDSPORE_CONFIG_PATH eg: MINDSPORE_CONFIG_PATH=/etc"; | ||||
| dump_enable_ = false; | dump_enable_ = false; | ||||
| @@ -132,7 +132,7 @@ bool Dump::SetDumpConfFromJsonFile() { | |||||
| auto id = context_ptr->device_id(); | auto id = context_ptr->device_id(); | ||||
| char real_path[PATH_MAX] = {0}; | char real_path[PATH_MAX] = {0}; | ||||
| if (nullptr == realpath(config_path_str, real_path)) { | if (nullptr == realpath(config_path_str, real_path)) { | ||||
| MS_LOG(ERROR) << "env e2e dump path error, " << config_path_str; | |||||
| MS_LOG(ERROR) << "Env e2e dump path error, " << config_path_str; | |||||
| dump_enable_ = false; | dump_enable_ = false; | ||||
| return false; | return false; | ||||
| } | } | ||||
| @@ -150,20 +150,20 @@ bool Dump::SetDumpConfFromJsonFile() { | |||||
| bool Dump::DumpToFile(const std::string& filename, const void* data, size_t len) { | bool Dump::DumpToFile(const std::string& filename, const void* data, size_t len) { | ||||
| if (filename.empty() || data == nullptr || len == 0) { | if (filename.empty() || data == nullptr || len == 0) { | ||||
| MS_LOG(ERROR) << "incorrect parameter."; | |||||
| MS_LOG(ERROR) << "Incorrect parameter."; | |||||
| return false; | return false; | ||||
| } | } | ||||
| std::string realpath; | std::string realpath; | ||||
| bool ret = GetRealPath(filename, &realpath); | bool ret = GetRealPath(filename, &realpath); | ||||
| if (!ret) { | if (!ret) { | ||||
| MS_LOG(ERROR) << "get real path failed."; | |||||
| MS_LOG(ERROR) << "Get real path failed."; | |||||
| return false; | return false; | ||||
| } | } | ||||
| std::ofstream fd; | std::ofstream fd; | ||||
| fd.open(realpath, std::ios::binary | std::ios::out); | fd.open(realpath, std::ios::binary | std::ios::out); | ||||
| if (!fd.is_open()) { | if (!fd.is_open()) { | ||||
| MS_LOG(ERROR) << "open file " << realpath << " fail."; | |||||
| MS_LOG(ERROR) << "Open file " << realpath << " fail."; | |||||
| return false; | return false; | ||||
| } | } | ||||
| (void)fd.write(reinterpret_cast<const char*>(data), SizeToLong(len)); | (void)fd.write(reinterpret_cast<const char*>(data), SizeToLong(len)); | ||||
| @@ -182,7 +182,7 @@ bool Dump::GetRealPath(const std::string& inpath, std::string* outpath) { | |||||
| if (path_split_pos != std::string::npos) { | if (path_split_pos != std::string::npos) { | ||||
| std::string prefix_path = inpath.substr(0, path_split_pos); | std::string prefix_path = inpath.substr(0, path_split_pos); | ||||
| if (prefix_path.length() >= PATH_MAX) { | if (prefix_path.length() >= PATH_MAX) { | ||||
| MS_LOG(ERROR) << "prefix path is too longer!"; | |||||
| MS_LOG(ERROR) << "Prefix path is too longer!"; | |||||
| return false; | return false; | ||||
| } | } | ||||
| std::string last_path = inpath.substr(path_split_pos, inpath.length() - path_split_pos); | std::string last_path = inpath.substr(path_split_pos, inpath.length() - path_split_pos); | ||||
| @@ -201,11 +201,11 @@ bool Dump::GetRealPath(const std::string& inpath, std::string* outpath) { | |||||
| if (path_split_pos == std::string::npos) { | if (path_split_pos == std::string::npos) { | ||||
| if (inpath.length() >= PATH_MAX) { | if (inpath.length() >= PATH_MAX) { | ||||
| MS_LOG(ERROR) << "prefix path is too longer!"; | |||||
| MS_LOG(ERROR) << "Prefix path is too longer!"; | |||||
| return false; | return false; | ||||
| } | } | ||||
| if (nullptr == realpath(inpath.c_str(), real_path)) { | if (nullptr == realpath(inpath.c_str(), real_path)) { | ||||
| MS_LOG(ERROR) << "file " << inpath << " does not exit, it will be created."; | |||||
| MS_LOG(ERROR) << "File " << inpath << " does not exit, it will be created."; | |||||
| } | } | ||||
| *outpath = std::string(real_path); | *outpath = std::string(real_path); | ||||
| } | } | ||||
| @@ -218,7 +218,7 @@ bool Dump::CreateNotExistDirs(const std::string& path) { | |||||
| MS_EXCEPTION_IF_NULL(fs); | MS_EXCEPTION_IF_NULL(fs); | ||||
| char temp_path[PATH_MAX] = {0}; | char temp_path[PATH_MAX] = {0}; | ||||
| if (path.length() > PATH_MAX) { | if (path.length() > PATH_MAX) { | ||||
| MS_LOG(ERROR) << "path lens is max than " << PATH_MAX; | |||||
| MS_LOG(ERROR) << "Path lens is max than " << PATH_MAX; | |||||
| return false; | return false; | ||||
| } | } | ||||
| for (uint32_t i = 0; i < path.length(); i++) { | for (uint32_t i = 0; i < path.length(); i++) { | ||||
| @@ -229,7 +229,7 @@ bool Dump::CreateNotExistDirs(const std::string& path) { | |||||
| temp_path[i] = '\0'; | temp_path[i] = '\0'; | ||||
| std::string path_handle(temp_path); | std::string path_handle(temp_path); | ||||
| if (!fs->FileExist(temp_path)) { | if (!fs->FileExist(temp_path)) { | ||||
| MS_LOG(INFO) << "dir " << path_handle << " does not exit, creating..."; | |||||
| MS_LOG(INFO) << "Dir " << path_handle << " does not exit, creating..."; | |||||
| if (!fs->CreateDir(temp_path)) { | if (!fs->CreateDir(temp_path)) { | ||||
| MS_LOG(ERROR) << "Create " << path_handle << " dir error"; | MS_LOG(ERROR) << "Create " << path_handle << " dir error"; | ||||
| return false; | return false; | ||||
| @@ -241,7 +241,7 @@ bool Dump::CreateNotExistDirs(const std::string& path) { | |||||
| } | } | ||||
| if (!fs->FileExist(path)) { | if (!fs->FileExist(path)) { | ||||
| MS_LOG(INFO) << "dir " << path << " does not exit, creating..."; | |||||
| MS_LOG(INFO) << "Dir " << path << " does not exit, creating..."; | |||||
| if (!fs->CreateDir(path)) { | if (!fs->CreateDir(path)) { | ||||
| MS_LOG(ERROR) << "Create " << path << " dir error"; | MS_LOG(ERROR) << "Create " << path << " dir error"; | ||||
| return false; | return false; | ||||
| @@ -193,7 +193,7 @@ void TraceManager::DebugTrace(const TraceInfoPtr& trace_info) { | |||||
| } | } | ||||
| TraceContextPtr context = std::make_shared<TraceContext>(trace_info); | TraceContextPtr context = std::make_shared<TraceContext>(trace_info); | ||||
| if (trace_info->debug_info() == nullptr) { | if (trace_info->debug_info() == nullptr) { | ||||
| MS_LOG(EXCEPTION) << "trace debug info is null"; | |||||
| MS_LOG(EXCEPTION) << "Trace debug info is null"; | |||||
| } | } | ||||
| TraceManager::trace_context_stack_.push(context); | TraceManager::trace_context_stack_.push(context); | ||||
| } | } | ||||
| @@ -205,7 +205,7 @@ void TraceManager::DebugTrace(const DebugInfoPtr& debug_info, const TraceInfoPtr | |||||
| auto cloned_info = trace_info->clone(); | auto cloned_info = trace_info->clone(); | ||||
| cloned_info->set_debug_info(debug_info); | cloned_info->set_debug_info(debug_info); | ||||
| if (cloned_info->debug_info() == nullptr) { | if (cloned_info->debug_info() == nullptr) { | ||||
| MS_LOG(EXCEPTION) << "trace debug info is null with cloned trace"; | |||||
| MS_LOG(EXCEPTION) << "Trace debug info is null with cloned trace"; | |||||
| } | } | ||||
| TraceContextPtr context = std::make_shared<TraceContext>(cloned_info); | TraceContextPtr context = std::make_shared<TraceContext>(cloned_info); | ||||
| TraceManager::trace_context_stack_.push(context); | TraceManager::trace_context_stack_.push(context); | ||||
| @@ -89,7 +89,7 @@ std::string GetDebugInfo(const DebugInfoPtr& info, SourceLineTip tip) { | |||||
| return ""; | return ""; | ||||
| } | } | ||||
| // a trace info identifys a node transform, so we can trace the node transform through | |||||
| // a trace info identifies a node transform, so we can trace the node transform through | |||||
| // a link of trace info and debug info | // a link of trace info and debug info | ||||
| std::string GetInfoWithAction(const std::vector<DebugInfoPtr>& info_vec, SourceLineTip tip) { | std::string GetInfoWithAction(const std::vector<DebugInfoPtr>& info_vec, SourceLineTip tip) { | ||||
| if (info_vec.size() < 1) { | if (info_vec.size() < 1) { | ||||
| @@ -173,7 +173,7 @@ void DumpInferStack(std::ostringstream& oss) { | |||||
| } | } | ||||
| auto graph_context = graph_infer->graph_context(); | auto graph_context = graph_infer->graph_context(); | ||||
| if (graph_context == nullptr) { | if (graph_context == nullptr) { | ||||
| MS_LOG(INFO) << "null context continue"; | |||||
| MS_LOG(INFO) << "Null context continue"; | |||||
| continue; | continue; | ||||
| } | } | ||||
| auto graph = graph_context->func_graph(); | auto graph = graph_context->func_graph(); | ||||
| @@ -264,7 +264,7 @@ void AnalyzedFuncGraphExporter::ExportFuncGraph(const std::string& filename, | |||||
| param_index = 1; | param_index = 1; | ||||
| auto tagged_func_graphs = CalcTaggedFuncGraphs(); | auto tagged_func_graphs = CalcTaggedFuncGraphs(); | ||||
| // first output grapn on the analysis stack | |||||
| // first output graph on the analysis stack | |||||
| for (const auto& node_cfg : node_cfgs) { | for (const auto& node_cfg : node_cfgs) { | ||||
| auto fg = node_cfg->context()->func_graph(); | auto fg = node_cfg->context()->func_graph(); | ||||
| // the graph is already output, skip it | // the graph is already output, skip it | ||||
| @@ -291,7 +291,7 @@ void AnalyzedFuncGraphExporter::ExportFuncGraph(const std::string& filename, | |||||
| ofs << "\n\n"; | ofs << "\n\n"; | ||||
| (void)func_graph_set.erase(fg); | (void)func_graph_set.erase(fg); | ||||
| } | } | ||||
| ofs << "# num of total funcgraphs: " << exported.size(); | |||||
| ofs << "# num of total function graphs: " << exported.size(); | |||||
| ofs.close(); | ofs.close(); | ||||
| } | } | ||||
| @@ -332,7 +332,7 @@ void GetInferStackInfo(std::ostringstream& oss) { | |||||
| MS_LOG(INFO) << "Get graph analysis information *end*"; | MS_LOG(INFO) << "Get graph analysis information *end*"; | ||||
| } | } | ||||
| // trace the graph evaluator statck | |||||
| // trace the graph evaluator stack | |||||
| static std::stack<std::pair<abstract::EvaluatorPtr, abstract::AnfNodeConfigPtr>> graph_infer_stack; | static std::stack<std::pair<abstract::EvaluatorPtr, abstract::AnfNodeConfigPtr>> graph_infer_stack; | ||||
| // trace the cnode infer debug info | // trace the cnode infer debug info | ||||
| static std::vector<abstract::AnfNodeConfigPtr> cnode_debug_stack{}; | static std::vector<abstract::AnfNodeConfigPtr> cnode_debug_stack{}; | ||||
| @@ -36,6 +36,6 @@ std::string TraceInfo::GetActionBetweenNode(const DebugInfoPtr& info) { | |||||
| } else if (debug_info()->trace_info() != nullptr) { | } else if (debug_info()->trace_info() != nullptr) { | ||||
| return act_name + debug_info()->trace_info()->GetActionBetweenNode(info); | return act_name + debug_info()->trace_info()->GetActionBetweenNode(info); | ||||
| } | } | ||||
| return "not in the traced info"; | |||||
| return "Not in the traced info"; | |||||
| } | } | ||||
| } // namespace mindspore | } // namespace mindspore | ||||
| @@ -83,7 +83,7 @@ class AnfVisitor; | |||||
| // Methods: | // Methods: | ||||
| // func_graph: return FuncGraph that this AnfNode belongs to. | // func_graph: return FuncGraph that this AnfNode belongs to. | ||||
| // scope: return the scope namespace of this AnfNode. Set it using set_scope. | // scope: return the scope namespace of this AnfNode. Set it using set_scope. | ||||
| // abstract: return the cached inferred abstract value. It cantains type, shape | |||||
| // abstract: return the cached inferred abstract value. It contains type, shape | |||||
| // value. Set New cache using set_abstract. | // value. Set New cache using set_abstract. | ||||
| // intermediate_abstract: return the cached inferring abstract value. | // intermediate_abstract: return the cached inferring abstract value. | ||||
| // Type/Shape: return the related info of this AnfNode. When this AnfNode is an | // Type/Shape: return the related info of this AnfNode. When this AnfNode is an | ||||
| @@ -284,7 +284,7 @@ class Parameter : public ANode { | |||||
| }; | }; | ||||
| using ParameterPtr = std::shared_ptr<Parameter>; | using ParameterPtr = std::shared_ptr<Parameter>; | ||||
| // Value is used to represent the atomic expression metioned in BNF. | |||||
| // Value is used to represent the atomic expression mentioned in BNF. | |||||
| // It mainly be stored in ValueNode. Value and ValueNode is related definition. | // It mainly be stored in ValueNode. Value and ValueNode is related definition. | ||||
| class Value : public Base { | class Value : public Base { | ||||
| public: | public: | ||||
| @@ -313,7 +313,7 @@ using ValuePtr = std::shared_ptr<Value>; | |||||
| using ValuePtrList = std::vector<ValuePtr>; | using ValuePtrList = std::vector<ValuePtr>; | ||||
| // ValueNode is used to hold value. Unlike CNode and Parameter, ValueNode | // ValueNode is used to hold value. Unlike CNode and Parameter, ValueNode | ||||
| // do not belong to any particular function graph. | |||||
| // does not belong to any particular function graph. | |||||
| class ValueNode : public ANode { | class ValueNode : public ANode { | ||||
| public: | public: | ||||
| explicit ValueNode(const ValuePtr &value) : value_(value) {} | explicit ValueNode(const ValuePtr &value) : value_(value) {} | ||||
| @@ -34,19 +34,19 @@ bool Number::operator==(const Type& other) const { | |||||
| Int::Int(const int nbits) : Number(IntBitsToTypeId(nbits), nbits, false) { | Int::Int(const int nbits) : Number(IntBitsToTypeId(nbits), nbits, false) { | ||||
| if (nbits != 8 && nbits != 16 && nbits != 32 && nbits != 64) { | if (nbits != 8 && nbits != 16 && nbits != 32 && nbits != 64) { | ||||
| MS_LOG(EXCEPTION) << "wrong number of bits."; | |||||
| MS_LOG(EXCEPTION) << "Wrong number of bits."; | |||||
| } | } | ||||
| } | } | ||||
| UInt::UInt(const int nbits) : Number(UIntBitsToTypeId(nbits), nbits, false) { | UInt::UInt(const int nbits) : Number(UIntBitsToTypeId(nbits), nbits, false) { | ||||
| if (nbits != 8 && nbits != 16 && nbits != 32 && nbits != 64) { | if (nbits != 8 && nbits != 16 && nbits != 32 && nbits != 64) { | ||||
| MS_LOG(EXCEPTION) << "wrong number of bits."; | |||||
| MS_LOG(EXCEPTION) << "Wrong number of bits."; | |||||
| } | } | ||||
| } | } | ||||
| Float::Float(const int nbits) : Number(FloatBitsToTypeId(nbits), nbits, false) { | Float::Float(const int nbits) : Number(FloatBitsToTypeId(nbits), nbits, false) { | ||||
| if (nbits != 16 && nbits != 32 && nbits != 64) { | if (nbits != 16 && nbits != 32 && nbits != 64) { | ||||
| MS_LOG(EXCEPTION) << "wrong number of bits."; | |||||
| MS_LOG(EXCEPTION) << "Wrong number of bits."; | |||||
| } | } | ||||
| } | } | ||||
| @@ -37,7 +37,7 @@ TypeId IntBitsToTypeId(const int nbits) { | |||||
| case 64: | case 64: | ||||
| return kNumberTypeInt64; | return kNumberTypeInt64; | ||||
| default: | default: | ||||
| MS_LOG(EXCEPTION) << "wrong number of bits."; | |||||
| MS_LOG(EXCEPTION) << "Wrong number of bits."; | |||||
| } | } | ||||
| } | } | ||||
| @@ -52,7 +52,7 @@ TypeId UIntBitsToTypeId(const int nbits) { | |||||
| case 64: | case 64: | ||||
| return kNumberTypeUInt64; | return kNumberTypeUInt64; | ||||
| default: | default: | ||||
| MS_LOG(EXCEPTION) << "wrong number of bits."; | |||||
| MS_LOG(EXCEPTION) << "Wrong number of bits."; | |||||
| } | } | ||||
| } | } | ||||
| @@ -65,7 +65,7 @@ TypeId FloatBitsToTypeId(const int nbits) { | |||||
| case 64: | case 64: | ||||
| return kNumberTypeFloat64; | return kNumberTypeFloat64; | ||||
| default: | default: | ||||
| MS_LOG(EXCEPTION) << "wrong number of bits."; | |||||
| MS_LOG(EXCEPTION) << "Wrong number of bits."; | |||||
| } | } | ||||
| } | } | ||||
| @@ -174,7 +174,7 @@ class FuncGraph : public FuncGraphBase { | |||||
| GraphDebugInfoPtr debug_info(); | GraphDebugInfoPtr debug_info(); | ||||
| void set_debug_info(const GraphDebugInfoPtr &info) { | void set_debug_info(const GraphDebugInfoPtr &info) { | ||||
| if (info == nullptr) { | if (info == nullptr) { | ||||
| MS_LOG(EXCEPTION) << "graph set null debug info"; | |||||
| MS_LOG(EXCEPTION) << "Graph set null debug info"; | |||||
| } | } | ||||
| this->debug_info_ = info; | this->debug_info_ = info; | ||||
| } | } | ||||
| @@ -817,7 +817,7 @@ void FuncGraphChildDirect::OnMoveAllCNode(FuncGraphPtr src, FuncGraphPtr dst) { | |||||
| void FuncGraphParentsDirectCollector::OnModEdge(AnfNodePtr node, int, AnfNodePtr inp, EdgeProcessDirection direction) { | void FuncGraphParentsDirectCollector::OnModEdge(AnfNodePtr node, int, AnfNodePtr inp, EdgeProcessDirection direction) { | ||||
| MS_EXCEPTION_IF_NULL(node); | MS_EXCEPTION_IF_NULL(node); | ||||
| FuncGraphPtr fg1 = node->func_graph(); | FuncGraphPtr fg1 = node->func_graph(); | ||||
| // possible chirld parent | |||||
| // possible child parent | |||||
| if (IsValueNode<FuncGraph>(inp)) { | if (IsValueNode<FuncGraph>(inp)) { | ||||
| FuncGraphPtr fg2 = GetValueNode<FuncGraphPtr>(inp); | FuncGraphPtr fg2 = GetValueNode<FuncGraphPtr>(inp); | ||||
| if (Mod(fg1, ParentProxy(fg2), direction)) { | if (Mod(fg1, ParentProxy(fg2), direction)) { | ||||
| @@ -1181,7 +1181,7 @@ bool FuncGraphJTotalComputer::SeekJ(const FuncGraphPtr& fg, const FuncGraphSetPt | |||||
| } | } | ||||
| path->add(fg); | path->add(fg); | ||||
| // checkg if func graphs used contains J(func_graph); | |||||
| // check if func graphs used contains J(func_graph); | |||||
| auto& used = this->manager_->func_graphs_used(); | auto& used = this->manager_->func_graphs_used(); | ||||
| for (auto& item : used[fg]) { | for (auto& item : used[fg]) { | ||||
| auto used_g = item.first; | auto used_g = item.first; | ||||
| @@ -650,7 +650,7 @@ class FuncGraphTransaction { | |||||
| explicit FuncGraphTransaction(FuncGraphManager* manager) : manager_(manager), changes_() { | explicit FuncGraphTransaction(FuncGraphManager* manager) : manager_(manager), changes_() { | ||||
| MS_EXCEPTION_IF_NULL(manager_); | MS_EXCEPTION_IF_NULL(manager_); | ||||
| if (!manager_->IsManaged()) { | if (!manager_->IsManaged()) { | ||||
| MS_LOG(DEBUG) << "the manager is not managed yet"; | |||||
| MS_LOG(DEBUG) << "The manager is not managed yet"; | |||||
| } | } | ||||
| } | } | ||||
| @@ -148,7 +148,7 @@ class MetaTensor : public Value { | |||||
| // | // | ||||
| // The constructed MetaTensor object has the same type and shape with meta_tensor. | // The constructed MetaTensor object has the same type and shape with meta_tensor. | ||||
| // | // | ||||
| // param meta_tensor An exisiting MetaTensor object. | |||||
| // param meta_tensor An existing MetaTensor object. | |||||
| virtual MetaTensor& operator=(const MetaTensor& meta_tensor); | virtual MetaTensor& operator=(const MetaTensor& meta_tensor); | ||||
| // brief Compares two MetaTensor objects. | // brief Compares two MetaTensor objects. | ||||
| @@ -166,7 +166,7 @@ class MetaTensor : public Value { | |||||
| TypeId data_type() const { return data_type_; } | TypeId data_type() const { return data_type_; } | ||||
| std::string ToString() const override; | std::string ToString() const override; | ||||
| std::string DumpText() const override; | std::string DumpText() const override; | ||||
| // bried Sets the data type of a tensor in its MetaTensor. | |||||
| // brief Sets the data type of a tensor in its MetaTensor. | |||||
| // | // | ||||
| // param data_type The data type of the tensor to be set. | // param data_type The data type of the tensor to be set. | ||||
| virtual TypeId set_data_type(const TypeId data_type) { | virtual TypeId set_data_type(const TypeId data_type) { | ||||
| @@ -314,7 +314,7 @@ class Tensor : public MetaTensor { | |||||
| // | // | ||||
| // The constructed Tensor object has the same type and shape with tensor. | // The constructed Tensor object has the same type and shape with tensor. | ||||
| // | // | ||||
| // param tensor An exisiting Tensor object. | |||||
| // param tensor An existing Tensor object. | |||||
| Tensor& operator=(const Tensor& tensor); | Tensor& operator=(const Tensor& tensor); | ||||
| // brief Compares two Tensor objects. | // brief Compares two Tensor objects. | ||||
| @@ -383,7 +383,7 @@ class Tensor : public MetaTensor { | |||||
| // return The [TypeId] of the tensor data. | // return The [TypeId] of the tensor data. | ||||
| TypeId GetDataType(const py::buffer_info& buf) const; | TypeId GetDataType(const py::buffer_info& buf) const; | ||||
| // bried Sets the data type of a tensor. | |||||
| // brief Sets the data type of a tensor. | |||||
| // | // | ||||
| // param data_type The data type of the tensor to be set. | // param data_type The data type of the tensor to be set. | ||||
| // | // | ||||
| @@ -43,14 +43,13 @@ VisitFuncType AnfVisitor::Match(const PrimitivePtr &prim, const std::vector<opt: | |||||
| } | } | ||||
| auto &inputs = node->cast<CNodePtr>()->inputs(); | auto &inputs = node->cast<CNodePtr>()->inputs(); | ||||
| // infact, funcs_size == inps_size - 1 | |||||
| auto funcs_size = funcs.size(); | auto funcs_size = funcs.size(); | ||||
| auto inps_size = inputs.size(); | |||||
| auto inputs_size = inputs.size(); | |||||
| // check the inputs are matched with the predicate functions | // check the inputs are matched with the predicate functions | ||||
| if (funcs_size > 0) { | if (funcs_size > 0) { | ||||
| // use the predicate function list to check the number of inputs | // use the predicate function list to check the number of inputs | ||||
| if (funcs_size != (inps_size - 1)) { | |||||
| if (funcs_size != (inputs_size - 1)) { | |||||
| return; | return; | ||||
| } | } | ||||
| @@ -63,7 +62,7 @@ VisitFuncType AnfVisitor::Match(const PrimitivePtr &prim, const std::vector<opt: | |||||
| } | } | ||||
| // visit the inputs | // visit the inputs | ||||
| for (size_t i = 1; i < inps_size; i++) { | |||||
| for (size_t i = 1; i < inputs_size; i++) { | |||||
| this->Visit(inputs[i]); | this->Visit(inputs[i]); | ||||
| } | } | ||||
| }; | }; | ||||
| @@ -36,7 +36,7 @@ | |||||
| #endif | #endif | ||||
| const char SINGLE_OP_GRAPH[] = "single_op_graph"; | const char SINGLE_OP_GRAPH[] = "single_op_graph"; | ||||
| // primitive unable to infer value for constant input in pynative mode | |||||
| // primitive unable to infer value for constant input in PyNative mode | |||||
| const std::unordered_set<std::string> vm_operators = {"partial", "depend"}; | const std::unordered_set<std::string> vm_operators = {"partial", "depend"}; | ||||
| namespace mindspore { | namespace mindspore { | ||||
| @@ -45,7 +45,7 @@ inline ValuePtr PyAttrValue(const py::object& obj) { | |||||
| ValuePtr converted_ret = nullptr; | ValuePtr converted_ret = nullptr; | ||||
| bool converted = parse::ConvertData(obj, &converted_ret); | bool converted = parse::ConvertData(obj, &converted_ret); | ||||
| if (!converted) { | if (!converted) { | ||||
| MS_LOG(EXCEPTION) << "attribute convert error with type:" << std::string(py::str(obj)); | |||||
| MS_LOG(EXCEPTION) << "Attribute convert error with type:" << std::string(py::str(obj)); | |||||
| } | } | ||||
| return converted_ret; | return converted_ret; | ||||
| } | } | ||||
| @@ -67,7 +67,7 @@ void PynativeInfer(const PrimitivePyPtr& prim, const py::tuple& py_args, OpExecI | |||||
| OpExecInfoPtr GenerateOpExecInfo(const py::args& args) { | OpExecInfoPtr GenerateOpExecInfo(const py::args& args) { | ||||
| if (args.size() != PY_ARGS_NUM) { | if (args.size() != PY_ARGS_NUM) { | ||||
| MS_LOG(ERROR) << "four args are needed by RunOp"; | |||||
| MS_LOG(ERROR) << "Four args are needed by RunOp"; | |||||
| return nullptr; | return nullptr; | ||||
| } | } | ||||
| auto op_exec_info = std::make_shared<OpExecInfo>(); | auto op_exec_info = std::make_shared<OpExecInfo>(); | ||||
| @@ -145,13 +145,13 @@ py::object RunOpInVM(const OpExecInfoPtr& op_exec_info, PynativeStatusCode* stat | |||||
| py::object RunOpInMs(const OpExecInfoPtr& op_exec_info, PynativeStatusCode* status) { | py::object RunOpInMs(const OpExecInfoPtr& op_exec_info, PynativeStatusCode* status) { | ||||
| MS_EXCEPTION_IF_NULL(op_exec_info); | MS_EXCEPTION_IF_NULL(op_exec_info); | ||||
| MS_LOG(INFO) << "start run op[" << op_exec_info->op_name << "] with backend policy ms"; | |||||
| MS_LOG(INFO) << "Start run op[" << op_exec_info->op_name << "] with backend policy ms"; | |||||
| auto ms_context = MsContext::GetInstance(); | auto ms_context = MsContext::GetInstance(); | ||||
| MS_EXCEPTION_IF_NULL(ms_context); | MS_EXCEPTION_IF_NULL(ms_context); | ||||
| ms_context->set_enable_pynative_infer(true); | ms_context->set_enable_pynative_infer(true); | ||||
| std::string device_target = ms_context->device_target(); | std::string device_target = ms_context->device_target(); | ||||
| if (device_target != kAscendDevice && device_target != kGPUDevice) { | if (device_target != kAscendDevice && device_target != kGPUDevice) { | ||||
| MS_EXCEPTION(ArgumentError) << "device target [" << device_target << "] is not supported in Pynative mode"; | |||||
| MS_EXCEPTION(ArgumentError) << "Device target [" << device_target << "] is not supported in Pynative mode"; | |||||
| } | } | ||||
| std::shared_ptr<session::SessionBasic> session = session::SessionFactory::Get().Create(device_target); | std::shared_ptr<session::SessionBasic> session = session::SessionFactory::Get().Create(device_target); | ||||
| MS_EXCEPTION_IF_NULL(session); | MS_EXCEPTION_IF_NULL(session); | ||||
| @@ -197,7 +197,7 @@ py::object RunOpWithBackendPolicy(MsBackendPolicy backend_policy, const OpExecIn | |||||
| break; | break; | ||||
| } | } | ||||
| default: | default: | ||||
| MS_LOG(ERROR) << "No backend configed for run op"; | |||||
| MS_LOG(ERROR) << "No backend configured for run op"; | |||||
| } | } | ||||
| return result; | return result; | ||||
| } | } | ||||
| @@ -240,7 +240,7 @@ py::tuple RunOp(const py::args& args) { | |||||
| } | } | ||||
| result = RunOpWithBackendPolicy(backend_policy, op_exec_info, &status); | result = RunOpWithBackendPolicy(backend_policy, op_exec_info, &status); | ||||
| if (status != PYNATIVE_SUCCESS) { | if (status != PYNATIVE_SUCCESS) { | ||||
| MS_LOG(ERROR) << "Fail to run " << op_exec_info->op_name; | |||||
| MS_LOG(ERROR) << "Failed to run " << op_exec_info->op_name; | |||||
| return err_ret; | return err_ret; | ||||
| } | } | ||||
| @@ -47,7 +47,7 @@ inline ValuePtr PyAttrValue(const py::object& obj) { | |||||
| ValuePtr converted_ret = nullptr; | ValuePtr converted_ret = nullptr; | ||||
| bool converted = parse::ConvertData(obj, &converted_ret); | bool converted = parse::ConvertData(obj, &converted_ret); | ||||
| if (!converted) { | if (!converted) { | ||||
| MS_LOG(EXCEPTION) << "attribute convert error with type:" << std::string(py::str(obj)); | |||||
| MS_LOG(EXCEPTION) << "Attribute convert error with type:" << std::string(py::str(obj)); | |||||
| } | } | ||||
| return converted_ret; | return converted_ret; | ||||
| } | } | ||||
| @@ -67,7 +67,7 @@ MeTensorPtr ConvertPyObjToTensor(const py::object& obj) { | |||||
| } else if (py::isinstance<py::array>(obj)) { | } else if (py::isinstance<py::array>(obj)) { | ||||
| me_tensor_ptr = std::make_shared<MeTensor>(py::cast<py::array>(obj), nullptr); | me_tensor_ptr = std::make_shared<MeTensor>(py::cast<py::array>(obj), nullptr); | ||||
| } else { | } else { | ||||
| MS_LOG(EXCEPTION) << "run op inputs type is invalid!"; | |||||
| MS_LOG(EXCEPTION) << "Run op inputs type is invalid!"; | |||||
| } | } | ||||
| return me_tensor_ptr; | return me_tensor_ptr; | ||||
| } | } | ||||
| @@ -97,7 +97,7 @@ bool SetInputsForSingleOpGraph(const OpExecInfoPtr& op_exec_info, const std::vec | |||||
| auto const_op_desc = | auto const_op_desc = | ||||
| transform::TransformUtil::GetGeTensorDesc(me_tensor_ptr->shape_c(), me_tensor_ptr->data_type(), kOpFormat_NCHW); | transform::TransformUtil::GetGeTensorDesc(me_tensor_ptr->shape_c(), me_tensor_ptr->data_type(), kOpFormat_NCHW); | ||||
| if (const_op_desc == nullptr) { | if (const_op_desc == nullptr) { | ||||
| MS_LOG(ERROR) << "Create variable " << op_name << " ouptut descriptor failed!"; | |||||
| MS_LOG(ERROR) << "Create variable " << op_name << " output descriptor failed!"; | |||||
| return false; | return false; | ||||
| } | } | ||||
| auto pointer_cast_const_op = std::static_pointer_cast<transform::Constant>(const_op); | auto pointer_cast_const_op = std::static_pointer_cast<transform::Constant>(const_op); | ||||
| @@ -108,7 +108,7 @@ bool SetInputsForSingleOpGraph(const OpExecInfoPtr& op_exec_info, const std::vec | |||||
| continue; | continue; | ||||
| } | } | ||||
| if (adapter->setInput(op, op_input_idx++, const_op)) { | if (adapter->setInput(op, op_input_idx++, const_op)) { | ||||
| MS_LOG(ERROR) << "fail to set params, index is " << op_input_idx; | |||||
| MS_LOG(ERROR) << "Failed to set params, index is " << op_input_idx; | |||||
| return false; | return false; | ||||
| } | } | ||||
| graph_input_nodes->push_back(*const_op); | graph_input_nodes->push_back(*const_op); | ||||
| @@ -178,7 +178,7 @@ void ToTensorPtr(const OpExecInfoPtr op_exec_info, std::vector<GeTensorPtr>* con | |||||
| MeTensorPtr me_tensor_ptr = ConvertPyObjToTensor(op_inputs[i]); | MeTensorPtr me_tensor_ptr = ConvertPyObjToTensor(op_inputs[i]); | ||||
| auto ge_tensor_ptr = transform::TransformUtil::ConvertTensor(me_tensor_ptr, kOpFormat_NCHW); | auto ge_tensor_ptr = transform::TransformUtil::ConvertTensor(me_tensor_ptr, kOpFormat_NCHW); | ||||
| if (ge_tensor_ptr == nullptr) { | if (ge_tensor_ptr == nullptr) { | ||||
| MS_LOG(EXCEPTION) << "convert inputs to GE tensor failed in op " << op_exec_info->op_name << "."; | |||||
| MS_LOG(EXCEPTION) << "Convert inputs to GE tensor failed in op " << op_exec_info->op_name << "."; | |||||
| } | } | ||||
| // set inputs for operator to build single node graph | // set inputs for operator to build single node graph | ||||
| inputs->push_back(ge_tensor_ptr); | inputs->push_back(ge_tensor_ptr); | ||||
| @@ -192,7 +192,7 @@ PynativeStatusCode ConvertAttributes(const OpExecInfoPtr& op_exec_info, const st | |||||
| for (auto& item : op_attrs) { | for (auto& item : op_attrs) { | ||||
| if (!py::isinstance<py::str>(item.first)) { | if (!py::isinstance<py::str>(item.first)) { | ||||
| MS_LOG(ERROR) << "type error in py dict convert"; | |||||
| MS_LOG(ERROR) << "Type error in py dict convert"; | |||||
| return PYNATIVE_OP_ATTRS_ERR; | return PYNATIVE_OP_ATTRS_ERR; | ||||
| } | } | ||||
| std::string name = py::cast<std::string>(item.first); | std::string name = py::cast<std::string>(item.first); | ||||
| @@ -203,7 +203,7 @@ PynativeStatusCode ConvertAttributes(const OpExecInfoPtr& op_exec_info, const st | |||||
| // build graph | // build graph | ||||
| GeGraphPtr graph = std::make_shared<GeGraph>(op_exec_info->op_name); | GeGraphPtr graph = std::make_shared<GeGraph>(op_exec_info->op_name); | ||||
| if (BuildSingleOpGraph(op_exec_info, inputs, attrs, graph) == false) { | if (BuildSingleOpGraph(op_exec_info, inputs, attrs, graph) == false) { | ||||
| MS_LOG(ERROR) << "Fail to BuildSingleOpGraph"; | |||||
| MS_LOG(ERROR) << "Failed to BuildSingleOpGraph"; | |||||
| return PYNATIVE_GRAPH_GE_BUILD_ERR; | return PYNATIVE_GRAPH_GE_BUILD_ERR; | ||||
| } | } | ||||
| @@ -211,7 +211,7 @@ PynativeStatusCode ConvertAttributes(const OpExecInfoPtr& op_exec_info, const st | |||||
| transform::Status ret = | transform::Status ret = | ||||
| transform::DfGraphManager::GetInstance().AddGraph(SINGLE_OP_GRAPH, std::shared_ptr<transform::DfGraph>(graph)); | transform::DfGraphManager::GetInstance().AddGraph(SINGLE_OP_GRAPH, std::shared_ptr<transform::DfGraph>(graph)); | ||||
| if (ret != transform::SUCCESS) { | if (ret != transform::SUCCESS) { | ||||
| MS_LOG(ERROR) << "Fail to AddGraph into graph manager"; | |||||
| MS_LOG(ERROR) << "Failed to AddGraph into graph manager"; | |||||
| return PYNATIVE_GRAPH_MANAGER_ERR; | return PYNATIVE_GRAPH_MANAGER_ERR; | ||||
| } | } | ||||
| @@ -289,7 +289,7 @@ py::object RunOpInGE(const OpExecInfoPtr& op_exec_info, PynativeStatusCode* stat | |||||
| run_ret = graph_runner->RunGraph(run_options, ge_inputs, &ge_outputs); | run_ret = graph_runner->RunGraph(run_options, ge_inputs, &ge_outputs); | ||||
| } | } | ||||
| if (run_ret != transform::Status::SUCCESS) { | if (run_ret != transform::Status::SUCCESS) { | ||||
| MS_LOG(ERROR) << "GraphRunner Fails to Run Graph"; | |||||
| MS_LOG(ERROR) << "GraphRunner fails to run graph"; | |||||
| *status = PYNATIVE_GRAPH_GE_RUN_ERR; | *status = PYNATIVE_GRAPH_GE_RUN_ERR; | ||||
| return std::move(err_ret); | return std::move(err_ret); | ||||
| } | } | ||||