|
- /**
- * Copyright 2019 Huawei Technologies Co., Ltd
- *
- * Licensed under the Apache License, Version 2.0 (the "License");
- * you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
- #ifndef MINDSPORE_CCSRC_PARALLEL_STEP_PARALLEL_H_
- #define MINDSPORE_CCSRC_PARALLEL_STEP_PARALLEL_H_
-
- #include <vector>
-
- #include <map>
- #include <memory>
- #include <string>
- #include <unordered_map>
- #include <utility>
- #include <set>
-
- #include "./common.h"
- #include "optimizer/opt.h"
- #include "parallel/strategy.h"
- #include "parallel/tensor_layout/tensor_redistribution.h"
-
- using OperatorInfoPtr = std::shared_ptr<mindspore::parallel::OperatorInfo>;
-
- namespace mindspore {
- namespace parallel {
- const uint64_t kUSecondInSecond = 1000000;
-
- struct LossNodeInfo {
- bool has_tuple_getitem = false;
- int dout_index = 0; // now don't support the sens is a tuple
- };
-
- std::vector<AnfNodePtr> CreateInput(const Operator &op, const AnfNodePtr &node, const std::string &instance_name);
- std::string CreateInstanceName(const CNodePtr &node, size_t index);
- void ForwardCommunication(OperatorVector forward_op, const CNodePtr &node);
-
- void InsertRedistribution(const RedistributionOpListPtr &redistribution_oplist_ptr, const CNodePtr &node,
- const FuncGraphPtr &func_graph, int pos, const CNodePtr &pre_node);
-
- TensorLayout GetTensorInLayout(const CNodePtr &pre_node, const PrimitivePtr &pre_prim,
- const OperatorInfoPtr &distribute_operator_pre);
-
- OperatorInfoPtr GetDistributeOperator(const CNodePtr &node);
-
- void Redistribution(const std::pair<AnfNodePtr, int> &node_pair, const OperatorInfoPtr &distribute_operator,
- const CNodePtr &middle_node, int index, TensorRedistribution tensor_redistribution,
- const CNodePtr &pre_node);
-
- bool StrategyFound(std::unordered_map<std::string, ValuePtr> attrs);
-
- bool IsParallelCareNode(const CNodePtr &cnode);
-
- void MarkForwardCNode(const FuncGraphPtr &root);
-
- bool FindCommunicationOp(const std::vector<AnfNodePtr> &all_nodes);
-
- void StepRedistribution(const CNodePtr &node, const OperatorInfoPtr &distribute_operator, const CNodePtr &insert_node,
- const TensorRedistribution &tensor_redistribution, const CNodePtr &pre_node);
-
- std::vector<AnfNodePtr> ReplaceOpInput(const Operator &replace_op, const std::string &instance_name,
- const CNodePtr &node);
-
- void StepReplaceOp(OperatorVector replace_op, const CNodePtr &node);
-
- void InsertVirtualDivOp(const VirtualDivOp &virtual_div_op, const CNodePtr &node);
-
- std::pair<AnfNodePtr, bool> FindParameter(const AnfNodePtr &node, const FuncGraphPtr &func_graph);
-
- std::pair<bool, CNodePtr> FindCNode(const AnfNodePtr &anode, const std::string &name, const FuncGraphPtr &func_graph);
-
- void InsertMirrorOps(const MirrorOps &mirror_ops, const CNodePtr &node);
-
- void BackwardCommunication(const OperatorInfoPtr &distribute_operator, const CNodePtr &node, bool is_loss_node);
-
- // Generate and init parallel operator
- OperatorInfoPtr OperatorInstance(const PrimitivePtr &prim, const PrimitiveAttrs &attrs,
- const std::vector<Shapes> &shape_list);
-
- // Generate without initing parallel operator
- OperatorInfoPtr NewOperatorInstance(const PrimitivePtr &prim, const PrimitiveAttrs &attrs,
- std::vector<Shapes> shape_list);
-
- // Extract strategy from attr
- StrategyPtr ExtractStrategy(std::unordered_map<std::string, ValuePtr> attrs);
-
- Shapes GetNodeShape(const AnfNodePtr &node);
-
- std::vector<AnfNodePtr> FindParameterByRefKeyNode(const AnfNodePtr &node, const FuncGraphPtr &func_graph);
-
- // Extract shape from anfnode
- std::vector<Shapes> ExtractShape(const CNodePtr &node);
-
- std::pair<AnfNodePtr, int> FindParallelCareNode(const AnfNodePtr &node);
-
- // Find finally sub graph
- std::pair<AnfNodePtr, int> FindSubGraph(const FuncGraphPtr &func_graph, const AnfNodePtr ¶meter);
-
- // Set distribute shape for parameters abstract
- void SetParallelShape(const AnfNodePtr ¶meter, const std::pair<AnfNodePtr, int> &res);
-
- // change parameters'shape in resource
- void CoverSliceShape(const FuncGraphPtr &root);
-
- void SetVirtualDatasetStrategy(const CNodePtr &node);
-
- // Creat parallel operator for primitive node(has strategy)
- void ExtractInformation(const std::vector<AnfNodePtr> &all_nodes);
-
- TensorLayout GetInputLayoutFromCNode(const std::pair<AnfNodePtr, int> &node_pair);
-
- std::shared_ptr<TensorLayout> FindNextLayout(const CNodePtr &node);
-
- std::shared_ptr<TensorLayout> GetOutputLayoutFromCNode(const CNodePtr &cnode, size_t output_index);
-
- std::shared_ptr<TensorLayout> FindPrevParallelCareNodeLayout(const AnfNodePtr &node, size_t output_index);
-
- std::shared_ptr<TensorLayout> FindPrevLayout(const AnfNodePtr &node);
-
- void ReshapeInit(const std::vector<AnfNodePtr> &all_nodes);
-
- // Add node for whole graph
- void ParallelCommunication(const FuncGraphPtr &root, const std::vector<AnfNodePtr> &all_nodes,
- const FuncGraphManagerPtr &manager);
-
- void RestoreStrategy(const FuncGraphPtr &func_graph);
-
- void CheckpointStrategy(const FuncGraphPtr &func_graph);
-
- // main step of Parallel
- bool StepParallel(const FuncGraphPtr &func_graph, const opt::OptimizerPtr &optimizer);
-
- int32_t GetTupleGetItemIndex(const CNodePtr &cnode);
-
- std::vector<CNodePtr> FindLossCNodeFromRoot(const FuncGraphPtr &root);
-
- Status ParallelInit();
-
- std::vector<std::string> ExtractInputsTensorName(const CNodePtr &node);
-
- std::set<FuncGraphPtr> ForwardGraph(const FuncGraphPtr &root);
- } // namespace parallel
- } // namespace mindspore
-
- #endif // MINDSPORE_CCSRC_PARALLEL_STEP_PARALLEL_H_
|