|
123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282 |
- /**
- * Copyright 2020 Huawei Technologies Co., Ltd
- *
- * Licensed under the Apache License, Version 2.0 (the "License");
- * you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
- #ifndef MINDSPORE_CCSRC_DEBUG_DEBUGGER_DEBUGGER_H_
- #define MINDSPORE_CCSRC_DEBUG_DEBUGGER_DEBUGGER_H_
-
- #include <list>
- #include <memory>
- #include <string>
- #include <utility>
- #include <vector>
- #include <map>
- #include "backend/session/kernel_graph.h"
- #include "debug/debugger/grpc_client.h"
- #include "debug/debug_services.h"
-
- using debugger::Chunk;
- using debugger::DataType;
- using debugger::EventReply;
- using debugger::GraphProto;
- using debugger::ModelProto;
- using debugger::TensorProto;
- using debugger::WatchCondition;
- using debugger::WatchCondition_Parameter;
- using debugger::WatchNode;
- using debugger::WatchpointHit;
-
- template <class T>
- using ProtoVector = google::protobuf::RepeatedPtrField<T>;
-
- namespace mindspore {
- // different types of command recieved by debugger
- // need to keep sync with client-side proto and server-side proto
- enum class DebuggerCommand {
- kExitCMD = 2,
- kRunCMD = 3,
- kSetCMD = 4,
- kViewCMD = 5,
- kVersionMatchedCMD = 6,
- kUnknownCMD = -1
- };
-
- class Debugger : public std::enable_shared_from_this<Debugger> {
- public:
- static std::shared_ptr<Debugger> GetInstance() {
- std::lock_guard<std::mutex> i_lock(instance_lock_);
- if (debugger_ == nullptr) {
- debugger_ = std::shared_ptr<Debugger>(new (std::nothrow) Debugger());
- }
- return debugger_;
- }
-
- // deconstructor
- ~Debugger() = default;
-
- // init
- // only save device_id
- void Init(const uint32_t device_id, const std::string device_target);
-
- // reset debugger
- void Reset();
-
- // enable debugger
- // send graph and wait for command
- // do nothing if graph is set already
- void PreExecute(const KernelGraphPtr &graph_ptr, uint32_t graph_sum = 1);
-
- // analyze tensors and wait for command
- // don't need a graph_ptr because it is saved during pre_execute
- void PostExecute();
-
- bool ReadNodeDataRequired(const CNodePtr &kernel);
-
- void PostExecuteNode(const CNodePtr &kernel);
-
- // suspend the execution after a debug_op
- void PostDebugOp();
-
- bool DumpTensorToFile(const std::string &tensor_name, bool trans_flag, const std::string &filepath,
- const std::string &host_fmt, const std::vector<int64_t> &host_shape, TypeId host_type,
- TypeId addr_type_id, const std::string &addr_format, size_t slot) const;
-
- bool DebugServicesIsWatchPoint(const std::string &kernel_name, const CNodePtr &kernel = nullptr) const;
-
- void EmptyTensor();
-
- void SetTensorLoaderIterNum(uint32_t iter_num);
-
- void EmptyPrevTensor();
-
- uint32_t GetTensorLoaderIterNum() const;
-
- bool LoadNewTensor(const std::shared_ptr<TensorData> &tensor, bool keep_prev);
-
- bool debugger_enabled() const;
-
- bool partial_memory();
-
- void SetCurNode(std::string cur_name);
-
- std::string run_level() const;
-
- void SetStepNum(int32_t cur_num_step);
-
- int32_t step_num() const;
-
- void SetStreamTaskToOpnameMap(const std::map<std::pair<uint32_t, uint32_t>, std::string> &mapping);
-
- // check if any feature that uses the debugger backend is enabled
- bool DebuggerBackendEnabled();
-
- void SetTrainingDone(bool training_done);
-
- // returns true if reply received and mindspore version matched with mindinsight version
- // version_check should be true if you want the function to do backend compability check with Mindinsight
- bool SendMetadata(bool version_check);
-
- void LoadParametersAndConst();
-
- void UpdateStepNum(const session::KernelGraph *graph);
-
- void ClearCurrentData();
-
- void LoadGraphOutputs();
-
- void CheckDatasetSinkMode();
-
- void LoadGraphs(const KernelGraphPtr &graph_ptr);
-
- uint32_t GetFirstRunGraphId();
-
- void SetGraphPtr(const KernelGraphPtr &graph_ptr) { graph_ptr_ = graph_ptr; }
-
- std::list<KernelGraphPtr> GetGraphPtrList() { return graph_ptr_list_; }
-
- bool TensorExistsInCurrent(std::string tensor_name);
-
- private:
- // private constructor for singleton
- Debugger();
-
- // enable debugger
- // instantiate class members
- // read env variable for grpc client
- void EnableDebugger();
-
- void SetOpOverflowBinPath(uint32_t graph_id);
-
- // check if dump using debugger backend is enabled
- bool CheckDebuggerDumpEnabled();
-
- // check if debugger enabled
- bool CheckDebuggerEnabled();
-
- void CheckDebuggerEnabledParam();
-
- bool CheckDebuggerPartialMemoryEnabled();
-
- // check and save graph pointer
- void CheckGraphPtr(const KernelGraphPtr &graph_ptr);
-
- // check if the graph is a dataset graph
- void CheckDatasetGraph();
-
- // serialize graph and get proto
- GraphProto GetGraphProto(const KernelGraphPtr &graph_ptr) const;
-
- // send graph and enter command wait loop
- void SendGraphAndSuspend(const GraphProto &graph_proto);
-
- void SendMultiGraphsAndSuspend(const std::list<GraphProto> &graph_proto_list, uint32_t graph_sum);
-
- // wait for command and process command
- // send command request and process reply in a loop
- // break if RunCMD
- void CommandLoop();
-
- // set what nodes and conditions to watch
- void SetWatchpoint(const ProtoVector<WatchNode> &nodes, const WatchCondition &condition, const int32_t id,
- const ProtoVector<WatchCondition_Parameter> ¶meters);
-
- // remove watchpoint with id
- void RemoveWatchpoint(const int32_t id);
-
- // load tensor for view command
- std::list<TensorProto> LoadTensors(const ProtoVector<TensorProto> &tensors) const;
-
- // terminate training process
- void Exit();
-
- // analyze tensors and check watchpoint conditions
- // return names of tensors and what condition they hit
- std::list<WatchpointHit> CheckWatchpoints(const std::string &watchnode = std::string(),
- const CNodePtr &kernel = nullptr, bool recheck = false);
-
- // send watchpoints that hit
- void SendWatchpoints(const std::list<WatchpointHit> &points);
-
- // Find if any operation overflow happened and return their names
- std::vector<std::string> CheckOpOverflow();
-
- // Check if the port is valid
- bool CheckPort(const char *port);
-
- // Check if the IP is valid
- bool CheckIp(const char *host);
-
- void LoadSingleAnfnode(const AnfNodePtr &anf_node, const size_t output_index);
-
- // class members
- std::unique_ptr<GrpcClient> grpc_client_;
- std::unique_ptr<DebugServices> debug_services_;
- KernelGraphPtr graph_ptr_;
- uint32_t device_id_;
- std::string device_target_;
- int32_t num_step_;
- bool debugger_enabled_;
- std::string run_level_;
- std::string node_name_;
- std::string cur_name_;
- bool training_done_;
- bool is_dataset_graph_;
- bool partial_memory_;
- std::mutex access_lock_;
- std::map<std::pair<uint32_t, uint32_t>, std::string> stream_task_to_opname_;
- std::map<uint32_t, std::vector<std::string>> overflow_ops_;
- double last_overflow_bin_;
- std::map<uint32_t, std::string> overflow_bin_path_;
- // flag to keep track of the very first suspension of debugger
- bool initial_suspend_;
- std::list<GraphProto> graph_proto_list_;
- std::list<KernelGraphPtr> graph_ptr_list_;
-
- // singleton
- static std::mutex instance_lock_;
- static std::shared_ptr<Debugger> debugger_;
- uint32_t not_dataset_graph_sum_;
- std::list<uint32_t> rungraph_id_list_;
- std::string version_;
- };
-
- using DebuggerPtr = std::shared_ptr<Debugger>;
-
- // get debugger ModelProto
- std::string GetDebuggerFuncGraphProtoString(const FuncGraphPtr &func_graph);
- ModelProto GetDebuggerFuncGraphProto(const FuncGraphPtr &func_graph);
-
- // for getting proto DataType from Type of Tensor
- DataType GetDebuggerNumberDataType(const TypePtr &type);
-
- // process reply and command type
- DebuggerCommand GetCommand(const EventReply &reply);
-
- // parse other data out of EventReply
- ProtoVector<WatchCondition_Parameter> GetParameters(const EventReply &reply);
- ProtoVector<WatchNode> GetWatchnodes(const EventReply &reply);
- std::string GetNodeName(const EventReply &reply);
- std::string GetRunLevel(const EventReply &reply);
- WatchCondition GetWatchcondition(const EventReply &reply);
- int32_t GetWatchpointID(const EventReply &reply);
- bool GetWatchpointDelete(const EventReply &reply);
- ProtoVector<TensorProto> GetTensors(const EventReply &reply);
- bool GetMiVersionMatched(const EventReply &reply);
-
- // get the full name of a tensor, which is the name used in TensorLoader
- std::string GetTensorFullName(const TensorProto &tensor);
-
- uint64_t BytestoInt64(const std::vector<char> &buffer);
- } // namespace mindspore
- #endif // MINDSPORE_CCSRC_DEBUG_DEBUGGER_DEBUGGER_H_
|