|
|
|
@@ -40,6 +40,8 @@ using debugger::WatchCondition_Condition_nan; |
|
|
|
using debugger::WatchNode; |
|
|
|
using debugger::WatchpointHit; |
|
|
|
|
|
|
|
#define CHUNK_SIZE 1024 * 1024 * 3 |
|
|
|
|
|
|
|
namespace mindspore { |
|
|
|
|
|
|
|
DebuggerPtr Debugger::debugger_ = nullptr; |
|
|
|
@@ -458,6 +460,16 @@ void Debugger::CommandLoop() { |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
void AddTensorProtoInfo(TensorProto *tensor_item, TensorProto tensor) { |
|
|
|
tensor_item->set_node_name(tensor.node_name()); |
|
|
|
tensor_item->set_slot(tensor.slot()); |
|
|
|
tensor_item->set_iter(tensor.iter()); |
|
|
|
tensor_item->set_truncate(tensor.truncate()); |
|
|
|
tensor_item->clear_tensor_content(); |
|
|
|
tensor_item->clear_data_type(); |
|
|
|
tensor_item->clear_dims(); |
|
|
|
} |
|
|
|
|
|
|
|
void Debugger::SetWatchpoint(const ProtoVector<WatchNode> &nodes, const WatchCondition &condition, const int32_t id) { |
|
|
|
std::vector<std::tuple<std::string, bool>> check_node_list; |
|
|
|
std::transform(nodes.begin(), nodes.end(), std::back_inserter(check_node_list), |
|
|
|
@@ -482,35 +494,40 @@ std::list<TensorProto> Debugger::LoadTensors(const ProtoVector<TensorProto> &ten |
|
|
|
// ret_name will contain tensor names that are found in TensorLoader |
|
|
|
// items in ret_name will be in the same order with tensors if found |
|
|
|
debug_services_->ReadNodesTensors(name, &ret_name, &data_ptr, &data_size, &dtype, &shape); |
|
|
|
|
|
|
|
std::list<TensorProto> tensor_list; |
|
|
|
unsigned int result_index = 0; |
|
|
|
|
|
|
|
for (auto tensor : tensors) { |
|
|
|
TensorProto tensor_item; |
|
|
|
tensor_item.set_node_name(tensor.node_name()); |
|
|
|
tensor_item.set_slot(tensor.slot()); |
|
|
|
tensor_item.set_iter(tensor.iter()); |
|
|
|
tensor_item.set_truncate(tensor.truncate()); |
|
|
|
tensor_item.clear_tensor_content(); |
|
|
|
tensor_item.clear_data_type(); |
|
|
|
tensor_item.clear_dims(); |
|
|
|
// always set finished to true before big tensor splitting is supported |
|
|
|
tensor_item.set_finished(true); |
|
|
|
|
|
|
|
// return empty tensor if didn't find the requested tensor |
|
|
|
int size_iter = 0; |
|
|
|
if (result_index >= ret_name.size() || ret_name[result_index] != GetTensorFullName(tensor)) { |
|
|
|
TensorProto tensor_item; |
|
|
|
tensor_item.set_finished(true); |
|
|
|
AddTensorProtoInfo(&tensor_item, tensor); |
|
|
|
tensor_list.push_back(tensor_item); |
|
|
|
continue; |
|
|
|
} |
|
|
|
int tensor_size = data_size[result_index]; |
|
|
|
while (size_iter < tensor_size) { |
|
|
|
int chunk_size = CHUNK_SIZE; |
|
|
|
TensorProto tensor_item; |
|
|
|
tensor_item.set_finished(false); |
|
|
|
if (tensor_size - size_iter <= CHUNK_SIZE) { |
|
|
|
chunk_size = tensor_size - size_iter; |
|
|
|
tensor_item.set_finished(true); |
|
|
|
} |
|
|
|
AddTensorProtoInfo(&tensor_item, tensor); |
|
|
|
// return empty tensor if didn't find the requested tensor |
|
|
|
|
|
|
|
tensor_item.set_tensor_content(data_ptr[result_index], data_size[result_index]); |
|
|
|
tensor_item.set_data_type(GetDebuggerNumberDataType(dtype[result_index])); |
|
|
|
for (auto &elem : shape[result_index]) { |
|
|
|
tensor_item.add_dims(elem); |
|
|
|
} |
|
|
|
tensor_item.set_tensor_content(data_ptr[result_index] + size_iter, chunk_size); |
|
|
|
|
|
|
|
// add tensor to result list and increment result_index to check next item in ret_name |
|
|
|
tensor_list.push_back(tensor_item); |
|
|
|
tensor_item.set_data_type(GetDebuggerNumberDataType(dtype[result_index])); |
|
|
|
for (auto &elem : shape[result_index]) { |
|
|
|
tensor_item.add_dims(elem); |
|
|
|
} |
|
|
|
// add tensor to result list and increment result_index to check next item in ret_name |
|
|
|
tensor_list.push_back(tensor_item); |
|
|
|
size_iter += CHUNK_SIZE; |
|
|
|
} |
|
|
|
result_index++; |
|
|
|
} |
|
|
|
return tensor_list; |
|
|
|
|