| @@ -470,41 +470,45 @@ void Somas::GenContiguousList(const session::KernelGraph *graph) { | |||
| if (node->GetType() != kCommunicationNode) { | |||
| continue; | |||
| } | |||
| std::vector<size_t> inputs; | |||
| auto input_before_gap = CreateGapTensor(gap_tensor_id); | |||
| input_before_gap->contiguous_ = true; | |||
| gap_tensor_id++; | |||
| inputs.push_back(input_before_gap->GetId()); | |||
| for (const auto &input_tensor : node->input_tensors_) { | |||
| comm_input_total_size_ += input_tensor->aligned_size_; | |||
| input_tensor->contiguous_ = true; | |||
| inputs.push_back(input_tensor->GetId()); | |||
| } | |||
| auto input_after_gap = CreateGapTensor(gap_tensor_id); | |||
| gap_tensor_id++; | |||
| input_after_gap->contiguous_ = true; | |||
| inputs.push_back(input_after_gap->GetId()); | |||
| contiguous_tensors_list_.push_back(inputs); | |||
| std::vector<size_t> outputs; | |||
| auto output_before_gap = CreateGapTensor(gap_tensor_id); | |||
| gap_tensor_id++; | |||
| output_before_gap->contiguous_ = true; | |||
| outputs.push_back(output_before_gap->GetId()); | |||
| for (const auto &output_tensor : node->output_tensors_) { | |||
| comm_output_total_size_ += output_tensor->aligned_size_; | |||
| output_tensor->contiguous_ = true; | |||
| outputs.push_back(output_tensor->GetId()); | |||
| } | |||
| auto output_after_gap = CreateGapTensor(gap_tensor_id); | |||
| gap_tensor_id++; | |||
| output_after_gap->contiguous_ = true; | |||
| outputs.push_back(output_after_gap->GetId()); | |||
| contiguous_tensors_list_.push_back(outputs); | |||
| if ((!node->input_tensors_.empty()) && (!node->input_tensors_[0]->contiguous_)) { | |||
| std::vector<size_t> inputs; | |||
| auto input_before_gap = CreateGapTensor(gap_tensor_id); | |||
| input_before_gap->contiguous_ = true; | |||
| gap_tensor_id++; | |||
| inputs.push_back(input_before_gap->GetId()); | |||
| for (const auto &input_tensor : node->input_tensors_) { | |||
| comm_input_total_size_ += input_tensor->aligned_size_; | |||
| input_tensor->contiguous_ = true; | |||
| inputs.push_back(input_tensor->GetId()); | |||
| } | |||
| auto input_after_gap = CreateGapTensor(gap_tensor_id); | |||
| gap_tensor_id++; | |||
| input_after_gap->contiguous_ = true; | |||
| inputs.push_back(input_after_gap->GetId()); | |||
| contiguous_tensors_list_.push_back(inputs); | |||
| } | |||
| if ((!node->output_tensors_.empty()) && (!node->output_tensors_[0]->contiguous_)) { | |||
| std::vector<size_t> outputs; | |||
| auto output_before_gap = CreateGapTensor(gap_tensor_id); | |||
| gap_tensor_id++; | |||
| output_before_gap->contiguous_ = true; | |||
| outputs.push_back(output_before_gap->GetId()); | |||
| for (const auto &output_tensor : node->output_tensors_) { | |||
| comm_output_total_size_ += output_tensor->aligned_size_; | |||
| output_tensor->contiguous_ = true; | |||
| outputs.push_back(output_tensor->GetId()); | |||
| } | |||
| auto output_after_gap = CreateGapTensor(gap_tensor_id); | |||
| gap_tensor_id++; | |||
| output_after_gap->contiguous_ = true; | |||
| outputs.push_back(output_after_gap->GetId()); | |||
| contiguous_tensors_list_.push_back(outputs); | |||
| } | |||
| } | |||
| } | |||