GitOrigin-RevId: ae98bc887f
tags/v1.7.0
| @@ -5,4 +5,4 @@ Use these files in the following way (replace `xxx` to specific file name, like | |||
| ``` | |||
| python -m megengine.tools.xxx | |||
| ``` | |||
| ``` | |||
| @@ -0,0 +1,235 @@ | |||
| #! /usr/bin/env python3 | |||
| # MegEngine is Licensed under the Apache License, Version 2.0 (the "License") | |||
| # | |||
| # Copyright (c) 2014-2021 Megvii Inc. All rights reserved. | |||
| # | |||
| # Unless required by applicable law or agreed to in writing, | |||
| # software distributed under the License is distributed on an | |||
| # "AS IS" BASIS, WITHOUT ARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
| import argparse | |||
| import json | |||
| import math | |||
| import os | |||
| from megengine.utils.module_stats import sizeof_fmt | |||
| from megengine.utils.tensorboard import SummaryWriterExtend | |||
| def load_single_graph(fpath): | |||
| with open(fpath) as fin: | |||
| data = json.load(fin) | |||
| for t in ["operator", "var"]: | |||
| data[t] = {int(i): j for i, j in data[t].items()} | |||
| gvars = data["var"] | |||
| for oid, i in data["operator"].items(): | |||
| i["input"] = list(map(int, i["input"])) | |||
| out = i["output"] = list(map(int, i["output"])) | |||
| for j in out: | |||
| gvars[j]["owner_opr"] = oid | |||
| for var in data["var"].values(): | |||
| mp = var.get("mem_plan", None) | |||
| if mp: | |||
| var["shape"] = "{" + ",".join(map(str, mp["layout"]["shape"])) + "}" | |||
| else: | |||
| var["shape"] = "<?>" | |||
| return data | |||
| def comp_graph_plotter(input, writer): | |||
| jgraph = load_single_graph(input) | |||
| all_oprs = jgraph["operator"] | |||
| all_vars = jgraph["var"] | |||
| for i in all_oprs: | |||
| opr = all_oprs[i] | |||
| if opr["type"] == "ImmutableTensor": | |||
| continue | |||
| inputlist = [] | |||
| for var in opr["input"]: | |||
| inpopr = all_oprs[all_vars[var]["owner_opr"]] | |||
| if inpopr["type"] == "ImmutableTensor": | |||
| continue | |||
| inputlist.append(all_oprs[all_vars[var]["owner_opr"]]["name"]) | |||
| writer.add_node_raw(opr["name"], opr["type"], inputlist) | |||
| writer.add_graph_by_node_raw_list() | |||
| def load_mem_info(fpath): | |||
| with open(fpath) as fin: | |||
| data = json.load(fin) | |||
| oprs = data["opr"] | |||
| for oid, i in oprs.items(): | |||
| i["size"] = 0 | |||
| for oid, i in data["chunk"].items(): | |||
| i["size"] = int(i["logic_addr_end"]) - int(i["logic_addr_begin"]) | |||
| data["peak_memory"] = 0 | |||
| data["weight_memory"] = 0 | |||
| for oid, i in data["chunk"].items(): | |||
| if i["type"] == "static_mem": | |||
| i["owner_opr"] = oprs[i["time_begin"]]["name"] | |||
| life_begin = int(i["time_begin"]) | |||
| life_end = int(i["time_end"]) | |||
| if i["overwrite_dest_id"] != "-1": | |||
| life_begin = life_begin + 1 | |||
| if data["peak_memory"] < int(i["logic_addr_end"]): | |||
| data["peak_memory"] = int(i["logic_addr_end"]) | |||
| for j in range(life_begin, life_end): | |||
| oprs[str(j)]["size"] = oprs[str(j)]["size"] + i["size"] | |||
| elif i["type"] == "weight_mem": | |||
| data["weight_memory"] += int(i["logic_addr_end"]) - int( | |||
| i["logic_addr_begin"] | |||
| ) | |||
| return data | |||
| def peak_mem_regist(input, writer): | |||
| jmem = load_mem_info(input) | |||
| writer.add_text( | |||
| "PEAK_MEMORY_SIZE", | |||
| [sizeof_fmt(jmem["peak_memory"]) + "(" + str(jmem["peak_memory"]) + " B)"], | |||
| ) | |||
| writer.add_text( | |||
| "WEIGHT_MEMORY_SIZE", | |||
| [sizeof_fmt(jmem["weight_memory"]) + "(" + str(jmem["weight_memory"]) + " B)"], | |||
| ) | |||
| all_oprs = jmem["opr"] | |||
| all_chunks = jmem["chunk"] | |||
| max_size = 0 | |||
| max_size_oprs = [] | |||
| # get oprs that reach the max memory | |||
| for oid, i in all_oprs.items(): | |||
| if i["size"] == max_size: | |||
| max_size_oprs.append(int(i["id"])) | |||
| elif i["size"] > max_size: | |||
| max_size = i["size"] | |||
| max_size_oprs.clear() | |||
| max_size_oprs.append(int(i["id"])) | |||
| # get component of chunks | |||
| max_size_oprs.sort() | |||
| opr2chunks = [] | |||
| num = len(max_size_oprs) | |||
| for i in range(num): | |||
| opr2chunks.append([]) | |||
| for oid, i in all_chunks.items(): | |||
| if i["type"] == "static_mem": | |||
| life_begin = int(i["time_begin"]) | |||
| life_end = int(i["time_end"]) | |||
| if i["overwrite_dest_id"] != "-1": | |||
| life_begin = life_begin + 1 | |||
| if max_size_oprs[0] >= life_end or max_size_oprs[-1] < life_begin: | |||
| continue | |||
| for j in range(num): | |||
| if max_size_oprs[j] >= life_end: | |||
| break | |||
| elif max_size_oprs[j] >= life_begin: | |||
| opr2chunks[j].append(i["id"]) | |||
| peak_num = 0 | |||
| for i in range(num): | |||
| suffix_1 = "PEAK" + str(peak_num) | |||
| if i - 1 > 0 and opr2chunks[i - 1] == opr2chunks[i]: | |||
| continue | |||
| max_num = 0 | |||
| opr2chunks[i] = sorted( | |||
| opr2chunks[i], | |||
| key=lambda chunk_id: all_chunks[chunk_id]["size"], | |||
| reverse=True, | |||
| ) | |||
| writer.add_text( | |||
| suffix_1 + "/" + "<SUMMARY_INFO>", | |||
| ["reached_max_opr_name: " + all_oprs[str(max_size_oprs[i])]["name"]], | |||
| 0, | |||
| ) | |||
| writer.add_text( | |||
| suffix_1 + "/" + "<SUMMARY_INFO>", | |||
| ["max_used_size: " + sizeof_fmt(max_size)], | |||
| 1, | |||
| ) | |||
| for j in opr2chunks[i]: | |||
| suffix_2 = "MAX" + str(max_num) | |||
| j_size = sizeof_fmt(all_chunks[j]["size"]) | |||
| j_percent = round(all_chunks[j]["size"] / max_size * 100, 3) | |||
| writer.add_text( | |||
| suffix_1 + "/" + suffix_2 + "_OPR", | |||
| ["percent: " + str(j_percent) + "%"], | |||
| 0, | |||
| ) | |||
| writer.add_text( | |||
| suffix_1 + "/" + suffix_2 + "_OPR", ["memory_size: " + j_size], 1, | |||
| ) | |||
| writer.add_text( | |||
| suffix_1 + "/" + suffix_2 + "_OPR", | |||
| ["owner_opr: " + all_chunks[j]["owner_opr"]], | |||
| 2, | |||
| ) | |||
| writer.add_node_raw_attributes( | |||
| all_chunks[j]["owner_opr"], | |||
| { | |||
| "memory_" + all_chunks[j]["id"]: j_size, | |||
| "memory_percent": str(j_percent) + "%", | |||
| "summary_memory_" + str(peak_num): sizeof_fmt(max_size), | |||
| }, | |||
| ) | |||
| writer.add_node_raw_name_suffix( | |||
| all_chunks[j]["owner_opr"], "_" + suffix_1 + "_" + suffix_2 | |||
| ) | |||
| max_num += 1 | |||
| peak_num += 1 | |||
| writer.add_graph_by_node_raw_list() | |||
| def convert(args): | |||
| file_process_order = { | |||
| "graph.json": comp_graph_plotter, | |||
| "StaticMemoryInfo.json": peak_mem_regist, | |||
| } | |||
| g = os.walk(args.input) | |||
| for path, dir_list, file_list in g: | |||
| out_path = path.replace(args.input, args.output) | |||
| writer = SummaryWriterExtend(out_path) | |||
| for key, value in file_process_order.items(): | |||
| if key in file_list: | |||
| value(os.path.join(path, key), writer) | |||
| def main(): | |||
| """`graph_info_analyze.py` is uesed to convert json dumped by `VisableDataSet` | |||
| class to logs which can be read by python `tensorboard`. | |||
| Now `get_static_memory_alloc_info()` support this feature,it will dump a dir | |||
| which can be convert by `graph_info_analyze.py`. | |||
| Examples:: | |||
| graph_info_analyze.py -i <input_dir_name> -o <output_dir_name> | |||
| tensorboard --logdir <output_dir_name> | |||
| """ | |||
| parser = argparse.ArgumentParser( | |||
| "convert json dumped by c to logs which can be read by python tensorboard", | |||
| formatter_class=argparse.ArgumentDefaultsHelpFormatter, | |||
| ) | |||
| parser.add_argument( | |||
| "-i", "--input", required=True, help="input dirctor name(c tensorboard info)" | |||
| ) | |||
| parser.add_argument( | |||
| "-o", | |||
| "--output", | |||
| required=True, | |||
| help="output dirctor name(python tensorboard info)", | |||
| ) | |||
| args = parser.parse_args() | |||
| convert(args) | |||
| if __name__ == "__main__": | |||
| main() | |||
| @@ -0,0 +1,131 @@ | |||
| /** | |||
| * \file src/core/impl/utils/tensorboard.cpp | |||
| * MegEngine is Licensed under the Apache License, Version 2.0 (the "License") | |||
| * | |||
| * Copyright (c) 2014-2021 Megvii Inc. All rights reserved. | |||
| * | |||
| * Unless required by applicable law or agreed to in writing, | |||
| * software distributed under the License is distributed on an | |||
| * "AS IS" BASIS, WITHOUT ARRANTIES OR CONDITIONS OF ANY KIND, either express or | |||
| * implied. | |||
| */ | |||
| #include "megbrain/utils/visable_data_set.h" | |||
| #if MGB_ENABLE_JSON | |||
| #include <fstream> | |||
| #include <iostream> | |||
| using namespace mgb; | |||
| #if WIN32 | |||
| #include <direct.h> | |||
| #include <fcntl.h> | |||
| #include <io.h> | |||
| #define getcwd _getcwd | |||
| namespace { | |||
| auto mkdir(const char* path, int) { | |||
| return _mkdir(path); | |||
| } | |||
| } // namespace | |||
| #else | |||
| #include <sys/stat.h> | |||
| #include <sys/types.h> | |||
| #include <unistd.h> | |||
| #endif | |||
| namespace { | |||
| #if defined(IOS) | |||
| #pragma message "build test on iOS; need ios_get_mgb_output_dir() to be defined" | |||
| extern "C" void ios_get_mgb_output_dir(char** dir); | |||
| #endif | |||
| std::string output_file(std::string dir_name, const std::string& fname) { | |||
| static std::string cwd; | |||
| static std::mutex cwd_mtx; | |||
| MGB_LOCK_GUARD(cwd_mtx); | |||
| if (cwd.empty()) { | |||
| #if defined(IOS) | |||
| char* buf = nullptr; | |||
| ios_get_mgb_output_dir(&buf); | |||
| #else | |||
| auto buf = getcwd(nullptr, 0); | |||
| #endif | |||
| mgb_assert(buf); | |||
| cwd = buf; | |||
| free(buf); | |||
| dir_name = dir_name + "/"; | |||
| for (size_t i = 0; i < dir_name.size(); i++) { | |||
| size_t pos = dir_name.find("/", i); | |||
| if (pos < dir_name.size() && pos - i > 1) { | |||
| cwd.append("/" + dir_name.substr(i, pos - i)); | |||
| mkdir(cwd.c_str(), 0755); | |||
| i = pos; | |||
| } | |||
| } | |||
| } | |||
| if (fname.empty()) | |||
| return cwd; | |||
| auto ret = cwd + "/" + fname; | |||
| FILE* fout = fopen(ret.c_str(), "w"); | |||
| mgb_assert(fout, "failed to open %s: %s", ret.c_str(), strerror(errno)); | |||
| fclose(fout); | |||
| return ret; | |||
| } | |||
| } // namespace | |||
| void VisableDataSet::draw_graph(std::shared_ptr<json::Value> graph_json) { | |||
| graph_json->writeto_fpath(output_file(m_logdir, "graph.json")); | |||
| } | |||
| void VisableDataSet::dump_info(Content& c) { | |||
| auto&& content_set = m_file2content[c.file_name()]; | |||
| content_set.insert(c.content_name()); | |||
| auto&& value_list = | |||
| m_filecontent2value[c.file_name() + "/" + c.content_name()]; | |||
| value_list.push_back(c.to_json()); | |||
| } | |||
| void VisableDataSet::write_to_file() { | |||
| for (auto& i : m_file2content) { | |||
| auto f_objptr = json::Object::make(); | |||
| auto&& f_obj = *f_objptr; | |||
| for (auto& c : i.second) { | |||
| auto c_objptr = json::Object::make(); | |||
| auto&& c_obj = *c_objptr; | |||
| for (auto& j : m_filecontent2value[i.first + "/" + c]) { | |||
| c_obj[(*j).cast_final_safe<json::Object>()["id"] | |||
| ->cast_final_safe<json::String>() | |||
| .get_impl()] = j; | |||
| } | |||
| f_obj[c] = c_objptr; | |||
| } | |||
| f_objptr->writeto_fpath(output_file(m_logdir, i.first)); | |||
| } | |||
| } | |||
| // const std::string PeakMemoryInfo::content() const {return } | |||
| std::shared_ptr<json::Value> Chunk::to_json() const { | |||
| auto objptr = json::Object::make(); | |||
| auto&& obj = *objptr; | |||
| obj["id"] = json::String::make(id()); | |||
| obj["type"] = json::String::make(m_type); | |||
| obj["time_begin"] = json::String::make(m_time_begin); | |||
| obj["time_end"] = json::String::make(m_time_end); | |||
| obj["logic_addr_begin"] = json::String::make(m_logic_addr_begin); | |||
| obj["logic_addr_end"] = json::String::make(m_logic_addr_end); | |||
| obj["overwrite_dest_id"] = json::String::make(m_overwrite_dest_id); | |||
| return objptr; | |||
| } | |||
| std::shared_ptr<json::Value> OprSeq::to_json() const { | |||
| auto objptr = json::Object::make(); | |||
| auto&& obj = *objptr; | |||
| obj["id"] = json::String::make(id()); | |||
| obj["name"] = json::String::make(m_name); | |||
| return objptr; | |||
| } | |||
| #endif | |||
| @@ -0,0 +1,88 @@ | |||
| /** | |||
| * \file src/core/include/megbrain/utils/visable_data_set.h | |||
| * MegEngine is Licensed under the Apache License, Version 2.0 (the "License") | |||
| * | |||
| * Copyright (c) 2014-2021 Megvii Inc. All rights reserved. | |||
| * | |||
| * Unless required by applicable law or agreed to in writing, | |||
| * software distributed under the License is distributed on an | |||
| * "AS IS" BASIS, WITHOUT ARRANTIES OR CONDITIONS OF ANY KIND, either express or | |||
| * implied. | |||
| */ | |||
| #pragma once | |||
| #include "megbrain/utils/json.h" | |||
| #if MGB_ENABLE_JSON | |||
| #include <set> | |||
| #include "megbrain/plugin/base.h" | |||
| #include "megbrain/plugin/static_mem_record.h" | |||
| namespace mgb { | |||
| class VisableDataSet : public NonCopyableObj { | |||
| private: | |||
| const std::string m_logdir; | |||
| std::unordered_map<std::string, std::set<std::string>> m_file2content; | |||
| std::unordered_map<std::string, std::vector<std::shared_ptr<json::Value>>> | |||
| m_filecontent2value; | |||
| public: | |||
| class Content { | |||
| private: | |||
| std::string m_file_name; | |||
| std::string m_content_name; | |||
| std::string m_id; | |||
| public: | |||
| Content(std::string file_name, std::string content_name, std::string id) | |||
| : m_file_name(file_name), | |||
| m_content_name(content_name), | |||
| m_id(id) {} | |||
| const std::string& file_name() const { return m_file_name; } | |||
| const std::string& content_name() const { return m_content_name; } | |||
| const std::string& id() const { return m_id; } | |||
| virtual std::shared_ptr<json::Value> to_json() const = 0; | |||
| virtual ~Content() = default; | |||
| }; | |||
| VisableDataSet(std::string logdir) : m_logdir(logdir) {} | |||
| void draw_graph(std::shared_ptr<json::Value> graph_json); | |||
| void dump_info(Content& c); | |||
| void write_to_file(); | |||
| }; | |||
| class Chunk : public VisableDataSet::Content { | |||
| private: | |||
| const char* enum_str[2] = {"static_mem", "weight_mem"}; | |||
| std::string m_type, m_time_begin, m_time_end, m_logic_addr_begin, | |||
| m_logic_addr_end, | |||
| m_overwrite_dest_id; // m_overwriter_dest_id = "-1" means no | |||
| // overwrite dest | |||
| public: | |||
| enum chunk_type { static_mem, weight_mem }; | |||
| Chunk(std::string id, chunk_type type, std::string time_begin, | |||
| std::string time_end, std::string logic_addr_begin, | |||
| std::string logic_addr_end, std::string overwrite_dest_id) | |||
| : Content("StaticMemoryInfo.json", "chunk", id), | |||
| m_type(enum_str[type]), | |||
| m_time_begin(time_begin), | |||
| m_time_end(time_end), | |||
| m_logic_addr_begin(logic_addr_begin), | |||
| m_logic_addr_end(logic_addr_end), | |||
| m_overwrite_dest_id(overwrite_dest_id) {} | |||
| std::shared_ptr<json::Value> to_json() const override; | |||
| }; | |||
| class OprSeq : public VisableDataSet::Content { | |||
| private: | |||
| std::string m_id, m_name; | |||
| public: | |||
| OprSeq(std::string id, std::string opr_name) | |||
| : Content("StaticMemoryInfo.json", "opr", id), m_name(opr_name) {} | |||
| std::shared_ptr<json::Value> to_json() const override; | |||
| }; | |||
| } // namespace mgb | |||
| #endif | |||