|
- /**
- * Copyright 2019 Huawei Technologies Co., Ltd
- *
- * Licensed under the Apache License, Version 2.0 (the "License");
- * you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
- #ifndef MINDSPORE_CCSRC_BACKEND_KERNEL_COMPILER_COMMON_UTILS_H_
- #define MINDSPORE_CCSRC_BACKEND_KERNEL_COMPILER_COMMON_UTILS_H_
-
- #include <dirent.h>
- #include <memory>
- #include <unordered_map>
- #include <unordered_set>
- #include <map>
- #include <string>
- #include <algorithm>
- #include <vector>
- #include <utility>
- #include <nlohmann/json.hpp>
- #include "backend/kernel_compiler/kernel.h"
- #include "backend/kernel_compiler/oplib/opinfo.h"
- #include "backend/kernel_compiler/kernel_build_info.h"
-
- namespace mindspore {
- namespace kernel {
- constexpr auto kAkgKernelMeta = "kernel_meta/";
- constexpr auto kProcessorAiCore = "aicore";
- constexpr auto kProcessorAiCpu = "aicpu";
- constexpr auto kProcessorCuda = "cuda";
- constexpr auto kProcessorCpu = "cpu";
- constexpr auto kProcessorUnknown = "unknown";
- constexpr auto kJsonSuffix = ".json";
- constexpr auto kInfoSuffix = ".info";
- constexpr unsigned int AUTODIFF_COMPILE_OVERTIME = 600;
- constexpr auto kCOMPILER_CACHE_PATH = "MS_COMPILER_CACHE_PATH";
-
- const std::vector<std::string> support_devices = {"aicore", "aicpu", "cuda"};
-
- struct KernelMetaInfo {
- uintptr_t func_stub_;
- uint32_t block_dim_;
- };
- using KernelMetaPtr = std::shared_ptr<KernelMetaInfo>;
-
- class KernelMeta {
- public:
- KernelMeta() = default;
- void Initialize();
- std::string Search(const std::string &kernel_name) const;
- bool Insert(const std::string &kernel_name, const std::string &kernel_json);
- std::string kernel_meta_path() const { return kernel_meta_path_; }
- bool initialized() const { return initialized_; }
- static KernelMeta *GetInstance() {
- static KernelMeta kernel_meta;
- return &kernel_meta;
- }
- ~KernelMeta() = default;
-
- private:
- bool initialized_ = false;
- std::string kernel_meta_path_;
- std::unordered_map<std::string, std::string> kernel_meta_map_;
- };
-
- std::string GetCompilerCachePath();
- bool CheckCache(const std::string &kernel_name);
- KernelPackPtr SearchCache(const std::string &kernel_name, const std::string &processor);
- KernelPackPtr InsertCache(const std::string &kernel_name, const std::string &processor);
- TypeId DtypeToTypeId(const std::string &dtypes);
- std::string Dtype2ShortType(const std::string &dtypes);
- size_t GetDtypeNbyte(const std::string &dtypes);
- bool GetShapeSize(const std::vector<size_t> &shape, const TypePtr &type_ptr, int64_t *size_i);
- bool ParseMetadata(const CNodePtr &kernel_node, const std::shared_ptr<const OpInfo> &op_info_ptr, Processor processor,
- std::vector<std::shared_ptr<KernelBuildInfo>> *const kernel_info_list);
- void SaveJsonInfo(const std::string &json_name, const std::string &info, const std::string &base_path);
- std::string GetProcessor(const AnfNodePtr &anf_node);
- Processor GetProcessor(const string &processor);
- bool IsSameShape(const std::vector<size_t> &shape_a, const std::vector<size_t> &shape_b);
- int Sign(float x);
- int GetReductionInt(const std::string &reduction);
- std::pair<AnfNodePtr, size_t> GetKernelInput(const AnfNodePtr &anf_node, size_t index);
- std::vector<std::pair<AnfNodePtr, std::pair<size_t, size_t>>> GetInputIndex(const std::vector<AnfNodePtr> &node_list,
- const std::vector<AnfNodePtr> &input_list);
- std::vector<std::pair<AnfNodePtr, size_t>> GetOutputIndex(const std::vector<AnfNodePtr> &node_list,
- const std::vector<AnfNodePtr> &input_list,
- const std::vector<AnfNodePtr> &output_list);
- void GetValidKernelNodes(const FuncGraphPtr &func_graph, std::vector<AnfNodePtr> *node_list);
- void GetValidKernelNodes(const FuncGraphPtr &func_graph, std::vector<AnfNodePtr> *node_list,
- std::vector<AnfNodePtr> *input_list, std::vector<AnfNodePtr> *output_list);
- void GetFuncGraphOutputNodes(const FuncGraphPtr &func_graph, std::vector<AnfNodePtr> *output_list);
- bool GetInputTensorValue(const AnfNodePtr &anf_node, size_t input_idx, nlohmann::json *const node_json);
- void GetGraphRealOutput(const FuncGraphPtr &func_graph, std::vector<std::pair<AnfNodePtr, size_t>> *node_list);
- bool IsWeightBoundary(const AnfNodePtr &node);
- std::vector<int64_t> GetReduceAttrAxis(const CNodePtr &cnode);
- std::string GetProcessorStr(const AnfNodePtr &anf_node);
- Processor GetProcessorFromContext();
- std::string GetStrProcessorFromContext();
- float Scaling(size_t in_size, size_t out_size, bool align_corners);
- float ScaleGrid(const int x, const float scale);
- FusionType GetFusionTypeByName(const std::string &name);
- std::string GetFusionNameByType(const kernel::FusionType &type);
- struct CachedInterpolation {
- size_t lower;
- size_t upper;
- float lerp;
- };
-
- void ComputeInterpolationWeights(const size_t out_size, const size_t in_size, const float scale,
- CachedInterpolation *interpolation);
-
- template <typename T>
- inline std::string Vector2Str(const std::vector<T> &inputs) {
- if (!inputs.empty()) {
- std::ostringstream oss;
- (void)std::copy(inputs.begin(), inputs.end() - 1, std::ostream_iterator<T>(oss, ", "));
- oss << inputs.back();
- return oss.str();
- }
- return "";
- }
-
- template <typename T>
- inline T ComputeLerp(T top_left, T top_right, T bottom_left, T bottom_right, T x_lerp, T y_lerp) {
- T top = top_left + (top_right - top_left) * x_lerp;
- T bottom = bottom_left + (bottom_right - bottom_left) * x_lerp;
- return top + (bottom - top) * y_lerp;
- }
-
- void CastShapeSizeToLong(const std::vector<size_t> &shape, std::vector<int64_t> *long_shape);
- void CheckSliceValid(const std::vector<int64_t> &start, const std::vector<int64_t> &stop,
- const std::vector<int64_t> &step, const std::vector<int64_t> &input_shape);
- size_t CalOffset(const std::vector<int64_t> &start, const std::vector<int64_t> &stop,
- const std::vector<int64_t> &dim_offset);
- std::vector<int64_t> CalDimOffset(const std::vector<int64_t> &input_shape);
- size_t GetCopySize(const std::vector<int64_t> &dim_offset, const std::vector<int64_t> &start,
- const std::vector<int64_t> &stop);
- size_t UnitSizeInBytes(const mindspore::TypeId &t);
-
- #define CHECK_KERNEL_INPUTS_NUM(actual_inputs_num, expect_inputs_num, kernel_name) \
- do { \
- if ((actual_inputs_num) != (expect_inputs_num)) { \
- MS_LOG(EXCEPTION) << (kernel_name) << " requires " << (expect_inputs_num) << " inputs, but got " \
- << (actual_inputs_num) << "."; \
- } \
- } while (0)
-
- #define CHECK_KERNEL_OUTPUTS_NUM(actual_outputs_num, expect_outputs_num, kernel_name) \
- do { \
- if ((actual_outputs_num) != (expect_outputs_num)) { \
- MS_LOG(EXCEPTION) << (kernel_name) << " should have " << (expect_outputs_num) << " outputs, but got " \
- << (actual_outputs_num) << "."; \
- } \
- } while (0)
-
- #define CHECK_KERNEL_WORKSPACE_SIZE(actual_size, expect_size, kernel_name) \
- do { \
- if ((actual_size) != (expect_size)) { \
- MS_LOG(EXCEPTION) << (kernel_name) << " requires " << (expect_size) << " workspace, but got " << (actual_size) \
- << "."; \
- } \
- } while (0)
- } // namespace kernel
- } // namespace mindspore
-
- #endif // MINDSPORE_CCSRC_BACKEND_KERNEL_COMPILER_COMMON_UTILS_H_
|