From: @ezphlow Reviewed-by: Signed-off-by:tags/v1.2.0-rc1
| @@ -19,16 +19,19 @@ | |||
| #include <fstream> | |||
| #include <unordered_set> | |||
| #include <utility> | |||
| #include <nlohmann/json.hpp> | |||
| #include "minddata/dataset/core/tensor.h" | |||
| #include "minddata/dataset/engine/runtime_context.h" | |||
| #include "minddata/dataset/include/iterator.h" | |||
| #include "minddata/dataset/include/samplers.h" | |||
| #include "minddata/dataset/include/transforms.h" | |||
| #include "minddata/dataset/util/path.h" | |||
| #include "minddata/dataset/util/status.h" | |||
| #include "minddata/dataset/include/type_id.h" | |||
| #include "minddata/dataset/core/client.h" | |||
| #include "minddata/dataset/core/type_id.h" | |||
| #include "minddata/dataset/engine/consumers/tree_consumer.h" | |||
| #include "minddata/dataset/engine/consumers/pull_based_tree_consumer.h" | |||
| #include "minddata/dataset/kernels/c_func_op.h" | |||
| #include "minddata/dataset/kernels/tensor_op.h" | |||
| @@ -985,54 +988,81 @@ ManifestDataset::ManifestDataset(const std::vector<char> &dataset_file, const st | |||
| MindDataDataset::MindDataDataset(const std::vector<char> &dataset_file, | |||
| const std::vector<std::vector<char>> &columns_list, | |||
| const std::shared_ptr<Sampler> &sampler, nlohmann::json padded_sample, | |||
| const std::shared_ptr<Sampler> &sampler, nlohmann::json *padded_sample, | |||
| int64_t num_padded) { | |||
| auto sampler_obj = sampler ? sampler->Parse() : nullptr; | |||
| nlohmann::json sample = nullptr; | |||
| if (padded_sample) { | |||
| sample = *padded_sample; | |||
| } | |||
| auto ds = std::make_shared<MindDataNode>(CharToString(dataset_file), VectorCharToString(columns_list), sampler_obj, | |||
| padded_sample, num_padded); | |||
| sample, num_padded); | |||
| ir_node_ = std::static_pointer_cast<DatasetNode>(ds); | |||
| } | |||
| MindDataDataset::MindDataDataset(const std::vector<char> &dataset_file, | |||
| const std::vector<std::vector<char>> &columns_list, Sampler *sampler, | |||
| nlohmann::json padded_sample, int64_t num_padded) { | |||
| nlohmann::json *padded_sample, int64_t num_padded) { | |||
| auto sampler_obj = sampler ? sampler->Parse() : nullptr; | |||
| nlohmann::json sample = nullptr; | |||
| if (padded_sample) { | |||
| sample = *padded_sample; | |||
| } | |||
| auto ds = std::make_shared<MindDataNode>(CharToString(dataset_file), VectorCharToString(columns_list), sampler_obj, | |||
| padded_sample, num_padded); | |||
| sample, num_padded); | |||
| ir_node_ = std::static_pointer_cast<DatasetNode>(ds); | |||
| } | |||
| MindDataDataset::MindDataDataset(const std::vector<char> &dataset_file, | |||
| const std::vector<std::vector<char>> &columns_list, | |||
| const std::reference_wrapper<Sampler> sampler, nlohmann::json padded_sample, | |||
| const std::reference_wrapper<Sampler> sampler, nlohmann::json *padded_sample, | |||
| int64_t num_padded) { | |||
| auto sampler_obj = sampler.get().Parse(); | |||
| nlohmann::json sample = nullptr; | |||
| if (padded_sample) { | |||
| sample = *padded_sample; | |||
| } | |||
| auto ds = std::make_shared<MindDataNode>(CharToString(dataset_file), VectorCharToString(columns_list), sampler_obj, | |||
| padded_sample, num_padded); | |||
| sample, num_padded); | |||
| ir_node_ = std::static_pointer_cast<DatasetNode>(ds); | |||
| } | |||
| MindDataDataset::MindDataDataset(const std::vector<std::vector<char>> &dataset_files, | |||
| const std::vector<std::vector<char>> &columns_list, | |||
| const std::shared_ptr<Sampler> &sampler, nlohmann::json padded_sample, | |||
| const std::shared_ptr<Sampler> &sampler, nlohmann::json *padded_sample, | |||
| int64_t num_padded) { | |||
| auto sampler_obj = sampler ? sampler->Parse() : nullptr; | |||
| nlohmann::json sample = nullptr; | |||
| if (padded_sample) { | |||
| sample = *padded_sample; | |||
| } | |||
| auto ds = std::make_shared<MindDataNode>(VectorCharToString(dataset_files), VectorCharToString(columns_list), | |||
| sampler_obj, padded_sample, num_padded); | |||
| sampler_obj, sample, num_padded); | |||
| ir_node_ = std::static_pointer_cast<DatasetNode>(ds); | |||
| } | |||
| MindDataDataset::MindDataDataset(const std::vector<std::vector<char>> &dataset_files, | |||
| const std::vector<std::vector<char>> &columns_list, Sampler *sampler, | |||
| nlohmann::json padded_sample, int64_t num_padded) { | |||
| nlohmann::json *padded_sample, int64_t num_padded) { | |||
| auto sampler_obj = sampler ? sampler->Parse() : nullptr; | |||
| nlohmann::json sample = nullptr; | |||
| if (padded_sample) { | |||
| sample = *padded_sample; | |||
| } | |||
| auto ds = std::make_shared<MindDataNode>(VectorCharToString(dataset_files), VectorCharToString(columns_list), | |||
| sampler_obj, padded_sample, num_padded); | |||
| sampler_obj, sample, num_padded); | |||
| ir_node_ = std::static_pointer_cast<DatasetNode>(ds); | |||
| } | |||
| MindDataDataset::MindDataDataset(const std::vector<std::vector<char>> &dataset_files, | |||
| const std::vector<std::vector<char>> &columns_list, | |||
| const std::reference_wrapper<Sampler> sampler, nlohmann::json padded_sample, | |||
| const std::reference_wrapper<Sampler> sampler, nlohmann::json *padded_sample, | |||
| int64_t num_padded) { | |||
| auto sampler_obj = sampler.get().Parse(); | |||
| nlohmann::json sample = nullptr; | |||
| if (padded_sample) { | |||
| sample = *padded_sample; | |||
| } | |||
| auto ds = std::make_shared<MindDataNode>(VectorCharToString(dataset_files), VectorCharToString(columns_list), | |||
| sampler_obj, padded_sample, num_padded); | |||
| sampler_obj, sample, num_padded); | |||
| ir_node_ = std::static_pointer_cast<DatasetNode>(ds); | |||
| } | |||
| #endif | |||
| @@ -22,7 +22,7 @@ | |||
| #include "minddata/dataset/core/device_tensor.h" | |||
| #include "minddata/dataset/core/tensor_row.h" | |||
| #include "minddata/dataset/core/tensor.h" | |||
| #include "minddata/dataset/include/type_id.h" | |||
| #include "minddata/dataset/core/type_id.h" | |||
| #include "minddata/dataset/kernels/ir/tensor_operation.h" | |||
| #include "minddata/dataset/kernels/tensor_op.h" | |||
| #ifndef ENABLE_ANDROID | |||
| @@ -51,6 +51,7 @@ | |||
| #include "minddata/dataset/core/config_manager.h" | |||
| #include "minddata/dataset/core/data_type.h" | |||
| #include "minddata/dataset/core/type_id.h" | |||
| #include "minddata/dataset/util/path.h" | |||
| #include "minddata/dataset/util/random.h" | |||
| #include "minddata/dataset/util/services.h" | |||
| @@ -17,6 +17,7 @@ | |||
| #include "minddata/dataset/api/python/pybind_conversion.h" | |||
| #include "minddata/dataset/api/python/pybind_register.h" | |||
| #include "minddata/dataset/core/type_id.h" | |||
| #include "minddata/dataset/include/execute.h" | |||
| namespace mindspore { | |||
| @@ -19,8 +19,9 @@ | |||
| #include "minddata/dataset/api/python/pybind_register.h" | |||
| #include "minddata/dataset/core/global_context.h" | |||
| #include "minddata/dataset/include/constants.h" | |||
| #include "minddata/dataset/api/python/pybind_conversion.h" | |||
| #include "minddata/dataset/core/type_id.h" | |||
| #include "minddata/dataset/include/constants.h" | |||
| #include "minddata/dataset/include/datasets.h" | |||
| namespace mindspore { | |||
| @@ -28,11 +28,12 @@ | |||
| #include "pybind11/pybind11.h" | |||
| #include "pybind11/stl.h" | |||
| #include "pybind11/stl_bind.h" | |||
| #include "minddata/dataset/include/datasets.h" | |||
| #include "minddata/dataset/include/samplers.h" | |||
| #include "minddata/dataset/api/python/pybind_register.h" | |||
| #include "minddata/dataset/core/type_id.h" | |||
| #include "minddata/dataset/engine/ir/cache/pre_built_dataset_cache.h" | |||
| #include "minddata/dataset/engine/ir/datasetops/source/csv_node.h" | |||
| #include "minddata/dataset/include/datasets.h" | |||
| #include "minddata/dataset/include/samplers.h" | |||
| #include "minddata/dataset/kernels/ir/data/transforms_ir.h" | |||
| #include "minddata/dataset/kernels/py_func_op.h" | |||
| namespace py = pybind11; | |||
| @@ -16,8 +16,8 @@ | |||
| #include "include/api/context.h" | |||
| #include "include/api/types.h" | |||
| #include "minddata/dataset/include/type_id.h" | |||
| #include "minddata/dataset/core/ascend_resource.h" | |||
| #include "minddata/dataset/core/type_id.h" | |||
| #include "minddata/dataset/kernels/image/image_utils.h" | |||
| namespace mindspore { | |||
| @@ -18,7 +18,7 @@ | |||
| #include "minddata/dataset/core/device_tensor.h" | |||
| #include "minddata/dataset/include/constants.h" | |||
| #include "minddata/dataset/core/data_type.h" | |||
| #include "minddata/dataset/include/type_id.h" | |||
| #include "minddata/dataset/core/type_id.h" | |||
| #include "mindspore/core/ir/dtype/type_id.h" | |||
| #include "utils/hashing.h" | |||
| #ifndef ENABLE_ANDROID | |||
| @@ -24,7 +24,7 @@ | |||
| #include <utility> | |||
| #include <functional> | |||
| #include "minddata/dataset/include/type_id.h" | |||
| #include "minddata/dataset/core/type_id.h" | |||
| #include "utils/ms_utils.h" | |||
| #include "minddata/dataset/include/constants.h" | |||
| @@ -46,7 +46,7 @@ Status SyncWaitNode::Build(std::vector<std::shared_ptr<DatasetOp>> *const node_o | |||
| // Right now barrier should only take num_rows_per_buffer = 1 | |||
| // The reason for this is because having it otherwise can lead to blocking issues | |||
| // See barrier_op.h for more details | |||
| int32_t rows_per_buffer = 1; | |||
| const int32_t rows_per_buffer = 1; | |||
| auto op = std::make_shared<BarrierOp>(rows_per_buffer, connector_que_size_, condition_name_, callback_); | |||
| op->set_total_repeats(GetTotalRepeats()); | |||
| op->set_num_repeats_per_epoch(GetNumRepeatsPerEpoch()); | |||
| @@ -15,6 +15,7 @@ | |||
| */ | |||
| #include <string> | |||
| #include <nlohmann/json.hpp> | |||
| #include "minddata/dataset/include/datasets.h" | |||
| #include "minddata/dataset/engine/opt/pre/deep_copy_pass.h" | |||
| #include "minddata/dataset/engine/ir/datasetops/root_node.h" | |||
| @@ -15,6 +15,7 @@ | |||
| */ | |||
| #include <string> | |||
| #include <nlohmann/json.hpp> | |||
| #include "minddata/dataset/include/datasets.h" | |||
| #include "minddata/dataset/engine/opt/pre/input_validation_pass.h" | |||
| @@ -21,6 +21,7 @@ | |||
| #include <unistd.h> | |||
| #include <algorithm> | |||
| #include <functional> | |||
| #include <map> | |||
| #include <memory> | |||
| #include <set> | |||
| @@ -36,8 +37,6 @@ | |||
| #include "minddata/dataset/include/json_fwd.hpp" | |||
| #include "minddata/dataset/include/samplers.h" | |||
| #include "minddata/dataset/include/text.h" | |||
| #include "minddata/dataset/include/type_id.h" | |||
| #include "minddata/dataset/engine/consumers/pull_based_tree_consumer.h" | |||
| namespace mindspore { | |||
| namespace dataset { | |||
| @@ -53,7 +52,7 @@ class DatasetCache; | |||
| class DatasetNode; | |||
| class Iterator; | |||
| class PullBasedIteratorConsumer; | |||
| class PullBasedIterator; | |||
| class TensorOperation; | |||
| class SchemaObj; | |||
| @@ -1243,21 +1242,21 @@ inline std::shared_ptr<ManifestDataset> Manifest(const std::string &dataset_file | |||
| class MindDataDataset : public Dataset { | |||
| public: | |||
| explicit MindDataDataset(const std::vector<char> &dataset_file, const std::vector<std::vector<char>> &columns_list, | |||
| const std::shared_ptr<Sampler> &sampler, nlohmann::json padded_sample, int64_t num_padded); | |||
| const std::shared_ptr<Sampler> &sampler, nlohmann::json *padded_sample, int64_t num_padded); | |||
| explicit MindDataDataset(const std::vector<char> &dataset_file, const std::vector<std::vector<char>> &columns_list, | |||
| Sampler *sampler, nlohmann::json padded_sample, int64_t num_padded); | |||
| Sampler *sampler, nlohmann::json *padded_sample, int64_t num_padded); | |||
| explicit MindDataDataset(const std::vector<char> &dataset_file, const std::vector<std::vector<char>> &columns_list, | |||
| const std::reference_wrapper<Sampler> sampler, nlohmann::json padded_sample, | |||
| const std::reference_wrapper<Sampler> sampler, nlohmann::json *padded_sample, | |||
| int64_t num_padded); | |||
| explicit MindDataDataset(const std::vector<std::vector<char>> &dataset_files, | |||
| const std::vector<std::vector<char>> &columns_list, const std::shared_ptr<Sampler> &sampler, | |||
| nlohmann::json padded_sample, int64_t num_padded); | |||
| nlohmann::json *padded_sample, int64_t num_padded); | |||
| explicit MindDataDataset(const std::vector<std::vector<char>> &dataset_files, | |||
| const std::vector<std::vector<char>> &columns_list, Sampler *sampler, | |||
| nlohmann::json padded_sample, int64_t num_padded); | |||
| nlohmann::json *padded_sample, int64_t num_padded); | |||
| explicit MindDataDataset(const std::vector<std::vector<char>> &dataset_files, | |||
| const std::vector<std::vector<char>> &columns_list, | |||
| const std::reference_wrapper<Sampler> sampler, nlohmann::json padded_sample, | |||
| const std::reference_wrapper<Sampler> sampler, nlohmann::json *padded_sample, | |||
| int64_t num_padded); | |||
| ~MindDataDataset() = default; | |||
| }; | |||
| @@ -1275,7 +1274,7 @@ class MindDataDataset : public Dataset { | |||
| /// \return Shared pointer to the current MindDataDataset | |||
| inline std::shared_ptr<MindDataDataset> MindData( | |||
| const std::string &dataset_file, const std::vector<std::string> &columns_list = {}, | |||
| const std::shared_ptr<Sampler> &sampler = std::make_shared<RandomSampler>(), nlohmann::json padded_sample = nullptr, | |||
| const std::shared_ptr<Sampler> &sampler = std::make_shared<RandomSampler>(), nlohmann::json *padded_sample = nullptr, | |||
| int64_t num_padded = 0) { | |||
| return std::make_shared<MindDataDataset>(StringToChar(dataset_file), VectorStringToChar(columns_list), sampler, | |||
| padded_sample, num_padded); | |||
| @@ -1292,7 +1291,7 @@ inline std::shared_ptr<MindDataDataset> MindData( | |||
| /// \return Shared pointer to the current MindDataDataset | |||
| inline std::shared_ptr<MindDataDataset> MindData(const std::string &dataset_file, | |||
| const std::vector<std::string> &columns_list, Sampler *sampler, | |||
| nlohmann::json padded_sample = nullptr, int64_t num_padded = 0) { | |||
| nlohmann::json *padded_sample = nullptr, int64_t num_padded = 0) { | |||
| return std::make_shared<MindDataDataset>(StringToChar(dataset_file), VectorStringToChar(columns_list), sampler, | |||
| padded_sample, num_padded); | |||
| } | |||
| @@ -1309,7 +1308,7 @@ inline std::shared_ptr<MindDataDataset> MindData(const std::string &dataset_file | |||
| inline std::shared_ptr<MindDataDataset> MindData(const std::string &dataset_file, | |||
| const std::vector<std::string> &columns_list, | |||
| const std::reference_wrapper<Sampler> sampler, | |||
| nlohmann::json padded_sample = nullptr, int64_t num_padded = 0) { | |||
| nlohmann::json *padded_sample = nullptr, int64_t num_padded = 0) { | |||
| return std::make_shared<MindDataDataset>(StringToChar(dataset_file), VectorStringToChar(columns_list), sampler, | |||
| padded_sample, num_padded); | |||
| } | |||
| @@ -1326,7 +1325,7 @@ inline std::shared_ptr<MindDataDataset> MindData(const std::string &dataset_file | |||
| /// \return Shared pointer to the current MindDataDataset | |||
| inline std::shared_ptr<MindDataDataset> MindData( | |||
| const std::vector<std::string> &dataset_files, const std::vector<std::string> &columns_list = {}, | |||
| const std::shared_ptr<Sampler> &sampler = std::make_shared<RandomSampler>(), nlohmann::json padded_sample = nullptr, | |||
| const std::shared_ptr<Sampler> &sampler = std::make_shared<RandomSampler>(), nlohmann::json *padded_sample = nullptr, | |||
| int64_t num_padded = 0) { | |||
| return std::make_shared<MindDataDataset>(VectorStringToChar(dataset_files), VectorStringToChar(columns_list), sampler, | |||
| padded_sample, num_padded); | |||
| @@ -1342,7 +1341,7 @@ inline std::shared_ptr<MindDataDataset> MindData( | |||
| /// \return Shared pointer to the current MindDataDataset | |||
| inline std::shared_ptr<MindDataDataset> MindData(const std::vector<std::string> &dataset_files, | |||
| const std::vector<std::string> &columns_list, Sampler *sampler, | |||
| nlohmann::json padded_sample = nullptr, int64_t num_padded = 0) { | |||
| nlohmann::json *padded_sample = nullptr, int64_t num_padded = 0) { | |||
| return std::make_shared<MindDataDataset>(VectorStringToChar(dataset_files), VectorStringToChar(columns_list), sampler, | |||
| padded_sample, num_padded); | |||
| } | |||
| @@ -1358,7 +1357,7 @@ inline std::shared_ptr<MindDataDataset> MindData(const std::vector<std::string> | |||
| inline std::shared_ptr<MindDataDataset> MindData(const std::vector<std::string> &dataset_files, | |||
| const std::vector<std::string> &columns_list, | |||
| const std::reference_wrapper<Sampler> sampler, | |||
| nlohmann::json padded_sample = nullptr, int64_t num_padded = 0) { | |||
| nlohmann::json *padded_sample = nullptr, int64_t num_padded = 0) { | |||
| return std::make_shared<MindDataDataset>(VectorStringToChar(dataset_files), VectorStringToChar(columns_list), sampler, | |||
| padded_sample, num_padded); | |||
| } | |||
| @@ -1114,6 +1114,11 @@ class RandomRotation(ImageTensorOperation): | |||
| @check_random_rotation | |||
| def __init__(self, degrees, resample=Inter.NEAREST, expand=False, center=None, fill_value=0): | |||
| if isinstance(degrees, numbers.Number): | |||
| degrees = degrees % 360 | |||
| if isinstance(degrees, (list, tuple)): | |||
| degrees = [degrees[0] % 360, degrees[1] % 360] | |||
| self.degrees = degrees | |||
| self.resample = resample | |||
| self.expand = expand | |||
| @@ -1121,6 +1126,8 @@ class RandomRotation(ImageTensorOperation): | |||
| self.fill_value = fill_value | |||
| def parse(self): | |||
| # pylint false positive | |||
| # pylint: disable=E1130 | |||
| degrees = (-self.degrees, self.degrees) if isinstance(self.degrees, numbers.Number) else self.degrees | |||
| interpolation = DE_C_INTER_MODE[self.resample] | |||
| expand = self.expand | |||
| @@ -289,7 +289,7 @@ TEST_F(MindDataTestPipeline, TestMindDataSuccess7) { | |||
| std::string file_path1 = datasets_root_path_ + "/../mindrecord/testMindDataSet/testImageNetData/imagenet.mindrecord0"; | |||
| std::vector<std::string> file_list = {file_path1}; | |||
| std::shared_ptr<Dataset> ds = | |||
| MindData(file_list, {"file_name", "label"}, std::make_shared<SequentialSampler>(), pad, 4); | |||
| MindData(file_list, {"file_name", "label"}, std::make_shared<SequentialSampler>(), &pad, 4); | |||
| EXPECT_NE(ds, nullptr); | |||
| // Create a Skip operation on ds, skip original data in mindrecord and get padded samples | |||
| @@ -341,7 +341,7 @@ TEST_F(MindDataTestPipeline, TestMindDataSuccess8) { | |||
| std::string file_path1 = datasets_root_path_ + "/../mindrecord/testMindDataSet/testImageNetData/imagenet.mindrecord0"; | |||
| std::vector<std::string> file_list = {file_path1}; | |||
| std::shared_ptr<Dataset> ds = | |||
| MindData(file_list, {"file_name", "label"}, std::make_shared<SequentialSampler>(), pad, 4); | |||
| MindData(file_list, {"file_name", "label"}, std::make_shared<SequentialSampler>(), &pad, 4); | |||
| EXPECT_NE(ds, nullptr); | |||
| std::vector<mindspore::dataset::DataType> types = ToDETypes(ds->GetOutputTypes()); | |||
| @@ -412,13 +412,13 @@ TEST_F(MindDataTestPipeline, TestMindDataSuccess9) { | |||
| std::string file_path1 = datasets_root_path_ + "/../mindrecord/testMindDataSet/testImageNetData/imagenet.mindrecord0"; | |||
| std::vector<std::string> file_list = {file_path1}; | |||
| std::shared_ptr<Dataset> ds1 = | |||
| MindData(file_list, {"file_name", "label"}, std::make_shared<SequentialSampler>(), pad, 4); | |||
| MindData(file_list, {"file_name", "label"}, std::make_shared<SequentialSampler>(), &pad, 4); | |||
| EXPECT_NE(ds1, nullptr); | |||
| ds1 = ds1->Skip(5); | |||
| EXPECT_NE(ds1, nullptr); | |||
| std::shared_ptr<Dataset> ds2 = | |||
| MindData(file_list, {"file_name", "label"}, std::make_shared<SequentialSampler>(), pad, 4); | |||
| MindData(file_list, {"file_name", "label"}, std::make_shared<SequentialSampler>(), &pad, 4); | |||
| EXPECT_NE(ds2, nullptr); | |||
| ds2 = ds2->Skip(5); | |||
| EXPECT_NE(ds2, nullptr); | |||
| @@ -588,7 +588,7 @@ TEST_F(MindDataTestPipeline, TestMindDataFail4) { | |||
| // Create a MindData Dataset | |||
| std::string file_path2 = datasets_root_path_ + "/../mindrecord/testMindDataSet/testImageNetData/imagenet.mindrecord0"; | |||
| std::shared_ptr<Dataset> ds2 = MindData(file_path2, {"label"}, std::make_shared<RandomSampler>(), pad, -2); | |||
| std::shared_ptr<Dataset> ds2 = MindData(file_path2, {"label"}, std::make_shared<RandomSampler>(), &pad, -2); | |||
| EXPECT_NE(ds2, nullptr); | |||
| // Create an iterator over the result of the above dataset | |||
| @@ -598,7 +598,7 @@ TEST_F(MindDataTestPipeline, TestMindDataFail4) { | |||
| // Create a MindData Dataset | |||
| std::string file_path3 = datasets_root_path_ + "/../mindrecord/testMindDataSet/testImageNetData/imagenet.mindrecord0"; | |||
| std::shared_ptr<Dataset> ds3 = MindData(file_path3, {}, std::make_shared<RandomSampler>(), pad, 1); | |||
| std::shared_ptr<Dataset> ds3 = MindData(file_path3, {}, std::make_shared<RandomSampler>(), &pad, 1); | |||
| EXPECT_NE(ds3, nullptr); | |||
| // Create an iterator over the result of the above dataset | |||
| @@ -614,7 +614,7 @@ TEST_F(MindDataTestPipeline, TestMindDataFail4) { | |||
| // Create a MindData Dataset | |||
| std::string file_path4 = datasets_root_path_ + "/../mindrecord/testMindDataSet/testImageNetData/imagenet.mindrecord0"; | |||
| std::shared_ptr<Dataset> ds4 = | |||
| MindData(file_path4, {"file_name", "label"}, std::make_shared<RandomSampler>(), pad2, 1); | |||
| MindData(file_path4, {"file_name", "label"}, std::make_shared<RandomSampler>(), &pad2, 1); | |||
| EXPECT_NE(ds4, nullptr); | |||
| // Create an iterator over the result of the above dataset | |||
| @@ -20,7 +20,8 @@ | |||
| #include "include/api/status.h" | |||
| #include "include/api/types.h" | |||
| #include "minddata/dataset/core/tensor_shape.h" | |||
| #include "minddata/dataset/include/type_id.h" | |||
| #include "minddata/dataset/core/type_id.h" | |||
| #include "minddata/dataset/core/de_tensor.h" | |||
| #include "utils/log_adapter.h" | |||
| using mindspore::Status; | |||