You can not select more than 25 topics Topics must start with a chinese character,a letter or number, can include dashes ('-') and can be up to 35 characters long.

test_serdes_dataset.py 21 kB

5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507
  1. # Copyright 2020-2021 Huawei Technologies Co., Ltd
  2. #
  3. # Licensed under the Apache License, Version 2.0 (the "License");
  4. # you may not use this file except in compliance with the License.
  5. # You may obtain a copy of the License at
  6. #
  7. # http://www.apache.org/licenses/LICENSE-2.0
  8. #
  9. # Unless required by applicable law or agreed to in writing, software
  10. # distributed under the License is distributed on an "AS IS" BASIS,
  11. # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  12. # See the License for the specific language governing permissions and
  13. # limitations under the License.
  14. # ==============================================================================
  15. """
  16. Testing dataset serialize and deserialize in DE
  17. """
  18. import filecmp
  19. import glob
  20. import json
  21. import os
  22. import pytest
  23. import numpy as np
  24. from test_minddataset_sampler import add_and_remove_cv_file, get_data, CV_DIR_NAME, CV_FILE_NAME
  25. from util import config_get_set_num_parallel_workers, config_get_set_seed
  26. import mindspore.common.dtype as mstype
  27. import mindspore.dataset as ds
  28. import mindspore.dataset.transforms.c_transforms as c
  29. import mindspore.dataset.transforms.py_transforms as py
  30. import mindspore.dataset.vision.c_transforms as vision
  31. import mindspore.dataset.vision.py_transforms as py_vision
  32. from mindspore import log as logger
  33. from mindspore.dataset.vision import Inter
  34. def test_serdes_imagefolder_dataset(remove_json_files=True):
  35. """
  36. Test simulating resnet50 dataset pipeline.
  37. """
  38. data_dir = "../data/dataset/testPK/data"
  39. ds.config.set_seed(1)
  40. # define data augmentation parameters
  41. rescale = 1.0 / 255.0
  42. shift = 0.0
  43. resize_height, resize_width = 224, 224
  44. weights = [1.0, 0.1, 0.02, 0.3, 0.4, 0.05, 1.2, 0.13, 0.14, 0.015, 0.16, 1.1]
  45. # Constructing DE pipeline
  46. sampler = ds.WeightedRandomSampler(weights, 11)
  47. child_sampler = ds.SequentialSampler()
  48. sampler.add_child(child_sampler)
  49. data1 = ds.ImageFolderDataset(data_dir, sampler=sampler)
  50. data1 = data1.repeat(1)
  51. data1 = data1.map(operations=[vision.Decode(True)], input_columns=["image"])
  52. rescale_op = vision.Rescale(rescale, shift)
  53. resize_op = vision.Resize((resize_height, resize_width), Inter.LINEAR)
  54. data1 = data1.map(operations=[rescale_op, resize_op], input_columns=["image"])
  55. data1 = data1.batch(2)
  56. # Serialize the dataset pre-processing pipeline.
  57. # data1 should still work after saving.
  58. ds.serialize(data1, "imagenet_dataset_pipeline.json")
  59. ds1_dict = ds.serialize(data1)
  60. assert validate_jsonfile("imagenet_dataset_pipeline.json") is True
  61. # Print the serialized pipeline to stdout
  62. ds.show(data1)
  63. # Deserialize the serialized json file
  64. data2 = ds.deserialize(json_filepath="imagenet_dataset_pipeline.json")
  65. # Serialize the pipeline we just deserialized.
  66. # The content of the json file should be the same to the previous serialize.
  67. ds.serialize(data2, "imagenet_dataset_pipeline_1.json")
  68. assert validate_jsonfile("imagenet_dataset_pipeline_1.json") is True
  69. assert filecmp.cmp('imagenet_dataset_pipeline.json', 'imagenet_dataset_pipeline_1.json')
  70. # Deserialize the latest json file again
  71. data3 = ds.deserialize(json_filepath="imagenet_dataset_pipeline_1.json")
  72. data4 = ds.deserialize(input_dict=ds1_dict)
  73. num_samples = 0
  74. # Iterate and compare the data in the original pipeline (data1) against the deserialized pipeline (data2)
  75. for item1, item2, item3, item4 in zip(data1.create_dict_iterator(num_epochs=1, output_numpy=True),
  76. data2.create_dict_iterator(num_epochs=1, output_numpy=True),
  77. data3.create_dict_iterator(num_epochs=1, output_numpy=True),
  78. data4.create_dict_iterator(num_epochs=1, output_numpy=True)):
  79. np.testing.assert_array_equal(item1['image'], item2['image'])
  80. np.testing.assert_array_equal(item1['image'], item3['image'])
  81. np.testing.assert_array_equal(item1['label'], item2['label'])
  82. np.testing.assert_array_equal(item1['label'], item3['label'])
  83. np.testing.assert_array_equal(item3['image'], item4['image'])
  84. np.testing.assert_array_equal(item3['label'], item4['label'])
  85. num_samples += 1
  86. logger.info("Number of data in data1: {}".format(num_samples))
  87. assert num_samples == 6
  88. # Remove the generated json file
  89. if remove_json_files:
  90. delete_json_files()
  91. def test_serdes_mnist_dataset(remove_json_files=True):
  92. """
  93. Test serdes on mnist dataset pipeline.
  94. """
  95. data_dir = "../data/dataset/testMnistData"
  96. ds.config.set_seed(1)
  97. data1 = ds.MnistDataset(data_dir, num_samples=100)
  98. one_hot_encode = c.OneHot(10) # num_classes is input argument
  99. data1 = data1.map(operations=one_hot_encode, input_columns="label")
  100. # batch_size is input argument
  101. data1 = data1.batch(batch_size=10, drop_remainder=True)
  102. ds.serialize(data1, "mnist_dataset_pipeline.json")
  103. assert validate_jsonfile("mnist_dataset_pipeline.json") is True
  104. data2 = ds.deserialize(json_filepath="mnist_dataset_pipeline.json")
  105. ds.serialize(data2, "mnist_dataset_pipeline_1.json")
  106. assert validate_jsonfile("mnist_dataset_pipeline_1.json") is True
  107. assert filecmp.cmp('mnist_dataset_pipeline.json', 'mnist_dataset_pipeline_1.json')
  108. data3 = ds.deserialize(json_filepath="mnist_dataset_pipeline_1.json")
  109. num = 0
  110. for data1, data2, data3 in zip(data1.create_dict_iterator(num_epochs=1, output_numpy=True),
  111. data2.create_dict_iterator(num_epochs=1, output_numpy=True),
  112. data3.create_dict_iterator(num_epochs=1, output_numpy=True)):
  113. np.testing.assert_array_equal(data1['image'], data2['image'])
  114. np.testing.assert_array_equal(data1['image'], data3['image'])
  115. np.testing.assert_array_equal(data1['label'], data2['label'])
  116. np.testing.assert_array_equal(data1['label'], data3['label'])
  117. num += 1
  118. logger.info("mnist total num samples is {}".format(str(num)))
  119. assert num == 10
  120. if remove_json_files:
  121. delete_json_files()
  122. def test_serdes_zip_dataset(remove_json_files=True):
  123. """
  124. Test serdes on zip dataset pipeline.
  125. """
  126. files = ["../data/dataset/testTFTestAllTypes/test.data"]
  127. schema_file = "../data/dataset/testTFTestAllTypes/datasetSchema.json"
  128. ds.config.set_seed(1)
  129. ds0 = ds.TFRecordDataset(files, schema=schema_file, shuffle=ds.Shuffle.GLOBAL)
  130. data1 = ds.TFRecordDataset(files, schema=schema_file, shuffle=ds.Shuffle.GLOBAL)
  131. data2 = ds.TFRecordDataset(files, schema=schema_file, shuffle=ds.Shuffle.FILES)
  132. data2 = data2.shuffle(10000)
  133. data2 = data2.rename(input_columns=["col_sint16", "col_sint32", "col_sint64", "col_float",
  134. "col_1d", "col_2d", "col_3d", "col_binary"],
  135. output_columns=["column_sint16", "column_sint32", "column_sint64", "column_float",
  136. "column_1d", "column_2d", "column_3d", "column_binary"])
  137. data3 = ds.zip((data1, data2))
  138. ds.serialize(data3, "zip_dataset_pipeline.json")
  139. assert validate_jsonfile("zip_dataset_pipeline.json") is True
  140. assert validate_jsonfile("zip_dataset_pipeline_typo.json") is False
  141. data4 = ds.deserialize(json_filepath="zip_dataset_pipeline.json")
  142. ds.serialize(data4, "zip_dataset_pipeline_1.json")
  143. assert validate_jsonfile("zip_dataset_pipeline_1.json") is True
  144. assert filecmp.cmp('zip_dataset_pipeline.json', 'zip_dataset_pipeline_1.json')
  145. rows = 0
  146. for d0, d3, d4 in zip(ds0.create_tuple_iterator(output_numpy=True), data3.create_tuple_iterator(output_numpy=True),
  147. data4.create_tuple_iterator(output_numpy=True)):
  148. num_cols = len(d0)
  149. offset = 0
  150. for t1 in d0:
  151. np.testing.assert_array_equal(t1, d3[offset])
  152. np.testing.assert_array_equal(t1, d3[offset + num_cols])
  153. np.testing.assert_array_equal(t1, d4[offset])
  154. np.testing.assert_array_equal(t1, d4[offset + num_cols])
  155. offset += 1
  156. rows += 1
  157. assert rows == 12
  158. if remove_json_files:
  159. delete_json_files()
  160. def test_serdes_random_crop():
  161. """
  162. Test serdes on RandomCrop pipeline.
  163. """
  164. logger.info("test_random_crop")
  165. DATA_DIR = ["../data/dataset/test_tf_file_3_images/train-0000-of-0001.data"]
  166. SCHEMA_DIR = "../data/dataset/test_tf_file_3_images/datasetSchema.json"
  167. original_seed = config_get_set_seed(1)
  168. original_num_parallel_workers = config_get_set_num_parallel_workers(1)
  169. # First dataset
  170. data1 = ds.TFRecordDataset(DATA_DIR, SCHEMA_DIR, columns_list=["image"])
  171. decode_op = vision.Decode()
  172. random_crop_op = vision.RandomCrop([512, 512], [200, 200, 200, 200])
  173. data1 = data1.map(operations=decode_op, input_columns="image")
  174. data1 = data1.map(operations=random_crop_op, input_columns="image")
  175. # Serializing into python dictionary
  176. ds1_dict = ds.serialize(data1)
  177. # Serializing into json object
  178. _ = json.dumps(ds1_dict, indent=2)
  179. # Reconstruct dataset pipeline from its serialized form
  180. data1_1 = ds.deserialize(input_dict=ds1_dict)
  181. # Second dataset
  182. data2 = ds.TFRecordDataset(DATA_DIR, SCHEMA_DIR, columns_list=["image"])
  183. data2 = data2.map(operations=decode_op, input_columns="image")
  184. for item1, item1_1, item2 in zip(data1.create_dict_iterator(num_epochs=1, output_numpy=True),
  185. data1_1.create_dict_iterator(num_epochs=1, output_numpy=True),
  186. data2.create_dict_iterator(num_epochs=1, output_numpy=True)):
  187. np.testing.assert_array_equal(item1['image'], item1_1['image'])
  188. _ = item2["image"]
  189. # Restore configuration num_parallel_workers
  190. ds.config.set_seed(original_seed)
  191. ds.config.set_num_parallel_workers(original_num_parallel_workers)
  192. def test_serdes_cifar10_dataset(remove_json_files=True):
  193. """
  194. Test serdes on Cifar10 dataset pipeline
  195. """
  196. data_dir = "../data/dataset/testCifar10Data"
  197. original_seed = config_get_set_seed(1)
  198. original_num_parallel_workers = config_get_set_num_parallel_workers(1)
  199. data1 = ds.Cifar10Dataset(data_dir, num_samples=10, shuffle=False)
  200. data1 = data1.take(6)
  201. trans = [
  202. vision.RandomCrop((32, 32), (4, 4, 4, 4)),
  203. vision.Resize((224, 224)),
  204. vision.Rescale(1.0 / 255.0, 0.0),
  205. vision.Normalize([0.4914, 0.4822, 0.4465], [0.2023, 0.1994, 0.2010]),
  206. vision.HWC2CHW()
  207. ]
  208. type_cast_op = c.TypeCast(mstype.int32)
  209. data1 = data1.map(operations=type_cast_op, input_columns="label")
  210. data1 = data1.map(operations=trans, input_columns="image")
  211. data1 = data1.batch(3, drop_remainder=True)
  212. data1 = data1.repeat(1)
  213. data2 = util_check_serialize_deserialize_file(data1, "cifar10_dataset_pipeline", remove_json_files)
  214. num_samples = 0
  215. # Iterate and compare the data in the original pipeline (data1) against the deserialized pipeline (data2)
  216. for item1, item2 in zip(data1.create_dict_iterator(num_epochs=1, output_numpy=True),
  217. data2.create_dict_iterator(num_epochs=1, output_numpy=True)):
  218. np.testing.assert_array_equal(item1['image'], item2['image'])
  219. num_samples += 1
  220. assert num_samples == 2
  221. # Restore configuration num_parallel_workers
  222. ds.config.set_seed(original_seed)
  223. ds.config.set_num_parallel_workers(original_num_parallel_workers)
  224. def test_serdes_celeba_dataset(remove_json_files=True):
  225. """
  226. Test serdes on Celeba dataset pipeline.
  227. """
  228. DATA_DIR = "../data/dataset/testCelebAData/"
  229. data1 = ds.CelebADataset(DATA_DIR, decode=True, num_shards=1, shard_id=0)
  230. # define map operations
  231. data1 = data1.repeat(2)
  232. center_crop = vision.CenterCrop((80, 80))
  233. pad_op = vision.Pad(20, fill_value=(20, 20, 20))
  234. data1 = data1.map(operations=[center_crop, pad_op], input_columns=["image"], num_parallel_workers=8)
  235. data2 = util_check_serialize_deserialize_file(data1, "celeba_dataset_pipeline", remove_json_files)
  236. num_samples = 0
  237. # Iterate and compare the data in the original pipeline (data1) against the deserialized pipeline (data2)
  238. for item1, item2 in zip(data1.create_dict_iterator(num_epochs=1, output_numpy=True),
  239. data2.create_dict_iterator(num_epochs=1, output_numpy=True)):
  240. np.testing.assert_array_equal(item1['image'], item2['image'])
  241. num_samples += 1
  242. assert num_samples == 8
  243. def test_serdes_csv_dataset(remove_json_files=True):
  244. """
  245. Test serdes on Csvdataset pipeline.
  246. """
  247. DATA_DIR = "../data/dataset/testCSV/1.csv"
  248. data1 = ds.CSVDataset(
  249. DATA_DIR,
  250. column_defaults=["1", "2", "3", "4"],
  251. column_names=['col1', 'col2', 'col3', 'col4'],
  252. shuffle=False)
  253. columns = ["col1", "col4", "col2"]
  254. data1 = data1.project(columns=columns)
  255. data2 = util_check_serialize_deserialize_file(data1, "csv_dataset_pipeline", remove_json_files)
  256. num_samples = 0
  257. # Iterate and compare the data in the original pipeline (data1) against the deserialized pipeline (data2)
  258. for item1, item2 in zip(data1.create_dict_iterator(num_epochs=1, output_numpy=True),
  259. data2.create_dict_iterator(num_epochs=1, output_numpy=True)):
  260. np.testing.assert_array_equal(item1['col1'], item2['col1'])
  261. np.testing.assert_array_equal(item1['col2'], item2['col2'])
  262. np.testing.assert_array_equal(item1['col4'], item2['col4'])
  263. num_samples += 1
  264. assert num_samples == 3
  265. def test_serdes_voc_dataset(remove_json_files=True):
  266. """
  267. Test serdes on VOC dataset pipeline.
  268. """
  269. data_dir = "../data/dataset/testVOC2012"
  270. original_seed = config_get_set_seed(1)
  271. original_num_parallel_workers = config_get_set_num_parallel_workers(1)
  272. # define map operations
  273. random_color_adjust_op = vision.RandomColorAdjust(brightness=(0.5, 0.5))
  274. random_rotation_op = vision.RandomRotation((0, 90), expand=True, resample=Inter.BILINEAR, center=(50, 50),
  275. fill_value=150)
  276. data1 = ds.VOCDataset(data_dir, task="Detection", usage="train", decode=True)
  277. data1 = data1.map(operations=random_color_adjust_op, input_columns=["image"])
  278. data1 = data1.map(operations=random_rotation_op, input_columns=["image"])
  279. data1 = data1.skip(2)
  280. data2 = util_check_serialize_deserialize_file(data1, "voc_dataset_pipeline", remove_json_files)
  281. num_samples = 0
  282. # Iterate and compare the data in the original pipeline (data1) against the deserialized pipeline (data2)
  283. for item1, item2 in zip(data1.create_dict_iterator(num_epochs=1, output_numpy=True),
  284. data2.create_dict_iterator(num_epochs=1, output_numpy=True)):
  285. np.testing.assert_array_equal(item1['image'], item2['image'])
  286. num_samples += 1
  287. assert num_samples == 7
  288. # Restore configuration num_parallel_workers
  289. ds.config.set_seed(original_seed)
  290. ds.config.set_num_parallel_workers(original_num_parallel_workers)
  291. def test_serdes_to_device(remove_json_files=True):
  292. """
  293. Test serdes on transfer dataset pipeline.
  294. """
  295. data_dir = ["../data/dataset/test_tf_file_3_images/train-0000-of-0001.data"]
  296. schema_file = "../data/dataset/test_tf_file_3_images/datasetSchema.json"
  297. data1 = ds.TFRecordDataset(data_dir, schema_file, columns_list=["image", "label"], shuffle=False)
  298. data1 = data1.to_device()
  299. util_check_serialize_deserialize_file(data1, "transfer_dataset_pipeline", remove_json_files)
  300. def test_serdes_pyvision(remove_json_files=True):
  301. """
  302. Test serdes on py_transform pipeline.
  303. """
  304. data_dir = ["../data/dataset/test_tf_file_3_images/train-0000-of-0001.data"]
  305. schema_file = "../data/dataset/test_tf_file_3_images/datasetSchema.json"
  306. data1 = ds.TFRecordDataset(data_dir, schema_file, columns_list=["image", "label"], shuffle=False)
  307. transforms = [
  308. py_vision.Decode(),
  309. py_vision.CenterCrop([32, 32]),
  310. py_vision.ToTensor()
  311. ]
  312. data1 = data1.map(operations=py.Compose(transforms), input_columns=["image"])
  313. # Current python function derialization will be failed for pickle, so we disable this testcase
  314. # as an exception testcase.
  315. try:
  316. util_check_serialize_deserialize_file(data1, "pyvision_dataset_pipeline", remove_json_files)
  317. assert False
  318. except NotImplementedError as e:
  319. assert "python function is not yet supported" in str(e)
  320. def test_serdes_uniform_augment(remove_json_files=True):
  321. """
  322. Test serdes on uniform augment.
  323. """
  324. data_dir = "../data/dataset/testPK/data"
  325. data = ds.ImageFolderDataset(dataset_dir=data_dir, shuffle=False)
  326. ds.config.set_seed(1)
  327. transforms_ua = [vision.RandomHorizontalFlip(),
  328. vision.RandomVerticalFlip(),
  329. vision.RandomColor(),
  330. vision.RandomSharpness(),
  331. vision.Invert(),
  332. vision.AutoContrast(),
  333. vision.Equalize()]
  334. transforms_all = [vision.Decode(), vision.Resize(size=[224, 224]),
  335. vision.UniformAugment(transforms=transforms_ua, num_ops=5)]
  336. data = data.map(operations=transforms_all, input_columns="image", num_parallel_workers=1)
  337. util_check_serialize_deserialize_file(data, "uniform_augment_pipeline", remove_json_files)
  338. def test_serdes_exception():
  339. """
  340. Test exception case in serdes
  341. """
  342. data_dir = ["../data/dataset/test_tf_file_3_images/train-0000-of-0001.data"]
  343. schema_file = "../data/dataset/test_tf_file_3_images/datasetSchema.json"
  344. data1 = ds.TFRecordDataset(data_dir, schema_file, columns_list=["image", "label"], shuffle=False)
  345. data1 = data1.filter(input_columns=["image", "label"], predicate=lambda data: data < 11, num_parallel_workers=4)
  346. data1_json = ds.serialize(data1)
  347. with pytest.raises(RuntimeError) as msg:
  348. ds.deserialize(input_dict=data1_json)
  349. assert "Filter is not yet supported by ds.engine.deserialize" in str(msg)
  350. def util_check_serialize_deserialize_file(data_orig, filename, remove_json_files):
  351. """
  352. Utility function for testing serdes files. It is to check if a json file is indeed created with correct name
  353. after serializing and if it remains the same after repeatedly saving and loading.
  354. :param data_orig: original data pipeline to be serialized
  355. :param filename: filename to be saved as json format
  356. :param remove_json_files: whether to remove the json file after testing
  357. :return: The data pipeline after serializing and deserializing using the original pipeline
  358. """
  359. file1 = filename + ".json"
  360. file2 = filename + "_1.json"
  361. ds.serialize(data_orig, file1)
  362. assert validate_jsonfile(file1) is True
  363. assert validate_jsonfile("wrong_name.json") is False
  364. data_changed = ds.deserialize(json_filepath=file1)
  365. ds.serialize(data_changed, file2)
  366. assert validate_jsonfile(file2) is True
  367. assert filecmp.cmp(file1, file2)
  368. # Remove the generated json file
  369. if remove_json_files:
  370. delete_json_files()
  371. return data_changed
  372. def validate_jsonfile(filepath):
  373. try:
  374. file_exist = os.path.exists(filepath)
  375. with open(filepath, 'r') as jfile:
  376. loaded_json = json.load(jfile)
  377. except IOError:
  378. return False
  379. return file_exist and isinstance(loaded_json, dict)
  380. def delete_json_files():
  381. file_list = glob.glob('*.json')
  382. for f in file_list:
  383. try:
  384. os.remove(f)
  385. except IOError:
  386. logger.info("Error while deleting: {}".format(f))
  387. # Test save load minddataset
  388. def skip_test_minddataset(add_and_remove_cv_file):
  389. """tutorial for cv minderdataset."""
  390. columns_list = ["data", "file_name", "label"]
  391. num_readers = 4
  392. indices = [1, 2, 3, 5, 7]
  393. sampler = ds.SubsetRandomSampler(indices)
  394. data_set = ds.MindDataset(CV_FILE_NAME + "0", columns_list, num_readers,
  395. sampler=sampler)
  396. # Serializing into python dictionary
  397. ds1_dict = ds.serialize(data_set)
  398. # Serializing into json object
  399. ds1_json = json.dumps(ds1_dict, sort_keys=True)
  400. # Reconstruct dataset pipeline from its serialized form
  401. data_set = ds.deserialize(input_dict=ds1_dict)
  402. ds2_dict = ds.serialize(data_set)
  403. # Serializing into json object
  404. ds2_json = json.dumps(ds2_dict, sort_keys=True)
  405. assert ds1_json == ds2_json
  406. _ = get_data(CV_DIR_NAME)
  407. assert data_set.get_dataset_size() == 5
  408. num_iter = 0
  409. for _ in data_set.create_dict_iterator(num_epochs=1, output_numpy=True):
  410. num_iter += 1
  411. assert num_iter == 5
  412. if __name__ == '__main__':
  413. test_serdes_imagefolder_dataset()
  414. test_serdes_mnist_dataset()
  415. test_serdes_cifar10_dataset()
  416. test_serdes_celeba_dataset()
  417. test_serdes_csv_dataset()
  418. test_serdes_voc_dataset()
  419. test_serdes_zip_dataset()
  420. test_serdes_random_crop()
  421. test_serdes_exception()