 added python api based on cpp api
1st draft of python iterator
Added Cifar10 and Cifar100 pybind port
Change pybind to use IR for Skip and Manifest
Signed-off-by: alex-yuyue <yue.yu1@huawei.com>
DatasetNode as a base for all IR nodes
namespace change
Fix the namespace issue and make ut tests work
Signed-off-by: alex-yuyue <yue.yu1@huawei.com>
Add VOCDataset
!63 Added RandomDataset
* Added RandomDataset
add imagefolder ir
Pybind switch: CelebA and UT
!61 CLUE example with class definition
* Merge branch 'python-api' of gitee.com:ezphlow/mindspore into clue_class_pybind
* Passing testcases
* Added CLUE, not working
add ManifestDataset IR
Signed-off-by: alex-yuyue <yue.yu1@huawei.com>
Update Coco & VOC & TFReader, Update clang-format, Reorder
datasets_binding
!69 Add Generator and move c_dataset.Iterator to dataset.Iterator
* Add GeneratorDataset to c_dataset
* Add GeneratorDataset to c_dataset
!67 Moving c_datasets and adding sampler wrapper
* Need to add create() method in datasets.py
* migration from c_dataset to dataset part 1
!71 Fix indent error
* Fix indentation error
!72 Fix c_api tests cases
* Fix c_api tests cases
!73 Added CSV Dataset
* Added CSVDataset
pybind switch: Take and CelebA fixes
!75 move c_dataset functionality to datasets
* Fixed existing testcases
* Added working clue and imagefolder
* Added sampler conversion from pybind
* Added sampler creation
!77 Add Python API tree
* Python API tree
add minddataset
TextFileDataset pybind
Rename to skip test_concat.py and test_minddataset_exception.py
!80 Add batch IR to python-api branch, most test cases work
* staging III
* staging, add pybind
Enable more c_api take and CelebA tests; delete util_c_api
!84 Schema changes in datasets.py
* Schema changes
!85 Remove input_indexes from sub-classes
* remove input_index from each subclass
!83 Remove C datasets
* Removed c_dataset package
* Remove c_datasets
!82 pybind switch: shuffle
* pybind switch: shuffle
!86 Add build_vocab
* Add build_vocab
Rebase with upstream/master
_shuffle conflict
BatchNode error
!88 Fix rebase problem
* fix rebase problem
Enable more unit tests; code typo/nit fixes
!91 Fix python vocag hang
* Fix python vocab hang
!89 Added BucketBatchByLength Pybind switch
* Added BucketBatchByLength
Update and enable more tet_c_api_*.py tests
!95 Add BuildSentencePeiceVocab
* - Add BuildSentencePeiceVocab
!96 Fix more tests
* - Fix some tests
- Enable more test_c_api_*
- Add syncwait
!99 pybind switch for device op
* pybind switch for device op
!93 Add getters to python API
* Add getters to python API
!101 Validate tree, error if graph
* - Add sync wait
!103 TFrecord/Random Datasets schema problem
* - TfRecord/Random schem aproblem
!102 Added filter pybind switch
* Added Filter pybind switch
!104 Fix num_samples
* - TfRecord/Random schem aproblem
!105 Fix to_device hang
* Fix to_device hang
!94 Adds Cache support for CLUE dataset
* Added cache for all dataset ops
* format change
* Added CLUE cache support
* Added Cache conversion
Add save pybind
fix compile err
init modify concat_node
!107 Fix some tests cases
* Fix tests cases
Enable and fix more tests
!109 pybind switch for get dataset size
* pybind_get_dataset_size
some check-code fixes for pylint, cpplint and clang-format
!113 Add callback
* revert
* dataset_sz 1 line
* fix typo
* get callback to work
!114 Make Android compile clean
* Make Android Compile Clean
Fix build issues due to rebase
!115 Fix more tests
* Fix tests cases
* !93 Add getters to python API
fix test_profiling.py
!116 fix get dataset size
* fix get dataset size
!117 GetColumnNames pybind switch
* Added GetColumnNames pybind switch
code-check fixes: clangformat, cppcheck, cpplint, pylint
Delete duplicate test_c_api_*.py files; more lint fixes
!121 Fix cpp tests
* Remove extra call to getNext in cpp tests
!122 Fix Schema with Generator
* Fix Schema with Generator
fix some cases of csv & mindrecord
!124 fix tfrecord get_dataset_size and add some UTs
* fix tfrecord get dataset size and add some ut for get_dataset_size
!125 getter separation
* Getter separation
!126 Fix sampler.GetNumSamples
* Fix sampler.GetNumSampler
!127 Assign runtime getter to each get function
* Assign runtime getter to each get function
Fix compile issues
!128 Match master code
* Match master code
!129 Cleanup DeviceOp/save code
* Cleanup ToDevice/Save code
!130 Add cache fix
* Added cache fix for map and image folder
!132 Fix testing team issues
* Pass queue_name from python to C++
* Add Schema.from_json
!131 Fix Cache op issues and delete de_pipeline
* Roll back C++ change
* Removed de_pipeline and passing all cache tests.
* fixed cache tests
!134 Cleanup datasets.py part1
* Cleanup dataset.py part1
!133 Updated validation for SentencePieceVocab.from_dataset
* Added type_check for column names in SentencePieceVocab.from_dataset
Rebase on master 181120 10:20
fix profiling
temporary solution of catching stauts from Node.Build()
!141 ToDevice Termination
* ToDevice termination
pylint fixes
!137 Fix test team issues and add some corresponding tests
* Fix test team issues and add some corresponding tests
!138 TreeGetter changes to use OptPass
* Getter changes to use OptPass (Zirui)
Rebase fix
!143 Fix cpplint issue
* Fix cpplint issue
pylint fixes in updated testcases
!145 Reset exceptions testcase
* reset exception test to master
!146 Fix Check_Pylint Error
* Fix Check_Pylint Error
!147 fix android
* fix android
!148 ToDevice changes
* Add ToDevice to the iterator List for cleanup at exit
!149 Pylint issue
* Add ToDevice to the iterator List for cleanup at exit
!150 Pylint 2
* Add ToDevice to the iterator List for cleanup at exit
!152 ExecutionTree error
* ET destructor error
!153 in getter_pass, only remove callback, without deleting map op
* getter pass no longer removes map
!156 early __del__ of iterator/to_device
* early __del__ of iterator
!155 Address review comments Eric 1
* Added one liner fix to validators.py
* roll back signature fix
* lint fix
* Eric Address comments 2
* C++ lint fix
* Address comments Eric 1
!158 Review rework for dataset bindings - part 1
* Reorder nodes repeat and rename
* Review rework for dataset bindings - part 1
!154 Fixing minor problems in the comments (datasets.py, python_tree_consumer.cc, iterators_bindings.cc, and iterators.py)
* Fixing minor problems in the comments (datasets.py, python_tree_consum…
!157 add replace none
* Add replace_none to datasets.py, address comments in tests
Trying to resolve copy
Override the deepcopy method of deviceop
Create_ir_tree method
Create_ir_tree method 2
Create_ir_tree method 2
del to_device if already exists
del to_device if already exists
cache getters shapes and types
Added yolov3 relaxation, to be rolled back
Get shapes and types together
bypass yolo
NumWorkers for MapOp
revert Yolo
revert Thor
Print more info
Debug code: Update LOG INFO to LOG ERROR
do not remove epochctrl for getter pass
Remove repeat(1)
pritn batch size
add log to tree_consumer and device_queue op
Revert PR 8744
Signed-off-by: alex-yuyue <yue.yu1@huawei.com>
__del__ toDEvice
__del__ toDevice2
!165 add ifndef ENABLE_ANDROID to device queue print
* Add ifndef ENABLE_ANDROID to device queue print
revert some changes
!166 getter: get_data_info
* getter: get_data_info
!168 add back tree print
* revert info to warnning in one log
* add back the missed print tree log
Release GIL in GetDataInfo
5 years ago |
123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349 |
- # Copyright 2019 Huawei Technologies Co., Ltd
- #
- # Licensed under the Apache License, Version 2.0 (the "License");
- # you may not use this file except in compliance with the License.
- # You may obtain a copy of the License at
- #
- # http://www.apache.org/licenses/LICENSE-2.0
- #
- # Unless required by applicable law or agreed to in writing, software
- # distributed under the License is distributed on an "AS IS" BASIS,
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- # See the License for the specific language governing permissions and
- # limitations under the License.
- # ==============================================================================
- """
- The configuration module provides various functions to set and get the supported
- configuration parameters, and read a configuration file.
- """
- import os
- import random
- import numpy
- import mindspore._c_dataengine as cde
-
- __all__ = ['set_seed', 'get_seed', 'set_prefetch_size', 'get_prefetch_size', 'set_num_parallel_workers',
- 'get_num_parallel_workers', 'set_monitor_sampling_interval', 'get_monitor_sampling_interval', 'load',
- 'get_callback_timeout', 'set_auto_num_workers', 'get_auto_num_workers', '_init_device_info']
-
- INT32_MAX = 2147483647
- UINT32_MAX = 4294967295
-
- _config = cde.GlobalContext.config_manager()
-
-
- def _init_device_info():
- """
- INTERNAL USE ONLY!
- As rank_id need to pass into deep layer for numa and device_queue.
- One process work with only one rank_id, In standalone scenario,
- rank_id may come from env 'CUDA_VISIBLE_DEVICES', For distribute
- scenario, rank_id come from _get_global_rank()
- """
- from mindspore import context
- from mindspore.parallel._auto_parallel_context import auto_parallel_context
- from mindspore.parallel._utils import _get_global_rank
- if context.get_context("device_target") == "GPU":
- rank_id = _get_global_rank()
- parallel_mode = auto_parallel_context().get_parallel_mode()
- if parallel_mode == "stand_alone":
- rank_id = context.get_context("device_id")
- _config.set_rank_id(rank_id)
- elif context.get_context("device_target") == "Ascend":
- # Ascend is a special scenario, we'd better get rank info from env
- env_rank_size = os.getenv("RANK_SIZE", None)
- env_rank_id = os.getenv("RANK_ID", None)
- if env_rank_size and env_rank_id:
- # Ascend only support multi-process scenario
- rank_size = int(env_rank_size.strip())
- rank_id = int(env_rank_id.strip())
- if rank_size > 1:
- _config.set_rank_id(rank_id)
- # Now single process under ascend mode doesn't support numa bind for performance consideration.
- if _config.get_numa_enable() is True and rank_size == 1:
- raise ValueError("single process under Ascend mode doesn't support numa bind for "
- "performance consideration.")
-
-
- def set_seed(seed):
- """
- Set the seed to be used in any random generator. This is used to produce deterministic results.
-
- Note:
- This set_seed function sets the seed in the Python random library and numpy.random library
- for deterministic Python augmentations using randomness. This set_seed function should
- be called with every iterator created to reset the random seed. In the pipeline, this
- does not guarantee deterministic results with num_parallel_workers > 1.
-
- Args:
- seed(int): Seed to be set.
-
- Raises:
- ValueError: If seed is invalid (< 0 or > MAX_UINT_32).
-
- Examples:
- >>> # Set a new global configuration value for the seed value.
- >>> # Operations with randomness will use the seed value to generate random values.
- >>> ds.config.set_seed(1000)
- """
- if seed < 0 or seed > UINT32_MAX:
- raise ValueError("Seed given is not within the required range.")
- _config.set_seed(seed)
- random.seed(seed)
- # numpy.random isn't thread safe
- numpy.random.seed(seed)
-
-
- def get_seed():
- """
- Get the seed.
-
- Returns:
- int, seed.
- """
- return _config.get_seed()
-
-
- def set_prefetch_size(size):
- """
- Set the number of rows to be prefetched.
-
- Args:
- size (int): Total number of rows to be prefetched.
-
- Raises:
- ValueError: If prefetch_size is invalid (<= 0 or > MAX_INT_32).
-
- Examples:
- >>> # Set a new global configuration value for the prefetch size.
- >>> ds.config.set_prefetch_size(1000)
- """
- if size <= 0 or size > INT32_MAX:
- raise ValueError("Prefetch size given is not within the required range.")
- _config.set_op_connector_size(size)
-
-
- def get_prefetch_size():
- """
- Get the prefetch size in number of rows.
-
- Returns:
- int, total number of rows to be prefetched.
- """
- return _config.get_op_connector_size()
-
-
- def set_num_parallel_workers(num):
- """
- Set the default number of parallel workers.
-
- Args:
- num (int): Number of parallel workers to be used as a default for each operation.
-
- Raises:
- ValueError: If num_parallel_workers is invalid (<= 0 or > MAX_INT_32).
-
- Examples:
- >>> # Set a new global configuration value for the number of parallel workers.
- >>> # Now parallel dataset operators will run with 8 workers.
- >>> ds.config.set_num_parallel_workers(8)
- """
- if num <= 0 or num > INT32_MAX:
- raise ValueError("Number of parallel workers given is not within the required range.")
- _config.set_num_parallel_workers(num)
-
-
- def get_num_parallel_workers():
- """
- Get the default number of parallel workers.
- This is the DEFAULT num_parallel_workers value used for each op, it is not related to AutoNumWorker feature.
-
- Returns:
- int, number of parallel workers to be used as a default for each operation.
- """
- return _config.get_num_parallel_workers()
-
-
- def set_numa_enable(numa_enable):
- """
- Set the default state of numa enabled.
-
- Args:
- numa_enable (bool): Whether to use numa bind feature.
-
- Raises:
- TypeError: If numa_enable is not a boolean data type.
-
- Examples:
- >>> # Set a new global configuration value for the state of numa enabled.
- >>> # Now parallel dataset operators will run with numa bind function
- >>> ds.config.set_numa_enable(True)
- """
- if not isinstance(numa_enable, bool):
- raise TypeError("numa_enable must be a boolean dtype.")
- _config.set_numa_enable(numa_enable)
-
-
- def get_numa_enable():
- """
- Get the default state of numa enabled.
- This is the DEFAULT numa enabled value used for the all process.
-
- Returns:
- bool, the default state of numa enabled.
- """
- return _config.get_numa_enable()
-
-
- def set_monitor_sampling_interval(interval):
- """
- Set the default interval (in milliseconds) for monitor sampling.
-
- Args:
- interval (int): Interval (in milliseconds) to be used for performance monitor sampling.
-
- Raises:
- ValueError: If interval is invalid (<= 0 or > MAX_INT_32).
-
- Examples:
- >>> # Set a new global configuration value for the monitor sampling interval.
- >>> ds.config.set_monitor_sampling_interval(100)
- """
- if interval <= 0 or interval > INT32_MAX:
- raise ValueError("Interval given is not within the required range.")
- _config.set_monitor_sampling_interval(interval)
-
-
- def get_monitor_sampling_interval():
- """
- Get the default interval of performance monitor sampling.
-
- Returns:
- int, interval (in milliseconds) for performance monitor sampling.
- """
- return _config.get_monitor_sampling_interval()
-
-
- def set_auto_num_workers(enable):
- """
- Set num_parallel_workers for each op automatically. (This feature is turned off by default)
- If turned on, the num_parallel_workers in each op will be adjusted automatically, possibly overwriting the
- num_parallel_workers passed in by user or the default value (if user doesn't pass anything) set by
- ds.config.set_num_parallel_workers().
- For now, this function is only optimized for Yolo3 dataset with per_batch_map (running map in batch).
- This feature aims to provide a baseline for optimized num_workers assignment for each op.
- Op whose num_parallel_workers is adjusted to a new value will be logged.
-
- Args:
- enable (bool): Whether to enable auto num_workers feature or not.
-
- Raises:
- TypeError: If enable is not of boolean type.
-
- Examples:
- >>> # Enable auto_num_worker feature, this might override the num_parallel_workers passed in by user
- >>> ds.config.set_auto_num_workers(True)
- """
- if not isinstance(enable, bool):
- raise TypeError("enable isn't of type bool.")
- _config.set_auto_num_workers(enable)
-
-
- def _set_auto_workers_config(option):
- """
- INTERNAL USE ONLY!
- Select the weight profile of auto_num_workers. currently these 7 options are supported.
- Option #0 leaf_num_workers:batch_num_workers:map_num_workers=1:1:1
- Option #1 leaf_num_workers:batch_num_workers:map_num_workers=2:1:1
- Option #2 leaf_num_workers:batch_num_workers:map_num_workers=1:2:1
- Option #3 leaf_num_workers:batch_num_workers:map_num_workers=1:1:2
- Option #4 leaf_num_workers:batch_num_workers:map_num_workers=2:2:1
- Option #5 leaf_num_workers:batch_num_workers:map_num_workers=2:1:2
- Option #6 leaf_num_workers:batch_num_workers:map_num_workers=1:2:2
- Args:
- option (int): The id of the profile to use.
- Raises:
- ValueError: If option is not int or not within the range of [0, 6]
- """
- if not isinstance(option, int):
- raise ValueError("option isn't of type int.")
- if option < 0 or option > 6:
- raise ValueError("option isn't within the required range of [0, 6].")
- _config.set_auto_worker_config(option)
-
-
- def get_auto_num_workers():
- """
- Get the setting (turned on or off) automatic number of workers.
-
- Returns:
- bool, whether auto num worker feature is turned on.
-
- Examples:
- >>> num_workers = ds.config.get_auto_num_workers()
- """
- return _config.get_auto_num_workers()
-
-
- def set_callback_timeout(timeout):
- """
- Set the default timeout (in seconds) for DSWaitedCallback.
- In case of a deadlock, the wait function will exit after the timeout period.
-
- Args:
- timeout (int): Timeout (in seconds) to be used to end the wait in DSWaitedCallback in case of a deadlock.
-
- Raises:
- ValueError: If timeout is invalid (<= 0 or > MAX_INT_32).
-
- Examples:
- >>> # Set a new global configuration value for the timeout value.
- >>> ds.config.set_callback_timeout(100)
- """
- if timeout <= 0 or timeout > INT32_MAX:
- raise ValueError("Timeout given is not within the required range.")
- _config.set_callback_timeout(timeout)
-
-
- def get_callback_timeout():
- """
- Get the default timeout for DSWaitedCallback.
- In case of a deadlock, the wait function will exit after the timeout period.
-
- Returns:
- int, the duration in seconds.
- """
- return _config.get_callback_timeout()
-
-
- def __str__():
- """
- String representation of the configurations.
-
- Returns:
- str, configurations.
- """
- return str(_config)
-
-
- def load(file):
- """
- Load configurations from a file.
-
- Args:
- file (str): Path of the configuration file to be loaded.
-
- Raises:
- RuntimeError: If file is invalid and parsing fails.
-
- Examples:
- >>> # Set new default configuration values according to values in the configuration file.
- >>> ds.config.load("/path/to/config_directory/config.cfg")
- >>> # example config file:
- >>> # {
- >>> # "logFilePath": "/tmp",
- >>> # "numParallelWorkers": 4,
- >>> # "seed": 5489,
- >>> # "monitorSamplingInterval": 30
- >>> # }
- """
- _config.load(file)
|