i-robot
f46208e2ce
!26214 Fix Complex of the Layers
Merge pull request !26214 from huangxinjing/fix_cco_layers
4 years ago
yangzhenzhang
ba99e4c505
add check for resize op
4 years ago
i-robot
07f5702e48
!26276 [Auto parallel] Fix the cycle complexity problem
Merge pull request !26276 from Xiaoda/100-adjusting-for-cycle-complexity
4 years ago
i-robot
6bdd38399a
!25811 fault_recover_by_mirror_group
Merge pull request !25811 from yao_yf/fault_recover_by_mirror_group
4 years ago
haoran.wang
9b13da25cc
Modify Softmax as eliminated op in D-Rec
4 years ago
Xiaoda Zhang
58451687f2
fix the high cycle complexity problem
4 years ago
lichenever
8abc711298
fix_codex_pclint_r1.5
4 years ago
yao_yf
501b978d16
find data parallel common group in auto parallel
4 years ago
i-robot
9916e5844d
!26170 Fix cyclomatic complexity
Merge pull request !26170 from hewei/fix_ccn
4 years ago
huangxinjing
8c2dec7fe2
Fix coo layers
4 years ago
i-robot
6c587dc2d3
!25391 Support to profiling parallel strategy
Merge pull request !25391 from ougongchang/profiling_stategy
4 years ago
He Wei
c7910cbd37
Fix cyclomatic complexity
4 years ago
i-robot
7a73bae5c3
!26036 add output strategy for matmul operator
Merge pull request !26036 from yangzhenzhang/add-output-strategy-for-op-init
4 years ago
ougongchang
9229f1c1ff
profiler support to collect parallel strategy info
If SetNodeOutputType functions forcibly splits into multiple functions, the readability decreases, so it blocks lizard scans
4 years ago
Xiaoda Zhang
a772767265
support reshape in sharding propagation:
1) using 'swc index of strategy_cost_' as reshape's selected strategy;
2) when encountering reshape in BFS, select the 'swc index' with zero communication cost;
3) when encountering a reshape that is already visited, check whether there exists communication between reshape and current operator. It is OK if communication happens between two configured operators;
4) currently, two consecutive reshapes are not supported;
5) adjusting BFS structure in graph_costmodel.cc;
6) adjusting some code in step_auto_parallel.cc to avoid cyclomatic complexity.
4 years ago
yangzhenzhang
8431ba616c
add output strategy for op init
4 years ago
i-robot
14c3517b9a
!25452 Add Gradient Sharding
Merge pull request !25452 from huangxinjing/pipeline_shard_mini_step
4 years ago
huangxinjing
f354ab22a3
add pipeline shard interface
Add support for no pipeline accugradient
Add delay tag for fusion op
Optimizer the visite order
add mirror for mini step control
Move the group to attributes
Add gradient_shard control for the mini step
Fix code stype
Fix ut description
Add interface
4 years ago
haoran.wang
418fdedbc8
remove MatMul HCCL restriction
4 years ago
i-robot
ded1c77bbf
!25765 neighborExchangeV2 & grad
Merge pull request !25765 from TuDouNi/neighborExchangeV2
4 years ago
ttudu
e953c15cd2
NeighborExchangeV2 & Grad
4 years ago
i-robot
7c14f984ed
!25793 add output strategy for shard
Merge pull request !25793 from yangzhenzhang/add-output-strategy
4 years ago
i-robot
17db972047
!25710 modify pangu model batchsize split
Merge pull request !25710 from lilei/modify_pangu_model_for_master
4 years ago
yangzhenzhang
6ad6304b77
add output strategy
4 years ago
i-robot
b96c0e0c01
!25836 dataset_shard_tmp_fix
Merge pull request !25836 from yao_yf/dataset_shard_tmp_fix
4 years ago
i-robot
97ca5e5059
!25516 [CT][MS][parallel] GatherV2P merged to Gather
Merge pull request !25516 from 王浩然/Gather_version_update
4 years ago
i-robot
27be612516
!25719 Add Device number check
Merge pull request !25719 from huangxinjing/add_device_number_check
4 years ago
yao_yf
d8bd59b03d
dataset strategy tmp fix
4 years ago
haoran.wang
e4eafa8d7c
Gather bug fixed and delete PrepareGatherV2 func which would never be used
4 years ago
lilei
6dbee87bb8
modify pangu model micro batch
4 years ago
yao_yf
a90a8ec607
dataset with repeat strategy fix
4 years ago
huangxinjing
741e3604eb
Add Device Number Check
4 years ago
lilei
2263e05bc7
modify pangu model micro batch
4 years ago
b00518648
ef715a54a0
clean code
4 years ago
i-robot
48699c939c
!25544 add parallel op for resizenearestneighbor
Merge pull request !25544 from yangzhenzhang/add-parallel-op-for-resizenearestneighbor
4 years ago
yangzhenzhang
c42081619e
add parallel op for resizenearestneighbor
4 years ago
yao_yf
0d6f8e0619
dataset shard strategy fix
4 years ago
zhunaipan
8ce4e62725
optimize the comment and log description
修改: ops/operations/_inner_ops.py
修改: ops/operations/_quant_ops.py
修改: ops/operations/array_ops.py
修改: ops/operations/comm_ops.py
修改: ops/operations/math_ops.py
修改: ops/operations/quantum_ops.py
修改: ops/operations/rl_ops.py
修改: ops/operations/sponge_ops.py
修改: ops/operations/sponge_update_ops.py
修改: train/__init__.py
修改: common/tensor.py
修改: train/serialization.py
修改: ccsrc/pipeline/jit/parse/parse.h
修改: explainer/benchmark/_attribution/metric.py
修改: ops/composite/multitype_ops/_constexpr_utils.py
修改: ops/operations/comm_ops.py
修改: RELEASE.md
修改: mindspore/_extends/parse/standard_method.py
修改: mindspore/ccsrc/backend/kernel_compiler/cpu/concat_offset_cpu_kernel.cc
修改: mindspore/ccsrc/backend/kernel_compiler/cpu/dynamic_shape_cpu_kernel.cc
修改: mindspore/ccsrc/frontend/parallel/ops_info/reshape_info.cc
修改: mindspore/ccsrc/frontend/parallel/ops_info/tile_info.cc
修改: mindspore/ccsrc/frontend/parallel/ops_info/transpose_info.cc
修改: mindspore/ccsrc/frontend/parallel/strategy.h
修改: mindspore/common/tensor.py
修改: mindspore/core/abstract/prim_arrays.cc
修改: mindspore/core/abstract/prim_nn.cc
修改: mindspore/core/ops/conv2d.cc
修改: mindspore/core/ops/logical_and.h
修改: mindspore/core/ops/logical_not.h
修改: mindspore/core/ops/logical_or.h
修改: mindspore/core/ops/reduce_all.h
修改: mindspore/core/ops/reduce_any.h
修改: mindspore/lite/src/runtime/kernel/arm/fp32_grad/sgd.cc
修改: mindspore/nn/layer/quant.py
修改: mindspore/nn/optim/sgd.py
修改: mindspore/nn/sparse/sparse.py
修改: mindspore/numpy/array_creations.py
修改: mindspore/numpy/array_ops.py
修改: mindspore/numpy/logic_ops.py
修改: mindspore/numpy/math_ops.py
修改: mindspore/ops/operations/_inner_ops.py
修改: mindspore/ops/operations/array_ops.py
修改: mindspore/ops/operations/rl_ops.py
修改: mindspore/train/_utils.py
修改: tests/ut/python/model/test_lenet_core_after_exception.py
修改: mindspore/_extends/parse/standard_method.py
修改: mindspore/ops/operations/rl_ops.py
修改: mindspore/core/abstract/prim_nn.cc
修改: mindspore/core/ops/conv2d.cc
修改: mindspore/ccsrc/backend/kernel_compiler/cpu/ctcloss_cpu_kernel.cc
修改: mindspore/ccsrc/backend/kernel_compiler/cpu/fl/fused_pull_weight_kernel.h
修改: mindspore/ccsrc/backend/kernel_compiler/cpu/fl/fused_push_weight_kernel.h
修改: mindspore/ccsrc/backend/kernel_compiler/cpu/mkldnn/conv2d_grad_filter_cpu_kernel.cc
修改: mindspore/ccsrc/backend/kernel_compiler/cpu/mkldnn/conv2d_grad_input_cpu_kernel.cc
修改: mindspore/ccsrc/backend/kernel_compiler/cpu/ps/sparse_apply_ftrl_ps_kernel.cc
修改: mindspore/ccsrc/backend/kernel_compiler/cpu/ps/sparse_apply_lazy_adam_ps_kernel.cc
修改: mindspore/ccsrc/backend/kernel_compiler/cpu/rolling_cpu_kernel.cc
修改: mindspore/ccsrc/backend/kernel_compiler/cpu/scatter_arithmetic_cpu_kernel.cc
修改: mindspore/ccsrc/backend/kernel_compiler/cpu/split_cpu_kernel.cc
修改: mindspore/ccsrc/backend/kernel_compiler/cpu/update_cache_cpu_kernel.cc
修改: mindspore/ccsrc/backend/kernel_compiler/gpu/arrays/split_gpu_kernel.h
修改: mindspore/ccsrc/backend/kernel_compiler/gpu/math/broadcast_gpu_kernel.h
修改: mindspore/ccsrc/backend/kernel_compiler/gpu/nn/conv2d_grad_input_gpu_kernel.h
修改: mindspore/ccsrc/fl/server/server.cc
修改: mindspore/ccsrc/frontend/optimizer/ad/kpynative.cc
修改: mindspore/ccsrc/frontend/optimizer/irpass/incorporate_getitem.h
修改: mindspore/ccsrc/frontend/optimizer/irpass/inline.h
修改: mindspore/ccsrc/minddata/dataset/core/device_tensor.cc
修改: mindspore/ccsrc/minddata/dataset/core/tensor.cc
修改: mindspore/ccsrc/minddata/dataset/engine/datasetops/source/emnist_op.cc
修改: mindspore/ccsrc/minddata/dataset/engine/datasetops/source/mnist_op.cc
修改: mindspore/ccsrc/minddata/dataset/engine/datasetops/source/qmnist_op.cc
修改: mindspore/ccsrc/minddata/dataset/engine/ir/datasetops/dataset_node.cc
修改: mindspore/ccsrc/minddata/dataset/engine/opt/pre/epoch_ctrl_pass.cc
修改: mindspore/ccsrc/minddata/dataset/kernels/image/lite_image_utils.cc
修改: mindspore/ccsrc/pipeline/jit/action.cc
修改: mindspore/ccsrc/pipeline/jit/static_analysis/evaluator.cc
修改: mindspore/ccsrc/runtime/device/ascend/executor/tiling/op_tiling_adapter.cc
修改: mindspore/compression/quant/quant_utils.py
修改: mindspore/core/abstract/prim_nn.cc
修改: mindspore/dataset/engine/validators.py
修改: mindspore/lite/micro/coder/opcoders/nnacl/fp32/affine_fp32_coder.cc
修改: mindspore/lite/micro/coder/opcoders/nnacl/int8/affine_int8_coder.cc
修改: mindspore/lite/src/runtime/kernel/ascend310/src/custom_kernel.cc
修改: mindspore/lite/src/runtime/kernel/opencl/kernel/matmul.cc
修改: mindspore/lite/src/runtime/kernel/opencl/kernel/strassen.cc
修改: mindspore/lite/tools/common/graph_util.h
修改: mindspore/lite/tools/optimizer/fisson/fisson_util.cc
修改: mindspore/ops/composite/math_ops.py
修改: mindspore/ops/operations/_inner_ops.py
修改: mindspore/ops/operations/array_ops.py
修改: mindspore/ops/operations/math_ops.py
修改: mindspore/ops/operations/other_ops.py
修改: mindspore/boost/boost_cell_wrapper.py
修改: mindspore/ccsrc/backend/kernel_compiler/cpu/update_cache_cpu_kernel.cc
修改: mindspore/ccsrc/common/trans.cc
修改: mindspore/ccsrc/frontend/parallel/cache_embedding/cache_embedding.cc
修改: mindspore/ccsrc/frontend/parallel/ops_info/gather_info.cc
修改: mindspore/lite/src/common/log_util.h
修改: mindspore/nn/wrap/loss_scale.py
修改: mindspore/parallel/nn/moe.py
修改: tests/mindspore_test_framework/mindspore_test.py
修改: mindspore/ccsrc/backend/kernel_compiler/cpu/split_cpu_kernel.cc
修改: mindspore/lite/tools/common/graph_util.h
修改: mindspore/ccsrc/frontend/parallel/ops_info/gather_info.cc
修改: mindspore/core/ops/conv2d.cc
修改: tests/ut/python/model/test_lenet_core_after_exception.py
4 years ago
lichenever
f9f515c93f
opt_pipeline_performance
4 years ago
i-robot
281fc7c2ce
!24898 [AutoParallel]fix pipeline find end node bug
Merge pull request !24898 from lichen/fix_pipeline_find_end_node_bug
4 years ago
lichenever
596923d74b
fix_pipeline_bug_r1.5
4 years ago
i-robot
f83070728d
!24790 support user define strategy gen method under auto parallel context
Merge pull request !24790 from zhuyuxiao/master
4 years ago
zhuyuxiao
1907246931
change api
4 years ago
i-robot
3fd94000c5
!24568 Apply batch parallel in auto_parallel mode when strategies are not specified
Merge pull request !24568 from zhuyuxiao/master
4 years ago
b00518648
ea50695cae
pclint
4 years ago
zhuyuxiao
cf76c76745
apply batch parallel in auto_parallel mode when strategies are not specified
4 years ago
lichenever
5ff1124e71
fix_pipeline_with_no_loss_bug
4 years ago
yao_yf
d38f1b7734
model parallel communication recompute rule change
4 years ago
i-robot
3af44a16ee
!24377 modify pangu model batchsize split
Merge pull request !24377 from lilei/modify_pangu_model_for_master
4 years ago
lilei
35dbb76214
modify pangu model batchsize split
4 years ago