i-robot
0d27e1e63c
!27189 [AUTOPARALLEL] Fix context raise error
Merge pull request !27189 from huangxinjing/error_raise
4 years ago
i-robot
d8e5200e39
!27245 modify parallel api note for master
Merge pull request !27245 from lilei/modify_parallel_API_note_for_master
4 years ago
huangxinjing
baefffd318
fix intent
4 years ago
q00596439
de36fdc169
add allreduce fusion size and unify the interface
4 years ago
lilei
6a1c537183
modify parallel api note for master
4 years ago
Xiaoda Zhang
04db51a528
In a previous PR ( https://gitee.com/mindspore/mindspore/pulls/26807/ ), we replaced 'auto_parallel_search_mode' by 'search_mode' directly.
However, to be forward compatitable, it is suitable to keep 'auto_parallel_search_mode' available. This PR recovers the 'auto_parallel_search_mode' interface and adds a warning when using this old interface.
This PR also deals with other codestyle things.
4 years ago
Xiaoda Zhang
ad5ac77ae8
1) 'auto_parallel_search_mode' changes to 'search_mode';
2) 'sharding_propagation' moves to 'search_mode';
4 years ago
lilei
f8827a09fb
modify parallel API note for master
4 years ago
huangxinjing
f354ab22a3
add pipeline shard interface
Add support for no pipeline accugradient
Add delay tag for fusion op
Optimizer the visite order
add mirror for mini step control
Move the group to attributes
Add gradient_shard control for the mini step
Fix code stype
Fix ut description
Add interface
4 years ago
i-robot
b8ec196d6e
!25708 Add mode check
Merge pull request !25708 from huangxinjing/add_mode_check
4 years ago
i-robot
27be612516
!25719 Add Device number check
Merge pull request !25719 from huangxinjing/add_device_number_check
4 years ago
huangxinjing
bd5cfafe2c
Add mode check
Add parallel mode check
4 years ago
huangxinjing
741e3604eb
Add Device Number Check
4 years ago
cononlly
cc83b8064f
Fix code for GPU Pipeline
4 years ago
lichenever
56b31fce42
add_no_elimilate_for_comm_op
4 years ago
i-robot
04c2e928c4
!21622 modify code check for Master
Merge pull request !21622 from lilei/modify_code_check_master
4 years ago
ms_yan
36a8886ca2
Revert "[feat] [assistant] [I3T96T] add new Dataset operator CMUARCTICDataset"
This reverts commit b077aa1cab .
Revert "[feat] [assistant] [I3T96X] add new Dataset operator LibriSpeechDataset"
This reverts commit 4e6f7dc97d .
delete pass_registry_test.cc
comment hiai_nlu_model_multi.pb related line
4 years ago
djc
4e6f7dc97d
[feat] [assistant] [I3T96X] add new Dataset operator LibriSpeechDataset
4 years ago
lilei
5714d68d1d
modify code check for master
4 years ago
yao_yf
a83bf73298
union auto_parallel_context interface dataset_strategy
4 years ago
yao_yf
dc7dc7d3fa
dataset strategy set
4 years ago
Xiaoda Zhang
bb5d4212f7
enable All2All in infering redistribution ops
4 years ago
Xiaoda Zhang
04381273b3
Add the sharding propagation function:
1) users configure sharding strategies for operators;
2) framework will propagate the strategies from configured-ops to
non-configured ops using BFS;
3) the propagation goal is to minimize redistribution communication
cost;
4 years ago
lichenever
cb438ce350
rectification_log
4 years ago
Ziyan
95ac0f6d58
fix optimizer weight shard config
4 years ago
huangxinjing
e79db658e8
Fix codex for python file
4 years ago
Ziyan
2a752f24bf
enable not fully use opt shard
5 years ago
Ziyan
d19d42ee44
modify grad accu and comm fusion api
5 years ago
liujunzhu
6541b96c40
Add communication parallel mode.
5 years ago
yangzhenzhang
7303c3d3b8
add group ckpt
5 years ago
yangzhenzhang
9da3f9bec9
mini step grad accumulation
5 years ago
lizhenyu
7eb49cfce7
[bugfix] server core dump after traning
5 years ago
jinyaohui
e6f9806cfb
add broadcast
5 years ago
lichenever
cfffff2875
add check for allreduce fusion
5 years ago
huangxinjing
2fa6a3b3c2
Fix doc error
5 years ago
mindspore-ci-bot
9bd34a1b29
!6673 Add stage information for ops and strategy
Merge pull request !6673 from huangxinjing/stage_strategy
5 years ago
huangxinjing
4ef439e27b
Add stage information for ops and strategy
5 years ago
lichenever
395d3f0848
add_limit_for_allreduce_fusion
5 years ago
huangxinjing
8ba1503135
Add default value for auto search parallel mode
5 years ago
yao_yf
b70204c080
auto parallel context add notes and func mv
5 years ago
Ziyan
8ea177e614
fix_api_problems
5 years ago
lichenever
f2d3fd34ce
rectification_allreduce_fusion_api
5 years ago
yao_yf
d4cfe55c04
rename mirror_mean to gradients_mean
5 years ago
yao_yf
8f7aa5bd5a
auto parallel context modify
5 years ago
Yi Huaijie
394be43492
raise RuntimeError when set different mode after Initializer created
5 years ago
Yi Huaijie
89a4ebf8a1
parallel mode must be set before create an initializer
5 years ago
yuchaojie
64a1560f1a
add allreduce group for resnet gpu version
5 years ago
yuchaojie
ed9cf2036c
add nccl default allreduce fusion group
5 years ago
Ziyan
39f08eb7dd
enable optimizer parallel
5 years ago
Ziyan
0925e35252
enable optimizer parallel with broadcast
5 years ago