wangjun
46612fabfb
add st for shard
4 years ago
i-robot
ad9757ccf0
!30661 [Auto parallel] [MoE] Fix an error of configuring MoE parallel
Merge pull request !30661 from Xiaoda/124-moe-changes
4 years ago
Xiaoda Zhang
81e5abe580
fix an error of configuring parallel
4 years ago
huangxinjing
896daee845
[AutoParallel]Fix insert error for the mirror
4 years ago
yangzhenzhang
43e6e16da3
check platform for resizebilinear
4 years ago
i-robot
0341d96dd6
!30469 add shard function to support part of the graph executed in auto_parallel under pynative mode
Merge pull request !30469 from wangjun/0223_pp
4 years ago
i-robot
cfe0f76d2b
!30491 ut for allgather fusion
Merge pull request !30491 from jiahongQian/master
4 years ago
wangjun
24d448239c
add pynative_parallel
4 years ago
i-robot
981eae461a
!30118 自动优化器并行特性
Merge pull request !30118 from zhuyuxiao/I4S85V
4 years ago
jiahongQian
25f57505bf
ut for allgather fusion
4 years ago
i-robot
bbcfbce9e0
!29997 [Auto parallel] [MoE] Support data_parallel + expert_parallel
Merge pull request !29997 from Xiaoda/124-moe-changes
4 years ago
zhuyuxiao
d0e0e305d3
good
4 years ago
i-robot
f2130e7434
!30483 [AutoParallel]Pipeline Automatic detection Opt
Merge pull request !30483 from lichen/pipeline_opt_detection
4 years ago
yao_yf
e21f878e14
adasum ut fix
4 years ago
Xiaoda Zhang
b714451937
implementing expert_parallel+data_parallel in MoE:
1) extending _Linear's input as 4-dimension tensor: [outer_batch, expert_dim, -1, hidden], and _Liner's BatchMatMul becomes BatchMatMul(4_dim_tensor, 3_dim_tensor);
2) configuring the _Linear's BatchMatMul sharding strategy as [[dp, ep, 1, 1], [ep, 1, mp]];
3) introducing a new parameter 'expert_parallel' in TransformerOpParallelConfig, creating a new class MoEParallelConfig to include 'data_parallel', 'model_parallel' and 'expert_parallel';
4) changing parallel config for FeedForward, TransformerEncoderLayer, TransformerDecoderLayer.
4 years ago
wangshengnan12@huawei.com
acbefd80ea
pipeline_opt_detection
4 years ago
i-robot
81260a2319
!30466 takedown test_auto_parallel_adasum.py to ensure stability, again
Merge pull request !30466 from yanghaoran/master
4 years ago
i-robot
14393503b7
!30431 allreduce allgather fusion
Merge pull request !30431 from jiahongQian/master
4 years ago
yanghaoran
71d6b7d506
takedown test_auto_parallel_adasum.py to ensure stability, again
4 years ago
i-robot
2e8eac8341
!30367 auto_parallel_adasum_support_data_parallel
Merge pull request !30367 from yao_yf/auto_parallel_adasum_support_data_parallel
4 years ago
jiahongQian
8a2151d8bb
allgather reducescatter fusion
4 years ago
i-robot
5bee7156b9
!30369 add_virtualdataset_ut
Merge pull request !30369 from lilei/add_virtualdataset_ut
4 years ago
yao_yf
19236b1a70
auto parallel adasum support data parallel and hybrid parallel
4 years ago
huangxinjing
092ba035e3
Add global norm parallel support
4 years ago
yanghaoran
bfe139b662
takedonw test_auto_parallel_adasum.py to ensure gate stability
4 years ago
lilei
bc62e24d94
add_virtualdataset_ut
4 years ago
i-robot
94c8c6355c
!30294 auto_parallel_adasum_checks_and_ut.
Merge pull request !30294 from yao_yf/auto_parallel_adasum_checks_and_ut
4 years ago
huangxinjing
5e325ac336
[AUTO_PARALLEL]Fix insert nodes error
4 years ago
i-robot
7386612515
!29820 moe_topk routing
Merge pull request !29820 from wangshengnan123/moe_topk_routing
4 years ago
yao_yf
4b79d4c425
auto parallel adasum uts and checks
4 years ago
wangshengnan123
7322426648
top_k routing
4 years ago
i-robot
48d4f34576
!30167 Add UT case for 'Convolution+Transformer' structure
Merge pull request !30167 from Bert0108/ut_conformer
4 years ago
Bert0108
25a9c73a08
add ut case for monitoring the conformer structure
4 years ago
liuluobin
b0b79ef8fb
Add backend check for RandomChoiceWithMask
4 years ago
i-robot
d5100190c9
!29612 Impementatin of CropAndResizeInfo and ROIAlignInfo.
Merge pull request !29612 from liuluobin/crop_resize_master
4 years ago
liuluobin
80fd640819
Implementation of ROIAlignInfo and CropAndResizeInfo
4 years ago
Margaret_wangrui
462c38813d
[ME] Add parameter name check.
4 years ago
yangzhenzhang
4061c11e18
support pad mode for conv2d transpose
4 years ago
i-robot
a55b1b5e05
!29453 support pad mode for conv2d parallel operator
Merge pull request !29453 from yangzhenzhang/support-pad-mode-for-conv2d
4 years ago
i-robot
ea391f6eb1
!29505 Fix AllGather Cast when the parameters are shared
Merge pull request !29505 from huangxinjing/fx_allgather_cast
4 years ago
i-robot
b0b3d7ede7
!29421 Implementation of BoundingBoxEncodeInfo, IOUInfo and RandomChoickWithMaskInfo
Merge pull request !29421 from liuluobin/parallel_ops_master
4 years ago
i-robot
55ba926a04
!29087 optimize sharding propagation&add ut on pangu
Merge pull request !29087 from bichaoyang/master
4 years ago
huangxinjing
34101d748d
Add cast ut and fix juge condition
4 years ago
yangzhenzhang
6df5e9aa5d
support pad mode for conv2d
4 years ago
liuluobin
fb79c508f2
Implementation of BoundingBoxEncodeInfo, IOUInfo, RandomChoiceWithMaskInfo.
4 years ago
b00518648
2ff3425c76
1.consider comm cost inside a op when select a strategy;2.deal with the ops that share same param;3.add ut to fix the perfermence at pangu_alpha
4 years ago
yangzhenzhang
4d0b1a47ee
support config group for batchnorm
4 years ago
yangzhenzhang
000c5b5394
support dilation for conv2d
4 years ago
i-robot
7bb5819889
!29265 fix resize_bilinear infer
Merge pull request !29265 from jiangzhenguang/resize_bilinear
4 years ago
jiangzhenguang
abddc993ea
fix resize_bilinear
4 years ago