mindspore-ci-bot
7d5db0e99b
!2840 Update fusion condition of reshape and transpose
Merge pull request !2840 from YuJianfeng/master
5 years ago
yujianfeng
3df6845803
Fix fusion condition of reshape and transpose
5 years ago
mindspore-ci-bot
17e1786890
!2820 [bugfix]: communication operator's output should not be swapped
Merge pull request !2820 from zyli2020/refine_mem_swap
5 years ago
mindspore-ci-bot
32405f9ab3
!2817 add internal output
Merge pull request !2817 from kisnwang/optimize-sub-graph-memcpy
5 years ago
lizhenyu
1b4a7cdeb7
fix mem swap bug
5 years ago
kswang
e9067b4a10
add internal output
5 years ago
mindspore-ci-bot
065b8cbdf1
!2700 use same batchnorm operation in PyNative mode and Graph mode
Merge pull request !2700 from chujinjin/change_pynative_batchnorm_same_as_graph_mode
5 years ago
chujinjin
c880774942
change pynative batchnorm same as graph mode
5 years ago
mindspore-ci-bot
2291213b8b
!2574 Clean codex
Merge pull request !2574 from TronZhang/codex_0624
5 years ago
laiyongqiang
33d1427a14
optimize is all nop node detect in mem reuse
5 years ago
tronzhang
673afad0ba
fixed codex's warnnings and errors
5 years ago
Wei Luning
ef13a4b6fb
adjust cse code when op has side effect.
5 years ago
mindspore-ci-bot
4c6bff75af
!1393 Gpu Support AdamWeightDecay optimizer fusion
Merge pull request !1393 from chenweifeng/adam_weight_decay
5 years ago
He Wei
43e0967024
Decouple ir::Tensor class from python
5 years ago
mindspore-ci-bot
c61d254c1f
!2408 fix bug to remove reshape when reshape is depend's input
Merge pull request !2408 from laiyongqiang/remove_reshape
5 years ago
wilfChen
034d2ea2aa
Gpu Adam Fusion
5 years ago
laiyongqiang
eb37669e3b
fix bug to remove reshape when reshape is depend's input
5 years ago
jjfeing
d535f3a289
fix atomic clean
5 years ago
mindspore-ci-bot
8b5166e569
!2393 fix bug of hccl kernel info and change cast's kernel info
Merge pull request !2393 from lianliguang/fix-bug-of-merge-cast-to-op-and-gene-hccl-kernel-info
5 years ago
gong chen
a6dfa281ea
Init GraphKernel.
- It provides a unified style to express graph and kernel for user.
- It provides a unified IR to represent graph and kernel for developer.
- It breaks the boundary between graph and kernel.
- It provides more opportunities to do compile optimization.
5 years ago
mindspore-ci-bot
a3e7b30457
!2384 Add split fission pass
Merge pull request !2384 from YuJianfeng/split
5 years ago
mindspore-ci-bot
24be3f82ad
!2336 Fix bug of not insert memcpy when hccl_op has more than one input
Merge pull request !2336 from huanghui/insert-memcpy-async-pass
5 years ago
WilliamLian
5f9d2759ee
fix bug of hccl kernel info and change cast's kernel info
5 years ago
yujianfeng
7ad877a948
Add Split fission pass
5 years ago
WilliamLian
ee79023e86
clean pclint warning
5 years ago
huanghui
c66fe00049
fix bug of not insert memcpy when hccl_op has more than one input
5 years ago
mindspore-ci-bot
230963d08e
!2332 Fix ConfusionSoftmaxGrad fusion pass work if the ReduceSum's attr of keep_dims set with False
Merge pull request !2332 from huanghui/fix-ConfusionSoftmaxGrad-fusion-pass
5 years ago
mindspore-ci-bot
a663f2066c
!2285 [Code Review] code review fix
Merge pull request !2285 from jjfeing/master
5 years ago
mindspore-ci-bot
2ad211df97
!2293 fix bug of transdata
Merge pull request !2293 from lianliguang/add-broadcast-and-reduce-kernel-select-rules
5 years ago
huanghui
c7bda5363d
fix ConfusionSoftmaxGrad fusion pass work if the ReduceSum's attr of keep_dims set False
5 years ago
huanghui
8463731bcc
make those AdamXX and LambXX fusion pass not work for unexpect data type
5 years ago
mindspore-ci-bot
2e002ab64c
!2292 gpu fix all nop node graph execute
Merge pull request !2292 from limingqi107/master
5 years ago
WilliamLian
e3a26c2229
fix transdata insert cast
5 years ago
jjfeing
c26274f324
fix code review bug
5 years ago
limingqi107
0f4397cece
fix all nop node graph execute
5 years ago
WilliamLian
d3ea1e4298
fix code review
5 years ago
mindspore-ci-bot
294449697a
!2179 Add sort by index for each group of AllReduce
Merge pull request !2179 from YuJianfeng/all_reduce_index
5 years ago
yujianfeng
f15cb6b7c9
Add sort by index for each group of AllReduce
5 years ago
mindspore-ci-bot
04a23d138f
!2164 add more needtrans format for transdata
Merge pull request !2164 from lianliguang/master
5 years ago
mindspore-ci-bot
4642df207a
!2210 gpu optimize the max device memory config
Merge pull request !2210 from limingqi107/max_device_memory_optimize
5 years ago
WilliamLian
5d25bf7ca2
add more transform format insert transdata
5 years ago
limingqi107
55b3557c0d
gpu optimize the max device memory config
5 years ago
huanghui
1243022f01
fix the back outputs of BNTrainingUpdateV3 may cover the front whitch is empty
5 years ago
mindspore-ci-bot
058bcec347
!2173 gpu memreuse supports summary node
Merge pull request !2173 from limingqi107/gpu_memreuse_support_summary_node
5 years ago
mindspore-ci-bot
8903b50042
!2153 Add transdata for output node in pynative hook mode
Merge pull request !2153 from JoyLvliang/pynative-insert-transdata-for-hook-mode
5 years ago
limingqi107
20083679a0
gpu memreuse supports summary node
5 years ago
lvliang
075da9a4b1
pynative-insert-transdata-for-hook-mode
5 years ago
limingqi107
b83f90a8d8
gpu optimize Nop node
5 years ago
mindspore-ci-bot
5da171a735
!2076 fix summary nodes memory reuse refcount
Merge pull request !2076 from laiyongqiang/summary_refcnt
5 years ago
mindspore-ci-bot
87371be65d
!2127 Fix exception when FusedBatchNorm's first input 's shape dims not equal 4
Merge pull request !2127 from huanghui/fix-fusebatchnorm-split
5 years ago