zengzitao
|
8064de7931
|
fix maximum_grad and minimum_grad input_shape not equal to output_shape bug
|
4 years ago |
i-robot
|
07f58b0b46
|
!17626 matmul to mul
Merge pull request !17626 from lingyunli63/matmul_to_mul
|
4 years ago |
lingyunli63
|
4f34e537a0
|
replace matmul/batchmatmul by mul when k is 1
|
4 years ago |
Yang Jiao
|
4d1f8d1d00
|
fix expander bugs
|
4 years ago |
mindspore-ci-bot
|
7e5d68991a
|
!17680 Avoid overflow of in realdiv
From: @wenfangpei
Reviewed-by: @gaoxiong1,@ckey_dou
Signed-off-by: @ckey_dou
|
4 years ago |
wenfangpei
|
fa531a473d
|
avoid overflow in realdiv
|
4 years ago |
Yang Jiao
|
d9a60e0d28
|
fix static warnings
|
5 years ago |
Gaoxiong
|
1d3da1789e
|
Acyclic checking bugfix of split model
|
4 years ago |
mindspore-ci-bot
|
84859aba23
|
!16695 bug fix in lamb
From: @wenfangpei
Reviewed-by: @gaoxiong1,@ckey_dou
Signed-off-by: @ckey_dou
|
4 years ago |
wenfangpei
|
fd84c20a6a
|
bug fix in lamb
|
4 years ago |
looop5
|
68f55e1e93
|
expand conv2d when input format is DefaultFormat but attr format is NHWC
|
4 years ago |
looop5
|
dd81f47271
|
add conv2d expander
use data_format instead of format in op_infer
add pad and unpad inside conv2d expander
fix pad value
add limit in conv2d expander
|
5 years ago |
mindspore-ci-bot
|
f91a365564
|
!16322 [GraphKernel] Enable matmul for gpu
From: @lingyunli63
Reviewed-by:
Signed-off-by:
|
4 years ago |
lingyunli63
|
afc69b16f7
|
enable gpu gk MatMul and insert pad/unpad
|
5 years ago |
zengzitao
|
516487c9cd
|
fix log(inputs), inputs less than 0.0, cause loss inf bug
|
4 years ago |
chenlei_autodiff
|
e0f882b5ce
|
[GraphKernel] fix op infer for elementwise.
|
4 years ago |
dayschan
|
54a0edb837
|
Limit the cluster size in costmodel
|
5 years ago |
mindspore-ci-bot
|
08959ddb74
|
!16224 [GraphKernel]Fix wide & deep ps mode precision error
From: @zengzitao
Reviewed-by: @gaoxiong1,@ckey_dou
Signed-off-by: @ckey_dou
|
5 years ago |
mindspore-ci-bot
|
f247cc2db7
|
!16047 [GraphKernel]adapt batchmatmul for GraphKernel in ascend
From: @wenfangpei
Reviewed-by: @gaoxiong1,@ckey_dou
Signed-off-by: @ckey_dou
|
5 years ago |
mindspore-ci-bot
|
c492d784bf
|
!16104 [GraphKernel] add stitch fusion test case in CI
From: @r1chardf1d0
Reviewed-by: @gaoxiong1,@ckey_dou
Signed-off-by: @ckey_dou
|
5 years ago |
zengzitao
|
15cf743ede
|
rewrite sigmoid_cross_entropy_with_logits expander to avoid overflow
|
5 years ago |
wenfangpei
|
2e2e850b4a
|
adapt batchmatmul for GraphKernel in ascend
|
5 years ago |
r1chardf1d0
|
b1be842d2b
|
add stitch fusion case in ci
|
5 years ago |
hanhuifeng2020
|
bc46d644fe
|
[GraphKernel]Support reshape/elewise/broadcast+transdata fusion
|
5 years ago |
mindspore-ci-bot
|
0887d35b1c
|
!15951 [GraphKernel]add the attribute reduce_output_fuse to enable fuse for the reduce_output on Ascend
From: @hanhuifeng2020
Reviewed-by: @gaoxiong1,@dylangeng
Signed-off-by: @dylangeng
|
5 years ago |
mindspore-ci-bot
|
ed539597c2
|
!15415 [GraphKernel]adapt for logsoftmax in ascend
From: @wenfangpei
Reviewed-by: @gaoxiong1,@ckey_dou,@gaoxiong1,@ckey_dou
Signed-off-by: @ckey_dou,@ckey_dou
|
5 years ago |
wenfangpei
|
4174a7b38f
|
expanders of some fusion ops
|
5 years ago |
wenfangpei
|
db8256e61f
|
adapt for logsoftmax in ascend
|
5 years ago |
mindspore-ci-bot
|
1827697642
|
!15961 Eliminate recursion call in split model
From: @gaoxiong1
Reviewed-by: @ckey_dou,@dylangeng
Signed-off-by: @dylangeng
|
5 years ago |
wenfangpei
|
c41875b318
|
adapt expanders of some ops from gpu to ascend
|
5 years ago |
Gaoxiong
|
4bc67f38de
|
eliminate recursion call
|
5 years ago |
hanhuifeng2020
|
425d401e85
|
[GraphKernel]add the attr reduce_output_fuse to enable fuse for reduce_output on Ascend
|
5 years ago |
mindspore-ci-bot
|
52e7f51970
|
!15741 [GraphKernel] batchnorm expander supports when first input is float16
From: @looop5
Reviewed-by: @gaoxiong1,@dylangeng
Signed-off-by: @dylangeng
|
5 years ago |
mindspore-ci-bot
|
ff75ce8ac4
|
!15097 [graph kernel] clean code for expanders.
From: @chenlei_autodiff
Reviewed-by: @gaoxiong1,@dylangeng
Signed-off-by: @dylangeng
|
5 years ago |
looop5
|
24f441ba33
|
batchnorm expander supports when first input is float16
|
5 years ago |
mindspore-ci-bot
|
d6f58cb765
|
!15658 Reduce recursion overhead of split model
From: @gaoxiong1
Reviewed-by: @anyrenwei,@dylangeng
Signed-off-by: @dylangeng
|
5 years ago |
chenlei_autodiff
|
fd227bb448
|
[graph kernel] clean code for expanders.
|
5 years ago |
mindspore-ci-bot
|
168c64b60d
|
!15648 [GraphKernel] negative axis in Squeeze expander.
From: @chenlei_autodiff
Reviewed-by: @gaoxiong1,@anyrenwei
Signed-off-by: @anyrenwei
|
5 years ago |
Gaoxiong
|
71002ed19d
|
reduce recursion overhead of split model
|
5 years ago |
zengzitao
|
8dcff8d83c
|
refactor tile op and in expander open on gpu
|
5 years ago |
chenlei_autodiff
|
b419f60b0d
|
[GraphKernel] negative axis in Squeeze expander.
|
5 years ago |
r1chardf1d0
|
5c5d125b1d
|
optimize stitch fusion strategy
|
5 years ago |
hanhuifeng2020
|
25505642ce
|
enable GraphKernel for TransData
|
5 years ago |
lingyunli63
|
c48c2430f0
|
fuse matmul and elementwise in graphkernel
|
5 years ago |
mindspore-ci-bot
|
cd002cb7f7
|
!14893 enable stitch fusion on bert
From: @r1chardf1d0
Reviewed-by: @gaoxiong1,@ckey_dou
Signed-off-by: @ckey_dou
|
5 years ago |
mindspore-ci-bot
|
5b4685c5ea
|
!14604 [GraphKernel] add some expander ops
From: @chenlei_autodiff
Reviewed-by:
Signed-off-by:
|
5 years ago |
r1chardf1d0
|
3b32995936
|
enable stitch fusion on bert
|
5 years ago |
chenlei_autodiff
|
13fbfca6b9
|
[graph kernel] add expander ops.
|
5 years ago |
wenfangpei
|
b9715db358
|
bugfix in expanders of layernorm
|
5 years ago |
wenfangpei
|
66d28af79e
|
adapt for layernorm in ascend
|
5 years ago |