Zichun Ye
|
22172f18bc
|
update graph kernel support for argmax/argmin
fix pylint problem
fix conflict
fix op list
fix check warning
fix code based on review comments
update akg commit
fix check warning
|
4 years ago |
tronzhang
|
94dda0c7c7
|
speed up bert profermance in ascend for graph kernel
|
4 years ago |
i-robot
|
ab78bcf16f
|
!19081 add ops IsNan
Merge pull request !19081 from lingyunli63/add_ops_IsNan
|
4 years ago |
Yang Jiao
|
582669e41c
|
add expander checkings
|
4 years ago |
lingyunli63
|
50a66ae476
|
isnan isfinite isinf squaresumall identity oneslik
|
4 years ago |
i-robot
|
e7b8cf2961
|
!20508 consider case that graph has no tensor input
Merge pull request !20508 from lingyunli63/fix_no_tensor_input
|
4 years ago |
yanglf1121
|
c30b1e6d06
|
add graph kernel div, floordiv, mod, floormod, floor
|
4 years ago |
lingyunli63
|
32571f5a0c
|
consider no tensor input in graph
|
4 years ago |
wangrao124
|
7cddde47b0
|
!215 add ops: LogicalNot, LogicalAnd, LogicalOr, NotEqual, EqualCount, Asinh, Acosh
* add ops: LogicalNot, LogicalAnd, LogicalOr, NotEqual, EqualCount, Asinh, Acosh
|
4 years ago |
huangmengxi
|
1bd448ef2a
|
add atan, atan2, expm1
|
4 years ago |
chenlei_autodiff
|
7d55cef106
|
[GraphKernel] add sponge ops.
|
4 years ago |
zengzitao
|
aa019a639f
|
support complex in gpu about graph_kernel
|
4 years ago |
lingyunli63
|
3679e99fe8
|
fix recompute_fuse with reshape user
|
4 years ago |
lingyunli63
|
a995bea507
|
recompute_fuse
|
4 years ago |
zengzitao
|
8064de7931
|
fix maximum_grad and minimum_grad input_shape not equal to output_shape bug
|
4 years ago |
i-robot
|
07f58b0b46
|
!17626 matmul to mul
Merge pull request !17626 from lingyunli63/matmul_to_mul
|
4 years ago |
lingyunli63
|
4f34e537a0
|
replace matmul/batchmatmul by mul when k is 1
|
4 years ago |
Yang Jiao
|
d9a60e0d28
|
fix static warnings
|
4 years ago |
Gaoxiong
|
1d3da1789e
|
Acyclic checking bugfix of split model
|
4 years ago |
looop5
|
68f55e1e93
|
expand conv2d when input format is DefaultFormat but attr format is NHWC
|
4 years ago |
looop5
|
dd81f47271
|
add conv2d expander
use data_format instead of format in op_infer
add pad and unpad inside conv2d expander
fix pad value
add limit in conv2d expander
|
4 years ago |
mindspore-ci-bot
|
f91a365564
|
!16322 [GraphKernel] Enable matmul for gpu
From: @lingyunli63
Reviewed-by:
Signed-off-by:
|
4 years ago |
lingyunli63
|
afc69b16f7
|
enable gpu gk MatMul and insert pad/unpad
|
5 years ago |
chenlei_autodiff
|
e0f882b5ce
|
[GraphKernel] fix op infer for elementwise.
|
4 years ago |
dayschan
|
54a0edb837
|
Limit the cluster size in costmodel
|
4 years ago |
mindspore-ci-bot
|
f247cc2db7
|
!16047 [GraphKernel]adapt batchmatmul for GraphKernel in ascend
From: @wenfangpei
Reviewed-by: @gaoxiong1,@ckey_dou
Signed-off-by: @ckey_dou
|
4 years ago |
mindspore-ci-bot
|
c492d784bf
|
!16104 [GraphKernel] add stitch fusion test case in CI
From: @r1chardf1d0
Reviewed-by: @gaoxiong1,@ckey_dou
Signed-off-by: @ckey_dou
|
4 years ago |
wenfangpei
|
2e2e850b4a
|
adapt batchmatmul for GraphKernel in ascend
|
4 years ago |
r1chardf1d0
|
b1be842d2b
|
add stitch fusion case in ci
|
4 years ago |
hanhuifeng2020
|
bc46d644fe
|
[GraphKernel]Support reshape/elewise/broadcast+transdata fusion
|
4 years ago |
mindspore-ci-bot
|
0887d35b1c
|
!15951 [GraphKernel]add the attribute reduce_output_fuse to enable fuse for the reduce_output on Ascend
From: @hanhuifeng2020
Reviewed-by: @gaoxiong1,@dylangeng
Signed-off-by: @dylangeng
|
4 years ago |
mindspore-ci-bot
|
1827697642
|
!15961 Eliminate recursion call in split model
From: @gaoxiong1
Reviewed-by: @ckey_dou,@dylangeng
Signed-off-by: @dylangeng
|
4 years ago |
wenfangpei
|
c41875b318
|
adapt expanders of some ops from gpu to ascend
|
4 years ago |
Gaoxiong
|
4bc67f38de
|
eliminate recursion call
|
4 years ago |
hanhuifeng2020
|
425d401e85
|
[GraphKernel]add the attr reduce_output_fuse to enable fuse for reduce_output on Ascend
|
4 years ago |
mindspore-ci-bot
|
d6f58cb765
|
!15658 Reduce recursion overhead of split model
From: @gaoxiong1
Reviewed-by: @anyrenwei,@dylangeng
Signed-off-by: @dylangeng
|
4 years ago |
Gaoxiong
|
71002ed19d
|
reduce recursion overhead of split model
|
4 years ago |
zengzitao
|
8dcff8d83c
|
refactor tile op and in expander open on gpu
|
4 years ago |
r1chardf1d0
|
5c5d125b1d
|
optimize stitch fusion strategy
|
5 years ago |
hanhuifeng2020
|
25505642ce
|
enable GraphKernel for TransData
|
5 years ago |
lingyunli63
|
c48c2430f0
|
fuse matmul and elementwise in graphkernel
|
5 years ago |
mindspore-ci-bot
|
cd002cb7f7
|
!14893 enable stitch fusion on bert
From: @r1chardf1d0
Reviewed-by: @gaoxiong1,@ckey_dou
Signed-off-by: @ckey_dou
|
5 years ago |
r1chardf1d0
|
3b32995936
|
enable stitch fusion on bert
|
5 years ago |
chenlei_autodiff
|
13fbfca6b9
|
[graph kernel] add expander ops.
|
5 years ago |
wenfangpei
|
83399c1b8d
|
adapt for layermorm C++ code
|
5 years ago |
mindspore-ci-bot
|
ddf75da542
|
!14085 [GraphKernel] add some expander ops
From: @chenlei_autodiff
Reviewed-by:
Signed-off-by:
|
5 years ago |
chenlei_autodiff
|
f4289d40f3
|
add graph kernel expander ops.
|
5 years ago |
tronzhang
|
87bf1ec80f
|
delete mark_interface_fusion and tensor reuse frontend pass for graph kernel
|
5 years ago |
lingyunli63
|
4b966ed40d
|
support matmul on D
|
5 years ago |
huangbingjian
|
72ae1799f3
|
remove control_depend from py file
|
5 years ago |