hanhuifeng2020
bc46d644fe
[GraphKernel]Support reshape/elewise/broadcast+transdata fusion
4 years ago
mindspore-ci-bot
0887d35b1c
!15951 [GraphKernel]add the attribute reduce_output_fuse to enable fuse for the reduce_output on Ascend
From: @hanhuifeng2020
Reviewed-by: @gaoxiong1,@dylangeng
Signed-off-by: @dylangeng
4 years ago
mindspore-ci-bot
ed539597c2
!15415 [GraphKernel]adapt for logsoftmax in ascend
From: @wenfangpei
Reviewed-by: @gaoxiong1,@ckey_dou,@gaoxiong1,@ckey_dou
Signed-off-by: @ckey_dou,@ckey_dou
4 years ago
wenfangpei
4174a7b38f
expanders of some fusion ops
4 years ago
wenfangpei
db8256e61f
adapt for logsoftmax in ascend
4 years ago
mindspore-ci-bot
1827697642
!15961 Eliminate recursion call in split model
From: @gaoxiong1
Reviewed-by: @ckey_dou,@dylangeng
Signed-off-by: @dylangeng
4 years ago
wenfangpei
c41875b318
adapt expanders of some ops from gpu to ascend
4 years ago
Gaoxiong
4bc67f38de
eliminate recursion call
4 years ago
hanhuifeng2020
425d401e85
[GraphKernel]add the attr reduce_output_fuse to enable fuse for reduce_output on Ascend
4 years ago
mindspore-ci-bot
52e7f51970
!15741 [GraphKernel] batchnorm expander supports when first input is float16
From: @looop5
Reviewed-by: @gaoxiong1,@dylangeng
Signed-off-by: @dylangeng
4 years ago
mindspore-ci-bot
ff75ce8ac4
!15097 [graph kernel] clean code for expanders.
From: @chenlei_autodiff
Reviewed-by: @gaoxiong1,@dylangeng
Signed-off-by: @dylangeng
4 years ago
looop5
24f441ba33
batchnorm expander supports when first input is float16
4 years ago
mindspore-ci-bot
d6f58cb765
!15658 Reduce recursion overhead of split model
From: @gaoxiong1
Reviewed-by: @anyrenwei,@dylangeng
Signed-off-by: @dylangeng
4 years ago
chenlei_autodiff
fd227bb448
[graph kernel] clean code for expanders.
4 years ago
mindspore-ci-bot
168c64b60d
!15648 [GraphKernel] negative axis in Squeeze expander.
From: @chenlei_autodiff
Reviewed-by: @gaoxiong1,@anyrenwei
Signed-off-by: @anyrenwei
4 years ago
Gaoxiong
71002ed19d
reduce recursion overhead of split model
4 years ago
zengzitao
8dcff8d83c
refactor tile op and in expander open on gpu
4 years ago
chenlei_autodiff
b419f60b0d
[GraphKernel] negative axis in Squeeze expander.
4 years ago
r1chardf1d0
5c5d125b1d
optimize stitch fusion strategy
5 years ago
hanhuifeng2020
25505642ce
enable GraphKernel for TransData
5 years ago
lingyunli63
c48c2430f0
fuse matmul and elementwise in graphkernel
5 years ago
mindspore-ci-bot
cd002cb7f7
!14893 enable stitch fusion on bert
From: @r1chardf1d0
Reviewed-by: @gaoxiong1,@ckey_dou
Signed-off-by: @ckey_dou
4 years ago
mindspore-ci-bot
5b4685c5ea
!14604 [GraphKernel] add some expander ops
From: @chenlei_autodiff
Reviewed-by:
Signed-off-by:
4 years ago
r1chardf1d0
3b32995936
enable stitch fusion on bert
5 years ago
chenlei_autodiff
13fbfca6b9
[graph kernel] add expander ops.
5 years ago
wenfangpei
b9715db358
bugfix in expanders of layernorm
5 years ago
wenfangpei
66d28af79e
adapt for layernorm in ascend
5 years ago
mindspore-ci-bot
b5bc938deb
!12914 [GraphKernel]expander lamb_apply_weight_assign
From: @wenfangpei
Reviewed-by: @anyrenwei,@gaoxiong1,@gaoxiong1
Signed-off-by: @anyrenwei
5 years ago
wenfangpei
83399c1b8d
adapt for layermorm C++ code
5 years ago
wenfangpei
a4ad6066b1
expander lamb_apply_weight_assign
5 years ago
mindspore-ci-bot
ddf75da542
!14085 [GraphKernel] add some expander ops
From: @chenlei_autodiff
Reviewed-by:
Signed-off-by:
5 years ago
mindspore-ci-bot
75fdaaa6aa
!14304 [GraphKernel] Dump GraphKernel split info as text; dump akg kernel launch fail message
From: @dayschan
Reviewed-by: @gaoxiong1,@gaoxiong1,@anyrenwei
Signed-off-by: @anyrenwei
5 years ago
chenlei_autodiff
f4289d40f3
add graph kernel expander ops.
5 years ago
dayschan
3c6c30024c
dump graph_kernel_split info
5 years ago
tronzhang
87bf1ec80f
delete mark_interface_fusion and tensor reuse frontend pass for graph kernel
5 years ago
lingyunli63
4b966ed40d
support matmul on D
5 years ago
huangbingjian
72ae1799f3
remove control_depend from py file
5 years ago
mindspore-ci-bot
5b95409022
!13512 add some expander ops
From: @zengzitao
Reviewed-by:
Signed-off-by:
5 years ago
mindspore-ci-bot
2fadad0875
!13121 expander lamb_apply_optimizer_assign
From: @wenfangpei
Reviewed-by:
Signed-off-by:
5 years ago
wenfangpei
043a558ae2
expander lamb_apply_optimizer_assign
5 years ago
zengzitao
d0a656f3cd
add some expander ops
5 years ago
dayschan
a2967330ea
Normalize the Reduce nodes' axis in GraphKernel
changed all reduce axis to non-negative value
5 years ago
tronzhang
7252ffb66b
pipeline optimization for parallel fusion
5 years ago
mindspore-ci-bot
d285692217
!12852 Change maketuple in coreops
From: @liangzhibo
Reviewed-by:
Signed-off-by:
5 years ago
l00591931
680324f225
Change make tuple in core.ops
5 years ago
dayschan
454500309c
add OpInfer for op Select
5 years ago
dayschan
7beca18f3c
Refactor GraphKernelExpander (3rd submission)
Rewrited the OpInfer as a class with functions "infer_shape","infer_type"
and "infer_format". Used op name to find the subclass that have implemented
these functions. Two common class "_Elemwise" and "_Reduce" was provided.
Op BiasAddGrad supported "FractalNZ" format.
5 years ago
dayschan
9d572f3963
Refactor GraphKernelExpander (2nd submission)
Changed the expander to class, and used the class name to identify the operator.
Moved the original op logic into the `_expand` function; added a `_check` function to check op inputs.
Use decorator to register the whitelist formats that operator supports,
and the decorator will change the `_check` function to check the formats.
5 years ago
dayschan
e0e6c39eae
Refactor GraphKernelExpander (1st submission)
Decoupled from the front-end interfaces.
1. Removed the call to "Renormalize".
Completed the infer-format in model_builder.
Only used the device shape and device format to
infer an abstract shape without considering padding.
2. Removed the call to python's Primitive interfaces.
The "Renormalize" relies on the PrimitivePy, so they can be
removed together. After that, the functions "ConstAttrToInput",
"DeleteAttrInInput" and related can be removed.
3. Reuse the AkgKernelJsonGenerator in GraphKernelExpander.
1) set the attribute "extract_opinfo_from_anf" to true, so that
the generator can handle the basic operator with anf info.
2) added a function "extract_expand_info" in expander.py
to convert the json into a more friendly format. The attrs
was converted to a dict instead of a list.
4. Scalars only support DefaultFormat.
Removed the argument "format" from graph_builder.value
5. Moved the expander op list from graph_kernel_helper.cc to graph_kernel_expander.cc
5 years ago
jinyaohui
30a27b2adb
modify Gelu、FastGelu to GeLU and FastGeLU
5 years ago