kswang
3247c00555
optimize heter memcpy
4 years ago
i-robot
4861711676
!18107 dump and offline debug fixes
Merge pull request !18107 from john_tzanakakis/jt_bug_fixes
4 years ago
i-robot
eaac4f47b3
!18058 Update graph input shape
Merge pull request !18058 from chenweifeng/graph-dynamic
4 years ago
John Tzanakakis
ac1847ffac
fix iter 0 and iter 1 being dumped in dir 0, make op_debug_mode optional for sync mode, read input files for offline debugger
4 years ago
wilfChen
27ed501716
graph input dynamic
4 years ago
wilfChen
2e6afc07ac
graph input dynamic
4 years ago
chujinjin
90feb6a6d2
fix bcewithlogitsloss op error in pynative
4 years ago
zhoufeng
0787efad03
move conv transpose pass from common to unify mindir, for pynative
Signed-off-by: zhoufeng <zhoufeng54@huawei.com>
4 years ago
lvchangquan
0b09fdf94c
fix an allreduce bug with two streams sync problem
5 years ago
limingqi107
22972a89a7
support the output address of graph reapply
5 years ago
kswang
2a48b2ecb8
reconstruct session code
5 years ago
tanghuikang
c8a14ba016
Clean code
5 years ago
kswang
97a97e02db
extract load input
5 years ago
caifubi
171b468bb3
PyNative AllReduce Bucket
5 years ago
wilfChen
a911b9ef9e
mindspore serving support gpu backend
5 years ago
jjfeing
1984cf8e20
unify mindir
5 years ago
Harshvardhan Gupta
dd0084c52b
improve perf, keep consistent tensor state, fix recheck, check weights at step end
5 years ago
mindspore-ci-bot
d38f8205dc
!8987 support getnext in pynative mode
From: @chujinjin
Reviewed-by:
Signed-off-by:
5 years ago
lvliang
8984cc9c03
pynative-support-dynamic-op-run-in-gpu
5 years ago
chujinjin
af031410bb
support getnext in pynative
5 years ago
caifubi
d44dd4f786
Move BuildOp into RunOp
5 years ago
HulkTang
c36b477568
Run ops one by one in pynative bp graph
5 years ago
Yi Huaijie
d7faa77b5e
support int64 shape
5 years ago
mindspore-ci-bot
c6246d7a7e
!7908 add reduce precision in pynative mode
Merge pull request !7908 from chujinjin/add_reduce_precision_in_pynative_mode
5 years ago
chujinjin
9197d9f2ee
add reduce precision in pynative mode
5 years ago
ZPaC
5059d8c3f9
Set gpu device id for multiple threads
5 years ago
kswang
11989b5e30
enable async run
5 years ago
John Tzanakakis
0e0d7eda19
code refactor
5 years ago
dayschan
37a48f6aac
GraphKernel supports GPU
1. Update akg submodule
2. Refactor akg_kernel_build, akg_ascend_kernel_build, akg_gpu_kernel_build
3. Add akg_kernel_json_decoder to support converting kernel_json to AnfNode.
4. Add GraphKernel Cost Model. (mindspore/_extends/graph_kernel)
5. Add some GraphKernel passes to GpuSession, move these passes to backend/optimizer/graph_kernel.
6. Add global id for ir files.
7. Fix bug in ConstInputToAttr.
5 years ago
lizhenyu
c3d6918649
add kernel select after optimize pass
5 years ago
limingqi107
341200ab97
gpu kernel_info_setter code review
5 years ago
lizhenyu
6fdd52080d
add mode black list checker
5 years ago
kswang
756bb6d53f
async run graph
5 years ago
kpy
570da089a8
set output value for dynamic graph
5 years ago
John Tzanakakis
b3c0eb61d5
GPU debugger - milestone 1 and GPU dump
Additonal Authors: Adel Shafiei, Harshvardhan Gupta
5 years ago
liubuyu
76dc80e7b7
Unified code style
5 years ago
liubuyu
43c79eb853
mindspore path adjust
5 years ago