i-robot
50a37e4e8e
!26102 Ascend Device Context for MindRT
Merge pull request !26102 from hwjiaorui/ascend_context
4 years ago
i-robot
30c0bc5a96
!26169 Add EnableAscendMIndRT Environment Variable
Merge pull request !26169 from hwjiaorui/enable_ascend_mindrt
4 years ago
ZPaC
87057fdc27
Dynamic load ompi and nccl
4 years ago
hwjiaorui
94cc26bf6b
add enable ascend mindrt environment
4 years ago
hwjiaorui
7d4d7c50da
Ascend device context
4 years ago
hwjiaorui
653cf2622b
fix compile
4 years ago
i-robot
1bfedfcb7f
!26007 Ascend Device Context for MindRT
Merge pull request !26007 from hwjiaorui/ascend_context
4 years ago
hwjiaorui
d6b2a34a69
MindRT Ascend Devcie Context
4 years ago
i-robot
11bec4d85e
!25995 Add nvidia collective lib implementation.
Merge pull request !25995 from ZPaC/dir-of-distributed
4 years ago
ZPaC
9e18bad126
Add nvidia collective lib implementation.
4 years ago
i-robot
4c76c0269e
!25996 Fix ompi dependency
Merge pull request !25996 from ZPaC/fix-mpi-dep
4 years ago
ZPaC
924c67c9b5
Fix ompi dependency
4 years ago
limingqi107
46d4a900c0
unified runtime support the auto monad in the subgraphs connecting scene
4 years ago
ZPaC
0b732431a1
Add MPI implementation.
4 years ago
i-robot
ca708af5fc
!25716 Add collective adapter
Merge pull request !25716 from ZPaC/dir-of-distributed
4 years ago
ZPaC
e5789757d6
Add collective adapter
4 years ago
LaiYongqiang
07e5ed9f16
ascend device context
4 years ago
limingqi107
2a5d9f964b
unified runtime support the sub graph sink
4 years ago
ZPaC
4c1ef4cef6
Fix ps cache broadcast error.
4 years ago
i-robot
b271aa7a25
!24969 device address add the key of device
Merge pull request !24969 from limingqi107/new_actor_runtime
4 years ago
maning202007
8ee33007b1
add const to the data_ptr definition
Fix pclint for debugger
fix bugs
fix duplicated code issue
Fix the depth of method over 4 issue
fix comments
fix CI errors
remove redundancy arguments
4 years ago
limingqi107
be100476d6
device address add the key of device
4 years ago
i-robot
4268cf5de0
!25026 support dynamic shape for cpu kernel of mae net
Merge pull request !25026 from fangzehua/dynamic_mae_1014
4 years ago
fangzehua
53009276d1
add cpu dynamic memory
4 years ago
i-robot
093aa3b7a0
!24867 Fix AllGather Recompute Cycled Graph for GPU
Merge pull request !24867 from Cononlly/master
4 years ago
cononlly
1b119ced2b
Add a pass for allgather_recompute
4 years ago
limingqi107
19330b6f82
unified runtime support graph sink output update
4 years ago
i-robot
c5585a9451
!24467 Add GraphKernelFlags into namespace mindspore::graphkernel
Merge pull request !24467 from DeshiChen/0918_namespace
4 years ago
dayschan
7502345c8f
Add GraphKernelFlags into namespace mindspore::graphkernel
4 years ago
LaiYongqiang
79599546ee
add ascend memory adapter for ascend memory management
4 years ago
i-robot
06b0beced7
!22731 GraphKernel supports CPU
Merge pull request !22731 from DeshiChen/0901_graphkernel_cpu
4 years ago
dayschan
32ecd8ee79
GraphKernel supports CPU
only Linux system is supported now.
change the default value of `ENABLE_AKG` to off, and controlled by option `-K`.
the `ENABLE_AKG` is auto enabled when `ENABLE_GPU` or `ENABLE_D` is on.
since now, we can use `ENABLE_AKG` to control the compilation of graphkernel
and akg codes.
fix usage description for option "-K", it should be "[-K on|off]".
LLVM is required by akg for cpu kernels, so AKG for cpu is default disabled now.
4 years ago
limingqi107
0a9cbd2220
imporve performance of CPU LSTMGrad
4 years ago
i-robot
5bd7e4def9
!24156 Move InsertCastCpu before BuildKernel
Merge pull request !24156 from DeshiChen/0926_insert_cast_cpu
4 years ago
i-robot
18960015b0
!24229 add dynamic shape attr pass
Merge pull request !24229 from hwjiaorui/dynamic-shape-pass
4 years ago
i-robot
0eccf93423
!23757 Add Support for AlltoAll in GPU
Merge pull request !23757 from Cononlly/master
4 years ago
hwjiaorui
eaa3dc57b7
add dynamic shape attr pass
4 years ago
dayschan
5ed932e5fd
Move InsertCastCpu before BuildKernel
It's necessary to immediately insert cast for the nodes that dtype was changed in SelectKernel,
instead of after BuildKernel. otherwise, some fusion passes cannot work as expected,
like Graph-Kernel optimization.
By now, the shape of LSTM's reserved output will be changed in InitKernel (called by BuildKernel),
and this output is only used by its gradient operator, so we can ignore the reserved output in this pass.
4 years ago
i-robot
811dd1778b
!23959 fix link error if compiled with -s on when import mindspore
Merge pull request !23959 from yanghaitao/yht_sync_shangyong
4 years ago
cononlly
a8088525d5
Add AlltoAll GPU operator and IR pass and tests
4 years ago
lizhenyu
0b6d7e4adb
fix code review warning
4 years ago
i-robot
b95a6d459d
!23906 add the interface of ascend supporting unified time
Merge pull request !23906 from limingqi107/new_actor_runtime
4 years ago
yanghaitao1
7f216aac41
fix link error if import mindspore
4 years ago
limingqi107
3e19107875
add the interface of ascend supporting unified time
4 years ago
dayschan
73ecb0c4ab
Add namespace mindspore::graphkernel
* change the graphkernel's passes code(backend/optimizer/graph_kernel/*) to the
new namespace `mindspore::graphkernel`, to decouple from `mindspore::opt`.
* change the original `mindspore::opt::graphkernel` to `mindspore::graphkernel::inner` (graph_kernel/model)
* change the original `mindspore::opt::expanders` to `mindspore::graphkernel::expanders` (graph_kernel/expanders)
TODO: modify graph_kernel_flags, kernel_compiler/akg/
4 years ago
limingqi107
16940f2d95
can not appear exception in the atexit
4 years ago
i-robot
fb58cfa273
!23414 update pynative profiling codes and delete some useless code
Merge pull request !23414 from lvchangquan/master
4 years ago
lvchangquan
b908f40d94
delete some useless codes and update pynative profiling codes
4 years ago
sabrinasun
1b8b453373
add remain isolation for dump
4 years ago
limingqi107
859aa42b41
unified runtime add data prepare actor and optimize code
4 years ago