90 Commits (a8d81c8b7bf9d54ae66ce8cab15836610e624724)

Author SHA1 Message Date
  mindspore-ci-bot 64abbeaa89 !705 add pynative cache 5 years ago
  kswang e6f36e050f change cpu kernel dir 5 years ago
  mindspore-ci-bot 63712848e2 !494 Split ccsrc cmake to individual sub-directories 5 years ago
  zhoufeng b681cec8f2 cmake refactor 5 years ago
  chujinjin e2b0a28142 add pynative cache 5 years ago
  mindspore-ci-bot da8c74b54c !688 [MS]fix kernel select bug 5 years ago
  chenjianping 6d47036f95 erase datatype raise kernel 5 years ago
  limingqi107 664f2628e5 optimize gpu allReduce alloc memory performance 5 years ago
  mindspore-ci-bot 2a31fce599 !647 change profiling iter_end point after last tbe kernel 5 years ago
  caifubi 3f69b5b518 insert profiling iteration end point after last tbe kernel 5 years ago
  mindspore-ci-bot ffdad1acd4 !617 Match format when kernel selecting using raise or reduce precision 5 years ago
  mindspore-ci-bot bfab0bc856 !618 clean ascend runtime codex 5 years ago
  caifubi 246fc290d0 clean runtime codex 5 years ago
  lianliguang c4aeb5a0b8 add format chooice when kernel selecting reduce or raise precision 5 years ago
  limingqi107 0f0e8fe874 gpu dynamic memory pool can not reuse allReduce in multi-stream 5 years ago
  liubuyu 0b6b5e5123 fix codedex warning 5 years ago
  caifubi 0cef4d8abd set fp_start profiling point after getnext node 5 years ago
  mindspore-ci-bot ca3aa6071a !527 gpu dynamic memory pool supports multi-allReduce 5 years ago
  mindspore-ci-bot ae31e8b516 !571 GPU change the strategy of computing capacity 5 years ago
  limingqi107 2891f0d20d gpu dynamic memory pool supports multi-allReduce 6 years ago
  VectorSL b8d7cd9775 gpu change compute capacity strategy 5 years ago
  mindspore-ci-bot 715c0735a8 !487 add dtype trans template 5 years ago
  liubuyu ac2d5df2a1 add dtype trans template 6 years ago
  mindspore-ci-bot 1f7f11a3df !553 insert profiling kernel for hccl kernel automaticly 5 years ago
  mindspore-ci-bot 77b62c7f16 !555 [MS][Wndows]support default log level 5 years ago
  chenjianping 18d8a1d2d3 support default log level on windows 10 5 years ago
  caifubi 94a455dace insert profiling kernel for hccl automaticly 6 years ago
  zhoufeng c2b3360d69 update clang format rule 6 years ago
  mindspore-ci-bot 475f62f680 !499 pynative support topk and print op 6 years ago
  lvliang 0e4824cd89 pynative-support-topk-and-print 6 years ago
  gukecai f8208c7c52 Support GetNext Parallel 6 years ago
  limingqi107 5e01b94ccd gpu dynamic memory pool suppoerts multi-graph 6 years ago
  mindspore-ci-bot 093c2caed4 !337 optimize execute order sort 6 years ago
  mindspore-ci-bot b418b18447 !443 add cpu one hot 6 years ago
  kswang 6775190e48 add cpu one hot 6 years ago
  mindspore-ci-bot 7c06d292c8 !387 auto mix precision 6 years ago
  liubuyu 852e61d46c bug fix 6 years ago
  liubuyu b1585f862d auto mix precision 6 years ago
  liubuyu fc07cd908e add 6d format transfer 6 years ago
  kswang 83eeac9310 optimize execute order sort 6 years ago
  jjfeing f9ef78609f add nc1hwc0_c04 format 6 years ago
  caifubi bce5f57752 use GraphId as key of DavinciModel in ascend_kernel_runtime.cc 6 years ago
  ZPaC b8a9121597 Add GPU send and recv controlling kernels. 6 years ago
  mindspore-ci-bot 58b013c319 !363 clear the warmming scan by package 6 years ago
  chenzomi b77f41d658 clear the warmming scan by package 6 years ago
  mindspore-ci-bot f9dd47620c !300 refactor kernel select priority scheme 6 years ago
  mindspore-ci-bot 285e258b3b !347 fix e2e dump shape not match 6 years ago
  mindspore-ci-bot 4c3969e12a !351 fix bug of TensorAddGrad single op run fail 6 years ago
  jojobugfree 762bf9ac25 fix tensoradd grad op run fail 6 years ago
  dengwentao 593c4fc700 fix shape used for dump 6 years ago