169 Commits (a64920c46802a3030bfdf8754870ffb1e5c6ba78)

Author SHA1 Message Date
  zengzitao 8064de7931 fix maximum_grad and minimum_grad input_shape not equal to output_shape bug 4 years ago
  i-robot 07f58b0b46 !17626 matmul to mul 4 years ago
  lingyunli63 4f34e537a0 replace matmul/batchmatmul by mul when k is 1 4 years ago
  Yang Jiao 4d1f8d1d00 fix expander bugs 4 years ago
  mindspore-ci-bot 7e5d68991a !17680 Avoid overflow of in realdiv 4 years ago
  wenfangpei fa531a473d avoid overflow in realdiv 4 years ago
  Yang Jiao d9a60e0d28 fix static warnings 5 years ago
  Gaoxiong 1d3da1789e Acyclic checking bugfix of split model 4 years ago
  mindspore-ci-bot 84859aba23 !16695 bug fix in lamb 4 years ago
  wenfangpei fd84c20a6a bug fix in lamb 4 years ago
  looop5 68f55e1e93 expand conv2d when input format is DefaultFormat but attr format is NHWC 4 years ago
  looop5 dd81f47271 add conv2d expander 5 years ago
  mindspore-ci-bot f91a365564 !16322 [GraphKernel] Enable matmul for gpu 4 years ago
  lingyunli63 afc69b16f7 enable gpu gk MatMul and insert pad/unpad 5 years ago
  zengzitao 516487c9cd fix log(inputs), inputs less than 0.0, cause loss inf bug 4 years ago
  chenlei_autodiff e0f882b5ce [GraphKernel] fix op infer for elementwise. 4 years ago
  dayschan 54a0edb837 Limit the cluster size in costmodel 5 years ago
  mindspore-ci-bot 08959ddb74 !16224 [GraphKernel]Fix wide & deep ps mode precision error 5 years ago
  mindspore-ci-bot f247cc2db7 !16047 [GraphKernel]adapt batchmatmul for GraphKernel in ascend 5 years ago
  mindspore-ci-bot c492d784bf !16104 [GraphKernel] add stitch fusion test case in CI 5 years ago
  zengzitao 15cf743ede rewrite sigmoid_cross_entropy_with_logits expander to avoid overflow 5 years ago
  wenfangpei 2e2e850b4a adapt batchmatmul for GraphKernel in ascend 5 years ago
  r1chardf1d0 b1be842d2b add stitch fusion case in ci 5 years ago
  hanhuifeng2020 bc46d644fe [GraphKernel]Support reshape/elewise/broadcast+transdata fusion 5 years ago
  mindspore-ci-bot 0887d35b1c !15951 [GraphKernel]add the attribute reduce_output_fuse to enable fuse for the reduce_output on Ascend 5 years ago
  mindspore-ci-bot ed539597c2 !15415 [GraphKernel]adapt for logsoftmax in ascend 5 years ago
  wenfangpei 4174a7b38f expanders of some fusion ops 5 years ago
  wenfangpei db8256e61f adapt for logsoftmax in ascend 5 years ago
  mindspore-ci-bot 1827697642 !15961 Eliminate recursion call in split model 5 years ago
  wenfangpei c41875b318 adapt expanders of some ops from gpu to ascend 5 years ago
  Gaoxiong 4bc67f38de eliminate recursion call 5 years ago
  hanhuifeng2020 425d401e85 [GraphKernel]add the attr reduce_output_fuse to enable fuse for reduce_output on Ascend 5 years ago
  mindspore-ci-bot 52e7f51970 !15741 [GraphKernel] batchnorm expander supports when first input is float16 5 years ago
  mindspore-ci-bot ff75ce8ac4 !15097 [graph kernel] clean code for expanders. 5 years ago
  looop5 24f441ba33 batchnorm expander supports when first input is float16 5 years ago
  mindspore-ci-bot d6f58cb765 !15658 Reduce recursion overhead of split model 5 years ago
  chenlei_autodiff fd227bb448 [graph kernel] clean code for expanders. 5 years ago
  mindspore-ci-bot 168c64b60d !15648 [GraphKernel] negative axis in Squeeze expander. 5 years ago
  Gaoxiong 71002ed19d reduce recursion overhead of split model 5 years ago
  zengzitao 8dcff8d83c refactor tile op and in expander open on gpu 5 years ago
  chenlei_autodiff b419f60b0d [GraphKernel] negative axis in Squeeze expander. 5 years ago
  r1chardf1d0 5c5d125b1d optimize stitch fusion strategy 5 years ago
  hanhuifeng2020 25505642ce enable GraphKernel for TransData 5 years ago
  lingyunli63 c48c2430f0 fuse matmul and elementwise in graphkernel 5 years ago
  mindspore-ci-bot cd002cb7f7 !14893 enable stitch fusion on bert 5 years ago
  mindspore-ci-bot 5b4685c5ea !14604 [GraphKernel] add some expander ops 5 years ago
  r1chardf1d0 3b32995936 enable stitch fusion on bert 5 years ago
  chenlei_autodiff 13fbfca6b9 [graph kernel] add expander ops. 5 years ago
  wenfangpei b9715db358 bugfix in expanders of layernorm 5 years ago
  wenfangpei 66d28af79e adapt for layernorm in ascend 5 years ago