|
- from __future__ import absolute_import
- from .executor import wrapped_mpi_nccl_init, Executor, gradients, scheduler_init,\
- scheduler_finish, get_worker_communicate, worker_init, worker_finish, server_init, server_finish, HetuConfig, new_group_comm
-
- from .AddConst import addbyconst_op
- from .AddElewise import add_op
- from .AvgPool import avg_pool2d_op, avg_pool2d_gradient_op
- from .BatchNorm import batch_normalization_op, batch_normalization_gradient_op, batch_normalization_gradient_of_data_op, batch_normalization_gradient_of_scale_op, batch_normalization_gradient_of_bias_op
- from .Broadcast import broadcastto_op
- from .BinaryCrossEntropy import binarycrossentropy_op
- from .Concat import concat_op, concat_gradient_op
- from .Conv2d import conv2d_op, conv2d_gradient_of_data_op, conv2d_gradient_of_filter_op
- from .Conv2dBroadcast import conv2d_broadcastto_op
- from .Conv2dReduceSum import conv2d_reducesum_op
- from .CuSparse import csrmv_op, csrmm_op
- from .Division import div_op, div_const_op
- from .Dropout import dropout_op, dropout_gradient_op
- from .Dropout2d import dropout2d_op, dropout2d_gradient_op
- from .MatrixMult import matmul_op
- from .MaxPool import max_pool2d_op, max_pool2d_gradient_op
- from .MultiplyConst import mul_byconst_op
- from .MultiplyElewise import mul_op
- from .OnesLike import oneslike_op
- from .Opposite import opposite_op
- from .Pad import pad_op, pad_gradient_op
- from .ReduceSumAxisZero import reducesumaxiszero_op
- from .Relu import relu_op, relu_gradient_op
- from .LeakyRelu import leaky_relu_op, leaky_relu_gradient_op
- from .Reshape import array_reshape_op, array_reshape_gradient_op
- from .Sigmoid import sigmoid_op
- from .Slice import slice_op, slice_gradient_op
- from .Softmax import softmax_func, softmax_op
- from .SoftmaxCrossEntropy import softmaxcrossentropy_op
- from .Split import split_op, split_gradient_op
- from .Sqrt import sqrt_op, rsqrt_op
- from .Tanh import tanh_op
- from .Transpose import transpose_op
- from .Variable import Variable, placeholder_op
- from .ZerosLike import zeroslike_op
- from .EmbeddingLookUp import embedding_lookup_op, embedding_lookup_gradient_op
- from .Where import where_op
- from .BatchMatrixMult import batch_matmul_op
- from .LayerNorm import layer_normalization_op
- from .InstanceNorm2d import instance_normalization2d_op
- from .BroadcastShape import broadcast_shape_op
- from .ReduceSum import reduce_sum_op
- from .ReduceMean import reduce_mean_op
- from .OneHot import one_hot_op
- from .AllReduceCommunicate import allreduceCommunicate_op, groupallreduceCommunicate_op
- from .ParameterServerCommunicate import parameterServerCommunicate_op, parameterServerSparsePull_op
- from .DataTransfer import datah2d_op, datad2h_op
- from .MatrixDot import matrix_dot_op
- from .DistGCN_15d import distgcn_15d_op
- from .PipelineSend import pipeline_send_op
- from .PipelineReceive import pipeline_receive_op
- from .Dispatch import dispatch
-
- __all__ = [
- 'Executor',
- 'gradients',
- 'wrapped_mpi_nccl_init',
- 'scheduler_init',
- 'scheduler_finish',
- 'get_worker_communicate',
- 'worker_init',
- 'worker_finish',
- 'server_init',
- 'server_finish',
- 'HetuConfig',
- 'new_group_comm',
-
- 'addbyconst_op',
- 'add_op',
- 'avg_pool2d_op',
- 'avg_pool2d_gradient_op',
- 'batch_normalization_op',
- 'batch_normalization_gradient_op',
- 'batch_normalization_gradient_of_data_op',
- 'batch_normalization_gradient_of_scale_op',
- 'batch_normalization_gradient_of_bias_op',
- 'broadcastto_op',
- 'concat_op',
- 'concat_gradient_op',
- 'conv2d_op',
- 'conv2d_gradient_of_data_op',
- 'conv2d_gradient_of_filter_op',
- 'conv2d_broadcastto_op',
- 'conv2d_reducesum_op',
- 'csrmv_op',
- 'csrmm_op',
- 'div_op',
- 'div_const_op',
- 'dropout_op',
- 'dropout_gradient_op',
- 'dropout2d_op',
- 'dropout2d_gradient_op',
- 'matmul_op',
- 'max_pool2d_op',
- 'max_pool2d_gradient_op',
- 'mul_byconst_op',
- 'mul_op',
- 'oneslike_op',
- 'opposite_op',
- 'pad_op',
- 'pad_gradient_op',
- 'reducesumaxiszero_op',
- 'relu_op',
- 'relu_gradient_op',
- 'leaky_relu_op',
- 'leaky_relu_gradient_op',
- 'array_reshape_op',
- 'array_reshape_gradient_op',
- 'sigmoid_op',
- 'slice_op',
- 'slice_gradient_op',
- 'softmax_func',
- 'softmax_op',
- 'softmaxcrossentropy_op',
- 'split_op',
- 'split_gradient_op',
- 'sqrt_op',
- 'scheduler_init',
- 'scheduler_finish',
- 'server_init',
- 'server_finish',
- 'rsqrt_op',
- 'tanh_op',
- 'transpose_op',
- 'Variable',
- 'worker_init',
- 'worker_finish',
- 'placeholder_op',
- 'zeroslike_op',
- "embedding_lookup_op",
- "embedding_lookup_gradient_op",
- 'where_op',
- 'batch_matmul_op',
- 'layer_normalization_op',
- 'instance_normalization2d_op',
- 'broadcast_shape_op',
- 'reduce_sum_op',
- 'reduce_mean_op',
- 'one_hot_op',
- 'allreduceCommunicate_op',
- 'parameterServerCommunicate_op',
- 'datah2d_op',
- 'datad2h_op',
- 'binarycrossentropy_op',
- 'matrix_dot_op',
- 'parameterServerSparsePull_op',
- 'distgcn_15d_op',
- 'groupallreduceCommunicate_op',
- 'pipeline_send_op',
- 'pipeline_receive_op',
- 'dispatch',
- ]
|