Skip to content
This repository has been archived by the owner on Nov 17, 2023. It is now read-only.

Commit

Permalink
format
Browse files Browse the repository at this point in the history
  • Loading branch information
Yiyan66 committed Feb 9, 2020
1 parent ccf7a1e commit 4969ea6
Showing 1 changed file with 9 additions and 9 deletions.
18 changes: 9 additions & 9 deletions src/operator/numpy/random/np_exponential_op.h
Original file line number Diff line number Diff line change
Expand Up @@ -146,11 +146,11 @@ void NumpyExponentialForward(const nnvm::NodeAttrs &attrs,

template<typename xpu, int ndim, typename DType>
inline void ScalarExponentialReparamBackwardImpl(const OpContext& ctx,
const std::vector<TBlob>& inputs,
const std::vector<OpReqType>& req,
const std::vector<TBlob>& outputs,
const mxnet::TShape& new_ishape,
const mxnet::TShape& new_oshape) {
const std::vector<TBlob>& inputs,
const std::vector<OpReqType>& req,
const std::vector<TBlob>& outputs,
const mxnet::TShape& new_ishape,
const mxnet::TShape& new_oshape) {
using namespace mshadow;
using namespace mshadow::expr;
using namespace broadcast;
Expand All @@ -172,10 +172,10 @@ inline void ScalarExponentialReparamBackwardImpl(const OpContext& ctx,

template<typename xpu>
void ExponentialReparamBackward(const nnvm::NodeAttrs& attrs,
const OpContext& ctx,
const std::vector<TBlob>& inputs,
const std::vector<OpReqType>& req,
const std::vector<TBlob>& outputs) {
const OpContext& ctx,
const std::vector<TBlob>& inputs,
const std::vector<OpReqType>& req,
const std::vector<TBlob>& outputs) {
// skip kernel launch for zero-size tensors
if (inputs[0].shape_.Size() == 0U) {
return;
Expand Down

0 comments on commit 4969ea6

Please sign in to comment.