Skip to content
This repository has been archived by the owner on Nov 17, 2023. It is now read-only.

Commit

Permalink
apply review
Browse files Browse the repository at this point in the history
  • Loading branch information
Bartlomiej Gawrych committed Feb 9, 2022
1 parent ced7e26 commit 45f631c
Show file tree
Hide file tree
Showing 2 changed files with 7 additions and 6 deletions.
1 change: 1 addition & 0 deletions src/operator/nn/dnnl/dnnl_masked_softmax.cc
Original file line number Diff line number Diff line change
Expand Up @@ -133,6 +133,7 @@ void DNNLMaskedSoftmaxFwd::Execute(const Tensors& tensors,
using namespace mxnet::op::mxnet_op;
using namespace dnnl;
/*
Three steps of masked softmax:
1. out = input * [(mask - 1) * inf]
2. out = softmax(out)
3. out = out * mask
Expand Down
12 changes: 6 additions & 6 deletions src/operator/nn/masked_softmax.cc
Original file line number Diff line number Diff line change
Expand Up @@ -21,13 +21,13 @@
* \file masked_softmax.cc
*/

#include "./softmax-inl.h"
#include "../tensor/elemwise_unary_op.h"
#include "../tensor/elemwise_binary_op.h"
#include "../operator_common.h"
#include "softmax-inl.h"
#include "operator/tensor/elemwise_unary_op.h"
#include "operator/tensor/elemwise_binary_op.h"
#include "operator/operator_common.h"
#if MXNET_USE_ONEDNN == 1
#include "dnnl/dnnl_base-inl.h"
#include "dnnl/dnnl_ops-inl.h"
#include "operator/nn/dnnl/dnnl_base-inl.h"
#include "operator/nn/dnnl/dnnl_ops-inl.h"
#endif

namespace mxnet {
Expand Down

0 comments on commit 45f631c

Please sign in to comment.