From b3967d1d48c8045cdb5b642ac92f9e1e9280d34c Mon Sep 17 00:00:00 2001 From: Pierre Nodet Date: Sun, 30 May 2021 18:08:24 +0200 Subject: [PATCH] Fix issue #654 Avoid two softmax calls in LogisticRegression --- pl_bolts/models/regression/logistic_regression.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/pl_bolts/models/regression/logistic_regression.py b/pl_bolts/models/regression/logistic_regression.py index 21178f5e17..f283c8182b 100644 --- a/pl_bolts/models/regression/logistic_regression.py +++ b/pl_bolts/models/regression/logistic_regression.py @@ -53,7 +53,7 @@ def training_step(self, batch, batch_idx): # flatten any input x = x.view(x.size(0), -1) - y_hat = self(x) + y_hat = self.linear(x) # PyTorch cross_entropy function combines log_softmax and nll_loss in single function loss = F.cross_entropy(y_hat, y, reduction='sum')