From c0794c3915a636c03aad7efde20396b115402712 Mon Sep 17 00:00:00 2001
From: "Daniel.Frisinghelli" <daniel.frisinghelli@eurac.edu>
Date: Fri, 15 Oct 2021 14:34:06 +0200
Subject: [PATCH] Changed computation of Weibull NLL.

---
 climax/core/loss.py | 14 +++++---------
 1 file changed, 5 insertions(+), 9 deletions(-)

diff --git a/climax/core/loss.py b/climax/core/loss.py
index 9a6e99f..c2cc811 100644
--- a/climax/core/loss.py
+++ b/climax/core/loss.py
@@ -148,16 +148,12 @@ class BernoulliWeibullLoss(BernoulliLoss):
         loss_bern = torch.log(1 - p_pred[mask_p] + self.epsilon)
 
         # Weibull contribution
-        # loss_weib = (torch.log(p_pred[~mask_p] + self.epsilon) +
-        #              torch.log(shape + self.epsilon) -
-        #              shape * torch.log(scale + self.epsilon) +
-        #              (shape - 1) * torch.log(y_weib + self.epsilon) -
-        #              torch.pow(y_weib / (scale + self.epsilon), shape)
-        #              )
         loss_weib = (torch.log(p_pred[~mask_p] + self.epsilon) +
-                     torch.log(shape / scale) -
-                     (shape - 1) * torch.log(y_weib / scale) -
-                     torch.pow(y_weib / scale, shape))
+                     torch.log(shape + self.epsilon) -
+                     shape * torch.log(scale + self.epsilon) +
+                     (shape - 1) * torch.log(y_weib + self.epsilon) -
+                     torch.pow(y_weib / (scale + self.epsilon), shape)
+                     )
 
         # fill loss array
         loss[torch.where(mask_p)] = - loss_bern
-- 
GitLab