From c0794c3915a636c03aad7efde20396b115402712 Mon Sep 17 00:00:00 2001 From: "Daniel.Frisinghelli" <daniel.frisinghelli@eurac.edu> Date: Fri, 15 Oct 2021 14:34:06 +0200 Subject: [PATCH] Changed computation of Weibull NLL. --- climax/core/loss.py | 14 +++++--------- 1 file changed, 5 insertions(+), 9 deletions(-) diff --git a/climax/core/loss.py b/climax/core/loss.py index 9a6e99f..c2cc811 100644 --- a/climax/core/loss.py +++ b/climax/core/loss.py @@ -148,16 +148,12 @@ class BernoulliWeibullLoss(BernoulliLoss): loss_bern = torch.log(1 - p_pred[mask_p] + self.epsilon) # Weibull contribution - # loss_weib = (torch.log(p_pred[~mask_p] + self.epsilon) + - # torch.log(shape + self.epsilon) - - # shape * torch.log(scale + self.epsilon) + - # (shape - 1) * torch.log(y_weib + self.epsilon) - - # torch.pow(y_weib / (scale + self.epsilon), shape) - # ) loss_weib = (torch.log(p_pred[~mask_p] + self.epsilon) + - torch.log(shape / scale) - - (shape - 1) * torch.log(y_weib / scale) - - torch.pow(y_weib / scale, shape)) + torch.log(shape + self.epsilon) - + shape * torch.log(scale + self.epsilon) + + (shape - 1) * torch.log(y_weib + self.epsilon) - + torch.pow(y_weib / (scale + self.epsilon), shape) + ) # fill loss array loss[torch.where(mask_p)] = - loss_bern -- GitLab