Skip to content

Instantly share code, notes, and snippets.

@KornelDylski
Created May 10, 2019 11:54
Show Gist options
  • Star 0 You must be signed in to star a gist
  • Fork 0 You must be signed in to fork a gist
  • Save KornelDylski/53495e62b72bb277f55e5498eb020ef4 to your computer and use it in GitHub Desktop.
Save KornelDylski/53495e62b72bb277f55e5498eb020ef4 to your computer and use it in GitHub Desktop.
# borrowed from https://github.com/clcarwin/focal_loss_pytorch/blob/master/focalloss.py
# added "reduction" param for fastai
import torch
import torch.nn as nn
import torch.nn.functional as F
class FocalLoss(nn.Module):
def __init__(self, gamma=0.0, alpha=None, reduction='mean'):
super(FocalLoss, self).__init__()
self.gamma = gamma
self.reduction = reduction
if isinstance(alpha,list): self.alpha = torch.Tensor(alpha)
elif isinstance(alpha,(float,int,long)): self.alpha = torch.Tensor([alpha, 1-alpha])
def forward(self, input, target):
if input.dim()>2:
input = input.view(input.size(0),input.size(1),-1) # N,C,H,W => N,C,H*W
input = input.transpose(1,2) # N,C,H*W => N,H*W,C
input = input.contiguous().view(-1,input.size(2)) # N,H*W,C => N*H*W,C
target = target.view(-1,1)
logpt = F.log_softmax(input, dim=1)
logpt = logpt.gather(1, target)
logpt = logpt.view(-1)
pt = logpt.data.exp()
if self.alpha is not None:
if self.alpha.type()!=input.data.type():
self.alpha = self.alpha.type_as(input.data)
at = self.alpha.gather(0,target.data.view(-1))
logpt = logpt * at
loss = -1 * (1-pt)**self.gamma * logpt
if self.reduction == 'mean':
return loss.mean()
elif self.reduction == 'sum':
return loss.sum()
else:
return loss
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment