4

I am looking to try different loss functions for a hierarchical multi-label classification problem. So far, I have been training different models or submodels like multilayer perceptron (MLP) branch inside a bigger model which deals with different levels of classification, yielding a binary vector. I have been also using Binary Cross-Entropy (BCE) and summing all the losses existing in the model before backpropagating.

I am considering trying other losses like MultiLabelSoftMarginLoss and MultiLabelMarginLoss.

What other loss functions are worth trying? Hamming loss perhaps or a variation? Is it better to sum all the losses and backpropagate or do multiple backpropagations?

nbro
  • 39,006
  • 12
  • 98
  • 176
Skinish
  • 153
  • 1
  • 1
  • 9

1 Answers1

0

What hierarchical loss functions are worth trying There are a number of hierarchical loss functions that have been described (see articles #1 and #2), such as hierarchical win and normalized LCA. The loss functions that you noted will not by themselves leverage the hierarchical information.

Snehal Patel
  • 912
  • 1
  • 1
  • 25