1

For a project, I've trained multiple networks for multiclass classification all ending with a ReLU activation at the output.

Now the output logits are not probabilities.

Is it valid to get the probability of each class by applying a softmax function at the end, after training?

user452306
  • 21
  • 3
  • Hello. Welcome to Artificial Intelligence Stack Exchange! Can you please explain 1. why you trained more than one neural network to solve the same problem, 2. why you used the ReLU at the last layers of these neural nets, 3. how exactly you trained the neural nets? Did you compare the (vector) labels with the output produced by the last ReLU, 3.1 you got good accuracy with this approach? – nbro Nov 22 '21 at 12:08
  • Hi, thank you. 1. It's not the exact same problem, it's for slightly differing datasets and slight differing architectures. 2. This is a constraint set for research purposes. 3. As usual, using backprop - output decision was decided using argmax over the last relu output vector; and yes test/validation accuracy and loss are satisfying – user452306 Nov 22 '21 at 12:23
  • My main and only question, is if we can convert post-hoc some relu-output to some 'propabilities' in a valid way – user452306 Nov 22 '21 at 12:24

0 Answers0