HomeMachine LearningDeep LearningNeural Network – Softmax Activation

Neural Network – Softmax Activation

Softmax is a special kind of activation function which use mostly used as the last layer. The final output of an NN needs to be human understandable so in the output we would need to be able to get probabilities in case of multi classifier problem.

Multi classifier problem, means a NN in which we are trying to predict if a given input is lets say “cat”, “dog”, “rat” etc or something similar that we are trying to classify the data. In such a case for a given input we would need a probability distribution in the output to be able to understand this.

Means suppose our neural networks gives an output

[10, 31,15]

this won’t make much sense to us but if we get an output

[.9833, .001, .002312] // ignore the actual values

this make more sense as this means there is a 98% change this is a cat.

This is what a softmax function does!

We can also use Sigmoid  as our last activation function in case is a binary classification problem.

To implement this in our neural network



Ok, again very simple!


Read more here



Leave a Reply

Your email address will not be published. Required fields are marked *

%d bloggers like this: