In the following TensorFlow function, we must feed the activation of artificial neurons in the final layer. That I understand. But I don't understand why it is called logits? Isn't that a mathematical function?
loss_function = tf.nn.softmax_cross_entropy_with_logits(
logits = last_layer,
labels = target_output
)
Logit is a function that maps probabilities
[0, 1]
to[-inf, +inf]
.Softmax is a function that maps
[-inf, +inf]
to[0, 1]
similar as Sigmoid. But Softmax also normalizes the sum of the values(output vector) to be 1.Tensorflow "with logit": It means that you are applying a softmax function to logit numbers to normalize it. The input_vector/logit is not normalized and can scale from [-inf, inf].
This normalization is used for multiclass classification problems. And for multilabel classification problems sigmoid normalization is used i.e.
tf.nn.sigmoid_cross_entropy_with_logits