In the following TensorFlow function, we must feed the activation of artificial neurons in the final layer. That I understand. But I don't understand why it is called logits? Isn't that a mathematical function?
loss_function = tf.nn.softmax_cross_entropy_with_logits(
logits = last_layer,
labels = target_output
)

Logit is a function that maps probabilities
[0, 1]to[-inf, +inf].Softmax is a function that maps
[-inf, +inf]to[0, 1]similar as Sigmoid. But Softmax also normalizes the sum of the values(output vector) to be 1.Tensorflow "with logit": It means that you are applying a softmax function to logit numbers to normalize it. The input_vector/logit is not normalized and can scale from [-inf, inf].
This normalization is used for multiclass classification problems. And for multilabel classification problems sigmoid normalization is used i.e.
tf.nn.sigmoid_cross_entropy_with_logits