Fonction exponentielle normalisée
Domaine
Vocabulary Apprentissage profond
Définition
Termes privilégiés
Anglais
Softmax
The softmax function is typically used to convert a vector of raw scores into class probabilities at the output layer of a Neural Network used for classification. It normalizes the scores by exponentiating and dividing by a normalization constant. If we are dealing with a large number of classes, a large vocabulary in Machine Translation for example, the normalization constant is expensive to compute. There exist various alternatives to make the computation more efficient, including Hierarchical Softmax or using a sampling-based loss such as NCE.
Contributeurs: Claude Coulombe, Jacques Barolet, Julie Roy, Patrick Drouin, wiki, Robert Meloche