Softmax function, known as softargmax or normalized exponential function, converts a vector of real numbers into a probability distribution of possible outcomes.
| FactSnippet No. 1,224,042 |
Softmax function, known as softargmax or normalized exponential function, converts a vector of real numbers into a probability distribution of possible outcomes.
| FactSnippet No. 1,224,042 |
The softmax function is often used as the last activation function of a neural network to normalize the output of a network to a probability distribution over predicted output classes, based on Luce's choice axiom.
| FactSnippet No. 1,224,043 |
Softmax function takes as input a vector of real numbers, and normalizes it into a probability distribution consisting of probabilities proportional to the exponentials of the input numbers.
| FactSnippet No. 1,224,044 |
Name "softmax" is misleading; the function is not a smooth maximum, but is rather a smooth approximation to the arg max function: the function whose value is which index has the maximum.
| FactSnippet No. 1,224,045 |
In probability theory, the output of the softargmax Softmax function can be used to represent a categorical distribution – that is, a probability distribution over different possible outcomes.
| FactSnippet No. 1,224,046 |
In statistical mechanics, the softargmax Softmax function is known as the Boltzmann distribution : the index set are the microstates of the system; the inputs are the energies of that state; the denominator is known as the partition Softmax function, often denoted by ; and the factor is called the coldness .
| FactSnippet No. 1,224,047 |
Softmax function is used in various multiclass classification methods, such as multinomial logistic regression [1], multiclass linear discriminant analysis, naive Bayes classifiers, and artificial neural networks.
| FactSnippet No. 1,224,048 |
Standard softmax function is often used in the final layer of a neural network-based classifier.
| FactSnippet No. 1,224,049 |
Standard logistic Softmax function is the special case for a 1-dimensional axis in 2-dimensional space, say the x-axis in the plane.
| FactSnippet No. 1,224,050 |
Softmax function is the gradient of the LogSumExp function, a smooth maximum:.
| FactSnippet No. 1,224,051 |
Softmax function was used in statistical mechanics as the Boltzmann distribution in the foundational paper Boltzmann, formalized and popularized in the influential textbook Gibbs .
| FactSnippet No. 1,224,052 |