Translations:Softmax Function/30/en: Difference between revisions
(Importing a new version from external source) |
(Importing a new version from external source) Tag: Manual revert |
| (One intermediate revision by the same user not shown) | |
(No difference)
| |
Latest revision as of 23:34, 27 April 2026
- A neural network produces raw logits $ \mathbf{z} $ from its final linear layer.
- Softmax converts logits to probabilities: $ \hat{\mathbf{y}} = \sigma(\mathbf{z}) $.
- The predicted class is $ \hat{c} = \arg\max_k \hat{y}_k $.
- Training uses Cross-Entropy Loss applied to the predicted distribution and the true labels.