Translations:Softmax Function/34/en: Difference between revisions
(Importing a new version from external source) |
(Importing a new version from external source) Tag: Manual revert |
| (One intermediate revision by the same user not shown) | |
(No difference)
| |
Latest revision as of 23:34, 27 April 2026
- attention mechanisms: Softmax normalizes alignment scores into attention weights in the Transformer architecture.
- Reinforcement learning: Softmax over action-value estimates produces a stochastic policy (Boltzmann exploration).
- Mixture models: Softmax parameterizes mixing coefficients in mixture-of-experts architectures.