Translations:Softmax Function/34/en: Difference between revisions

    From Marovi AI
    (Importing a new version from external source)
    Tag: Manual revert
    (Importing a new version from external source)
    Tag: Manual revert
     
    Line 1: Line 1:
    * '''Attention mechanisms''': Softmax normalizes alignment scores into attention weights in the [[Attention Mechanisms|Transformer]] architecture.
    * '''{{Term|attention}} mechanisms''': Softmax normalizes alignment scores into {{Term|attention}} weights in the [[Attention Mechanisms|Transformer]] architecture.
    * '''Reinforcement learning''': Softmax over action-value estimates produces a stochastic policy (Boltzmann exploration).
    * '''Reinforcement learning''': Softmax over action-value estimates produces a stochastic policy (Boltzmann exploration).
    * '''Mixture models''': Softmax parameterizes mixing coefficients in mixture-of-experts architectures.
    * '''Mixture models''': Softmax parameterizes mixing coefficients in {{Term|mixture of experts|mixture-of-experts}} architectures.

    Latest revision as of 23:34, 27 April 2026

    Information about message (contribute)
    This message has no documentation. If you know where or how this message is used, you can help other translators by adding documentation to this message.
    Message definition (Softmax Function)
    * '''{{Term|attention}} mechanisms''': Softmax normalizes alignment scores into {{Term|attention}} weights in the [[Attention Mechanisms|Transformer]] architecture.
    * '''Reinforcement learning''': Softmax over action-value estimates produces a stochastic policy (Boltzmann exploration).
    * '''Mixture models''': Softmax parameterizes mixing coefficients in {{Term|mixture of experts|mixture-of-experts}} architectures.
    • attention mechanisms: Softmax normalizes alignment scores into attention weights in the Transformer architecture.
    • Reinforcement learning: Softmax over action-value estimates produces a stochastic policy (Boltzmann exploration).
    • Mixture models: Softmax parameterizes mixing coefficients in mixture-of-experts architectures.