Translations:Attention Mechanisms/7/en: Difference between revisions
(Importing a new version from external source) |
(Importing a new version from external source) Tag: Manual revert |
||
| (2 intermediate revisions by the same user not shown) | |||
| Line 1: | Line 1: | ||
where <math>W_s</math>, <math>W_h</math>, and <math>v</math> are learned parameters. The attention weights are obtained by applying softmax: | where <math>W_s</math>, <math>W_h</math>, and <math>v</math> are learned parameters. The attention weights are obtained by applying {{Term|softmax}}: | ||
Latest revision as of 23:33, 27 April 2026
where $ W_s $, $ W_h $, and $ v $ are learned parameters. The attention weights are obtained by applying softmax: