Translations:Word Embeddings/41/en: Difference between revisions
(Importing a new version from external source) Tag: Manual revert |
(Importing a new version from external source) Tag: Manual revert |
||
| Line 1: | Line 1: | ||
* '''ELMo''' (Peters et al., 2018) — uses a bidirectional LSTM to generate context-dependent word representations. | * '''ELMo''' (Peters et al., 2018) — uses a bidirectional {{Term|long short-term memory|LSTM}} to generate context-dependent word representations. | ||
* '''BERT''' (Devlin et al., 2019) — uses a | * '''BERT''' (Devlin et al., 2019) — uses a {{Term|transformer}} encoder trained with masked language modelling. | ||
* '''GPT''' series (Radford et al., 2018–) — uses a | * '''GPT''' series (Radford et al., 2018–) — uses a {{Term|transformer}} decoder trained autoregressively. | ||
Latest revision as of 23:34, 27 April 2026
- ELMo (Peters et al., 2018) — uses a bidirectional LSTM to generate context-dependent word representations.
- BERT (Devlin et al., 2019) — uses a transformer encoder trained with masked language modelling.
- GPT series (Radford et al., 2018–) — uses a transformer decoder trained autoregressively.