Translations:Word Embeddings/41/en: Difference between revisions

    From Marovi AI
    (Importing a new version from external source)
    (Importing a new version from external source)
    Tag: Manual revert
    Line 1: Line 1:
    * '''ELMo''' (Peters et al., 2018) — uses a bidirectional {{Term|long short-term memory|LSTM}} to generate context-dependent word representations.
    * '''ELMo''' (Peters et al., 2018) — uses a bidirectional LSTM to generate context-dependent word representations.
    * '''BERT''' (Devlin et al., 2019) — uses a {{Term|transformer}} encoder trained with masked language modelling.
    * '''BERT''' (Devlin et al., 2019) — uses a Transformer encoder trained with masked language modelling.
    * '''GPT''' series (Radford et al., 2018–) — uses a {{Term|transformer}} decoder trained autoregressively.
    * '''GPT''' series (Radford et al., 2018–) — uses a Transformer decoder trained autoregressively.

    Revision as of 22:01, 27 April 2026

    Information about message (contribute)
    This message has no documentation. If you know where or how this message is used, you can help other translators by adding documentation to this message.
    Message definition (Word Embeddings)
    * '''ELMo''' (Peters et al., 2018) — uses a bidirectional {{Term|long short-term memory|LSTM}} to generate context-dependent word representations.
    * '''BERT''' (Devlin et al., 2019) — uses a {{Term|transformer}} encoder trained with masked language modelling.
    * '''GPT''' series (Radford et al., 2018–) — uses a {{Term|transformer}} decoder trained autoregressively.
    • ELMo (Peters et al., 2018) — uses a bidirectional LSTM to generate context-dependent word representations.
    • BERT (Devlin et al., 2019) — uses a Transformer encoder trained with masked language modelling.
    • GPT series (Radford et al., 2018–) — uses a Transformer decoder trained autoregressively.