Translations:BERT Pre-training of Deep Bidirectional Transformers/14/en: Difference between revisions
(Importing a new version from external source) |
(Importing a new version from external source) |
||
| Line 1: | Line 1: | ||
Input representation combines token embeddings, segment embeddings (indicating sentence A or B), and positional embeddings. BERT uses WordPiece tokenization with a 30,000-token vocabulary. | Input representation combines token {{Term|embedding|embeddings}}, segment {{Term|embedding|embeddings}} (indicating sentence A or B), and positional {{Term|embedding|embeddings}}. BERT uses WordPiece {{Term|tokenization}} with a 30,000-token vocabulary. | ||
Latest revision as of 21:37, 27 April 2026
Input representation combines token embeddings, segment embeddings (indicating sentence A or B), and positional embeddings. BERT uses WordPiece tokenization with a 30,000-token vocabulary.