<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://marovi.ai/index.php?action=history&amp;feed=atom&amp;title=Word_Embeddings%2Fes</id>
	<title>Word Embeddings/es - Revision history</title>
	<link rel="self" type="application/atom+xml" href="https://marovi.ai/index.php?action=history&amp;feed=atom&amp;title=Word_Embeddings%2Fes"/>
	<link rel="alternate" type="text/html" href="https://marovi.ai/index.php?title=Word_Embeddings/es&amp;action=history"/>
	<updated>2026-04-24T13:01:40Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.39.1</generator>
	<entry>
		<id>https://marovi.ai/index.php?title=Word_Embeddings/es&amp;diff=2162&amp;oldid=prev</id>
		<title>DeployBot: [deploy-bot] Deploy from CI (8c92aeb)</title>
		<link rel="alternate" type="text/html" href="https://marovi.ai/index.php?title=Word_Embeddings/es&amp;diff=2162&amp;oldid=prev"/>
		<updated>2026-04-24T07:09:01Z</updated>

		<summary type="html">&lt;p&gt;[deploy-bot] Deploy from CI (8c92aeb)&lt;/p&gt;
&lt;table style=&quot;background-color: #fff; color: #202122;&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;Revision as of 07:09, 24 April 2026&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l119&quot;&gt;Line 119:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 119:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[Category:NLP]]&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[Category:NLP]]&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[Category:Intermediate]]&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[Category:Intermediate]]&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&amp;lt;!--v1.2.0 cache-bust--&gt;&lt;/del&gt;&lt;/div&gt;&lt;/td&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-side-added&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&amp;lt;!-- pass 2 --&gt;&lt;/del&gt;&lt;/div&gt;&lt;/td&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-side-added&quot;&gt;&lt;/td&gt;&lt;/tr&gt;

&lt;!-- diff cache key mediawiki:diff::1.12:old-2119:rev-2162 --&gt;
&lt;/table&gt;</summary>
		<author><name>DeployBot</name></author>
	</entry>
	<entry>
		<id>https://marovi.ai/index.php?title=Word_Embeddings/es&amp;diff=2119&amp;oldid=prev</id>
		<title>DeployBot: Pass 2 force re-parse</title>
		<link rel="alternate" type="text/html" href="https://marovi.ai/index.php?title=Word_Embeddings/es&amp;diff=2119&amp;oldid=prev"/>
		<updated>2026-04-24T07:01:27Z</updated>

		<summary type="html">&lt;p&gt;Pass 2 force re-parse&lt;/p&gt;
&lt;table style=&quot;background-color: #fff; color: #202122;&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;Revision as of 07:01, 24 April 2026&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l120&quot;&gt;Line 120:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 120:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[Category:Intermediate]]&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[Category:Intermediate]]&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--v1.2.0 cache-bust--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--v1.2.0 cache-bust--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-side-deleted&quot;&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&amp;lt;!-- pass 2 --&gt;&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;

&lt;!-- diff cache key mediawiki:diff::1.12:old-2082:rev-2119 --&gt;
&lt;/table&gt;</summary>
		<author><name>DeployBot</name></author>
	</entry>
	<entry>
		<id>https://marovi.ai/index.php?title=Word_Embeddings/es&amp;diff=2082&amp;oldid=prev</id>
		<title>DeployBot: Force re-parse after Math source-mode rollout (v1.2.0)</title>
		<link rel="alternate" type="text/html" href="https://marovi.ai/index.php?title=Word_Embeddings/es&amp;diff=2082&amp;oldid=prev"/>
		<updated>2026-04-24T06:58:49Z</updated>

		<summary type="html">&lt;p&gt;Force re-parse after Math source-mode rollout (v1.2.0)&lt;/p&gt;
&lt;table style=&quot;background-color: #fff; color: #202122;&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;Revision as of 06:58, 24 April 2026&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l119&quot;&gt;Line 119:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 119:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[Category:NLP]]&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[Category:NLP]]&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[Category:Intermediate]]&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[[Category:Intermediate]]&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-side-deleted&quot;&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&amp;lt;!--v1.2.0 cache-bust--&gt;&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;

&lt;!-- diff cache key mediawiki:diff::1.12:old-2011:rev-2082 --&gt;
&lt;/table&gt;</summary>
		<author><name>DeployBot</name></author>
	</entry>
	<entry>
		<id>https://marovi.ai/index.php?title=Word_Embeddings/es&amp;diff=2011&amp;oldid=prev</id>
		<title>DeployBot: [deploy-bot] Deploy from CI (775ba6e)</title>
		<link rel="alternate" type="text/html" href="https://marovi.ai/index.php?title=Word_Embeddings/es&amp;diff=2011&amp;oldid=prev"/>
		<updated>2026-04-24T04:01:51Z</updated>

		<summary type="html">&lt;p&gt;[deploy-bot] Deploy from CI (775ba6e)&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;{{LanguageBar | page = Word Embeddings}}&lt;br /&gt;
{{ArticleInfobox | topic_area = NLP | difficulty = Intermediate | prerequisites = [[Neural Networks]]}}&lt;br /&gt;
{{ContentMeta | generated_by = claude-opus | model_used = claude-opus-4-6 | generated_date = 2026-03-13}}&lt;br /&gt;
&lt;br /&gt;
Los &amp;#039;&amp;#039;&amp;#039;word embeddings&amp;#039;&amp;#039;&amp;#039; son representaciones vectoriales densas y de baja dimensionalidad de palabras en las que las palabras semanticamente similares se mapean a puntos cercanos en el espacio vectorial. Son un componente fundamental del procesamiento del lenguaje natural (PLN) moderno, reemplazando las codificaciones dispersas one-hot con representaciones que capturan significado, analogia y relaciones sintacticas.&lt;br /&gt;
&lt;br /&gt;
== La hipotesis distribucional ==&lt;br /&gt;
&lt;br /&gt;
Los word embeddings se fundamentan en la &amp;#039;&amp;#039;&amp;#039;hipotesis distribucional&amp;#039;&amp;#039;&amp;#039;, enunciada de forma celebre por J. R. Firth (1957): &amp;quot;Conoceras una palabra por la compania que mantiene.&amp;quot; La idea es que las palabras que aparecen en contextos similares tienden a tener significados similares. Por ejemplo, &amp;quot;perro&amp;quot; y &amp;quot;gato&amp;quot; aparecen frecuentemente cerca de palabras como &amp;quot;mascota&amp;quot;, &amp;quot;pelo&amp;quot; y &amp;quot;veterinario&amp;quot;, por lo que deberian tener representaciones similares.&lt;br /&gt;
&lt;br /&gt;
Los enfoques tempranos para explotar la informacion distribucional incluyen matrices de coocurrencia, informacion mutua puntual (PMI) y analisis semantico latente (LSA). Los metodos modernos de word embeddings aprenden vectores densos directamente utilizando redes neuronales.&lt;br /&gt;
&lt;br /&gt;
== Representaciones one-hot vs. densas ==&lt;br /&gt;
&lt;br /&gt;
=== Codificacion one-hot ===&lt;br /&gt;
&lt;br /&gt;
En un vocabulario de &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt; palabras, un vector one-hot para la &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt;-esima palabra es un vector de &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt; dimensiones con un 1 en la posicion &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; y 0 en el resto. Esta representacion tiene dos deficiencias criticas:&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Dimensionalidad&amp;#039;&amp;#039;&amp;#039; — los vectores son de dimension extremadamente alta (tipicamente &amp;lt;math&amp;gt;V &amp;gt; 100{,}000&amp;lt;/math&amp;gt;).&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sin similitud&amp;#039;&amp;#039;&amp;#039; — cada par de vectores one-hot es igualmente distante: &amp;lt;math&amp;gt;\mathbf{e}_i^\top \mathbf{e}_j = 0&amp;lt;/math&amp;gt; para &amp;lt;math&amp;gt;i \neq j&amp;lt;/math&amp;gt;. &amp;quot;Gato&amp;quot; esta tan lejos de &amp;quot;perro&amp;quot; como lo esta de &amp;quot;democracia.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
=== Embeddings densos ===&lt;br /&gt;
&lt;br /&gt;
Un word embedding mapea cada palabra a un vector de valores reales de &amp;lt;math&amp;gt;d&amp;lt;/math&amp;gt; dimensiones (tipicamente &amp;lt;math&amp;gt;d = 100&amp;lt;/math&amp;gt;–&amp;lt;math&amp;gt;300&amp;lt;/math&amp;gt;):&lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt;\mathbf{w}_i \in \mathbb{R}^d, \quad d \ll V&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Las palabras similares tienen una alta similitud coseno:&lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt;\text{sim}(\mathbf{w}_a, \mathbf{w}_b) = \frac{\mathbf{w}_a \cdot \mathbf{w}_b}{\|\mathbf{w}_a\|\;\|\mathbf{w}_b\|}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Word2Vec ==&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Word2Vec&amp;#039;&amp;#039;&amp;#039; (Mikolov et al., 2013) introdujo dos arquitecturas eficientes para aprender word embeddings a partir de grandes corpus.&lt;br /&gt;
&lt;br /&gt;
=== Bolsa continua de palabras (CBOW) ===&lt;br /&gt;
&lt;br /&gt;
CBOW predice una palabra objetivo a partir de sus palabras de contexto circundantes. Dada una ventana de palabras de contexto &amp;lt;math&amp;gt;\{w_{t-c}, \ldots, w_{t-1}, w_{t+1}, \ldots, w_{t+c}\}&amp;lt;/math&amp;gt;, el modelo maximiza:&lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt;P(w_t \mid w_{t-c}, \ldots, w_{t+c})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Los vectores de las palabras de contexto se promedian y se pasan a traves de una capa softmax. CBOW es mas rapido de entrenar y funciona bien para palabras frecuentes.&lt;br /&gt;
&lt;br /&gt;
=== Skip-gram ===&lt;br /&gt;
&lt;br /&gt;
Skip-gram invierte la prediccion: dada una palabra objetivo, predice las palabras de contexto circundantes. Para cada par &amp;lt;math&amp;gt;(w_t, w_{t+j})&amp;lt;/math&amp;gt; donde &amp;lt;math&amp;gt;j \in [-c, c] \setminus \{0\}&amp;lt;/math&amp;gt;, el modelo maximiza:&lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt;P(w_{t+j} \mid w_t) = \frac{\exp(\mathbf{v}&amp;#039;_{w_{t+j}}{}^\top \mathbf{v}_{w_t})}{\sum_{w=1}^{V}\exp(\mathbf{v}&amp;#039;_w{}^\top \mathbf{v}_{w_t})}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
donde &amp;lt;math&amp;gt;\mathbf{v}_w&amp;lt;/math&amp;gt; y &amp;lt;math&amp;gt;\mathbf{v}&amp;#039;_w&amp;lt;/math&amp;gt; son los vectores de embedding de entrada y salida. Calcular el softmax completo sobre el vocabulario es costoso, por lo que se utilizan dos aproximaciones comunes:&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Muestreo negativo&amp;#039;&amp;#039;&amp;#039; — en lugar de calcular el softmax completo, el modelo contrasta la palabra de contexto verdadera contra &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; palabras &amp;quot;negativas&amp;quot; muestreadas aleatoriamente.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Softmax jerarquico&amp;#039;&amp;#039;&amp;#039; — organiza el vocabulario en un arbol binario, reduciendo el coste del softmax de &amp;lt;math&amp;gt;O(V)&amp;lt;/math&amp;gt; a &amp;lt;math&amp;gt;O(\log V)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Skip-gram funciona bien con palabras infrecuentes y captura relaciones sutiles. La famosa analogia &amp;quot;rey - hombre + mujer ≈ reina&amp;quot; surgio de embeddings Skip-gram.&lt;br /&gt;
&lt;br /&gt;
== GloVe ==&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;GloVe&amp;#039;&amp;#039;&amp;#039; (Global Vectors, Pennington et al., 2014) combina las fortalezas de la factorizacion de matrices globales y los metodos de ventana de contexto local. Construye una matriz de coocurrencia de palabras &amp;lt;math&amp;gt;X&amp;lt;/math&amp;gt; a partir del corpus, donde &amp;lt;math&amp;gt;X_{ij}&amp;lt;/math&amp;gt; cuenta con que frecuencia la palabra &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; aparece en el contexto de la palabra &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt;, y luego optimiza:&lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt;J = \sum_{i,j=1}^{V} f(X_{ij})\bigl(\mathbf{w}_i^\top \tilde{\mathbf{w}}_j + b_i + \tilde{b}_j - \log X_{ij}\bigr)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
donde &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; es una funcion de ponderacion que limita la influencia de coocurrencias muy frecuentes. Los embeddings de GloVe a menudo igualan o superan la calidad de Word2Vec, y el uso explicito de estadisticas globales puede mejorar el rendimiento en tareas de analogia.&lt;br /&gt;
&lt;br /&gt;
== fastText ==&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;fastText&amp;#039;&amp;#039;&amp;#039; (Bojanowski et al., 2017) extiende Word2Vec representando cada palabra como una bolsa de n-gramas de caracteres. Por ejemplo, la palabra &amp;quot;donde&amp;quot; con &amp;lt;math&amp;gt;n = 3&amp;lt;/math&amp;gt; se representa por los n-gramas {&amp;quot;&amp;lt;do&amp;quot;, &amp;quot;don&amp;quot;, &amp;quot;ond&amp;quot;, &amp;quot;nde&amp;quot;, &amp;quot;de&amp;gt;&amp;quot;} mas la palabra completa &amp;quot;&amp;lt;donde&amp;gt;&amp;quot;. El embedding de una palabra es la suma de sus vectores de n-gramas.&lt;br /&gt;
&lt;br /&gt;
Este enfoque tiene dos ventajas clave:&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Manejo de palabras raras y no vistas&amp;#039;&amp;#039;&amp;#039; — incluso las palabras que no estan en el vocabulario de entrenamiento pueden recibir embeddings al sumar sus vectores de n-gramas de caracteres.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Conciencia morfologica&amp;#039;&amp;#039;&amp;#039; — las palabras que comparten subcadenas (por ejemplo, &amp;quot;ensenar&amp;quot;, &amp;quot;ensenanza&amp;quot;, &amp;quot;ensenante&amp;quot;) comparten automaticamente componentes del embedding.&lt;br /&gt;
&lt;br /&gt;
== Evaluacion de embeddings ==&lt;br /&gt;
&lt;br /&gt;
Los word embeddings se evaluan mediante:&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Tipo de evaluacion !! Ejemplos !! Que mide&lt;br /&gt;
|-&lt;br /&gt;
| &amp;#039;&amp;#039;&amp;#039;Intrinseca: analogia&amp;#039;&amp;#039;&amp;#039; || &amp;quot;rey : reina :: hombre : ?&amp;quot; || Estructura lineal del espacio&lt;br /&gt;
|-&lt;br /&gt;
| &amp;#039;&amp;#039;&amp;#039;Intrinseca: similitud&amp;#039;&amp;#039;&amp;#039; || Correlacion con juicios de similitud humanos (SimLex-999, WS-353) || Calidad semantica&lt;br /&gt;
|-&lt;br /&gt;
| &amp;#039;&amp;#039;&amp;#039;Extrinseca: tarea posterior&amp;#039;&amp;#039;&amp;#039; || Reconocimiento de entidades nombradas, analisis de sentimiento, parsing || Utilidad practica&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
Las evaluaciones intrinsecas son rapidas pero no siempre predicen el rendimiento en tareas posteriores. La evaluacion extrinseca en la tarea objetivo es, en ultima instancia, la medida mas fiable.&lt;br /&gt;
&lt;br /&gt;
== Embeddings contextuales ==&lt;br /&gt;
&lt;br /&gt;
Los word embeddings tradicionales asignan un unico vector por palabra independientemente del contexto — la palabra &amp;quot;banco&amp;quot; tiene el mismo embedding ya sea que se refiera a un banco de rio o a una institucion financiera. Los &amp;#039;&amp;#039;&amp;#039;embeddings contextuales&amp;#039;&amp;#039;&amp;#039; abordan esta limitacion produciendo representaciones diferentes segun el texto circundante.&lt;br /&gt;
&lt;br /&gt;
Los modelos de embeddings contextuales mas notables incluyen:&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;ELMo&amp;#039;&amp;#039;&amp;#039; (Peters et al., 2018) — utiliza un LSTM bidireccional para generar representaciones de palabras dependientes del contexto.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;BERT&amp;#039;&amp;#039;&amp;#039; (Devlin et al., 2019) — utiliza un codificador Transformer entrenado con modelado de lenguaje enmascarado.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Serie GPT&amp;#039;&amp;#039;&amp;#039; (Radford et al., 2018–) — utiliza un decodificador Transformer entrenado de forma autorregresiva.&lt;br /&gt;
&lt;br /&gt;
Estos modelos han reemplazado en gran medida a los embeddings estaticos para la mayoria de las tareas de PLN, aunque los embeddings estaticos siguen siendo utiles por su eficiencia, interpretabilidad y en entornos con recursos limitados.&lt;br /&gt;
&lt;br /&gt;
== Vease tambien ==&lt;br /&gt;
&lt;br /&gt;
* [[Neural Networks]]&lt;br /&gt;
* [[Recurrent Neural Networks]]&lt;br /&gt;
* [[Loss Functions]]&lt;br /&gt;
* [[Backpropagation]]&lt;br /&gt;
&lt;br /&gt;
== Referencias ==&lt;br /&gt;
&lt;br /&gt;
* Firth, J. R. (1957). &amp;quot;A synopsis of linguistic theory, 1930–1955&amp;quot;. In &amp;#039;&amp;#039;Studies in Linguistic Analysis&amp;#039;&amp;#039;.&lt;br /&gt;
* Mikolov, T. et al. (2013). &amp;quot;Efficient Estimation of Word Representations in Vector Space&amp;quot;. &amp;#039;&amp;#039;arXiv:1301.3781&amp;#039;&amp;#039;.&lt;br /&gt;
* Pennington, J., Socher, R. and Manning, C. D. (2014). &amp;quot;GloVe: Global Vectors for Word Representation&amp;quot;. &amp;#039;&amp;#039;EMNLP&amp;#039;&amp;#039;.&lt;br /&gt;
* Bojanowski, P. et al. (2017). &amp;quot;Enriching Word Vectors with Subword Information&amp;quot;. &amp;#039;&amp;#039;TACL&amp;#039;&amp;#039;, 5, 135–146.&lt;br /&gt;
* Peters, M. E. et al. (2018). &amp;quot;Deep contextualized word representations&amp;quot;. &amp;#039;&amp;#039;NAACL&amp;#039;&amp;#039;.&lt;br /&gt;
* Devlin, J. et al. (2019). &amp;quot;BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding&amp;quot;. &amp;#039;&amp;#039;NAACL&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
[[Category:NLP]]&lt;br /&gt;
[[Category:Intermediate]]&lt;/div&gt;</summary>
		<author><name>DeployBot</name></author>
	</entry>
</feed>