<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://marovi.ai/index.php?action=history&amp;feed=atom&amp;title=Language_Modeling_with_Gated_Convolutional_Networks</id>
	<title>Language Modeling with Gated Convolutional Networks - Revision history</title>
	<link rel="self" type="application/atom+xml" href="https://marovi.ai/index.php?action=history&amp;feed=atom&amp;title=Language_Modeling_with_Gated_Convolutional_Networks"/>
	<link rel="alternate" type="text/html" href="https://marovi.ai/index.php?title=Language_Modeling_with_Gated_Convolutional_Networks&amp;action=history"/>
	<updated>2026-04-27T14:37:01Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.39.1</generator>
	<entry>
		<id>https://marovi.ai/index.php?title=Language_Modeling_with_Gated_Convolutional_Networks&amp;diff=8720&amp;oldid=prev</id>
		<title>DeployBot: Marked this version for translation</title>
		<link rel="alternate" type="text/html" href="https://marovi.ai/index.php?title=Language_Modeling_with_Gated_Convolutional_Networks&amp;diff=8720&amp;oldid=prev"/>
		<updated>2026-04-27T06:50:12Z</updated>

		<summary type="html">&lt;p&gt;Marked this version for translation&lt;/p&gt;
&lt;table style=&quot;background-color: #fff; color: #202122;&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;Revision as of 06:50, 27 April 2026&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l20&quot;&gt;Line 20:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 20:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;#039;&amp;#039;&amp;#039;Language Modeling with Gated Convolutional Networks&amp;#039;&amp;#039;&amp;#039; is a 2016 paper by Yann N. Dauphin, Angela Fan, Michael Auli, and David Grangier of Facebook AI Research that introduces the gated convolutional neural network (GCNN) for language modeling and the gated linear unit (GLU) activation. The paper challenges the prevailing assumption that recurrent networks are necessary to achieve state-of-the-art perplexity on large-scale language modeling benchmarks, showing that a finite-context, parallelizable convolutional stack equipped with multiplicative gating can match or exceed strong LSTM baselines while running an order of magnitude faster at inference. It was published at ICML 2017.&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;#039;&amp;#039;&amp;#039;Language Modeling with Gated Convolutional Networks&amp;#039;&amp;#039;&amp;#039; is a 2016 paper by Yann N. Dauphin, Angela Fan, Michael Auli, and David Grangier of Facebook AI Research that introduces the gated convolutional neural network (GCNN) for language modeling and the gated linear unit (GLU) activation. The paper challenges the prevailing assumption that recurrent networks are necessary to achieve state-of-the-art perplexity on large-scale language modeling benchmarks, showing that a finite-context, parallelizable convolutional stack equipped with multiplicative gating can match or exceed strong LSTM baselines while running an order of magnitude faster at inference. It was published at ICML 2017.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--T:2--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;== Overview == &lt;/ins&gt;&amp;lt;!--T:2--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;== Overview ==&lt;/del&gt;&lt;/div&gt;&lt;/td&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-side-added&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--T:3--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--T:3--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l29&quot;&gt;Line 29:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 28:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;The contribution is twofold. First, an architecture: stacked causal 1-D convolutions arranged into pre-activation residual bottleneck blocks and capped with an adaptive softmax output layer. Second, an activation function: the gated linear unit, which preserves a linear gradient path while retaining gating&amp;#039;s ability to modulate information flow. The combination yields convergence and final perplexity that compare favourably with strongly tuned LSTM baselines, and it does so with markedly better inference latency.&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;The contribution is twofold. First, an architecture: stacked causal 1-D convolutions arranged into pre-activation residual bottleneck blocks and capped with an adaptive softmax output layer. Second, an activation function: the gated linear unit, which preserves a linear gradient path while retaining gating&amp;#039;s ability to modulate information flow. The combination yields convergence and final perplexity that compare favourably with strongly tuned LSTM baselines, and it does so with markedly better inference latency.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--T:5--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;== Key Contributions == &lt;/ins&gt;&amp;lt;!--T:5--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;== Key Contributions ==&lt;/del&gt;&lt;/div&gt;&lt;/td&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-side-added&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--T:6--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--T:6--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l40&quot;&gt;Line 40:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 38:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;* Provides a controlled empirical comparison of gating mechanisms, showing GLU outperforms the tanh-based GTU of van den Oord et al. (2016), as well as plain ReLU and Tanh networks.&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;* Provides a controlled empirical comparison of gating mechanisms, showing GLU outperforms the tanh-based GTU of van den Oord et al. (2016), as well as plain ReLU and Tanh networks.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--T:7--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;== Methods == &lt;/ins&gt;&amp;lt;!--T:7--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;== Methods ==&lt;/del&gt;&lt;/div&gt;&lt;/td&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-side-added&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--T:8--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--T:8--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l64&quot;&gt;Line 64:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 61:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Architecturally each block contains up to five layers in a pre-activation bottleneck pattern (a wider &amp;lt;math&amp;gt;k&amp;gt;1&amp;lt;/math&amp;gt; convolution sandwiched between two &amp;lt;math&amp;gt;k=1&amp;lt;/math&amp;gt; projections) and is wrapped in a residual addition. Models in the paper range from 8 to 14 blocks, with hidden widths of 800–2048 and embedding sizes of 128–280. Training uses Nesterov momentum with momentum 0.99, gradient clipping at 0.1, weight normalization, Kaiming initialization, and learning rates between 1.0 and 2.0. The use of gradient clipping — usually motivated by recurrent gradient explosion — is justified here from a trust-region perspective and substantially speeds up training. Implementation is in Torch on Tesla M40 GPUs, with the largest models trained on 8 GPUs via synchronous data-parallel SGD.&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Architecturally each block contains up to five layers in a pre-activation bottleneck pattern (a wider &amp;lt;math&amp;gt;k&amp;gt;1&amp;lt;/math&amp;gt; convolution sandwiched between two &amp;lt;math&amp;gt;k=1&amp;lt;/math&amp;gt; projections) and is wrapped in a residual addition. Models in the paper range from 8 to 14 blocks, with hidden widths of 800–2048 and embedding sizes of 128–280. Training uses Nesterov momentum with momentum 0.99, gradient clipping at 0.1, weight normalization, Kaiming initialization, and learning rates between 1.0 and 2.0. The use of gradient clipping — usually motivated by recurrent gradient explosion — is justified here from a trust-region perspective and substantially speeds up training. Implementation is in Torch on Tesla M40 GPUs, with the largest models trained on 8 GPUs via synchronous data-parallel SGD.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--T:15--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;== Results == &lt;/ins&gt;&amp;lt;!--T:15--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;== Results ==&lt;/del&gt;&lt;/div&gt;&lt;/td&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-side-added&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--T:16--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--T:16--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l73&quot;&gt;Line 73:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 69:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;For computational efficiency, a GCNN-8 Bottleneck matches the throughput of a heavily cuDNN-optimized LSTM-2048 (about 45,800 tokens/s on GPU) at the same 43.9 perplexity operating point, while delivering 20× better responsiveness (per-token sequential latency) because each token can be evaluated independently rather than waiting on a recurrent hidden state. The ablation over gating mechanisms shows GLU converges faster and to lower perplexity than GTU, ReLU, or Tanh, and that learning-curve gaps between gated and ungated variants are large and consistent across both datasets. Increasing the receptive field beyond roughly 20 tokens of context yields diminishing returns, supporting the claim that finite contexts suffice for the bulk of practical language modeling.&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;For computational efficiency, a GCNN-8 Bottleneck matches the throughput of a heavily cuDNN-optimized LSTM-2048 (about 45,800 tokens/s on GPU) at the same 43.9 perplexity operating point, while delivering 20× better responsiveness (per-token sequential latency) because each token can be evaluated independently rather than waiting on a recurrent hidden state. The ablation over gating mechanisms shows GLU converges faster and to lower perplexity than GTU, ReLU, or Tanh, and that learning-curve gaps between gated and ungated variants are large and consistent across both datasets. Increasing the receptive field beyond roughly 20 tokens of context yields diminishing returns, supporting the claim that finite contexts suffice for the bulk of practical language modeling.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--T:18--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;== Impact == &lt;/ins&gt;&amp;lt;!--T:18--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;== Impact ==&lt;/del&gt;&lt;/div&gt;&lt;/td&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-side-added&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--T:19--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--T:19--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l85&quot;&gt;Line 85:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 80:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;A secondary methodological legacy is the paper&amp;#039;s clean separation of &amp;#039;&amp;#039;&amp;#039;throughput&amp;#039;&amp;#039;&amp;#039; (tokens/second under batching) from &amp;#039;&amp;#039;&amp;#039;responsiveness&amp;#039;&amp;#039;&amp;#039; (per-token sequential latency). This distinction has since become standard when evaluating sequence models for production deployment, where a model with the same training-time throughput as a baseline may nevertheless be unusable if it cannot decode a single sequence quickly enough.&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;A secondary methodological legacy is the paper&amp;#039;s clean separation of &amp;#039;&amp;#039;&amp;#039;throughput&amp;#039;&amp;#039;&amp;#039; (tokens/second under batching) from &amp;#039;&amp;#039;&amp;#039;responsiveness&amp;#039;&amp;#039;&amp;#039; (per-token sequential latency). This distinction has since become standard when evaluating sequence models for production deployment, where a model with the same training-time throughput as a baseline may nevertheless be unusable if it cannot decode a single sequence quickly enough.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--T:22--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;== See also == &lt;/ins&gt;&amp;lt;!--T:22--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;== See also ==&lt;/del&gt;&lt;/div&gt;&lt;/td&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-side-added&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--T:23--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--T:23--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l97&quot;&gt;Line 97:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 91:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;* [[WikiText-103]]&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;* [[WikiText-103]]&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--T:24--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;== References == &lt;/ins&gt;&amp;lt;!--T:24--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;== References ==&lt;/del&gt;&lt;/div&gt;&lt;/td&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-side-added&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--T:25--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&amp;lt;!--T:25--&amp;gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;

&lt;!-- diff cache key mediawiki:diff::1.12:old-8719:rev-8720 --&gt;
&lt;/table&gt;</summary>
		<author><name>DeployBot</name></author>
	</entry>
	<entry>
		<id>https://marovi.ai/index.php?title=Language_Modeling_with_Gated_Convolutional_Networks&amp;diff=8719&amp;oldid=prev</id>
		<title>DeployBot: [deploy-bot] Claude-authored from arxiv:1612.08083</title>
		<link rel="alternate" type="text/html" href="https://marovi.ai/index.php?title=Language_Modeling_with_Gated_Convolutional_Networks&amp;diff=8719&amp;oldid=prev"/>
		<updated>2026-04-27T06:50:12Z</updated>

		<summary type="html">&lt;p&gt;[deploy-bot] Claude-authored from arxiv:1612.08083&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;&amp;lt;languages /&amp;gt;&lt;br /&gt;
{{PaperTabs}}&lt;br /&gt;
{{PaperInfobox&lt;br /&gt;
 | topic_area  = NLP&lt;br /&gt;
 | difficulty  = Research&lt;br /&gt;
 | authors     = Yann N. Dauphin; Angela Fan; Michael Auli; David Grangier&lt;br /&gt;
 | year        = 2016&lt;br /&gt;
 | arxiv_id    = 1612.08083&lt;br /&gt;
 | source_url  = https://arxiv.org/abs/1612.08083&lt;br /&gt;
 | pdf_url     = https://arxiv.org/pdf/1612.08083.pdf&lt;br /&gt;
}}&lt;br /&gt;
{{ContentMeta&lt;br /&gt;
 | generated_by   = claude-code-direct&lt;br /&gt;
 | model_used     = claude-opus-4-7&lt;br /&gt;
 | generated_date = 2026-04-27&lt;br /&gt;
}}&lt;br /&gt;
&lt;br /&gt;
&amp;lt;translate&amp;gt;&lt;br /&gt;
&amp;lt;!--T:1--&amp;gt;&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Language Modeling with Gated Convolutional Networks&amp;#039;&amp;#039;&amp;#039; is a 2016 paper by Yann N. Dauphin, Angela Fan, Michael Auli, and David Grangier of Facebook AI Research that introduces the gated convolutional neural network (GCNN) for language modeling and the gated linear unit (GLU) activation. The paper challenges the prevailing assumption that recurrent networks are necessary to achieve state-of-the-art perplexity on large-scale language modeling benchmarks, showing that a finite-context, parallelizable convolutional stack equipped with multiplicative gating can match or exceed strong LSTM baselines while running an order of magnitude faster at inference. It was published at ICML 2017.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:2--&amp;gt;&lt;br /&gt;
== Overview ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:3--&amp;gt;&lt;br /&gt;
Statistical language models estimate the probability of a token sequence by factoring it into a product of next-word conditionals. Until 2016 the dominant neural approach used recurrent networks — typically LSTMs — whose strength was attributed to their unbounded effective context. The authors argue that this property is not strictly required: a sufficiently deep convolutional stack can represent contexts large enough for practical language modeling, and the absence of a temporal recurrence makes such a model trivially parallel over the time dimension.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:4--&amp;gt;&lt;br /&gt;
The contribution is twofold. First, an architecture: stacked causal 1-D convolutions arranged into pre-activation residual bottleneck blocks and capped with an adaptive softmax output layer. Second, an activation function: the gated linear unit, which preserves a linear gradient path while retaining gating&amp;#039;s ability to modulate information flow. The combination yields convergence and final perplexity that compare favourably with strongly tuned LSTM baselines, and it does so with markedly better inference latency.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:5--&amp;gt;&lt;br /&gt;
== Key Contributions ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:6--&amp;gt;&lt;br /&gt;
* Proposes the &amp;#039;&amp;#039;&amp;#039;Gated Linear Unit (GLU)&amp;#039;&amp;#039;&amp;#039;, a multiplicative activation in which one linear convolution is gated by the sigmoid of a parallel convolution, yielding a non-vanishing linear gradient path.&lt;br /&gt;
* Introduces the &amp;#039;&amp;#039;&amp;#039;Gated Convolutional Network (GCNN)&amp;#039;&amp;#039;&amp;#039; — stacked causal 1-D convolutions with residual bottleneck blocks and adaptive softmax — as the first non-recurrent model competitive with LSTMs on large-scale language modeling.&lt;br /&gt;
* Achieves a new single-model state of the art of 37.2 perplexity on &amp;#039;&amp;#039;&amp;#039;WikiText-103&amp;#039;&amp;#039;&amp;#039;, surpassing the previous LSTM-1024 baseline at 48.7.&lt;br /&gt;
* Establishes a new best single-GPU result on the &amp;#039;&amp;#039;&amp;#039;Google Billion Word&amp;#039;&amp;#039;&amp;#039; benchmark and reaches 31.9 perplexity with an 8-GPU configuration trained for 2 weeks (versus 30.6 perplexity for the LSTM of Jozefowicz et al. trained for 3 weeks on 32 GPUs).&lt;br /&gt;
* Demonstrates a &amp;#039;&amp;#039;&amp;#039;20× improvement in responsiveness&amp;#039;&amp;#039;&amp;#039; (sequential per-token latency) over a comparable LSTM, by exploiting the parallel structure of convolutions.&lt;br /&gt;
* Provides a controlled empirical comparison of gating mechanisms, showing GLU outperforms the tanh-based GTU of van den Oord et al. (2016), as well as plain ReLU and Tanh networks.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:7--&amp;gt;&lt;br /&gt;
== Methods ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:8--&amp;gt;&lt;br /&gt;
The model takes a sequence of word embeddings &amp;lt;math&amp;gt;\mathbf{E} = [\mathbf{D}_{w_0}, \ldots, \mathbf{D}_{w_N}]&amp;lt;/math&amp;gt; and feeds it through a stack of residual blocks, each computing the gated linear unit&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:9--&amp;gt;&lt;br /&gt;
:&amp;lt;math&amp;gt;h_l(\mathbf{X}) = (\mathbf{X} \ast \mathbf{W} + \mathbf{b}) \otimes \sigma(\mathbf{X} \ast \mathbf{V} + \mathbf{c})&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:10--&amp;gt;&lt;br /&gt;
where &amp;lt;math&amp;gt;\ast&amp;lt;/math&amp;gt; is 1-D convolution along the time axis, &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt; is the sigmoid, and &amp;lt;math&amp;gt;\otimes&amp;lt;/math&amp;gt; is element-wise multiplication. Causality is enforced by zero-padding the left of the input by &amp;lt;math&amp;gt;k-1&amp;lt;/math&amp;gt; positions so that the kernel never sees future tokens. The output of the stack is fed to an &amp;#039;&amp;#039;&amp;#039;adaptive softmax&amp;#039;&amp;#039;&amp;#039; that allocates more capacity to frequent words, dramatically reducing the cost of the output distribution for vocabularies of hundreds of thousands of types.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:11--&amp;gt;&lt;br /&gt;
The choice of activation is the central methodological contribution. The gradient of the GLU,&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:12--&amp;gt;&lt;br /&gt;
:&amp;lt;math&amp;gt;\nabla[\mathbf{X} \otimes \sigma(\mathbf{X})] = \nabla\mathbf{X} \otimes \sigma(\mathbf{X}) + \mathbf{X} \otimes \sigma&amp;#039;(\mathbf{X})\nabla\mathbf{X}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:13--&amp;gt;&lt;br /&gt;
contains an undamped term &amp;lt;math&amp;gt;\nabla\mathbf{X} \otimes \sigma(\mathbf{X})&amp;lt;/math&amp;gt; for active gating units, in contrast to the LSTM-style gated tanh unit (GTU) whose gradient is scaled by both &amp;lt;math&amp;gt;\tanh&amp;#039;&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\sigma&amp;#039;&amp;lt;/math&amp;gt; and therefore vanishes more rapidly with depth. The authors describe the GLU as a multiplicative skip connection: the network can still multiplicatively gate information flow, but the linear path keeps gradients well-conditioned in deep stacks.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:14--&amp;gt;&lt;br /&gt;
Architecturally each block contains up to five layers in a pre-activation bottleneck pattern (a wider &amp;lt;math&amp;gt;k&amp;gt;1&amp;lt;/math&amp;gt; convolution sandwiched between two &amp;lt;math&amp;gt;k=1&amp;lt;/math&amp;gt; projections) and is wrapped in a residual addition. Models in the paper range from 8 to 14 blocks, with hidden widths of 800–2048 and embedding sizes of 128–280. Training uses Nesterov momentum with momentum 0.99, gradient clipping at 0.1, weight normalization, Kaiming initialization, and learning rates between 1.0 and 2.0. The use of gradient clipping — usually motivated by recurrent gradient explosion — is justified here from a trust-region perspective and substantially speeds up training. Implementation is in Torch on Tesla M40 GPUs, with the largest models trained on 8 GPUs via synchronous data-parallel SGD.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:15--&amp;gt;&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:16--&amp;gt;&lt;br /&gt;
On the &amp;#039;&amp;#039;&amp;#039;Google Billion Word&amp;#039;&amp;#039;&amp;#039; benchmark, the GCNN-13 reaches 38.1 test perplexity on a single GPU, beating the comparable LSTM-2048 at 39.8. Scaled to 8 GPUs, the GCNN-14 Bottleneck reaches 31.9 perplexity, approaching the 30.6 of the much larger 2-layer LSTM-8192-1024 of Jozefowicz et al. while requiring roughly one third of the GPU-time. On &amp;#039;&amp;#039;&amp;#039;WikiText-103&amp;#039;&amp;#039;&amp;#039;, whose entries are full Wikipedia paragraphs averaging 4000 tokens, the GCNN-14 achieves 37.2 perplexity, a substantial improvement over the LSTM-1024 baseline at 48.7 and the first non-recurrent state of the art on this benchmark. The model also reaches 29.4 perplexity on &amp;#039;&amp;#039;&amp;#039;Gigaword&amp;#039;&amp;#039;&amp;#039; (versus 55.6 for a fully connected baseline) but underperforms on the small &amp;#039;&amp;#039;&amp;#039;Penn Treebank&amp;#039;&amp;#039;&amp;#039;, where the authors observe overfitting and conclude the architecture is better suited to large-scale problems.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:17--&amp;gt;&lt;br /&gt;
For computational efficiency, a GCNN-8 Bottleneck matches the throughput of a heavily cuDNN-optimized LSTM-2048 (about 45,800 tokens/s on GPU) at the same 43.9 perplexity operating point, while delivering 20× better responsiveness (per-token sequential latency) because each token can be evaluated independently rather than waiting on a recurrent hidden state. The ablation over gating mechanisms shows GLU converges faster and to lower perplexity than GTU, ReLU, or Tanh, and that learning-curve gaps between gated and ungated variants are large and consistent across both datasets. Increasing the receptive field beyond roughly 20 tokens of context yields diminishing returns, supporting the claim that finite contexts suffice for the bulk of practical language modeling.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:18--&amp;gt;&lt;br /&gt;
== Impact ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:19--&amp;gt;&lt;br /&gt;
The paper is a foundational reference for the broader shift away from purely recurrent sequence models. The gated linear unit it introduced is now a standard activation: GLU and its variants — particularly SwiGLU and GeGLU from the family analysed by Shazeer (2020) — are used in the feed-forward sublayers of large language models such as PaLM, LLaMA, and many open-source transformers, where they consistently outperform plain ReLU or GeLU on perplexity at fixed parameter count.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:20--&amp;gt;&lt;br /&gt;
The architectural argument that parallelizable, finite-context models can rival recurrent ones also helped clear the conceptual ground for the convolutional sequence-to-sequence work of Gehring et al. (2017) at the same lab and ultimately for the &amp;#039;&amp;#039;&amp;#039;Transformer&amp;#039;&amp;#039;&amp;#039; (Vaswani et al., 2017), which replaced both convolution and recurrence with self-attention while inheriting the parallelism argument and (in many later variants) the GLU activation. Within speech and translation pipelines the latency advantage was directly exploited by subsequent convolutional and gated systems before attention-based decoders became dominant.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:21--&amp;gt;&lt;br /&gt;
A secondary methodological legacy is the paper&amp;#039;s clean separation of &amp;#039;&amp;#039;&amp;#039;throughput&amp;#039;&amp;#039;&amp;#039; (tokens/second under batching) from &amp;#039;&amp;#039;&amp;#039;responsiveness&amp;#039;&amp;#039;&amp;#039; (per-token sequential latency). This distinction has since become standard when evaluating sequence models for production deployment, where a model with the same training-time throughput as a baseline may nevertheless be unusable if it cannot decode a single sequence quickly enough.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:22--&amp;gt;&lt;br /&gt;
== See also ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:23--&amp;gt;&lt;br /&gt;
* [[Long short-term memory]]&lt;br /&gt;
* [[Convolutional neural network]]&lt;br /&gt;
* [[Recurrent neural network]]&lt;br /&gt;
* [[Language model]]&lt;br /&gt;
* [[Transformer (machine learning model)]]&lt;br /&gt;
* [[Attention Is All You Need]]&lt;br /&gt;
* [[WikiText-103]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:24--&amp;gt;&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!--T:25--&amp;gt;&lt;br /&gt;
* Dauphin, Y. N.; Fan, A.; Auli, M.; Grangier, D. (2017). &amp;quot;Language Modeling with Gated Convolutional Networks&amp;quot;. &amp;#039;&amp;#039;Proceedings of the 34th International Conference on Machine Learning&amp;#039;&amp;#039; (ICML).&lt;br /&gt;
* Jozefowicz, R.; Vinyals, O.; Schuster, M.; Shazeer, N.; Wu, Y. (2016). &amp;quot;Exploring the Limits of Language Modeling&amp;quot;.&lt;br /&gt;
* Chelba, C. &amp;#039;&amp;#039;et al.&amp;#039;&amp;#039; (2013). &amp;quot;One Billion Word Benchmark for Measuring Progress in Statistical Language Modeling&amp;quot;.&lt;br /&gt;
* Merity, S.; Xiong, C.; Bradbury, J.; Socher, R. (2016). &amp;quot;Pointer Sentinel Mixture Models&amp;quot; (introduces the WikiText-103 corpus).&lt;br /&gt;
* van den Oord, A. &amp;#039;&amp;#039;et al.&amp;#039;&amp;#039; (2016). &amp;quot;Conditional Image Generation with PixelCNN Decoders&amp;quot; (the LSTM-style gated tanh unit baseline).&lt;br /&gt;
* Grave, E.; Joulin, A.; Cissé, M.; Grangier, D.; Jégou, H. (2017). &amp;quot;Efficient Softmax Approximation for GPUs&amp;quot; (adaptive softmax).&lt;br /&gt;
* Gehring, J.; Auli, M.; Grangier, D.; Yarats, D.; Dauphin, Y. N. (2017). &amp;quot;Convolutional Sequence to Sequence Learning&amp;quot;.&lt;br /&gt;
* Vaswani, A. &amp;#039;&amp;#039;et al.&amp;#039;&amp;#039; (2017). &amp;quot;Attention Is All You Need&amp;quot;.&lt;br /&gt;
* Shazeer, N. (2020). &amp;quot;GLU Variants Improve Transformer&amp;quot; (later analysis of GLU-family activations in transformer feed-forward layers).&lt;br /&gt;
* Hochreiter, S.; Schmidhuber, J. (1997). &amp;quot;Long Short-Term Memory&amp;quot;. &amp;#039;&amp;#039;Neural Computation&amp;#039;&amp;#039; 9(8): 1735–1780.&lt;br /&gt;
&amp;lt;/translate&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[Category:NLP]]&lt;br /&gt;
[[Category:Research]]&lt;br /&gt;
[[Category:Research Papers]]&lt;/div&gt;</summary>
		<author><name>DeployBot</name></author>
	</entry>
</feed>