<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://marovi.ai/index.php?action=history&amp;feed=atom&amp;title=Translations%3AStochastic_Gradient_Descent%2F25%2Fzh</id>
	<title>Translations:Stochastic Gradient Descent/25/zh - Revision history</title>
	<link rel="self" type="application/atom+xml" href="https://marovi.ai/index.php?action=history&amp;feed=atom&amp;title=Translations%3AStochastic_Gradient_Descent%2F25%2Fzh"/>
	<link rel="alternate" type="text/html" href="https://marovi.ai/index.php?title=Translations:Stochastic_Gradient_Descent/25/zh&amp;action=history"/>
	<updated>2026-04-27T22:02:11Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.39.1</generator>
	<entry>
		<id>https://marovi.ai/index.php?title=Translations:Stochastic_Gradient_Descent/25/zh&amp;diff=5470&amp;oldid=prev</id>
		<title>DeployBot: Batch translate Stochastic Gradient Descent unit 25 → zh</title>
		<link rel="alternate" type="text/html" href="https://marovi.ai/index.php?title=Translations:Stochastic_Gradient_Descent/25/zh&amp;diff=5470&amp;oldid=prev"/>
		<updated>2026-04-27T03:38:16Z</updated>

		<summary type="html">&lt;p&gt;Batch translate Stochastic Gradient Descent unit 25 → zh&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! 方法 !! 核心思想 !! 文献&lt;br /&gt;
|-&lt;br /&gt;
| &amp;#039;&amp;#039;&amp;#039;{{Term|momentum|Momentum}}&amp;#039;&amp;#039;&amp;#039; || 对历史梯度累积指数衰减的移动平均 || Polyak, 1964&lt;br /&gt;
|-&lt;br /&gt;
| &amp;#039;&amp;#039;&amp;#039;Nesterov 加速梯度&amp;#039;&amp;#039;&amp;#039; || 在“前瞻”位置上计算梯度 || Nesterov, 1983&lt;br /&gt;
|-&lt;br /&gt;
| &amp;#039;&amp;#039;&amp;#039;Adagrad&amp;#039;&amp;#039;&amp;#039; || 为每个参数设置学习率，对频繁更新的特征逐步减小 || Duchi et al., 2011&lt;br /&gt;
|-&lt;br /&gt;
| &amp;#039;&amp;#039;&amp;#039;RMSProp&amp;#039;&amp;#039;&amp;#039; || 利用平方梯度的移动平均修正 Adagrad 学习率不断衰减的问题 || Hinton（讲义），2012&lt;br /&gt;
|-&lt;br /&gt;
| &amp;#039;&amp;#039;&amp;#039;{{Term|Adam}}&amp;#039;&amp;#039;&amp;#039; || 将 {{Term|momentum}} 与 RMSProp 风格的自适应学习率结合 || Kingma 与 Ba, 2015&lt;br /&gt;
|-&lt;br /&gt;
| &amp;#039;&amp;#039;&amp;#039;AdamW&amp;#039;&amp;#039;&amp;#039; || 将权重衰减与自适应梯度更新解耦 || Loshchilov 与 Hutter, 2019&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>DeployBot</name></author>
	</entry>
</feed>