Skip to content

Commit 8839c06

Browse files
committed
Updated on 2024-10-26
1 parent 549ecb9 commit 8839c06

File tree

2 files changed

+21
-2
lines changed

2 files changed

+21
-2
lines changed

papers/list.json

Lines changed: 9 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,13 @@
11
[
2+
{
3+
"title": "Why Does Unsupervised Pre-Training Help Deep Learning?",
4+
"author": "Dumitru Erhan et al",
5+
"year": "2010",
6+
"topic": "pretraining, unsupervised",
7+
"venue": "JMLR",
8+
"description": "This paper argues that standard training schemes place parameters in regions of the parameter space that generalize poorly, while greedy layer-wise unsupervised pre-training allows each layer to learn a nonlinear transformation of its input that captures the main variations in the input, which acts as a regularizer: minimizing variance and introducing bias towards good initializations for the parameters. They argue that defining particular initialization points implicitly imposes constraints on the parameters in that it specifies which minima (out of many possible minima) of the cost function are allowed. They further argue that small perturbations in the trajectory of the parameters have a larger effect early on, and hint that early examples have larger influence and may trap model parameters in particular regions of parameter space corresponding to the arbitrary ordering of training examples (similar to the \"critical period\" in developmental psychology).",
9+
"link": "https://jmlr.org/papers/volume11/erhan10a/erhan10a.pdf"
10+
},
211
{
312
"title": "Improving Language Understanding by Generative Pre-Training",
413
"author": "Alec Radford et al",

papers_read.html

Lines changed: 12 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -75,10 +75,10 @@ <h1>Here's where I keep a list of papers I have read.</h1>
7575
I typically use this to organize papers I found interesting. Please feel free to do whatever you want with it. Note that this is not every single paper I have ever read, just a collection of ones that I remember to put down.
7676
</p>
7777
<p id="paperCount">
78-
So far, we have read 144 papers. Let's keep it up!
78+
So far, we have read 145 papers. Let's keep it up!
7979
</p>
8080
<small id="searchCount">
81-
Your search returned 144 papers. Nice!
81+
Your search returned 145 papers. Nice!
8282
</small>
8383

8484
<div class="search-inputs">
@@ -105,6 +105,16 @@ <h1>Here's where I keep a list of papers I have read.</h1>
105105
</thead>
106106
<tbody>
107107

108+
<tr>
109+
<td>Why Does Unsupervised Pre-Training Help Deep Learning?</td>
110+
<td>Dumitru Erhan et al</td>
111+
<td>2010</td>
112+
<td>pretraining, unsupervised</td>
113+
<td>JMLR</td>
114+
<td>This paper argues that standard training schemes place parameters in regions of the parameter space that generalize poorly, while greedy layer-wise unsupervised pre-training allows each layer to learn a nonlinear transformation of its input that captures the main variations in the input, which acts as a regularizer: minimizing variance and introducing bias towards good initializations for the parameters. They argue that defining particular initialization points implicitly imposes constraints on the parameters in that it specifies which minima (out of many possible minima) of the cost function are allowed. They further argue that small perturbations in the trajectory of the parameters have a larger effect early on, and hint that early examples have larger influence and may trap model parameters in particular regions of parameter space corresponding to the arbitrary ordering of training examples (similar to the &quot;critical period&quot; in developmental psychology).</td>
115+
<td><a href="https://jmlr.org/papers/volume11/erhan10a/erhan10a.pdf" target="_blank">Link</a></td>
116+
</tr>
117+
108118
<tr>
109119
<td>Improving Language Understanding by Generative Pre-Training</td>
110120
<td>Alec Radford et al</td>

0 commit comments

Comments
 (0)