Skip to content

Commit a8148c5

Browse files
committed
Updated on 2024-10-28
1 parent 05637bd commit a8148c5

File tree

2 files changed

+12
-12
lines changed

2 files changed

+12
-12
lines changed

papers/list.json

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,12 @@
11
[
22
{
3-
"title": "Simple Statistical Gradient-Following Algorithms for Connectionist Reinforcement Learning",
4-
"author": "Ronald J Williams et al",
3+
"title": "Improved Precision and Recall Metric for Assessing Generative Models",
4+
"author": "Tuomas Kynkaanniemi et al",
55
"year": "1192",
6-
"topic": "reinforcement learning",
7-
"venue": "Machine Learning",
8-
"description": "This paper introduces REINFORCE algorithms, a family of reinforcement learning algorithms for stochastic connectionist networks that perform gradient ascent on expected reinforcement without explicitly computing gradient estimates. The core contribution is the weight update rule Δwij = αij(r - bij)eij, where eij = ∂ln gi/∂wij represents characteristic eligibility, which naturally guides the network toward maximizing expected rewards in both immediate-reinforcement tasks and certain delayed-reinforcement scenarios. The paper demonstrates how these algorithms can be seamlessly integrated with backpropagation, extending their applicability to more complex neural architectures. Finally, the work provides theoretical guarantees about the algorithms' convergence properties and establishes connections to existing methods while introducing novel variants that could lead to more powerful reinforcement learning approaches.",
9-
"link": "https://people.cs.umass.edu/~barto/courses/cs687/williams92simple.pdf"
6+
"topic": "generative models, precision, recall",
7+
"venue": "NeurIPS 2019",
8+
"description": "This paper introduces an improved metric for evaluating generative models by separately measuring precision (quality of generated samples) and recall (coverage/diversity of generated distribution) using k-nearest neighbors to construct non-parametric manifold approximations of real and generated data distributions. The authors demonstrate their metric's effectiveness using StyleGAN and BigGAN, showing how it provides more nuanced insights than existing metrics like FID, particularly in revealing tradeoffs between image quality and variation that other metrics obscure. They use their metric to analyze and improve StyleGAN's architecture and training configurations, identifying new variants that achieve state-of-the-art results, and perform the first principled analysis of truncation methods. Finally, they extend their metric to evaluate individual sample quality, enabling quality assessment of interpolations and providing insights into the shape of the latent space that produces realistic images.",
9+
"link": ""
1010
},
1111
{
1212
"title": "Generative Pretraining from Pixels",

papers_read.html

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -106,13 +106,13 @@ <h1>Here's where I keep a list of papers I have read.</h1>
106106
<tbody>
107107

108108
<tr>
109-
<td>Simple Statistical Gradient-Following Algorithms for Connectionist Reinforcement Learning</td>
110-
<td>Ronald J Williams et al</td>
109+
<td>Improved Precision and Recall Metric for Assessing Generative Models</td>
110+
<td>Tuomas Kynkaanniemi et al</td>
111111
<td>1192</td>
112-
<td>reinforcement learning</td>
113-
<td>Machine Learning</td>
114-
<td>This paper introduces REINFORCE algorithms, a family of reinforcement learning algorithms for stochastic connectionist networks that perform gradient ascent on expected reinforcement without explicitly computing gradient estimates. The core contribution is the weight update rule Δwij = αij(r - bij)eij, where eij = ∂ln gi/∂wij represents characteristic eligibility, which naturally guides the network toward maximizing expected rewards in both immediate-reinforcement tasks and certain delayed-reinforcement scenarios. The paper demonstrates how these algorithms can be seamlessly integrated with backpropagation, extending their applicability to more complex neural architectures. Finally, the work provides theoretical guarantees about the algorithms&#x27; convergence properties and establishes connections to existing methods while introducing novel variants that could lead to more powerful reinforcement learning approaches.</td>
115-
<td><a href="https://people.cs.umass.edu/~barto/courses/cs687/williams92simple.pdf" target="_blank">Link</a></td>
112+
<td>generative models, precision, recall</td>
113+
<td>NeurIPS 2019</td>
114+
<td>This paper introduces an improved metric for evaluating generative models by separately measuring precision (quality of generated samples) and recall (coverage/diversity of generated distribution) using k-nearest neighbors to construct non-parametric manifold approximations of real and generated data distributions. The authors demonstrate their metric&#x27;s effectiveness using StyleGAN and BigGAN, showing how it provides more nuanced insights than existing metrics like FID, particularly in revealing tradeoffs between image quality and variation that other metrics obscure. They use their metric to analyze and improve StyleGAN&#x27;s architecture and training configurations, identifying new variants that achieve state-of-the-art results, and perform the first principled analysis of truncation methods. Finally, they extend their metric to evaluate individual sample quality, enabling quality assessment of interpolations and providing insights into the shape of the latent space that produces realistic images.</td>
115+
<td>N/A</td>
116116
</tr>
117117

118118
<tr>

0 commit comments

Comments
 (0)