Skip to content

Commit c812dee

Browse files
committed
Updated on 2024-09-03
1 parent 1b0138f commit c812dee

File tree

1 file changed

+10
-2
lines changed

1 file changed

+10
-2
lines changed

papers/list.json

Lines changed: 10 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,13 @@
11
[
2-
2+
{
3+
"title": "GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium",
4+
"author": "Martin Heusel et al",
5+
"year": "2017",
6+
"topic": "gan, equilibrium, FID, IS",
7+
"venue": "NeurIPS",
8+
"description": "This paper introduces a two time-scale update rule (TTUR) for games, and proves that this makes GANs converge to a local Nash equilibrium. More cited is the FID score introduced here. FID improves on IS by comparing the distributions of real and generated images directly. This is done by using the Inception model to extract features from images and then assuming these features follow a multidimensional Gaussian distribution. FID measures the difference between the Gaussians (representing the real and generated images) using the Frechet distance, which effectively captures differences in the mean and covariance (the first two moments) of the distributions. FID makes sense as it directly compares the distributions of real and generated images by using the extracted features from Inception. These features are assumed to follow some multidimensional Gaussian, which simplifies the comparison. The Guassian is chosen as it is the maximum entropy distribution for a given mean and covariance (proof: https://medium.com/mathematical-musings/how-gaussian-distribution-maximizes-entropy-the-proof-7f7dcb2caf4d) -- maximum entropy is important, because this means that the Gaussian makes the fewest additional assumptions about the data, making sure the model is as non-committal as possible given the available information. Then, we calculate the statistics between the real and generated image features, like their mean and covariances. Finally, we compute the FID score using Frechet AKA Wasserstein-2 distance.",
9+
"link": "https://arxiv.org/pdf/2212.09748"
10+
},
311
{
412
"title": "Scalable Diffusion Models with Transformers",
513
"author": "William Peebles et al",
@@ -212,7 +220,7 @@
212220
"author": "Dominic Zhao et al",
213221
"year": "2020",
214222
"topic": "hypernetworks, meta-learning",
215-
"venue": "NeurIPS Workshop",
223+
"venue": "NeurIPS ",
216224
"description": "The authors propose a soft weight-sharing hypernet architecture that performs well on meta-learning tasks. A good paper to show efforts in meta-learning with regards to hypernets, and comparing them to SOTA methods like Model-Agnostic Meta-Learning (MAML).",
217225
"link": "https://neurips.cc/virtual/2020/20189"
218226
},

0 commit comments

Comments
 (0)