Skip to content

Commit a01d200

Browse files
committed
Updated on 2024-09-06
1 parent c812dee commit a01d200

File tree

2 files changed

+58
-4
lines changed

2 files changed

+58
-4
lines changed

index.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@ <h3>
3939
When?
4040
</h3>
4141
<p>
42-
Last time this was edited was 2024-09-03 (YYYY/MM/DD).
42+
Last time this was edited was 2024-09-06 (YYYY/MM/DD).
4343
</p>
4444
<small><a href="misc.html">misc</a></small>
4545
</body>

papers/list.json

Lines changed: 57 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,63 @@
11
[
2+
{
3+
"title": "Diffusion Models Beat GANs on Image Synthesis",
4+
"author": "Prafulla Dhariwal et al",
5+
"year": "2021",
6+
"topic": "diffusion, gan",
7+
"venue": "Arxiv",
8+
"description": "This work demonstrates that diffusion models surpass the current state-of-the-art generative models in image quality, achieved through architecture improvements and classifier guidance, which balances diversity and fidelity. The model attains FID scores of 2.97 on ImageNet 128×128 and 4.59 on ImageNet 256×256, matching BigGAN-deep with as few as 25 forward passes while maintaining better distribution coverage. Additionally, combining classifier guidance with upsampling diffusion models further enhances FID scores to 3.94 on ImageNet 256×256 and 3.85 on ImageNet 512×512.",
9+
"link": "https://arxiv.org/pdf/2105.05233"
10+
},
11+
{
12+
"title": "Progressive Distillation for Fast Sampling of Diffusion Models",
13+
"author": "Tim Salimans et al",
14+
"year": "2022",
15+
"topic": "diffusion, distillation, sampling",
16+
"venue": "ICLR",
17+
"description": "Diffusion models excel in generative modeling, surpassing GANs in perceptual quality and autoregressive models in density estimation, but they suffer from slow sampling times. This paper introduces two key contributions: new parameterizations that improve stability with fewer sampling steps and a distillation method that progressively reduces the number of required steps by half each time. Applied to benchmarks like CIFAR-10 and ImageNet, the approach distills models from 8192 steps down to as few as 4 steps, maintaining high image quality while offering a more efficient solution for both training and inference.",
18+
"link": "https://arxiv.org/pdf/2202.00512"
19+
},
20+
{
21+
"title": "On Distillation of Guided Diffusion Models",
22+
"author": "Chenlin Meng et al",
23+
"year": "2023",
24+
"topic": "diffusion, classifier-free guidance",
25+
"venue": "Arxiv",
26+
"description": "Classifier-free guided diffusion models are effective for high-resolution image generation but are computationally expensive during inference due to the need to evaluate both conditional and unconditional models many times. This paper proposes a method to distill these models into faster ones by learning a single model that approximates the combined outputs, then progressively reducing the number of sampling steps. The approach significantly accelerates inference, generating images with comparable quality to the original model using as few as 1-4 denoising steps, achieving up to 256× speedup on datasets like ImageNet and LAION.",
27+
"link": "https://arxiv.org/pdf/2210.03142"
28+
},
29+
{
30+
"title": "Diffusion Probabilistic Models Made Slim",
31+
"author": "Xingyi Yang et al",
32+
"year": "2022",
33+
"topic": "diffusion, dpms, spectral diffusion",
34+
"venue": "Arxiv",
35+
"description": "Diffusion Probabilistic Models (DPMs) produce impressive visual results but suffer from high computational costs, limiting their use on resource-limited platforms. This paper introduces Spectral Diffusion (SD), a lightweight model designed to address DPMs' bias against high-frequency generation, which smaller networks struggle to capture. SD incorporates wavelet gating for frequency dynamics and spectrum-aware distillation to enhance high-frequency recovery, achieving 8-18× computational efficiency while maintaining competitive image fidelity.",
36+
"link": "https://arxiv.org/pdf/2211.17106"
37+
},
38+
{
39+
"title": "Structural Pruning for Diffusion Models",
40+
"author": "Gongfan Fang et al",
41+
"year": "2023",
42+
"topic": "diffusion, pruning",
43+
"venue": "NeurIPS",
44+
"description": "Generative modeling has advanced significantly with Diffusion Probabilistic Models (DPMs), but these models often require substantial computational resources. To address this, Diff-Pruning is introduced as a compression method that reduces the computational load by pruning unnecessary diffusion steps, using a Taylor expansion to identify key weights without extensive re-training. Empirical results show that Diff-Pruning can cut FLOPs by around 50%, while maintaining consistent generative performance at only 10-20% of the original training cost.",
45+
"link": "https://arxiv.org/pdf/2305.10924"
46+
},
47+
{
48+
"title": "Diffusion Models: A Comprehensive Survey of Methods and Applications",
49+
"author": "Ling Yang et al",
50+
"year": "2024",
51+
"topic": "diffusion, survey",
52+
"venue": "ACM",
53+
"description": "Diffusion models are a powerful class of deep generative models known for their success in tasks like image synthesis, video generation, and molecule design. This survey categorizes diffusion model research into efficient sampling, improved likelihood estimation, and handling specialized data structures, while also discussing the potential for combining them with other generative models. The review highlights their broad applications across fields such as computer vision, NLP, temporal data modeling, and interdisciplinary sciences, suggesting areas for further exploration.",
54+
"link": "https://arxiv.org/pdf/2209.00796"
55+
},
256
{
357
"title": "GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium",
458
"author": "Martin Heusel et al",
559
"year": "2017",
6-
"topic": "gan, equilibrium, FID, IS",
60+
"topic": "gan, equilibrium, fid, is",
761
"venue": "NeurIPS",
862
"description": "This paper introduces a two time-scale update rule (TTUR) for games, and proves that this makes GANs converge to a local Nash equilibrium. More cited is the FID score introduced here. FID improves on IS by comparing the distributions of real and generated images directly. This is done by using the Inception model to extract features from images and then assuming these features follow a multidimensional Gaussian distribution. FID measures the difference between the Gaussians (representing the real and generated images) using the Frechet distance, which effectively captures differences in the mean and covariance (the first two moments) of the distributions. FID makes sense as it directly compares the distributions of real and generated images by using the extracted features from Inception. These features are assumed to follow some multidimensional Gaussian, which simplifies the comparison. The Guassian is chosen as it is the maximum entropy distribution for a given mean and covariance (proof: https://medium.com/mathematical-musings/how-gaussian-distribution-maximizes-entropy-the-proof-7f7dcb2caf4d) -- maximum entropy is important, because this means that the Gaussian makes the fewest additional assumptions about the data, making sure the model is as non-committal as possible given the available information. Then, we calculate the statistics between the real and generated image features, like their mean and covariances. Finally, we compute the FID score using Frechet AKA Wasserstein-2 distance.",
963
"link": "https://arxiv.org/pdf/2212.09748"
@@ -12,7 +66,7 @@
1266
"title": "Scalable Diffusion Models with Transformers",
1367
"author": "William Peebles et al",
1468
"year": "2023",
15-
"topic": "Diffusion, DDPM, DiT",
69+
"topic": "diffusion,ddpm, dit",
1670
"venue": "CVPR",
1771
"description": "The authors explore using transformers in the latent space, rather than U-Nets. They find that their methods can lead to lower FID scores compared to prior SOTA. In this paper, their image generation pipeline is roughly: 1) Input high resolution image x 2) Encoder z = E(x), where E is a pre-trained frozen VAE encoder, and z is the latent representation 3) The DiT model operates on z 4) New latent representation z’ is sampled from the diffusion model 5) We then decode the z’ using the pre-trained frozen VAE decoder D, and x’ is now the generated high resolution image.",
1872
"link": "https://arxiv.org/pdf/2212.09748"
@@ -390,7 +444,7 @@
390444
"title": "1D convolution neural networks and applications: A survey",
391445
"author": "Serkan Kiranyaz et al",
392446
"year": "2020",
393-
"topic": "cnn",
447+
"topic": "cnn, survey",
394448
"venue": "Mechanical Systems and Signal Processing",
395449
"description": "A brief overview of applications of 1D CNNs is performed. It is largely focused on medicine (for instance, ECG) and fault detection (for instance, vibration based structural damage).",
396450
"link": "https://arxiv.org/pdf/1905.03554"

0 commit comments

Comments
 (0)