Skip to content

Commit 6bc378b

Browse files
committed
Updated on 2025-04-13
1 parent 9808d39 commit 6bc378b

File tree

2 files changed

+59
-2
lines changed

2 files changed

+59
-2
lines changed

papers/list.json

Lines changed: 27 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,31 @@
11
[
2+
{
3+
"title": "d1: Scaling Reasoning in Diffusion Large Language Models via Reinforcement Learning",
4+
"author": "Siyan Zhao et al",
5+
"year": "2025",
6+
"topic":"rl, diffusion, llm",
7+
"venue": "Arxiv",
8+
"description": "Zhao et al. introduce d1, a novel framework that enhances reasoning capabilities in diffusion Large Language Models (dLLMs) through a two-stage post-training approach combining supervised fine-tuning (SFT) and reinforcement learning. While traditional autoregressive models like Claude or GPT generate text left-to-right sequentially, diffusion models like LLaDA generate through iterative denoising from masked tokens in a non-autoregressive manner. The authors' key innovation is diffu-GRPO, a policy gradient algorithm specifically adapted for masked dLLMs that efficiently estimates log-probabilities using a one-step random prompt masking technique. This approach serves as a regularization mechanism, allowing more gradient updates per batch and reducing computational demands. Applied to LLaDA-8B-Instruct, the full d1 pipeline consistently outperforms both the base model and individual SFT or RL approaches across four mathematical and logical reasoning benchmarks, with particularly significant gains on complex tasks like Countdown (+21.5%) and Sudoku (+10.4%). The results demonstrate that masked dLLMs can benefit from RL-based reasoning enhancement techniques previously limited to autoregressive models, opening new directions for improving non-autoregressive language models.",
9+
"link": "https://arxiv.org/pdf/2211.16750"
10+
},
11+
{
12+
"title": "Score-Based Continuous-Time Discrete Diffusion Models",
13+
"author": "Haoran Sun et al",
14+
"year": "2023",
15+
"topic":"continuous, llm, diffusion, discrete",
16+
"venue": "ICLR",
17+
"description": "Sun et al. introduce a breakthrough approach extending score-based diffusion models to discrete categorical data through a continuous-time Markov chain formulation. Their key innovation is developing \"categorical ratio matching,\" a discrete analog to score matching that effectively characterizes how discrete distributions evolve over time without requiring gradients. This theoretical advancement enables the formulation of a coherent stochastic jump process for the reverse (denoising) process with an analytical sampling method as an alternative to numerical simulation. The authors further introduce three architectural approaches—energy-based models, masked models, and a novel \"hollow Transformer\"—to parameterize conditional distributions while avoiding information leakage issues. Their experiments on synthetic data, CIFAR10 images, and music generation demonstrate superior performance compared to previous discrete diffusion approaches, particularly when sampling with fewer steps, highlighting the flexibility advantages of the continuous-time formulation.",
18+
"link": "https://arxiv.org/pdf/2211.16750"
19+
},
20+
{
21+
"title": "A Continuous Time Framework for Discrete Denoising Models",
22+
"author": "Andrew Campbell et al",
23+
"year": "2022",
24+
"topic":"continuous, llm, diffusion, discrete",
25+
"venue": "NeurIPS",
26+
"description": "The paper \"A Continuous Time Framework for Discrete Denoising Models\" presents the first complete theoretical foundation for extending diffusion models to discrete data using continuous time. The authors formulate both the forward noising process and generative reverse process as Continuous Time Markov Chains (CTMCs), providing a mathematically rigorous alternative to previous discrete-time approaches. This continuous framework enables the derivation of a continuous time evidence lower bound (ELBO) for efficient training, and unlocks high-performance sampling strategies like tau-leaping—borrowed from chemical physics—which allows multiple dimensions to change simultaneously in a single simulation step. Particularly innovative is their predictor-corrector scheme that alternates between sampling steps guided by the reverse rate and corrector steps that push the distribution toward the true marginal, significantly improving sample quality on image generation tasks. The authors also contribute a novel theoretical error bound between generated and true data distributions, demonstrating how their approach maintains accuracy even in high-dimensional discrete spaces while offering greater flexibility than discrete time methods.",
27+
"link": "https://arxiv.org/pdf/2205.14987"
28+
},
229
{
330
"title": "PLANNER: Generating Diversified Paragraph via Latent Language Diffusion Model",
431
"author": "Yizhe Zhang et al",

papers_read.html

Lines changed: 32 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -16,10 +16,10 @@ <h1>Here's where I keep a list of papers I have read.</h1>
1616
I typically use this to organize papers I found interesting. Please feel free to do whatever you want with it. Note that this is not every single paper I have ever read, just a collection of ones that I remember to put down.
1717
</p>
1818
<p id="paperCount">
19-
So far, we have read 262 papers. Let's keep it up!
19+
So far, we have read 265 papers. Let's keep it up!
2020
</p>
2121
<small id="searchCount">
22-
Your search returned 262 papers. Nice!
22+
Your search returned 265 papers. Nice!
2323
</small>
2424

2525
<div class="search-inputs">
@@ -46,6 +46,36 @@ <h1>Here's where I keep a list of papers I have read.</h1>
4646
</thead>
4747
<tbody>
4848

49+
<tr>
50+
<td>d1: Scaling Reasoning in Diffusion Large Language Models via Reinforcement Learning</td>
51+
<td>Siyan Zhao et al</td>
52+
<td>2025</td>
53+
<td>rl, diffusion, llm</td>
54+
<td>Arxiv</td>
55+
<td>Zhao et al. introduce d1, a novel framework that enhances reasoning capabilities in diffusion Large Language Models (dLLMs) through a two-stage post-training approach combining supervised fine-tuning (SFT) and reinforcement learning. While traditional autoregressive models like Claude or GPT generate text left-to-right sequentially, diffusion models like LLaDA generate through iterative denoising from masked tokens in a non-autoregressive manner. The authors&#x27; key innovation is diffu-GRPO, a policy gradient algorithm specifically adapted for masked dLLMs that efficiently estimates log-probabilities using a one-step random prompt masking technique. This approach serves as a regularization mechanism, allowing more gradient updates per batch and reducing computational demands. Applied to LLaDA-8B-Instruct, the full d1 pipeline consistently outperforms both the base model and individual SFT or RL approaches across four mathematical and logical reasoning benchmarks, with particularly significant gains on complex tasks like Countdown (+21.5%) and Sudoku (+10.4%). The results demonstrate that masked dLLMs can benefit from RL-based reasoning enhancement techniques previously limited to autoregressive models, opening new directions for improving non-autoregressive language models.</td>
56+
<td><a href="https://arxiv.org/pdf/2211.16750" target="_blank">Link</a></td>
57+
</tr>
58+
59+
<tr>
60+
<td>Score-Based Continuous-Time Discrete Diffusion Models</td>
61+
<td>Haoran Sun et al</td>
62+
<td>2023</td>
63+
<td>continuous, llm, diffusion, discrete</td>
64+
<td>ICLR</td>
65+
<td>Sun et al. introduce a breakthrough approach extending score-based diffusion models to discrete categorical data through a continuous-time Markov chain formulation. Their key innovation is developing &quot;categorical ratio matching,&quot; a discrete analog to score matching that effectively characterizes how discrete distributions evolve over time without requiring gradients. This theoretical advancement enables the formulation of a coherent stochastic jump process for the reverse (denoising) process with an analytical sampling method as an alternative to numerical simulation. The authors further introduce three architectural approaches—energy-based models, masked models, and a novel &quot;hollow Transformer&quot;—to parameterize conditional distributions while avoiding information leakage issues. Their experiments on synthetic data, CIFAR10 images, and music generation demonstrate superior performance compared to previous discrete diffusion approaches, particularly when sampling with fewer steps, highlighting the flexibility advantages of the continuous-time formulation.</td>
66+
<td><a href="https://arxiv.org/pdf/2211.16750" target="_blank">Link</a></td>
67+
</tr>
68+
69+
<tr>
70+
<td>A Continuous Time Framework for Discrete Denoising Models</td>
71+
<td>Andrew Campbell et al</td>
72+
<td>2022</td>
73+
<td>continuous, llm, diffusion, discrete</td>
74+
<td>NeurIPS</td>
75+
<td>The paper &quot;A Continuous Time Framework for Discrete Denoising Models&quot; presents the first complete theoretical foundation for extending diffusion models to discrete data using continuous time. The authors formulate both the forward noising process and generative reverse process as Continuous Time Markov Chains (CTMCs), providing a mathematically rigorous alternative to previous discrete-time approaches. This continuous framework enables the derivation of a continuous time evidence lower bound (ELBO) for efficient training, and unlocks high-performance sampling strategies like tau-leaping—borrowed from chemical physics—which allows multiple dimensions to change simultaneously in a single simulation step. Particularly innovative is their predictor-corrector scheme that alternates between sampling steps guided by the reverse rate and corrector steps that push the distribution toward the true marginal, significantly improving sample quality on image generation tasks. The authors also contribute a novel theoretical error bound between generated and true data distributions, demonstrating how their approach maintains accuracy even in high-dimensional discrete spaces while offering greater flexibility than discrete time methods.</td>
76+
<td><a href="https://arxiv.org/pdf/2205.14987" target="_blank">Link</a></td>
77+
</tr>
78+
4979
<tr>
5080
<td>PLANNER: Generating Diversified Paragraph via Latent Language Diffusion Model</td>
5181
<td>Yizhe Zhang et al</td>

0 commit comments

Comments
 (0)