Skip to content

Commit 39244a5

Browse files
committed
Updated on 2025-03-04
1 parent b6df1cc commit 39244a5

File tree

2 files changed

+40
-2
lines changed

2 files changed

+40
-2
lines changed

papers/list.json

Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,22 @@
11
[
2+
{
3+
"title": "DORY: Deliberative Prompt Recovery for LLM",
4+
"author": "Lirong Gao et al",
5+
"year": "2024",
6+
"topic": "inversion",
7+
"venue": "Arxiv",
8+
"description": "This paper introduces a novel approach to recover original prompts from limited outputs of large language models. The authors discover a strong negative correlation between output probability-based uncertainty and prompt recovery success, showing that tokens with lower uncertainty are more likely to have appeared in the original prompt. Building on this insight, DORY recovers prompts through a three-step process: reconstructing a draft from output text, generating hints based on uncertainty, and reducing noise by comparing draft output with actual output. Unlike previous methods, DORY requires only a single LLM without any external resources or model training, making it a cost-effective solution for prompt recovery.",
9+
"link": "https://arxiv.org/pdf/2405.20657"
10+
},
11+
{
12+
"title": "Weak-to-Strong Reasoning",
13+
"author": "Yuqing Yang et al",
14+
"year": "2024",
15+
"topic": "reasoning",
16+
"venue": "Arxiv",
17+
"description": "This paper introduces a progressive learning framework for weak-to-strong reasoning, addressing the challenge of improving large language models (LLMs) without high-quality supervision. The authors demonstrate that naively fine-tuning a stronger model (like Llama2-70b) on outputs from weaker models (like Llama2-7b or Gemma-2b) is insufficient for complex reasoning tasks. Their proposed two-stage approach first uses selective data curation through a \"final answer consistency\" method to identify potentially correct examples, then applies preference optimization that enables the model to learn from contrasting examples. Experiments on mathematical reasoning datasets show substantial improvements over baseline approaches, with the framework proving particularly effective when the strong model learns to distinguish between correct and incorrect reasoning paths.",
18+
"link": "https://arxiv.org/pdf/2407.13647"
19+
},
220
{
321
"title": "The False Promise of Imitating Proprietary LLMs",
422
"author": "Arnav Gudibande et al",

papers_read.html

Lines changed: 22 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -16,10 +16,10 @@ <h1>Here's where I keep a list of papers I have read.</h1>
1616
I typically use this to organize papers I found interesting. Please feel free to do whatever you want with it. Note that this is not every single paper I have ever read, just a collection of ones that I remember to put down.
1717
</p>
1818
<p id="paperCount">
19-
So far, we have read 236 papers. Let's keep it up!
19+
So far, we have read 238 papers. Let's keep it up!
2020
</p>
2121
<small id="searchCount">
22-
Your search returned 236 papers. Nice!
22+
Your search returned 238 papers. Nice!
2323
</small>
2424

2525
<div class="search-inputs">
@@ -46,6 +46,26 @@ <h1>Here's where I keep a list of papers I have read.</h1>
4646
</thead>
4747
<tbody>
4848

49+
<tr>
50+
<td>DORY: Deliberative Prompt Recovery for LLM</td>
51+
<td>Lirong Gao et al</td>
52+
<td>2024</td>
53+
<td>inversion</td>
54+
<td>Arxiv</td>
55+
<td>This paper introduces a novel approach to recover original prompts from limited outputs of large language models. The authors discover a strong negative correlation between output probability-based uncertainty and prompt recovery success, showing that tokens with lower uncertainty are more likely to have appeared in the original prompt. Building on this insight, DORY recovers prompts through a three-step process: reconstructing a draft from output text, generating hints based on uncertainty, and reducing noise by comparing draft output with actual output. Unlike previous methods, DORY requires only a single LLM without any external resources or model training, making it a cost-effective solution for prompt recovery.</td>
56+
<td><a href="https://arxiv.org/pdf/2405.20657" target="_blank">Link</a></td>
57+
</tr>
58+
59+
<tr>
60+
<td>Weak-to-Strong Reasoning</td>
61+
<td>Yuqing Yang et al</td>
62+
<td>2024</td>
63+
<td>reasoning</td>
64+
<td>Arxiv</td>
65+
<td>This paper introduces a progressive learning framework for weak-to-strong reasoning, addressing the challenge of improving large language models (LLMs) without high-quality supervision. The authors demonstrate that naively fine-tuning a stronger model (like Llama2-70b) on outputs from weaker models (like Llama2-7b or Gemma-2b) is insufficient for complex reasoning tasks. Their proposed two-stage approach first uses selective data curation through a &quot;final answer consistency&quot; method to identify potentially correct examples, then applies preference optimization that enables the model to learn from contrasting examples. Experiments on mathematical reasoning datasets show substantial improvements over baseline approaches, with the framework proving particularly effective when the strong model learns to distinguish between correct and incorrect reasoning paths.</td>
66+
<td><a href="https://arxiv.org/pdf/2407.13647" target="_blank">Link</a></td>
67+
</tr>
68+
4969
<tr>
5070
<td>The False Promise of Imitating Proprietary LLMs</td>
5171
<td>Arnav Gudibande et al</td>

0 commit comments

Comments
 (0)