Skip to content

Commit a4548b0

Browse files
committed
Updated on 2025-01-16
1 parent 20b7cd2 commit a4548b0

File tree

3 files changed

+41
-3
lines changed

3 files changed

+41
-3
lines changed

index.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -64,7 +64,7 @@ <h1>Where?</h1>
6464
</section>
6565
<section>
6666
<h1>When?</h1>
67-
Last time this was edited was 2025-01-11 (YYYY/MM/DD).
67+
Last time this was edited was 2025-01-16 (YYYY/MM/DD).
6868
</section>
6969
<footer>
7070
<small><a href="misc.html">misc</a></small>

papers/list.json

Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,22 @@
11
[
2+
{
3+
"title": "Think Before You Speak: Training Language Models with Pause Tokens",
4+
"author": "Sachin Goyal et al",
5+
"year": "2024",
6+
"topic": "test-time compute, meta-tokens",
7+
"venue": "Arxiv",
8+
"description": "This paper introduces \"Pause Tokens\" which are a way of appending a sequence of tokens to the input prefix, and then delaying the output until the last pause token is seen.",
9+
"link": "https://arxiv.org/pdf/2310.02226"
10+
},
11+
{
12+
"title": "Scaling LLM Test-Time Compute Optimally can be More Effective than Scaling Model Parameters",
13+
"author": "Charlie Snell et al",
14+
"year": "2024",
15+
"topic": "test-time compute",
16+
"venue": "Arxiv",
17+
"description": "This paper explores the question of \"If an LLM is allowed to use a fixed but non-trivial amount of inference-time compute, how much can it improve its performance on a challenge prompt?\". Good for references on various test-time compute strategies.",
18+
"link": "https://arxiv.org/pdf/2408.03314"
19+
},
220
{
321
"title": "Medusa: Simple LLM Inference Acceleration Framework with Multiple Decoding Heads",
422
"author": "Tianle Cai et al",

papers_read.html

Lines changed: 22 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -16,10 +16,10 @@ <h1>Here's where I keep a list of papers I have read.</h1>
1616
I typically use this to organize papers I found interesting. Please feel free to do whatever you want with it. Note that this is not every single paper I have ever read, just a collection of ones that I remember to put down.
1717
</p>
1818
<p id="paperCount">
19-
So far, we have read 206 papers. Let's keep it up!
19+
So far, we have read 208 papers. Let's keep it up!
2020
</p>
2121
<small id="searchCount">
22-
Your search returned 206 papers. Nice!
22+
Your search returned 208 papers. Nice!
2323
</small>
2424

2525
<div class="search-inputs">
@@ -46,6 +46,26 @@ <h1>Here's where I keep a list of papers I have read.</h1>
4646
</thead>
4747
<tbody>
4848

49+
<tr>
50+
<td>Think Before You Speak: Training Language Models with Pause Tokens</td>
51+
<td>Sachin Goyal et al</td>
52+
<td>2024</td>
53+
<td>test-time compute, meta-tokens</td>
54+
<td>Arxiv</td>
55+
<td>This paper introduces &quot;Pause Tokens&quot; which are a way of appending a sequence of tokens to the input prefix, and then delaying the output until the last pause token is seen.</td>
56+
<td><a href="https://arxiv.org/pdf/2310.02226" target="_blank">Link</a></td>
57+
</tr>
58+
59+
<tr>
60+
<td>Scaling LLM Test-Time Compute Optimally can be More Effective than Scaling Model Parameters</td>
61+
<td>Charlie Snell et al</td>
62+
<td>2024</td>
63+
<td>test-time compute</td>
64+
<td>Arxiv</td>
65+
<td>This paper explores the question of &quot;If an LLM is allowed to use a fixed but non-trivial amount of inference-time compute, how much can it improve its performance on a challenge prompt?&quot;. Good for references on various test-time compute strategies.</td>
66+
<td><a href="https://arxiv.org/pdf/2408.03314" target="_blank">Link</a></td>
67+
</tr>
68+
4969
<tr>
5070
<td>Medusa: Simple LLM Inference Acceleration Framework with Multiple Decoding Heads</td>
5171
<td>Tianle Cai et al</td>

0 commit comments

Comments
 (0)