You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: papers/list.json
+18Lines changed: 18 additions & 0 deletions
Original file line number
Diff line number
Diff line change
@@ -1,4 +1,22 @@
1
1
[
2
+
{
3
+
"title": "Think Before You Speak: Training Language Models with Pause Tokens",
4
+
"author": "Sachin Goyal et al",
5
+
"year": "2024",
6
+
"topic": "test-time compute, meta-tokens",
7
+
"venue": "Arxiv",
8
+
"description": "This paper introduces \"Pause Tokens\" which are a way of appending a sequence of tokens to the input prefix, and then delaying the output until the last pause token is seen.",
9
+
"link": "https://arxiv.org/pdf/2310.02226"
10
+
},
11
+
{
12
+
"title": "Scaling LLM Test-Time Compute Optimally can be More Effective than Scaling Model Parameters",
13
+
"author": "Charlie Snell et al",
14
+
"year": "2024",
15
+
"topic": "test-time compute",
16
+
"venue": "Arxiv",
17
+
"description": "This paper explores the question of \"If an LLM is allowed to use a fixed but non-trivial amount of inference-time compute, how much can it improve its performance on a challenge prompt?\". Good for references on various test-time compute strategies.",
Copy file name to clipboardExpand all lines: papers_read.html
+22-2Lines changed: 22 additions & 2 deletions
Original file line number
Diff line number
Diff line change
@@ -16,10 +16,10 @@ <h1>Here's where I keep a list of papers I have read.</h1>
16
16
I typically use this to organize papers I found interesting. Please feel free to do whatever you want with it. Note that this is not every single paper I have ever read, just a collection of ones that I remember to put down.
17
17
</p>
18
18
<pid="paperCount">
19
-
So far, we have read 206 papers. Let's keep it up!
19
+
So far, we have read 208 papers. Let's keep it up!
20
20
</p>
21
21
<smallid="searchCount">
22
-
Your search returned 206 papers. Nice!
22
+
Your search returned 208 papers. Nice!
23
23
</small>
24
24
25
25
<divclass="search-inputs">
@@ -46,6 +46,26 @@ <h1>Here's where I keep a list of papers I have read.</h1>
46
46
</thead>
47
47
<tbody>
48
48
49
+
<tr>
50
+
<td>Think Before You Speak: Training Language Models with Pause Tokens</td>
51
+
<td>Sachin Goyal et al</td>
52
+
<td>2024</td>
53
+
<td>test-time compute, meta-tokens</td>
54
+
<td>Arxiv</td>
55
+
<td>This paper introduces "Pause Tokens" which are a way of appending a sequence of tokens to the input prefix, and then delaying the output until the last pause token is seen.</td>
<td>Scaling LLM Test-Time Compute Optimally can be More Effective than Scaling Model Parameters</td>
61
+
<td>Charlie Snell et al</td>
62
+
<td>2024</td>
63
+
<td>test-time compute</td>
64
+
<td>Arxiv</td>
65
+
<td>This paper explores the question of "If an LLM is allowed to use a fixed but non-trivial amount of inference-time compute, how much can it improve its performance on a challenge prompt?". Good for references on various test-time compute strategies.</td>
0 commit comments