Skip to content

Commit a453a28

Browse files
committed
Updated on 2024-08-23
1 parent 7c194aa commit a453a28

File tree

1 file changed

+9
-0
lines changed

1 file changed

+9
-0
lines changed

papers/list.json

Lines changed: 9 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,13 @@
11
[
2+
{
3+
"title": "Learning to Compress Prompts with Gist Tokens",
4+
"author": "Jesse Mu et al",
5+
"year": "2023",
6+
"topic": "llms, prompting, compression, tokens",
7+
"venue": "NeurIPS",
8+
"description": "The authors describe a method of using a distilling function G (similar to a hypernet) that is able to compress LM prompts into a smaller set of \"gist\" tokens. These tokens can then be cached and reused. The neat trick is that they reuse the LM itself as G, so gisting itself incurs no additional training cost. Note that in their \"Failure Cases\" section, they mention \"... While it is unclear why only the gist models exhibit this behavior (i.e. the fail example behavior), these issues can likely be mitigated with more careful sampling techniques...",
9+
"link": "https://arxiv.org/pdf/2304.08467"
10+
},
211
{
312
"title": "Once-For-All: Train One Network and Specialize it For Efficient Deployment",
413
"author": "Han Cai et al",

0 commit comments

Comments
 (0)