Skip to content

Commit 31629ea

Browse files
committed
Updated on 2024-09-07
1 parent a01d200 commit 31629ea

File tree

2 files changed

+19
-1
lines changed

2 files changed

+19
-1
lines changed

index.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@ <h3>
3939
When?
4040
</h3>
4141
<p>
42-
Last time this was edited was 2024-09-06 (YYYY/MM/DD).
42+
Last time this was edited was 2024-09-07 (YYYY/MM/DD).
4343
</p>
4444
<small><a href="misc.html">misc</a></small>
4545
</body>

papers/list.json

Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,22 @@
11
[
2+
{
3+
"title": "LLM Inference Unveiled: Survey and Roofline Model Insights",
4+
"author": "Roger Waleffe et al",
5+
"year": "2024",
6+
"topic": "llms, survey",
7+
"venue": "Arxiv",
8+
"description": "This paper surveys some recent advancements in LLC inference, like speculative decoding or operator fusion. They also analyze the findings using the Roofline model, which is likely the first paper to do such a thing for LLM inference. Good for checking out other papers that have recently been published.",
9+
"link": "https://arxiv.org/pdf/2402.16363"
10+
},
11+
{
12+
"title": "An Empircal Study of Mamba-based Language Models",
13+
"author": "Roger Waleffe et al",
14+
"year": "2024",
15+
"topic": "mamba, llms, transformer",
16+
"venue": "Arxiv",
17+
"description": "This paper compares Mamba-based, Transformer-based, and hybrid-based language models in a controlled setting where sizes and datasets are larger than the past (8B-params / 3.5T tokens). They find that Mamba and Mamba-2 lag behind Transformer models on copying and in-context learning tasks. They then see that a hybrid architecture of 43% Mamba, 7% self attention, and 50% MLP layers performs better than all others.",
18+
"link": "https://arxiv.org/pdf/2406.07887"
19+
},
220
{
321
"title": "Diffusion Models Beat GANs on Image Synthesis",
422
"author": "Prafulla Dhariwal et al",

0 commit comments

Comments
 (0)