Skip to content

Commit 03c1f75

Browse files
committed
Updated on 2024-08-30
1 parent 0a66c03 commit 03c1f75

File tree

2 files changed

+10
-1
lines changed

2 files changed

+10
-1
lines changed

index.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@ <h3>
3939
When?
4040
</h3>
4141
<p>
42-
Last time this was edited was 2024-08-28 (YYYY/MM/DD).
42+
Last time this was edited was 2024-08-30 (YYYY/MM/DD).
4343
</p>
4444
<small><a href="misc.html">misc</a></small>
4545
</body>

papers/list.json

Lines changed: 9 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,14 @@
11
[
22

3+
{
4+
"title": "Learning both Weights and Connections for Efficient Neural Networks",
5+
"author": "Song Han et al",
6+
"year": "2015",
7+
"topic": "pruning, compression, regularization",
8+
"venue": "NeurIPS",
9+
"description": "The authors show a method of pruning neural networks in three steps: 1) train the network to learn what connections are important, 2) prune unimportant connections, 3) retrain and fine-tune. In order to train for learning what connections are important, they do not focus on learning the final weight values, but rather just focus on the importance of connections. They don't explicitly mention how this is done, but one could look at the Hessian of the loss or the magnitude of the weights. I'd imagine you could do this within only a few training iterations. In their \"Regularization\" section, it is interesting to note that L1 regularization (penalizes non-zero params resulting in more params near zero) gave better accuracy after pruning, but before retraining. But, these remaining connections are not as good as with using L2. The authors also present a discussion of what dropout rate to use.",
10+
"link": "https://arxiv.org/pdf/1506.02626"
11+
},
312
{
413
"title": "Quest: Query-Aware Sparsity for Efficient Long-Context LLM Inference",
514
"author": "Jiaming Tang et al",

0 commit comments

Comments
 (0)