Skip to content

Commit f745446

Browse files
committed
Updated on 2024-11-11
1 parent d625554 commit f745446

File tree

3 files changed

+22
-3
lines changed

3 files changed

+22
-3
lines changed

index.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -74,7 +74,7 @@ <h1>Where?</h1>
7474
</p>
7575
<h1>When?</h1>
7676
<p>
77-
Last time this was edited was 2024-11-10 (YYYY/MM/DD).
77+
Last time this was edited was 2024-11-11 (YYYY/MM/DD).
7878
</p>
7979
<small><a href="misc.html">misc</a></small>
8080
</div>

papers/list.json

Lines changed: 9 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,13 @@
11
[
2+
{
3+
"title": "One Weight Bitwidth to Rule Them All",
4+
"author": "Ting-Wu Chin et al",
5+
"year": "2020",
6+
"topic": "quantization, bitwidth",
7+
"venue": "Arxiv",
8+
"description": "This paper examines weight quantization in deep neural networks and challenges the common assumption that using the lowest possible bitwidth without accuracy loss is optimal. The key insight is that when considering model size as a constraint and allowing network width to vary, some bitwidths consistently outperform others - specifically, networks with standard convolutions work better with binary weights while networks with depthwise convolutions prefer higher bitwidths. The authors discover that this difference is related to the number of input channels (fan-in) per convolutional kernel, with higher fan-in making networks more resilient to aggressive quantization. Most surprisingly, they demonstrate that using a single well-chosen bitwidth throughout the network can outperform more complex mixed-precision quantization approaches when comparing networks of equal size, suggesting that the traditional focus on minimizing bitwidth without considering network width may be suboptimal.",
9+
"link": "https://arxiv.org/pdf/2008.09916"
10+
},
211
{
312
"title": "Consistency Models",
413
"author": "Yang Song et al",

papers_read.html

Lines changed: 12 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -75,10 +75,10 @@ <h1>Here's where I keep a list of papers I have read.</h1>
7575
I typically use this to organize papers I found interesting. Please feel free to do whatever you want with it. Note that this is not every single paper I have ever read, just a collection of ones that I remember to put down.
7676
</p>
7777
<p id="paperCount">
78-
So far, we have read 163 papers. Let's keep it up!
78+
So far, we have read 164 papers. Let's keep it up!
7979
</p>
8080
<small id="searchCount">
81-
Your search returned 163 papers. Nice!
81+
Your search returned 164 papers. Nice!
8282
</small>
8383

8484
<div class="search-inputs">
@@ -105,6 +105,16 @@ <h1>Here's where I keep a list of papers I have read.</h1>
105105
</thead>
106106
<tbody>
107107

108+
<tr>
109+
<td>One Weight Bitwidth to Rule Them All</td>
110+
<td>Ting-Wu Chin et al</td>
111+
<td>2020</td>
112+
<td>quantization, bitwidth</td>
113+
<td>Arxiv</td>
114+
<td>This paper examines weight quantization in deep neural networks and challenges the common assumption that using the lowest possible bitwidth without accuracy loss is optimal. The key insight is that when considering model size as a constraint and allowing network width to vary, some bitwidths consistently outperform others - specifically, networks with standard convolutions work better with binary weights while networks with depthwise convolutions prefer higher bitwidths. The authors discover that this difference is related to the number of input channels (fan-in) per convolutional kernel, with higher fan-in making networks more resilient to aggressive quantization. Most surprisingly, they demonstrate that using a single well-chosen bitwidth throughout the network can outperform more complex mixed-precision quantization approaches when comparing networks of equal size, suggesting that the traditional focus on minimizing bitwidth without considering network width may be suboptimal.</td>
115+
<td><a href="https://arxiv.org/pdf/2008.09916" target="_blank">Link</a></td>
116+
</tr>
117+
108118
<tr>
109119
<td>Consistency Models</td>
110120
<td>Yang Song et al</td>

0 commit comments

Comments
 (0)