Skip to content

Commit 818703c

Browse files
committed
Updated on 2024-08-20
1 parent f42e1e4 commit 818703c

File tree

2 files changed

+11
-1
lines changed

2 files changed

+11
-1
lines changed

index.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@ <h3>
3939
When?
4040
</h3>
4141
<p>
42-
Last time this was edited was 2024-08-18 (YYYY/MM/DD).
42+
Last time this was edited was 2024-08-20 (YYYY/MM/DD).
4343
</p>
4444
<small><a href="misc.html">misc</a></small>
4545
</body>

papers/list.json

Lines changed: 10 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,14 @@
11
[
2+
3+
{
4+
"title": "Estimating or Propagating Gradients Through Stochastic Neurons for Conditional Computation",
5+
"author": "Yoshua Bengio et al",
6+
"year": "2013",
7+
"topic": "gradients, stochasticy, backpropagation",
8+
"venue": "Arxiv",
9+
"description": "The authors introduce a several methods of estimation / propagation for networks that have stochastic neurons. This is used often in networks that are quantization-aware, as they sometimes have decision-boundaries in the neurons that are not differentiable regularly. The paper also introduces the \"Straight Through Estimator\", which was actually first introduced in one of Hinton's lectures. One interesting idea they present (that I think may have also been introduced in Kingma's VAE paper?) is that we can model the output h_{i} of some stochastic neuron as the application of a deterministic function that also depends on some noise source z_{i}: h_{i} = f(a_{i},z_{i}). TLDR: Straight through units are typically the go-to due to ease of use and good performance.",
10+
"link": "https://arxiv.org/pdf/1308.3432"
11+
},
212
{
313
"title": "DoReFaNet: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients",
414
"author": "Shuchang Zhou et al",

0 commit comments

Comments
 (0)