Skip to content

Commit 4c2bbce

Browse files
committed
Updated on 2024-08-13
1 parent 6dfb939 commit 4c2bbce

File tree

4 files changed

+73
-5
lines changed

4 files changed

+73
-5
lines changed

index.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@ <h3>
3939
When?
4040
</h3>
4141
<p>
42-
Last time this was edited was 2024-08-08 (YYYY/MM/DD).
42+
Last time this was edited was 2024-08-13 (YYYY/MM/DD).
4343
</p>
4444
<small><a href="misc.html">misc</a></small>
4545
</body>

misc.html

Lines changed: 8 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -15,8 +15,13 @@ <h4>
1515
You have discovered the secret page!
1616
</h4>
1717
<p>
18-
You weren't supposed to be here! But as you are, welcome! Here is where I store some pictures that I have a good memory attached to.
19-
Some of the dates are guesses.
18+
You weren't suppose to be here! Here's where I store some things.
2019
</p>
20+
<h6>
21+
Quotes I like:
22+
<p>
23+
孫子: 故舉秋毫,不為多力;見日月,不為明目;聞雷霆,不為聰耳。
24+
</p>
25+
</h6>
2126
</body>
22-
</html>
27+
</html>

papers/list.json

Lines changed: 63 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,67 @@
11
[
2+
{
3+
"title": "SMASH: One-Shot Model Architecture Search through Hypernetworks",
4+
"author": "Tomer Volk et al",
5+
"year": "2023",
6+
"topic": "hypernetworks, multi-source adaptation, unseen domains, NLP",
7+
"venue": "EMNLP",
8+
"description": "The authors apply hypernets to unsupervised domain adaptation in NLP. They use example-based adaptation. The main idea is that they use an encoder-decoder to initially create the unique signatures from an input example, and then they embed it within the source domain's semantic space. The signature is then used by a hypernet to generate the task classifier's weights. The paper focuses on improving generalization to unseen domains by explicitly modeling the shared and domain specific characteristics of the input. To allow for parameter sharing, they propose modeling based on hypernets, which allow soft weight sharing. ",
9+
"link": "https://aclanthology.org/2023.findings-emnlp.610.pdf"
10+
},
11+
{
12+
"title": "Example-based Hypernetworks for Multi-source Adaptation to Unseen Domains",
13+
"author": "Tomer Volk et al",
14+
"year": "2023",
15+
"topic": "hypernetworks, multi-source adaptation, unseen domains, NLP",
16+
"venue": "EMNLP",
17+
"description": "The authors apply hypernets to unsupervised domain adaptation in NLP. They use example-based adaptation. The main idea is that they use an encoder-decoder to initially create the unique signatures from an input example, and then they embed it within the source domain's semantic space. The signature is then used by a hypernet to generate the task classifier's weights. The paper focuses on improving generalization to unseen domains by explicitly modeling the shared and domain specific characteristics of the input. To allow for parameter sharing, they propose modeling based on hypernets, which allow soft weight sharing. ",
18+
"link": "https://aclanthology.org/2023.findings-emnlp.610.pdf"
19+
},
20+
{
21+
"title": "Meta-Learning via Hypernetworks",
22+
"author": "Dominic Zhao et al",
23+
"year": "2020",
24+
"topic": "hypernetworks, meta-learning",
25+
"venue": "NeurIps Workshop",
26+
"description": "The authors propose a soft weight-sharing hypernet architecture that performs well on meta-learning tasks. A good paper to show efforts in meta-learning with regards to hypernets, and comparing them to SOTA methods like Model-Agnostic Meta-Learning (MAML).",
27+
"link": "https://neurips.cc/virtual/2020/20189"
28+
},
29+
{
30+
"title": "HyperDynamics: Meta-Learning Object and Agent Dynamics with Hypernetworks",
31+
"author": "Zhou Xian et al",
32+
"year": "2021",
33+
"topic": "hypernetworks, meta-learning, dynamics",
34+
"venue": "ICLR",
35+
"description": "The authors present a dynamics meta-learning framework which conditions on an agent's interations w/ env and (optionally) the visual input from it. From this, they can generate params of a neural dynamics model. The three modules they use are 1) an encoding module that encodes a few agent-env interations / agent's visual observations into a feature code, 2) a hypernet that conditions on the latent feature code to generate params of a dynamic model dedicated to this observed system, and 3) a target dynamics model that is made using the generated parameters, and takes input as a low-dim system state / agent action and outputs the prediction of next system state.",
36+
"link": "https://arxiv.org/pdf/2103.09439"
37+
},
38+
{
39+
"title": "Principled Weight Initialization for Hypernetworks",
40+
"author": "Oscar Chang et al",
41+
"year": "2020",
42+
"topic": "hypernetworks, weight initialization",
43+
"venue": "ICLR",
44+
"description": "Classical weight initialization techniques don't really work on hypernets, because they fail to produce weights for the mainnet in the correct scale. The authors derive formulas for hyperfan-out and hyperfan-in weight initialization, and show that it works well for the mainnet.",
45+
"link": "https://arxiv.org/pdf/2312.08399"
46+
},
47+
{
48+
"title": "Continual Learning with Hypernetworks",
49+
"author": "Johannes von Oswald et al",
50+
"year": "2020",
51+
"topic": "hypernetworks, continual learning, meta learning",
52+
"venue": "ICLR",
53+
"description": "The authors present a method of preventing catastrophic forgetting, by using task-conditioned hypernets (i.e., hypernets that generate weights of target model based on some task embedding). Thus, rather than memorizing many data characteristics, we can split the problem into just learning a single point per task, given the task embedding.",
54+
"link": "https://arxiv.org/pdf/1906.00695"
55+
},
56+
{
57+
"title": "Stochastic Hyperparameter Optimization through Hypernetworks",
58+
"author": "Jonathan Lorraine et al",
59+
"year": "2018",
60+
"topic": "hypernetworks, hyperparameters",
61+
"venue": "ICLR",
62+
"description": "Using hypernetworks to learn hyperparameters. They replace the training optimization loop in favor of a differentiable hypernetwork to allow for tuning of hyperparameters using grad descent.",
63+
"link": "https://arxiv.org/pdf/1802.09419"
64+
},
265
{
366
"title": "Playing Atari with Deep Reinforcement Learning",
467
"author": "Volodymyr Mnih et al",

papers_read.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -33,7 +33,7 @@ <h1>Here's where I keep a list of papers I have read.</h1>
3333
This list was curated by <a href="index.html">Lexington Whalen</a>, beginning from his first year of PhD to end. As he is me, I hope he keeps going!
3434
</p>
3535
<p>
36-
I typically use this to organize papers I found interesting. Please feel free to do whatever you want with it.
36+
I typically use this to organize papers I found interesting. Please feel free to do whatever you want with it. Note that this is not every single paper I have ever read, just a collection of ones that I remember to put down.
3737
</p>
3838
<p id="paperCount">
3939
So far, we have read 0 papers. Let's keep it up!

0 commit comments

Comments
 (0)