Skip to content

Commit 5dcaefe

Browse files
committed
Updated on 2024-11-25
1 parent edf963b commit 5dcaefe

File tree

3 files changed

+41
-3
lines changed

3 files changed

+41
-3
lines changed

index.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -74,7 +74,7 @@ <h1>Where?</h1>
7474
</p>
7575
<h1>When?</h1>
7676
<p>
77-
Last time this was edited was 2024-11-23 (YYYY/MM/DD).
77+
Last time this was edited was 2024-11-25 (YYYY/MM/DD).
7878
</p>
7979
<small><a href="misc.html">misc</a></small>
8080
</div>

papers/list.json

Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,22 @@
11
[
2+
{
3+
"title": "Finding needles in a haystack: A Black-Box Approach to Invisible Watermark Detection",
4+
"author": "Minzhou Pan, et al",
5+
"year": "2024",
6+
"topic": "watermark, offset learning",
7+
"venue": "Arxiv",
8+
"description": "The key insight of this paper centers on using \"offset learning\" to detect invisible watermarks in images. The intuition is that by having a clean reference dataset of similar images, you can effectively \"cancel out\" the normal image features that are common between clean and watermarked images, leaving only the watermark perturbations. They design an asymmetric loss function where clean images use exponential/softmax loss (to focus on hard examples) while detection dataset uses linear loss (to give equal weight to all examples), helping isolate the watermark signal. This is combined with an iterative pruning strategy that gradually removes likely-clean images from the detection set, allowing the model to better focus on and learn the watermark patterns. By formulating watermark detection this way, they avoid needing any prior knowledge of watermarking techniques or labeled data, making it a truly black-box approach.",
9+
"link": "https://arxiv.org/pdf/2403.15955"
10+
},
11+
{
12+
"title": "Mitigating the Alignment Tax of RLHF",
13+
"author": "Yong Lin, et al",
14+
"year": "2024",
15+
"topic": "rlhf, alignment",
16+
"venue": "Arxiv",
17+
"description": "This paper investigates the \"alignment tax\" problem where large language models lose some of their pre-trained abilities when aligned with human preferences through RLHF. The key insight is that model averaging (interpolating between pre-RLHF and post-RLHF model weights) is surprisingly effective at mitigating this trade-off because tasks share overlapping feature spaces, particularly in lower layers of the model. Building on this understanding, they propose Heterogeneous Model Averaging (HMA) which applies different averaging ratios to different layers of the transformer model, allowing optimization of the alignment-forgetting trade-off. The intuition is that since different layers capture different levels of features and task similarities, they should not be averaged equally, and finding optimal layer-specific averaging ratios can better preserve both alignment and pre-trained capabilities.",
18+
"link": "https://arxiv.org/pdf/2309.06256"
19+
},
220
{
321
"title": "AsyncDiff: Parallelizing Diffusion Models by Asynchronous Denoising",
422
"author": "Zigeng Chen, et al",

papers_read.html

Lines changed: 22 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -75,10 +75,10 @@ <h1>Here's where I keep a list of papers I have read.</h1>
7575
I typically use this to organize papers I found interesting. Please feel free to do whatever you want with it. Note that this is not every single paper I have ever read, just a collection of ones that I remember to put down.
7676
</p>
7777
<p id="paperCount">
78-
So far, we have read 176 papers. Let's keep it up!
78+
So far, we have read 178 papers. Let's keep it up!
7979
</p>
8080
<small id="searchCount">
81-
Your search returned 176 papers. Nice!
81+
Your search returned 178 papers. Nice!
8282
</small>
8383

8484
<div class="search-inputs">
@@ -105,6 +105,26 @@ <h1>Here's where I keep a list of papers I have read.</h1>
105105
</thead>
106106
<tbody>
107107

108+
<tr>
109+
<td>Finding needles in a haystack: A Black-Box Approach to Invisible Watermark Detection</td>
110+
<td>Minzhou Pan, et al</td>
111+
<td>2024</td>
112+
<td>watermark, offset learning</td>
113+
<td>Arxiv</td>
114+
<td>The key insight of this paper centers on using &quot;offset learning&quot; to detect invisible watermarks in images. The intuition is that by having a clean reference dataset of similar images, you can effectively &quot;cancel out&quot; the normal image features that are common between clean and watermarked images, leaving only the watermark perturbations. They design an asymmetric loss function where clean images use exponential/softmax loss (to focus on hard examples) while detection dataset uses linear loss (to give equal weight to all examples), helping isolate the watermark signal. This is combined with an iterative pruning strategy that gradually removes likely-clean images from the detection set, allowing the model to better focus on and learn the watermark patterns. By formulating watermark detection this way, they avoid needing any prior knowledge of watermarking techniques or labeled data, making it a truly black-box approach.</td>
115+
<td><a href="https://arxiv.org/pdf/2403.15955" target="_blank">Link</a></td>
116+
</tr>
117+
118+
<tr>
119+
<td>Mitigating the Alignment Tax of RLHF</td>
120+
<td>Yong Lin, et al</td>
121+
<td>2024</td>
122+
<td>rlhf, alignment</td>
123+
<td>Arxiv</td>
124+
<td>This paper investigates the &quot;alignment tax&quot; problem where large language models lose some of their pre-trained abilities when aligned with human preferences through RLHF. The key insight is that model averaging (interpolating between pre-RLHF and post-RLHF model weights) is surprisingly effective at mitigating this trade-off because tasks share overlapping feature spaces, particularly in lower layers of the model. Building on this understanding, they propose Heterogeneous Model Averaging (HMA) which applies different averaging ratios to different layers of the transformer model, allowing optimization of the alignment-forgetting trade-off. The intuition is that since different layers capture different levels of features and task similarities, they should not be averaged equally, and finding optimal layer-specific averaging ratios can better preserve both alignment and pre-trained capabilities.</td>
125+
<td><a href="https://arxiv.org/pdf/2309.06256" target="_blank">Link</a></td>
126+
</tr>
127+
108128
<tr>
109129
<td>AsyncDiff: Parallelizing Diffusion Models by Asynchronous Denoising</td>
110130
<td>Zigeng Chen, et al</td>

0 commit comments

Comments
 (0)