[ENHANCEMENT] Dynamic binning + distance-aware label smoothing #251
Killer3048
started this conversation in
Ideas
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Dynamic binning + distance-aware label smoothing
Why this matters
[-15..+15]
for quantization after mean scaling. This can cause severe overflow (if the real data extends beyond +15) or underflow (if data is confined way below ±15) in zero-shot scenarios.chronos.py
, the classMeanScaleUniformBins
specifically relies onlow_limit
andhigh_limit
to buildself.centers
andself.boundaries
. While that works fine for data roughly within[-15..+15]
, new domains can push well outside that range.torch.bucketize
without distinguishing “close” vs. “far” bins.Proposed enhancement
Dynamic range determination
MeanScaleUniformBins
to compute (at inference or per training batch) a local min/max (or percentiles, like p10/p90) for the scaled data.self.centers
andself.boundaries
so that we cover[actual_min..actual_max]
, clamping if needed (e.g.,[-50..50]
).Distance-aware label smoothing
b
is the correct bin, allocate ~80% probability tob
, ~10% tob-1
, and ~10% tob+1
(splitting among neighbors).ChronosModel.forward
, if we have direct control of the loss.Estimated impact on accuracy & inference
Zero-shot improvements:
Few-shot improvements:
Inference overhead:
Why it improves zero-/few-shot
Beta Was this translation helpful? Give feedback.
All reactions