Official implementation of ADMIRE-BayesOpt: Accelerated Data MIxture RE-weighting for Language Models with Bayesian Optimization. Please feel free to email us or raise an issue with this repository and we'll get back to you as soon as possible.
- Create a virtual environment (we use conda)
- Activate the virtual environment
- Install the repository
conda env create --name admire_bayesopt conda activate admire_bayesopt pip install -r requirements.txt
This implementaion is based on an official BoTorch tutorial: Multi-fidelity Bayesian optimization with discrete fidelities using KG. We followed its comparasions between BayesOpt and MFBayesOpt.
We opensource the data mixture dataset: admire_ift_runs
and use the mixture dataset on the Pile regmix-data
from RegMix.
We run experiments of different mixtures with Qwen2.5 0.5B / 3B / 7B.
Choose the index of target domain: --idx
.
Choose the dataset [admire_ift_runs/pile]: --dataset
.
Results will be saved in saved_logs
.
python bayesopt_admire_ift_runs.py --idx -3 #average of ood+id
python bayesopt_thepile.py --idx -1 #average
python mfbayesopt_maxvalue.py --dataset admire_ift_runs --idx -3 #average of ood+id
python mfbayesopt_maxvalue.py --dataset pile --idx -1 #average
Please use the following to cite this work:
@misc{chen2025admirebayesoptaccelerateddatamixture,
title={ADMIRE-BayesOpt: Accelerated Data MIxture RE-weighting for Language Models with Bayesian Optimization},
author={Shengzhuang Chen and Xu Ouyang and Michael Arthur Leopold Pearce and Thomas Hartvigsen and Jonathan Richard Schwarz},
year={2025},
eprint={2508.11551},
archivePrefix={arXiv},
primaryClass={stat.ML},
url={https://arxiv.org/abs/2508.11551},
}
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.