This repository is the official implementation of paper
Nhu-Thuat Tran and Hady W. Lauw. 2025. Parameter-Efficient Variational AutoEncoder for Multimodal Multi-Interest Recommendation. Accepted by The 33rd ACM International Conference on Multimedia (ACMMM'2025), Dublin, Ireland, October 27-31, 2025.
- Anaconda: 4.12.0
- Python: 3.7.5
- OS: MacOS
Please follow the instruction in README.md
file under data folder
Create virtual environment
conda create --prefix ./maplevae python=3.7.5 -y
Activate environment
conda activate ./maplevae
Install requirements
pip install -r requirements.txt
-
Create a YAML config file under
configs
folder as samples. -
Prepare
run.sh
file as follows
python run_maplevae.py --dataset <dataset_name> --config_file <your_config_file> --device_id <ID of GPU machine>
- To run training and evaluation
bash run.sh
We follow instructions from VALID (https://github.com/PreferredAI/VALID) for hyper-parameter tuning.
Then, we tune the key hyper-parameters in MapleVAE
n_gcn_layers
: the number of GCN layers1, 2, 3, 4
alpha
: multimodality graph combination weight (Section 3.2.3)0.6, 0.7, 0.8, 0.9
Q
: the number of quantizing levels andW
: the size of each codebook in RQ-VAE by changingtext_code_file
andimage_code_file
.
If you find our work useful for your research, please cite our paper as
@inproceedings{MapleVAE,
author = {Nhu{-}Thuat Tran and
Hady W. Lauw},
title = {Parameter-Efficient Variational AutoEncoder for Multimodal Multi-Interest Recommendation},
booktitle = {Proceedings of The 33rd ACM International Conference on Multimedia, Dublin, Ireland, October 27-31, 2025},
year = {2025}
}