VMem
is a plug-and-play memory mechanism of image-set models for consistent scene generation.
Existing methods either rely on inpainting with explicit geometry estimation, which suffers from inaccuracies, or use limited context windows in video-based approaches, leading to poor long-term coherence. To overcome these issues, we introduce Surfel Memory of Views (VMem), which anchors past views to surface elements (surfels) they observed. This enables conditioning novel view generation on the most relevant past views rather than just the most recent ones, enhancing long-term scene consistency while reducing computational cost.
conda create -n vmem python=3.10
conda activate vmem
pip install -r requirements.txt
You need to properly authenticate with Hugging Face to download our model weights. Once set up, our code will handle it automatically at your first run. You can authenticate by running
# This will prompt you to enter your Hugging Face credentials.
huggingface-cli login
Once authenticated, go to our model card here and enter your information for access.
We provide a demo for you to interact with VMem
. Simply run
python app.py
This work is built on top of CUT3R, DUSt3R and Stable Virtual Camera. We thank them for their great works.
If you find this repository useful, please consider giving a star ⭐ and citation.
@article{li2025vmem,
title={VMem: Consistent Interactive Video Scene Generation with Surfel-Indexed View Memory},
author={Li, Runjia and Torr, Philip and Vedaldi, Andrea and Jakab, Tomas},
journal={arXiv preprint arXiv:2506.18903},
year={2025}
}