InternNav is an All-in-one open-source toolbox for embodied navigation based on PyTorch, Habitat and Isaac Sim.
- Modular Support of the Entire Navigation System
We support modular customization and study of the entire navigation system, including vision-language navigation with discrete action space (VLN-CE), visual navigation (VN) given point/image/trajectory goals, and the whole VLN system with continuous trajectory outputs.
- Compatibility with Mainstream Simulation Platforms
The toolbox is compatible with different training and evaluation requirements, supporting different environments for the usage of mainstream simulation platforms such as Habitat and Isaac Sim.
- Comprehensive Datasets, Models and Benchmarks
The toolbox supports the most comprehensive 6 datasets & benchmarks and 10+ popular baselines, including both mainstream and our established brand new ones.
- State of the Art
The toolbox supports the most advanced high-quality navigation dataset, InternData-N1, which includes 3k+ scenes and 830k VLN data covering diverse embodiments and scenes, and the first dual-system navigation foundation model with leading performance on all the benchmarks and zero-shot generalization capability in the real world, InternVLA-N1.
- [2025/07] We are hosting 🏆IROS 2025 Grand Challenge, stay tuned at official website.
- [2025/07] InternNav v0.1.0 released.
- 🏠 Introduction
- 🔥 News
- 📚 Getting Started
- 📦 Overview of Benchmark & Model Zoo
- 🔧 Customization
- 👥 Contribute
- 🔗 Citation
- 📄 License
- 👏 Acknowledgements
Please refer to the documentation for quick start with InternNav, from installation to training or evaluating supported models.
System2 (VLN-CE) | System1 (VN) | Whole-system (VLN) |
System2 (VLN-CE) | System1 (VN) | Whole-system (VLN) |
NOTE:
- The detailed benchmark results will be updated in the next few days.
- VLN-CE RxR benchmark and StreamVLN will be supported soon.
Please refer to the tutorial for advanced usage of InternNav, including customization of datasets, models and experimental settings.
If you would like to contribute to InternNav, please check out our contribution guide. For example, raising issues, fixing bugs in the framework, and adapting or adding new policies and data to the framework.
Note: We welcome the feedback of the model's zero-shot performance when deploying in your own environment. Please show us your results and offer us your future demands regarding the model's capability. We will select the most valuable ones and collaborate with users together to solve them in the next few months :)
If you find our work helpful, please cite:
@misc{internnav2025,
title = {{InternNav: InternRobotics'} open platform for building generalized navigation foundation models},
author = {InternNav Contributors},
howpublished={\url{https://github.com/InternRobotics/InternNav}},
year = {2025}
}
If you use the specific pretrained models and benchmarks, please kindly cite the original papers involved in our work. Related BibTex entries of our papers are provided below.
Related Work BibTex
@misc{internvla-n1,
title = {{InternVLA-N1: An} Open Dual-System Navigation Foundation Model with Learned Latent Plans},
author = {InternNav Team},
year = {2025},
booktitle={arXiv},
}
@inproceedings{vlnpe,
title={Rethinking the Embodied Gap in Vision-and-Language Navigation: A Holistic Study of Physical and Visual Disparities},
author={Wang, Liuyi and Xia, Xinyuan and Zhao, Hui and Wang, Hanqing and Wang, Tai and Chen, Yilun and Liu, Chengju and Chen, Qijun and Pang, Jiangmiao},
booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)},
year={2025}
}
@misc{streamvln,
title = {StreamVLN: Streaming Vision-and-Language Navigation via SlowFast Context Modeling},
author = {Wei, Meng and Wan, Chenyang and Yu, Xiqian and Wang, Tai and Yang, Yuqiang and Mao, Xiaohan and Zhu, Chenming and Cai, Wenzhe and Wang, Hanqing and Chen, Yilun and Liu, Xihui and Pang, Jiangmiao},
booktitle={arXiv},
year = {2025}
}
@misc{navdp,
title = {NavDP: Learning Sim-to-Real Navigation Diffusion Policy with Privileged Information Guidance},
author = {Wenzhe Cai, Jiaqi Peng, Yuqiang Yang, Yujian Zhang, Meng Wei, Hanqing Wang, Yilun Chen, Tai Wang and Jiangmiao Pang},
year = {2025},
booktitle={arXiv},
}
InternNav's codes are MIT licensed.
The open-sourced InternData-N1 data are under the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License .
Other datasets like VLN-CE inherit their own distribution licenses.
- InternUtopia (Previously
GRUtopia
): The closed-loop evaluation and GRScenes-100 data in this framework relies on the InternUtopia framework. - Diffusion Policy: Diffusion policy implementation.
- LongCLIP: Long-text CLIP model.
- VLN-CE: Vision-and-Language Navigation in Continuous Environments based on Habitat.
- Qwen2.5-VL: The pretrained vision-language foundation model.
- LeRobot: The data format used in this project largely follows the conventions of LeRobot.