--- license: apache-2.0 pipeline_tag: robotics library_name: transformers --- # Mantis > This is the official checkpoint of **Mantis: A Versatile Vision-Language-Action Model with Disentangled Visual Foresight** - **Paper:** https://arxiv.org/pdf/2511.16175 - **Code:** https://github.com/zhijie-group/Mantis ### 🔥 Highlights - **Disentangled Visual Foresight** augments action learning without overburdening the backbone. - **Progressive Training** preserves the understanding capabilities of the backbone. - **Adaptive Temporal Ensemble** reduces inference cost while maintaining stable control. ### How to use This is the Mantis model trained on the [LIBERO](https://huggingface.co/datasets/Yysrc/mantis_libero_lerobot/tree/main) object dataset. For detailed usage please refer to [our repository](https://github.com/zhijie-group/Mantis). ### 📝 Citation If you find our code or models useful in your work, please cite [our paper](https://arxiv.org/pdf/2511.16175): ``` @article{yang2025mantis, title={Mantis: A Versatile Vision-Language-Action Model with Disentangled Visual Foresight}, author={Yang, Yi and Li, Xueqi and Chen, Yiyang and Song, Jin and Wang, Yihan and Xiao, Zipeng and Su, Jiadi and Qiaoben, You and Liu, Pengfei and Deng, Zhijie}, journal={arXiv preprint arXiv:2511.16175}, year={2025} } ```