ViNT: A Foundation Model for Visual Navigation
Published in Conference on Robot Learning (Oral), 2023
We present a multi-robot visual navigation foundation model for effective fine-tuning and novel task adaptation.
Also appeared as: Oral Presentation at BayLearn, Live Demo at CoRL, Workshop Paper at NeurIPS, YouTube Video, and Project Site
Recommended citation: Dhruv Shah*, Ajay Sridhar*, Nitish Dashora*, Kyle Stachowicz, Kevin Black, Noriaki Hirose, Sergey Levine. (2023). "ViNT: A Large-Scale, Multi-Task Visual Navigation Backbone with Cross-Robot Generalization." In the proceedings of 2023 Conference on Robot Learning (CoRL).
Download Paper