LangNav: Language as a Perceptual Representation for Navigation
About LangNav
LangNav is an LLM-based navigation agent which performs multi-step navigation end-to-end via textual descriptions of the scene. The language-based perceptual representation makes LangNav more data efficient compared to VL models. With only a few language-based trajectories from a R2R environment, we use GPT-4 to efficiently generate a huge amount of synthetic training data. A smaller language model (LLaMA2-7B) can then be trained on these synthetic data and do the task. In this repo, we provide the inference code, the model, and the training dataset we used for the paper:
LangNav: Language as a Perceptual Representation for Navigation
Bowen Pan, Rameswar Panda, SouYoung Jin, Rogerio Feris, Aude Oliva, Phillip Isola, Yoon Kim
NAACL 2024 (Findings)
Prerequisites
We don't have to install the Matterport3D Simulator as we have pre-extracted the caption of each viewpoint.
But we still need to prepare the data in directories
- MP3D navigability graphs:
connectivity
- Download the connectivity maps [23.8MB].
- R2R data:
data
- Download the R2R data [5.8MB].
- BLIP caption of the scene:
img_features
- Download the caption data [113MB] (r2r_blip_DETR_vis2text).
Install the Pytorch-Transformers.
Multi-step Navigation with Language-based Representation
Evaluate our LangNav-Sim2k-Llama2
model on the R2R datasets.
sh eval_scripts/eval_langnav_2k_synthetic_100_real.sh
We will also release the synthetic training dataset and the other models. Stay tuned!
Citation
If you use or discuss our LangNav, please cite our paper:
@article{pan2023langnav,
title={Langnav: Language as a perceptual representation for navigation},
author={Pan, Bowen and Panda, Rameswar and Jin, SouYoung and Feris, Rogerio and Oliva, Aude and Isola, Phillip and Kim, Yoon},
journal={arXiv preprint arXiv:2310.07889},
year={2023}
}
- Downloads last month
- 9