Skip to content

linglingxiansen/MapNav

Repository files navigation

Mapnav: A novel memory representation via annotated semantic maps for vlm-based vision-and-language navigation (ACL 2025)

Repository for Mapnav: A novel memory representation via annotated semantic maps for vlm-based vision-and-language navigation (ACL 2025)

License hf_checkpoint arXiv

Installation

The code has been tested only with Python 3.8 on Ubuntu 20.04.

  1. Environments Setup
  • Follow L3MVN to install Habitat-lab, Habitat-sim, rednet, torch and other independences.
  • Install the LLaVA.
  1. Dataset
  1. Path
  • Change the dataset path and habitat path in the config_utils.py

Training

You can download the huggingface dataset to generate you QA pairs to train your own model using LLaVA-NeXT.

Evaluation

CUDA_VISIBLE_DEVICES=1 python r2rnav_benchmark.py --split val1 --eval 1 --auto_gpu_config 0 -n 1 --num_local_steps 10 --print_images 1 --model_dir model_path --exp_name nohis_rgb --eval_episodes 1839 --collect 0 --stop_th 300

ASM Generation

You can look for the generation and annotation pipeline in the r2rnav_benchmark.py and huatu3.py.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages