Skip to content

yang-zj1026/VLN-CE-Isaac

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

VLN-CE-Isaac Benchmark

Overview

The VLN-CE-Isaac Benchmark is a framework for evaluating Visual Language Navigation in Isaac Lab. This repository contains the code and instructions to set up the environment, download the required data, and run the benchmark.

VLN-CE with Go2

Installation

Prerequisites

  • Ubuntu 20.04 or higher
  • NVIDIA GPU with CUDA support (check here for more detailed requirements)

Steps

  1. Create a virtual environment with python 3.10:

    conda create -n vlnce-isaac python=3.10
    conda activate vlnce-isaac
  2. Make sure that Isaac Sim is installed on your machine. Otherwise follow this guideline to install it. If installing via the Omniverse Launcher, please ensure that Isaac Sim 4.1.0 is selected and installed. On Ubuntu 22.04 or higher, you could install it via pip:

    pip install isaacsim-rl==4.1.0 isaacsim-replicator==4.1.0 isaacsim-extscache-physics==4.1.0 isaacsim-extscache-kit-sdk==4.1.0 isaacsim-extscache-kit==4.1.0 isaacsim-app==4.1.0 --extra-index-url https://pypi.nvidia.com
  3. Clone Isaac Lab and link the extensions.

    Note: This codebase was tested with Isaac Lab 1.1.0 and may not be compatible with newer versions. Please make sure to use the modified version of Isaac Lab provided below, which includes important bug fixes and updates. As Isaac Lab is under active development, we will consider supporting newer versions in the future.

    git clone [email protected]:yang-zj1026/IsaacLab.git
    cd IsaacLab
    cd source/extensions
    ln -s {THIS_REPO_DIR}/isaaclab_exts/omni.isaac.vlnce .
    ln -s {THIS_REPO_DIR}/isaaclab_exts/omni.isaac.matterport .
    cd ../..
  4. Run the Isaac Lab installer script and additionally install rsl rl in this repo.

    ./isaaclab.sh -i none
    ./isaaclab.sh -p -m pip install -e {THIS_REPO_DIR}/scripts/rsl_rl
    cd ..

Data

Download the data from huggingface and put them under isaaclab_exts/omni.isaac.vlnce/assets directory. The expected file structure should be like:

isaaclab_exts/omni.isaac.vlnce
├─ assets
|   ├─ vln_ce_isaac_v1.json.gz
|   ├─ matterport_usd

Code Usage

Run the demo with a PD path planner

python scripts/demo_planner.py --task=go2_matterport_vision --history_length=9 --load_run=2024-09-25_23-22-02

python scripts/demo_planner.py --task=h1_matterport_vision --load_run=2024-11-03_15-08-09_height_scan_obst

To train your own low-level policies, please refer to the legged-loco repo.

Citation

If you use VLN-CE-Isaac in your work please consider citing our paper:

@article{cheng2024navila,
  title={NaVILA: Legged Robot Vision-Language-Action Model for Navigation},
  author={Cheng, An-Chieh and Ji, Yandong and Yang, Zhaojing and Zou, Xueyan and Kautz, Jan and B{\i}y{\i}k, Erdem and Yin, Hongxu and Liu, Sifei and Wang, Xiaolong},
  journal={arXiv preprint arXiv:2412.04453},
  year={2024}
}

Acknowledgements

This project makes use of the following open source codebase:

License

This project is licensed under the MIT License. See the LICENSE file for details.

About

Vision-Language Navigation Benchmark in Isaac Lab

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages