Skip to content

lx709/VRSBench

Repository files navigation

VRSBench: A Versatile Vision-Language Benchmark Dataset for Remote Sensing Image Understanding

Xiang Li, Jian Ding, Mohamed Elhoseiny

Related Projects

RSGPT: A Remote Sensing Vision Language Model and Benchmark

Yuan Hu, Jianlong Yuan, Congcong Wen, Xiaonan Lu, Xiang Li☨

Vision-language models in remote sensing: Current progress and future trends

Xiang Li☨, Congcong Wen, Yuan Hu, Zhenghang Yuan, Xiao Xiang Zhu

VRSBench

VRSBench is a Versatile Vision-Language Benchmark for Remote Sensing Image Understanding.

VRSBench is a Versatile Vision-Language Benchmark for Remote Sensing Image Understanding. It consists of 29,614 remote sensing images with detailed captions, 52,472 object refers, and 3123,221 visual question-answer pairs. It facilitates the training and evaluation of vision-language models across a broad spectrum of remote sensing image understanding tasks.

🗓️ TODO

  • [2024.10.15] Release evaluation code.
  • [2024.10.15] Release code and models of baseline models.
  • [2024.06.19] We release the instructions and code for calling GPT-4V to get initial annotations.
  • [2024.06.19] We release the VRSBench, A Versatile Vision-Language Benchmark Dataset for Remote Sensing Image Understanding. VRSBench contains 29,614 images, with 29,614 human-verified detailed captions, 52,472 object references, and 123,221 question-answer pairs. check VRSBench Project Page.

Using datasets

The dataset can be downloaded from link and used via the Hugging Face datasets library. To load the dataset, you can use the following code snippet:

from datasets import load_dataset
fw = load_dataset("xiang709/VRSBench", streaming=True)

Dataset curation

To construct our VRSBench dataset, we employed multiple data engineering steps, including attribute extraction, prompting engineering, GPT-4 inference, and human verification.

  • Attribute Extraction: we extract image information, including the source and resolution, as well as object information—such as the object category, bounding box, color, position (absolute and relative), and size (absolute and relative)—from existing object detection datasets. Please check extract_patch_json.py for attribute extraction details.
  • Prompting Engineering: We carefully design instructions to prompt GPT-4V to create detailed image captions, object referring, and question-answer pairs. Please check instruction.txt for detailed instructions.
  • GPT-4 inference: Given input prompts, we call OpenAI API to automatically generate image captions, object referring, and question-answer pairs based on the prompts. Use the extract_patch_json.py to get initial annotations for image captioning, visual grounding, and VQA tasks using GPT-4V.
  • Human verification: To improve the quality of the dataset, we engage human annotators to validate each annotation generated by GPT-4V.

Model Training

For the above three tasks, we benchmark state-of-the-art models, including LLaVA-1.5, MiniGPT-v2, Mini-Gemini, and GeoChat, to demonstrate the potential of LVMs for remote sensing image understanding. To ensure a fair comparison, we reload the models that are initially trained on large-scale image-text alignment datasets, and then finetune each method using the training set of our RSVBench dataset. For each comparing method, we finetune the model on the training set of our RSVBench dataset for 5 epochs. Following GeoChat, we use LoRA finetuning to finetune all comparing methods, with a rank of 64.

Use the prepare_geochat_eval_all.ipynb to prepare the VRSBench evaluation file for image captioning, visual grounding, and VQA tasks.

Benchmark Results

The code and checkpoints of baseline models can be found at Onedrive.

Image Captioning Performance

Method BLEU-1 BLEU-2 BLEU-3 BLEU-4 METEOR ROUGE_L CIDEr Avg_L
GeoChat w/o ft 13.9 6.6 3.0 1.4 7.8 13.2 0.4 36
GPT-4V 37.2 22.5 13.7 8.6 20.9 30.1 19.1 67
MiniGPT-v2 36.8 22.4 13.9 8.7 17.1 30.8 21.4 37
LLaVA-1.5 48.1 31.5 21.2 14.7 21.9 36.9 33.9 49
GeoChat 46.7 30.2 20.1 13.8 21.1 35.2 28.2 52
Mini-Gemini 47.6 31.1 20.9 14.3 21.5 36.8 33.5 47

Caption: Detailed image caption performance on the VRSBench dataset. Avg_L denotes the average word length of generated captions.

Visual Grounding Performance

Method [email protected] (Unique) [email protected] (Unique) [email protected] (Non Unique) [email protected] (Non Unique) [email protected] (All) [email protected] (All)
GeoChat w/o ft 20.7 5.4 7.3 1.7 12.9 3.2
GPT-4V 8.6 2.2 2.5 0.4 5.1 1.1
MiniGPT-v2 40.7 18.9 32.4 15.2 35.8 16.8
LLaVA-1.5 51.1 16.4 34.8 11.5 41.6 13.6
GeoChat 57.4 22.6 44.5 18.0 49.8 19.9
Mini-Gemini 41.1 9.6 22.3 4.9 30.1 6.8

Caption: Visual grounding performance on the papernameAbbrev dataset. Boldface indicates the best performance.

Visual Question Answering Performance

Method Category Presence Quantity Color Shape Size Position Direction Scene Reasoning All
# VQAs 5435 7789 6374 3550 1422 1011 5829 477 4620 902
GeoChat w/o ft 48.5 85.9 19.2 17.0 18.3 32.0 43.4 42.1 44.2 57.4 40.8
GPT-4V 67.0 87.6 45.6 71.0 70.8 54.3 67.2 50.7 69.8 72.4 65.6
MiniGPT-v2 61.3 26.0 46.1 51.0 41.8 11.2 17.1 12.4 49.3 21.9 38.2
LLaVA-1.5 86.9 91.8 58.2 69.9 72.2 61.5 69.5 56.7 83.9 73.4 76.4
GeoChat 86.5 92.1 56.3 70.1 73.8 60.4 69.3 53.5 83.7 73.5 76.0
Mini-Gemini 87.8 92.1 58.8 74.0 75.3 58.0 68.0 56.7 83.2 74.4 77.8

Caption: Visual question answering performance on the VRSBench dataset. Boldface indicates the best performance.

Licensing Information

The dataset is released under the CC-BY-4.0, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Related Projects

  • RSGPT. The first GPT-based Large Vision-Language Model in remote sensing. RSGPT
  • RS-CLIP. A CLIP-based Vision-Language Model for remote sensing scene classification. RS-CLIP
  • Survey. A comprehensive survey about vision-language models in remote sensing. RSVLM.
  • MiniGPT-v2. MiniGPT-v2

📜 Citation

@article{li2024vrsbench,
  title={VRSBench: A Versatile Vision-Language Benchmark Dataset for Remote Sensing Image Understanding},
  author={Xiang Li, Jian Ding, and Mohamed Elhoseiny},
  journal={arXiv:2406.12384},
  year={2024}
}

🙏 Acknowledgement

Our VRSBench dataset is built based on DOTA-v2 and DIOR datasets.

We are thankful to LLaVA-1.5, MiniGPT-v2, Mini-Gemini, and GeoChat for releasing their models and code as open-source contributions.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published