Skip to content

Latest commit

 

History

History
87 lines (57 loc) · 4.43 KB

README.md

File metadata and controls

87 lines (57 loc) · 4.43 KB

Mask-Guided Progressive Network for Joint Raindrop and Rain Streak Removal in Videos (ACM MM 2023)

Hongtao Wu, Yijun Yang, Haoyu Chen, Jingjing Ren, Lei Zhu

This repo is the official Pytorch implementation of Mask-Guided Progressive Network for Joint Raindrop and Rain Streak Removal in Videos. The first dataset and approach for video rain streaks and raindrops removal.


Abstract: Videos captured in rainy weather are unavoidably corrupted by both rain streaks and raindrops in driving scenarios, and it is desirable and challenging to recover background details obscured by rain streaks and raindrops. However, existing video rain removal methods often address either video rain streak removal or video raindrop removal, thereby suffer from degraded performance when deal with both simultaneously. The bottleneck is a lack of a video dataset, where each video frame contains both rain streaks and raindrops. To address this issue, we in this work generate a synthesized dataset, namely VRDS, with 102 rainy videos from diverse scenarios, and each video frame has the corresponding rain streak map, raindrop mask, and the underlying rain-free clean image (ground truth). Moreover, we devise a mask-guided progressive video deraining network (ViMP-Net) to remove both rain streaks and raindrops of each video frame. Specifically, we develop an intensity-guided alignment block to predict the rain streak intensity map and remove the rain streaks of the input rainy video at the first stage. Then, we predict a raindrop mask and pass it into a devised mask-guided dual transformer block to learn inter-frame and intra-frame transformer features, which are then fed into a decoder for further eliminating raindrops. Experimental results demonstrate that our ViMP-Net outperforms state-of-the-art methods on our synthetic dataset and real-world rainy videos.


Our Dataset

Our VRDS dataset can be downloaded here.

Installation

This implementation is based on MMEditing, which is an open-source image and video editing toolbox.

Below are quick steps for installation.

Step 1. Install PyTorch following official instructions.

Step 2. Install MMCV with MIM.

pip3 install openmim
mim install mmcv-full

Step 3. Install ViMP-Net from source.

git clone https://github.com/TonyHongtaoWu/ViMP-Net.git
cd ViMP-Net
pip3 install -e .

Please refer to MMEditing Installation for more detailed instruction.

Training and Testing

You may need to adjust the dataset path and dataloader before starting.

You can train ViMP-Net on VRDS dataset using the below command with 4 GPUs:

CUDA_VISIBLE_DEVICES=0,1,2,3 ./tools/dist_train.sh configs/derainers/ViMPNet/ViMPNet.py 4

You can use the following command with 1 GPU to test your trained model model.pth:

CUDA_VISIBLE_DEVICES=0 ./tools/dist_test.sh configs/derainers/ViMPNet/ViMPNet.py "model.pth" 1 --save-path './save_path/'

You can find one model checkpoint trained on VRDS dataset here.

Our Results

The visual results of ViMP-Net can be downloaded in Google Drive and Outlook.

Contact

Should you have any question or suggestion, please contact [email protected].

Acknowledgement

This code is based on MMEditing and FuseFormer.

Citation

If you find this repository helpful to your research, please consider citing the following:

@inproceedings{wu2023mask,
  title={Mask-Guided Progressive Network for Joint Raindrop and Rain Streak Removal in Videos},
  author={Wu, Hongtao and Yang, Yijun and Chen, Haoyu and Ren, Jingjing and Zhu, Lei},
  booktitle={Proceedings of the 31st ACM International Conference on Multimedia},
  pages={7216--7225},
  year={2023}
}