Image Deraining with Frequency-Enhanced State Space Model

Shugo Yamashita, Masaaki Ikehara; Proceedings of the Asian Conference on Computer Vision (ACCV), 2024, pp. 3655-3671

Abstract


Removing rain degradations in images is recognized as a significant issue. In this field, deep learning-based approaches, such as Convolutional Neural Networks (CNNs) and Transformers, have succeeded. Recently, State Space Models (SSMs) have exhibited superior performance across various tasks in both natural language processing and image processing due to their ability to model long-range dependencies. This study introduces SSM to image deraining and proposes a Deraining Frequency-Enhanced State Space Model (DFSSM). To effectively remove rain streaks, which produce high-intensity frequency components in specific directions, we employ frequency domain processing concurrently with SSM. Additionally, we develop a novel mixed-scale gated-convolutional block, which uses convolutions with multiple kernel sizes to capture various scale degradations effectively and integrates a gating mechanism to manage the flow of information. Finally, experiments on synthetic and real-world rainy image datasets show that our method surpasses state-of-the-art methods. Code is available at https://github.com/ShugoYamashita/DFSSM.

Related Material


[pdf] [supp] [arXiv]
[bibtex]
@InProceedings{Yamashita_2024_ACCV, author = {Yamashita, Shugo and Ikehara, Masaaki}, title = {Image Deraining with Frequency-Enhanced State Space Model}, booktitle = {Proceedings of the Asian Conference on Computer Vision (ACCV)}, month = {December}, year = {2024}, pages = {3655-3671} }