Dilated Convolutional Transformer for High-Quality Image Deraining

Yufeng Li, Jiyang Lu, Hongming Chen, Xianhao Wu, Xiang Chen; Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2023, pp. 4199-4207


Convolutional neural networks (CNNs) and Transformers have achieved significant success in image signal processing. However, little effort has been made to effectively combine the properties of these two architectures to satisfy image deraining. In this paper, we propose an effective deraining method, dilated convolutional Transformer (DCT), which can enlarge the receptive fields of the network to aggregate global information. The fundamental building unit of our approach is the dilformer block containing multi-dilconv sparse attention (MDSA) and multi-dilconv feed-forward network (MDFN). The MDSA calculates the multi-scale query to generate accurate similarity map so that rich multi-scale information can be better utilized for the high-quality image reconstruction. In addition, we adopt ReLU to replace the original softmax to enforce sparsity in the Transformer for better feature aggregation. The MDFN is further established to better integrate the rain information of different scales in the feature transformation. Extensive experiments on the benchmarks show the favorable performance against state-of-the-art approaches.

Related Material

@InProceedings{Li_2023_CVPR, author = {Li, Yufeng and Lu, Jiyang and Chen, Hongming and Wu, Xianhao and Chen, Xiang}, title = {Dilated Convolutional Transformer for High-Quality Image Deraining}, booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops}, month = {June}, year = {2023}, pages = {4199-4207} }