ShadowMaskFormer: Mask Augmented Patch Embeddings for Shadow Removal

29 Apr 2024  ·  Zhuohao Li, Guoyang Xie, Guannan Jiang, Zhichao Lu ·

Transformer recently emerged as the de facto model for computer vision tasks and has also been successfully applied to shadow removal. However, these existing methods heavily rely on intricate modifications to the attention mechanisms within the transformer blocks while using a generic patch embedding. As a result, it often leads to complex architectural designs requiring additional computation resources. In this work, we aim to explore the efficacy of incorporating shadow information within the early processing stage. Accordingly, we propose a transformer-based framework with a novel patch embedding that is tailored for shadow removal, dubbed ShadowMaskFormer. Specifically, we present a simple and effective mask-augmented patch embedding to integrate shadow information and promote the model's emphasis on acquiring knowledge for shadow regions. Extensive experiments conducted on the ISTD, ISTD+, and SRD benchmark datasets demonstrate the efficacy of our method against state-of-the-art approaches while using fewer model parameters.

PDF Abstract

Datasets


Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Shadow Removal ISTD+ ShadowMaskFormer (arXiv 2024) (512x512) RMSE 2.95 # 3
PSNR 28.34 # 3
SSIM 0.849 # 6
LPIPS 0.211 # 8
Shadow Removal ISTD+ ShadowMaskFormer (arXiv 2024) (256x256) RMSE 3.39 # 16
PSNR 26.57 # 16
SSIM 0.698 # 22
LPIPS 0.395 # 22
Shadow Removal SRD ShadowMaskFormer (arXiv 2024) (512x512) RMSE 4.15 # 10
PSNR 25.13 # 10
SSIM 0.798 # 9
LPIPS 0.249 # 6
Shadow Removal SRD ShadowMaskFormer (arXiv 2024) (256x256) RMSE 4.69 # 19
PSNR 23.85 # 20
SSIM 0.671 # 20
LPIPS 0.386 # 20

Methods


No methods listed for this paper. Add relevant methods here