Intra and Inter Parser-Prompted Transformers for Effective Image Restoration
By: Cong Wang , Jinshan Pan , Liyan Wang and more
Potential Business Impact:
Fixes blurry, noisy, and dark pictures.
We propose Intra and Inter Parser-Prompted Transformers (PPTformer) that explore useful features from visual foundation models for image restoration. Specifically, PPTformer contains two parts: an Image Restoration Network (IRNet) for restoring images from degraded observations and a Parser-Prompted Feature Generation Network (PPFGNet) for providing IRNet with reliable parser information to boost restoration. To enhance the integration of the parser within IRNet, we propose Intra Parser-Prompted Attention (IntraPPA) and Inter Parser-Prompted Attention (InterPPA) to implicitly and explicitly learn useful parser features to facilitate restoration. The IntraPPA re-considers cross attention between parser and restoration features, enabling implicit perception of the parser from a long-range and intra-layer perspective. Conversely, the InterPPA initially fuses restoration features with those of the parser, followed by formulating these fused features within an attention mechanism to explicitly perceive parser information. Further, we propose a parser-prompted feed-forward network to guide restoration within pixel-wise gating modulation. Experimental results show that PPTformer achieves state-of-the-art performance on image deraining, defocus deblurring, desnowing, and low-light enhancement.
Similar Papers
Progressive Focused Transformer for Single Image Super-Resolution
CV and Pattern Recognition
Makes blurry pictures sharp, fast.
Dual Prompting Image Restoration with Diffusion Transformers
CV and Pattern Recognition
Fixes blurry pictures using text and images.
DPFormer: Dynamic Prompt Transformer for Continual Learning
CV and Pattern Recognition
Helps AI remember old lessons while learning new ones.