Abstract
As a subjective concept, image quality assessment (IQA) is significantly affected by perceptual mechanisms. Two mutually influenced mechanisms, namely spatial attention and contrast sensitivity, are particularly important for IQA. This paper aims to explore a deep learning approach based on transformer for the two mechanisms. By converting contrast sensitivity to attention representation, a unified multi-head attention module is performed on spatial and channel features in transformer encoder to simulate the two mechanisms in IQA. Sequential spatial-channel self-attention is proposed to avoid expensive computation in the classical Transformer model. In addition, as image rescaling can potentially affect perceived quality, zero-padding and masking with assigning special attention weights are performed to handle arbitrary image resolutions without requiring image rescaling. The evaluation results on publicly available large-scale IQA databases have demonstrated outstanding performance and generalization of the proposed IQA model.
Original language | English |
---|---|
Title of host publication | QoEVMA '22 |
Subtitle of host publication | Proceedings of the 2nd Workshop on Quality of Experience in Visual Multimedia Applications |
Publisher | Association for Computing Machinery, Inc |
Pages | 13-21 |
Number of pages | 9 |
ISBN (Electronic) | 9781450394994 |
DOIs | |
Publication status | Published - 10 Oct 2022 |
Event | 2nd Workshop on Quality of Experience in Visual Multimedia Applications (QoEVMA) at ACM Multimedia - Lisbon, Portugal Duration: 10 Oct 2022 → 14 Oct 2022 https://2022.acmmm.org/ |
Conference
Conference | 2nd Workshop on Quality of Experience in Visual Multimedia Applications (QoEVMA) at ACM Multimedia |
---|---|
Country/Territory | Portugal |
City | Lisbon |
Period | 10/10/22 → 14/10/22 |
Internet address |
Keywords
- contrast sensitivity
- image quality assessment (IQA)
- sequential spatial-channel attention (SSCA)
- spatial attention
- transformer