Efficient transformer with compressed-attention for stereo image super-resolution
File version
Version of Record (VoR)
Author(s)
Sowmya, A
Zhang, W
Sun, C
Griffith University Author(s)
Primary Supervisor
Other Supervisors
Editor(s)
Date
Size
File type(s)
Location
Abstract
While self-attention mechanisms in transformers exhibit superior performance in image super-resolution tasks, improving efficiency remains a challenge. To enhance the efficiency of self-attention mechanisms for stereo image super-resolution, we propose an efficient transformer with compressed-attention for stereo image super-resolution (ETCASSR). Specifically, we propose a simple yet effective compressed-attention mechanism that organizes channels from partial to full for attention operations. Using this mechanism, we develop a compressed window-based self-attention block and a compressed transposed self-attention block, enabling efficient intra-view feature extraction. To further enrich feature representation, we introduce a spatial local feature branch and a channel global feature branch to complement these two blocks. Furthermore, a compressed cross-attention block for cross-view feature extraction is designed by extending the compressed-attention mechanism. Combining these blocks, ETCASSR achieves state-of-the-art performance on stereo image super-resolution while maintaining low computational complexity and fast running speed. Additionally, we introduce ETCASR for single-image super-resolution by omitting the cross-view components from ETCASSR, also achieving superior performance with high efficiency. The proposed transformers offer significant potential applications in other vision tasks. Source code is available at https://github.com/jianwensong/ETCASSR.
Journal Title
Knowledge-Based Systems
Conference Title
Book Title
Edition
Volume
331
Issue
Thesis Type
Degree Program
School
Publisher link
Patent number
Funder(s)
Grant identifier(s)
Rights Statement
Rights Statement
© 2025 The Author(s). Published by Elsevier B.V. This is an open access article under the CC BY-NC license (http://creativecommons.org/licenses/by-nc/4.0/).
Item Access Status
Note
Access the data
Related item(s)
Subject
Artificial intelligence
Data management and data science
Machine learning
Persistent link to this record
Citation
Song, J; Sowmya, A; Zhang, W; Sun, C, Efficient transformer with compressed-attention for stereo image super-resolution, Knowledge Based Systems, 2025, 331, pp. 114844