| Issue |
ITM Web Conf.
Volume 78, 2025
International Conference on Computer Science and Electronic Information Technology (CSEIT 2025)
|
|
|---|---|---|
| Article Number | 04016 | |
| Number of page(s) | 7 | |
| Section | Foundations and Frontiers in Multimodal AI, Large Models, and Generative Technologies | |
| DOI | https://doi.org/10.1051/itmconf/20257804016 | |
| Published online | 08 September 2025 | |
Srgan-Sa: Super-Resolution Generative Adversarial Network with Integrated Attention Mechanisms
Faculty of Printing, Packing and Digtial Media Technology, Xi'An University of Technology, Shaanxi, China
International Education School, Hebei University of Economics and Business, Shijiazhuang, Hebei, China
School of International Education, Guangxi University of Science and Technology, Liuzhou, Guangxi, China
With the development of the times, this paper proposes a dynamic spatial attention enhancement SRGAN-SA model to better solve the above problems, such as the loss of high-frequency details in traditional image super-resolution reconstruction (SISR) methods, and the insufficient attention of SRGAN global feature processing mechanism to key areas. Based on the deep learning module, the model designs a learnable 7 * 7 convolution to generate dynamic attention maps. It’s embedded in the residual module of SRResNet generator to realize adaptive focus and feature reweighting of regions. The model adopts the confrontation training paradigm, combines VGG perception loss and confrontation loss optimization generator, and uses depth convolution discriminator to enhance texture identification. Experimental results show that compared with SRGAN, the PSNR of this method on Set5 test set is increased by 1.747dB, reaching 31.6381, and the SSIM is increased by 0.0377, reaching 0.8863. The visualization results show that the improved model can effectively suppress artifacts. The ablation experiment further verified that the contribution rate of the dual pool strategy and the dynamic convolution design to the performance improvement reached 0.82 dB and 0.65 dB respectively. It provides a new solution for the deep integration of attention mechanism and confrontation training.
© The Authors, published by EDP Sciences, 2025
This is an Open Access article distributed under the terms of the Creative Commons Attribution License 4.0, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.
Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.
Initial download of the metrics may take a while.

