Adaptive Dynamic Shuffle Convolutional Parallel Network for Image Super-Resolution
Abstract
:1. Introduction
- We propose a novel lightweight network architecture, named ADSCPN, designed to enhance feature learning across both channel and spatial dimensions. This innovative approach markedly improves the performance of image SR reconstruction.
- We introduce an innovative feature processing module, incorporating multi-layer pointwise perceptrons and shuffling attention, which enhances feature extraction through grouped convolutions and channel shuffling. Additionally, the framework integrates large-kernel convolution groups with dynamic convolutions, effectively reducing model size while maintaining high computational efficiency. These advancements enable the model to capture deep semantic features with low complexity, leading to a significant improvement in the quality of reconstructed images.
- We conduct extensive experiments on multiple benchmark datasets to evaluate the performance of our method. The results demonstrate that our model achieves superior performance compared to several state-of-the-art approaches, while maintaining low computational complexity.
2. Related Work
2.1. Traditional Super-Resolution Methods
2.2. Deep Learning-Based Super-Resolution Methods
2.3. Lightweight Super-Resolution Methods
2.4. Channel and Spatial Attention Mechanisms
2.5. Application of Large-Kernel Convolutions
2.6. Summary
3. Methodology
3.1. Notations and Problem Definition
3.2. Architecture of ADSCPN
3.3. Feature Processing Block
3.3.1. Channel Processing Block
3.3.2. Feature Extraction Block
3.3.3. Point-Attention Convolution Group
3.3.4. Efficiently Enhanced Spatial Attention
3.4. Optimization
Algorithm 1 The pseudo-code of ADSCN Algorithm |
|
4. Experiment
4.1. Dataset
- DIV2K. The DIV2K dataset comprises 1000 high-resolution images, characterized by a rich variety of features and diverse scene categories, including environments, humans, and other objects. For experimental purposes, the first 800 images are designated for training, images 801–900 are allocated to the validation set, and the remaining 100 images are reserved for performance evaluation. Each image in the dataset has a native 2 K resolution, accompanied by corresponding low-resolution images generated through the application of various degradation kernels, such as Gaussian white noise. This dataset is widely utilized for training and benchmarking super-resolution models due to its diversity and high-quality annotations.
- Set5 and Set14. The Set5 and Set14 datasets are among the earliest small-scale benchmarks introduced in the field of image processing. Set5 contains 5 images, while Set14 consists of 14 images, both encompassing a mix of human subjects and natural scenes. Despite their limited size, these datasets remain widely used for testing and validation purposes, offering a convenient and efficient means of assessing model performance, particularly in super-resolution tasks.
- B100. The B100 dataset comprises 100 images depicting a variety of subjects, including animals, plants, and real-world scenes. While offering a diverse range of content, the images in this dataset are relatively small in resolution and exhibit less detailed textures compared to other high-resolution benchmarks, making it a useful resource for evaluating super-resolution models under more constrained conditions.
- Urban100. The Urban100 dataset contains 100 high-resolution images, specifically focused on capturing intricate details of urban architecture. These images provide complex structural patterns and fine-grained textures, making the dataset particularly challenging and suitable for evaluating the performance of super-resolution models on architectural and man-made scenes.
- DRealSR and RealSR. The RealSR and DRealSR datasets are created using DSLR cameras with multiple zoom levels, varying aperture settings, and different lens focal lengths to simulate signal noise introduced during the degradation process. An image registration algorithm is employed to precisely align high- and low-resolution image pairs. These datasets encompass a wide range of scenes, including natural landscapes, architectural structures, as well as human and animal subjects. Both datasets are widely used for real-world image super-resolution reconstruction tasks, providing realistic and challenging conditions for model evaluation.
4.2. Evaluation Metrics
4.3. Baselines and Implementation Details
4.4. Performance Evaluation
4.4.1. Model Parameter Analysis
4.4.2. Quantitative Experimental Results Comparison
4.4.3. Qualitative Results Comparison
4.4.4. Model Complexity
4.5. Ablation Analysis
4.5.1. Effectiveness of Large-Kernel Parallel Convolution Groups
4.5.2. Effectiveness of Dynamic Convolution
4.5.3. Effectiveness of the Point Attention Convolution Group
4.5.4. Effectiveness of the Efficient Enhanced Spatial Attention Block
4.5.5. Effectiveness of Combined Components
4.5.6. Effectiveness of Activation Functions
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
Abbreviations
SR | Super-resolution |
HR | High-resolution |
LR | Low-resolution |
ADSCPN | Adaptive dynamic shuffle convolutional parallel network |
CNNs | Convolutional neural networks |
LD | Feature processing block |
CPB | Channel processing block |
FEB | Feature extraction block |
PW-MLP | Point-wise multi-layer perceptron |
PACG | Point-attention convolution group |
EESA | Efficiently enhanced spatial attention |
ESA | Enhanced spatial attention |
PSNR | Peak signal-to-noise ratio |
SSIM | Structural similarity index |
References
- Chen, Z.; Zhang, Y.; Gu, J.; Kong, L.; Yang, X.; Yu, F. Dual aggregation transformer for image super-resolution. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 12312–12321. [Google Scholar]
- Chen, H.; He, X.; Qing, L.; Wu, Y.; Ren, C.; Sheriff, R.E.; Zhu, C. Real-world single image super-resolution: A brief review. Inf. Fusion 2022, 79, 124–145. [Google Scholar] [CrossRef]
- Zhang, Q.; Xiao, J.; Tian, C.; Chun-Wei Lin, J.; Zhang, S. A robust deformed convolutional neural network (CNN) for image denoising. CAAI Trans. Intell. Technol. 2023, 8, 331–342. [Google Scholar] [CrossRef]
- Wang, C.; Lv, X.; Shao, M.; Qian, Y.; Zhang, Y. A novel fuzzy hierarchical fusion attention convolution neural network for medical image super-resolution reconstruction. Inf. Sci. 2023, 622, 424–436. [Google Scholar] [CrossRef]
- Georgescu, M.I.; Ionescu, R.T.; Miron, A.I.; Savencu, O.; Ristea, N.C.; Verga, N.; Khan, F.S. Multimodal multi-head convolutional attention with various kernel sizes for medical image super-resolution. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–7 January 2023; pp. 2195–2205. [Google Scholar]
- Cornebise, J.; Oršolić, I.; Kalaitzis, F. Open high-resolution satellite imagery: The worldstrat dataset–with application to super-resolution. Adv. Neural Inf. Process. Syst. 2022, 35, 25979–25991. [Google Scholar]
- Arefin, M.R.; Michalski, V.; St-Charles, P.L.; Kalaitzis, A.; Kim, S.; Kahou, S.E.; Bengio, Y. Multi-image super-resolution for remote sensing using deep recurrent networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 206–207. [Google Scholar]
- Zhu, L.; Yu, W.; Zhu, X.; Zhang, C.; Li, Y.; Zhang, S. MvHAAN: Multi-view hierarchical attention adversarial network for person re-identification. World Wide Web 2024, 27, 59. [Google Scholar] [CrossRef]
- Zhang, C.; Xie, F.; Yu, H.; Zhang, J.; Zhu, L.; Li, Y. PPIS-JOIN: A novel privacy-preserving image similarity join method. Neural Process. Lett. 2022, 54, 2783–2801. [Google Scholar] [CrossRef]
- Zheng, M.; Xu, J.; Shen, Y.; Tian, C.; Li, J.; Fei, L.; Zong, M.; Liu, X. Attention-based CNNs for image classification: A survey. In Proceedings of the Journal of Physics: Conference Series; IOP Publishing: Bristol, UK, 2022; Volume 2171, p. 012068. [Google Scholar]
- Li, Q.; Chen, Y.; Zhang, A.; Jiang, Y.; Zou, L.; Xu, Z.; Muntean, G.M. A super-resolution flexible video coding solution for improving live streaming quality. IEEE Trans. Multimed. 2022, 25, 6341–6355. [Google Scholar] [CrossRef]
- Huang, K.; Tian, C.; Xu, Z.; Li, N.; Lin, J.C.W. Motion Context guided Edge-preserving network for video salient object detection. Expert Syst. Appl. 2023, 233, 120739. [Google Scholar] [CrossRef]
- Zhang, Y.; Zhang, Y.; Wu, Y.; Tao, Y.; Bian, K.; Zhou, P.; Song, L.; Tuo, H. Improving quality of experience by adaptive video streaming with super-resolution. In Proceedings of the IEEE INFOCOM 2020-IEEE Conference on Computer Communications, Virtual, 6–9 July 2020; pp. 1957–1966. [Google Scholar]
- Qiao, C.; Li, D.; Guo, Y.; Liu, C.; Jiang, T.; Dai, Q.; Li, D. Evaluation and development of deep neural networks for image super-resolution in optical microscopy. Nat. Methods 2021, 18, 194–202. [Google Scholar] [CrossRef]
- Gendy, G.; He, G.; Sabor, N. Lightweight image super-resolution based on deep learning: State-of-the-art and future directions. Inf. Fusion 2023, 94, 284–310. [Google Scholar] [CrossRef]
- Tian, C.; Fei, L.; Zheng, W.; Xu, Y.; Zuo, W.; Lin, C.W. Deep learning on image denoising: An overview. Neural Netw. 2020, 131, 251–275. [Google Scholar] [CrossRef] [PubMed]
- Tian, C.; Xu, Y.; Fei, L.; Yan, K. Deep learning for image denoising: A survey. In Proceedings of the Genetic and Evolutionary Computing: Proceedings of the Twelfth International Conference on Genetic and Evolutionary Computing, Changzhou, China, 14–17 December 2018; Springer: Berlin/Heidelberg, Germany, 2019; pp. 563–572. [Google Scholar]
- Zhu, L.; Zhang, C.; Song, J.; Liu, L.; Zhang, S.; Li, Y. Multi-graph based hierarchical semantic fusion for cross-modal representation. In Proceedings of the 2021 IEEE International Conference on Multimedia and Expo (ICME), Virtual, 5–9 July 2021; pp. 1–6. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Tian, C.; Xu, Y.; Li, Z.; Zuo, W.; Fei, L.; Liu, H. Attention-guided CNN for image denoising. Neural Netw. 2020, 124, 117–129. [Google Scholar] [CrossRef] [PubMed]
- Zhu, L.; Song, J.; Wei, X.; Yu, H.; Long, J. CAESAR: Concept augmentation based semantic representation for cross-modal retrieval. Multimed. Tools Appl. 2022, 81, 34213–34243. [Google Scholar] [CrossRef]
- Tian, C.; Xu, Y.; Zuo, W. Image denoising using deep CNN with batch renormalization. Neural Netw. 2020, 121, 461–473. [Google Scholar] [CrossRef]
- Tian, C.; Yuan, Y.; Zhang, S.; Lin, C.W.; Zuo, W.; Zhang, D. Image super-resolution with an enhanced group convolutional neural network. Neural Netw. 2022, 153, 373–385. [Google Scholar] [CrossRef]
- Wang, Z.; Chen, J.; Hoi, S.C. Deep learning for image super-resolution: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 3365–3387. [Google Scholar] [CrossRef] [PubMed]
- Tian, C.; Zhang, X.; Lin, J.C.W.; Zuo, W.; Zhang, Y.; Lin, C.W. Generative adversarial networks for image super-resolution: A survey. arXiv 2022, arXiv:2204.13620. [Google Scholar]
- Dong, C.; Loy, C.C.; Tang, X. Accelerating the super-resolution convolutional neural network. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part II 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 391–407. [Google Scholar]
- Dong, C.; Loy, C.C.; He, K.; Tang, X. Image super-resolution using deep convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 295–307. [Google Scholar] [CrossRef]
- Kim, J.; Lee, J.K.; Lee, K.M. Deeply-recursive convolutional network for image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1637–1645. [Google Scholar]
- Tai, Y.; Yang, J.; Liu, X. Image super-resolution via deep recursive residual network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3147–3155. [Google Scholar]
- Lim, B.; Son, S.; Kim, H.; Nah, S.; Mu Lee, K. Enhanced deep residual networks for single image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 136–144. [Google Scholar]
- Ledig, C.; Theis, L.; Huszar, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; Volume 2017, pp. 4681–4690. [Google Scholar]
- Tian, C.; Zhang, X.; Zhang, Q.; Yang, M.; Ju, Z. Image super-resolution via dynamic network. CAAI Trans. Intell. Technol. 2024, 9, 837–849. [Google Scholar] [CrossRef]
- Tian, C.; Zhang, X.; Ren, J.; Zuo, W.; Zhang, Y.; Lin, C.W. A Heterogeneous Dynamic Convolutional Neural Network for Image Super-resolution. arXiv 2024, arXiv:2402.15704. [Google Scholar]
- Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
- Ahn, N.; Kang, B.; Sohn, K.A. Fast, accurate, and lightweight super-resolution with cascading residual network. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 252–268. [Google Scholar]
- Hui, Z.; Wang, X.; Gao, X. Fast and accurate single image super-resolution via information distillation network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 723–731. [Google Scholar]
- Hui, Z.; Gao, X.; Yang, Y.; Wang, X. Lightweight image super-resolution with information multi-distillation network. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 2024–2032. [Google Scholar]
- Liu, J.; Tang, J.; Wu, G. Residual feature distillation network for lightweight image super-resolution. In Proceedings of the Computer Vision–ECCV 2020 Workshops, Glasgow, UK, 23–28 August 2020; Proceedings, Part III 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 41–55. [Google Scholar]
- Kong, F.; Li, M.; Liu, S.; Liu, D.; He, J.; Bai, Y.; Chen, F.; Fu, L. Residual local feature network for efficient super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022; pp. 766–776. [Google Scholar]
- Zhang, A.; Ren, W.; Liu, Y.; Cao, X. Lightweight image super-resolution with superpixel token interaction. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–6 October 2023; pp. 12728–12737. [Google Scholar]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
- Sun, L.; Pan, J.; Tang, J. Shufflemixer: An efficient convnet for image super-resolution. Adv. Neural Inf. Process. Syst. 2022, 35, 17314–17326. [Google Scholar]
- Tian, C.; Zhuge, R.; Wu, Z.; Xu, Y.; Zuo, W.; Chen, C.; Lin, C.W. Lightweight image super-resolution with enhanced CNN. Knowl. Based Syst. 2020, 205, 106235. [Google Scholar] [CrossRef]
- Conde, M.V.; Choi, U.J.; Burchi, M.; Timofte, R. Swin2sr: Swinv2 transformer for compressed image super-resolution and restoration. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; pp. 669–687. [Google Scholar]
- Hou, H.; Andrews, H. Cubic splines for image interpolation and digital filtering. IEEE Trans. Acoust. Speech Signal Process. 1978, 26, 508–517. [Google Scholar]
- Keys, R. Cubic convolution interpolation for digital image processing. IEEE Trans. Acoust. Speech Signal Process. 1981, 29, 1153–1160. [Google Scholar] [CrossRef]
- Li, J.; Wu, J.; Deng, H.; Liu, J. A self-learning image super-resolution method via sparse representation and non-local similarity. Neurocomputing 2016, 184, 196–206. [Google Scholar] [CrossRef]
- Howard, A.G. MobileNets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
- Ma, N.; Zhang, X.; Zheng, H.T.; Sun, J. Shufflenet v2: Practical guidelines for efficient cnn architecture design. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 116–131. [Google Scholar]
- Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
- Hu, Y.; Li, J.; Huang, Y.; Gao, X. Channel-wise and spatial feature modulation network for single image super-resolution. IEEE Trans. Circuits Syst. Video Technol. 2019, 30, 3911–3927. [Google Scholar] [CrossRef]
- Li, Z.; Yang, J.; Liu, Z.; Yang, X.; Jeon, G.; Wu, W. Feedback network for image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 3867–3876. [Google Scholar]
- Tian, C.; Zhang, Y.; Zuo, W.; Lin, C.W.; Zhang, D.; Yuan, Y. A heterogeneous group CNN for image super-resolution. In Proceedings of the IEEE Transactions on Neural Networks and Learning Systems, Virtual Event, 23–28 May 2022. [Google Scholar]
- Luo, Z.; Huang, H.; Yu, L.; Li, Y.; Fan, H.; Liu, S. Deep constrained least squares for blind image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 21–24 June 2022; pp. 17642–17652. [Google Scholar]
- Ma, C.; Yang, C.Y.; Yang, X.; Yang, M.H. Learning a no-reference quality metric for single-image super-resolution. Comput. Vis. Image Underst. 2017, 158, 1–16. [Google Scholar] [CrossRef]
- Bevilacqua, M.; Roumy, A.; Guillemot, C.; Alberi-Morel, M.L. Low-complexity single-image super-resolution based on nonnegative neighbor embedding. In Proceedings of the 23rd British Machine Vision Conference (BMVC), Surrey, UK, 3–7 September 2012; BMVA Press: Durham, UK, 2012; pp. 135.1–135.10, ISBN 1-901725-46-4. [Google Scholar] [CrossRef]
- Zeyde, R.; Elad, M.; Protter, M. On single image scale-up using sparse-representations. In Proceedings of the Curves and Surfaces: 7th International Conference, Avignon, France, 24–30 June 2010; Revised Selected Papers 7. Springer: Berlin/Heidelberg, Germany, 2012; pp. 711–730. [Google Scholar]
- Martin, D.; Fowlkes, C.; Tal, D.; Malik, J. A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In Proceedings of the Eighth IEEE international Conference on Computer Vision (ICCV), Vancouver, BC, Canada, 7–14 July 2001; Volume 2, pp. 416–423. [Google Scholar]
- Huang, J.B.; Singh, A.; Ahuja, N. Single image super-resolution from transformed self-exemplars. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, , 7–12 June 2015; pp. 5197–5206. [Google Scholar]
- Chen, X.; Wang, X.; Zhou, J.; Qiao, Y.; Dong, C. Activating more pixels in image super-resolution transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–23 June 2023; pp. 22367–22377. [Google Scholar]
- Huang, Y.; Li, S.; Wang, L.; Tan, T.; Luo, Z. Unfolding the alternating optimization for blind super resolution. Adv. Neural Inf. Process. Syst. 2020, 33, 5632–5643. [Google Scholar]
- Fujimoto, A.; Ogawa, T.; Yamamoto, K.; Matsui, Y.; Yamasaki, T.; Aizawa, K. Manga109 dataset and creation of metadata. In Proceedings of the 1st International Workshop on Comics Analysis, Processing and Understanding, Cancun, Mexico, 4–7 December 2016; pp. 1–5. [Google Scholar]
- Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
- Tai, Y.; Yang, J.; Liu, X.; Xu, C. Memnet: A persistent memory network for image restoration. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. pp. 4539–4547.
- Lai, W.S.; Huang, J.B.; Ahuja, N.; Yang, M.H. Deep laplacian pyramid networks for fast and accurate super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 624–632. [Google Scholar]
- Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A convnet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 11976–11986. [Google Scholar]
- Ding, X.; Zhang, X.; Han, J.; Ding, G. Scaling up your kernels to 31x31: Revisiting large kernel design in cnns. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 11963–11975. [Google Scholar]
- Chu, X.; Zhang, B.; Ma, H.; Xu, R.; Li, Q. Fast, accurate and lightweight super-resolution with neural architecture search. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 59–64. [Google Scholar]
- Gou, Y.; Hu, P.; Lv, J.; Zhu, H.; Peng, X. Rethinking image super resolution from long-tailed distribution learning perspective. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 14327–14336. [Google Scholar]
- Yao, J.E.; Tsao, L.Y.; Lo, Y.C.; Tseng, R.; Chang, C.C.; Lee, C.Y. Local implicit normalizing flow for arbitrary-scale image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 1776–1785. [Google Scholar]
- Agustsson, E.; Timofte, R.N. Challenge on single image super-resolution: Dataset and study. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2016; pp. 21–26. [Google Scholar]
Notation | Definition |
---|---|
Low-resolution input image | |
High-resolution target image | |
Super-resolution output image | |
r | Scaling factor |
Height, width, and channels of the image | |
Shallow feature maps extracted from | |
Deep features after processing through K blocks | |
Features after channel projection | |
Features after spatial attention block | |
Attention weights for spatial features |
Group Number | Parameters | MACs | Set5 | Set14 | B100 |
---|---|---|---|---|---|
(K) | (G) | PSNR/SSIM | PSNR/SSIM | PSNR/SSIM | |
2 | 762 | 651 | 37.92/0.9603 | 33.45/0.9165 | 32.10/0.8987 |
4 | 703 | 597 | 37.88/0.9602 | 33.44/0.9170 | 32.10/0.8989 |
6 | 684 | 579 | 37.90/0.9602 | 33.44/0.9169 | 32.10/0.8988 |
8 | 674 | 570 | 37.90/0.9602 | 33.51/0.9174 | 32.11/0.8990 |
Method | Scale Factor | Set5 | Set14 | B100 | Urban100 |
---|---|---|---|---|---|
PSNR/SSIM | PSNR/SSIM | PSNR/SSIM | PSNR/SSIM | ||
Bicubic | 33.66/0.9299 | 30.24/0.8688 | 29.56/0.8431 | 26.88/0.8403 | |
SRCNN | 36.66/0.9542 | 32.42/0.9063 | 31.36/0.8964 | 29.50/0.8946 | |
FSRCNN | 37.00/0.9558 | 32.63/0.9088 | 31.53/0.8920 | 29.88/0.9009 | |
VDSR | 37.53/0.9590 | 33.05/0.9130 | 31.90/0.8960 | 30.77/0.9140 | |
LapSRN | 37.52/0.9591 | 32.99/0.9134 | 31.80/0.8952 | 30.41/0.9100 | |
DRCN | 37.63/0.9588 | 32.98/0.9130 | 31.85/0.8942 | 30.75/0.9133 | |
DRRN | 37.74/0.9591 | 33.23/0.9145 | 32.05/0.8973 | 31.23/0.9188 | |
CARN | 37.76/0.9590 | 33.52/0.9166 | 32.09/0.8978 | 31.92/0.9256 | |
MemNet | 37.78/0.9597 | 33.28/0.9142 | 32.08/0.8978 | 31.31/0.9195 | |
IDN | 37.83/0.9600 | 33.30/0.9148 | 32.08/0.8985 | 31.27/0.9196 | |
FALSR-A | 37.82/0.9595 | 33.55/0.9168 | 32.12/0.8987 | 31.93/0.9256 | |
LINF | 37.49/— | 33.38/— | 32.16/— | 31.22/— | |
FPL | 35.73/ 0.9509 | 31.59/0.9059 | 30.75/0.9022 | 30.46/0.9257 | |
ADSCPN | 37.84/0.9600 | 33.42/0.9165 | 32.06/0.8983 | 31.67/0.9238 | |
ADSCPN-plus | 37.90/0.9602 | 33.51/0.9174 | 32.11/0.8990 | 31.93/0.9262 |
Method | Scale Factor | Set5 | Set14 | B100 | Urban100 |
---|---|---|---|---|---|
PSNR/SSIM | PSNR/SSIM | PSNR/SSIM | PSNR/SSIM | ||
Bicubic | 30.39/0.8682 | 27.55/0.7742 | 27.21/0.7385 | 24.46/0.7349 | |
SRCNN | 32.75/0.9090 | 29.30/0.8203 | 28.41/0.7863 | 26.24/0.8090 | |
FSRCNN | 33.06/0.9140 | 29.43/0.8242 | 28.53/0.7910 | 26.43/0.8080 | |
VDSR | 33.66/0.9213 | 29.77/0.8314 | 28.82/0.7976 | 27.14/0.8279 | |
DRCN | 33.82/0.9226 | 29.76/0.8311 | 28.80/0.7963 | 27.15/0.8276 | |
DRRN | 34.03/0.9244 | 29.96/0.8349 | 28.95/0.8004 | 27.53/0.8378 | |
IDN | 34.11/0.9253 | 30.13/0.8360 | 29.01/0.8013 | 27.40/0.8359 | |
CARN-M | 33.99/0.9245 | 30.08/0.8352 | 28.91/0.8000 | 27.47/0.8371 | |
MemNet | 34.09/0.9248 | 30.00/0.8350 | 28.96/0.8001 | 27.56/0.8376 | |
MADNet | 34.16/0.9253 | 30.21/0.8398 | 29.08/0.8023 | 27.82/0.8423 | |
LINF | 33.94/— | 29.84/— | 28.55/— | 28.39/— | |
ADSCPN | 34.21/0.9252 | 30.22/0.8398 | 28.99/0.8024 | 27.81/0.8444 | |
ADSCPN-plus | 34.23/0.9258 | 30.24/0.8399 | 29.10/0.8033 | 28.78/0.8463 |
Method | Scale Factor | Set5 | Set14 | B100 | Urban100 |
---|---|---|---|---|---|
PSNR/SSIM | PSNR/SSIM | PSNR/SSIM | PSNR/SSIM | ||
Bicubic | 28.42/0.8104 | 26.00/0.7027 | 25.96/0.6675 | 23.14/0.6577 | |
SRCNN | 30.48/0.8628 | 27.50/0.7513 | 26.90/0.7101 | 24.52/0.7221 | |
FSRCNN | 30.72/0.8660 | 27.61/0.7550 | 26.98/0.7150 | 24.62/0.7280 | |
VDSR | 31.35/0.8838 | 28.01/0.7674 | 27.29/0.7251 | 25.18/0.7524 | |
LapSRN | 31.54/0.8852 | 28.09/0.7700 | 27.32/0.7275 | 25.21/0.7562 | |
DRCN | 31.53/0.8854 | 28.02/0.7670 | 27.23/0.7233 | 25.14/0.7510 | |
DRRN | 31.68/0.8888 | 28.21/0.7720 | 27.38/0.7284 | 25.44/0.7638 | |
IDN | 31.82/0.8903 | 28.25/0.7730 | 27.41/0.7297 | 25.41/0.7632 | |
CARN-M | 31.92/0.8903 | 28.42/0.7762 | 27.44/0.7304 | 25.62/0.7694 | |
MemNet | 31.74/0.8893 | 28.26/0.7723 | 27.40/0.7281 | 25.50/0.7630 | |
MADNet | 31.93/0.8917 | 28.44/0.7780 | 27.47/0.7327 | 25.76/0.7746 | |
ECBSR | 31.92/0.8946 | 28.34/0.7817 | 27.48/0.7393 | 25.53/0.7773 | |
LINF | 31.70/— | 27.54/— | 26.62/— | 25.15/— | |
FPL | 30.39/0.8805 | 26.91/0.7688 | 26.29/0.7393 | 24.78/0.7880 | |
ADSCPN | 32.04/0.8929 | 28.50/0.7793 | 27.52/0.7339 | 25.95/0.7812 | |
ADSCPN-plus | 32.12/0.8940 | 28.49/0.7794 | 27.53/0.7345 | 25.98/0.7823 |
Method | ||
---|---|---|
PSNR/SSIM | PSNR/SSIM | |
Bicubic | 28.69/0.8058 | 25.38/0.6822 |
ADSCPN-plus | 29.19/0.8227 | 25.65/0.6940 |
EDSR(192) | 29.21/0.8234 | 25.66/0.6945 |
ADRBN-plus | 29.21/0.8233 | 25.66/0.6944 |
Method | ×2 | ×3 | ×4 |
---|---|---|---|
PSNR/SSIM | PSNR/SSIM | PSNR/SSIM | |
Bicubic | 32.67/0.9060 | 31.51/0.8700 | 30.56/0.8595 |
ADSCPN-plus | 32.82/0.9089 | 31.60/0.8727 | 30.62/0.8611 |
Method | ×2 | ×3 | ×4 |
---|---|---|---|
PSNR/SSIM | PSNR/SSIM | PSNR/SSIM | |
Bicubic | 31.67/0.8865 | 28.63/0.8085 | 27.23/0.7637 |
ADSCPN-plus | 34.01/0.9225 | 30.88/0.8465 | 29.27/0.8255 |
Method | Parameters (K) | MACs (G) |
---|---|---|
SRCNN | 57 | 53 |
VDSR | 665 | 613 |
LapSRN | 813 | 30 |
DRCN | 1774 | 17,974 |
DRRN | 297 | 6797 |
MemNet | 677 | 2662 |
CARN | 1592 | 223 |
MADNet | 878 | 187 |
FALSR-A | 1021 | 235 |
Ours | 674 | 570 |
Convolution | Scale Factor | Set5 | Set14 | B100 | Urban100 |
---|---|---|---|---|---|
Structure | PSNR/SSIM | PSNR/SSIM | PSNR/SSIM | PSNR/SSIM | |
Large Kernel Parallel Depthwise Separable Conv Group | ×2 | 37.90/0.9602 | 33.51/0.9174 | 32.11/0.8990 | 31.93/0.9262 |
7 × 7 Depthwise Separable Conv | ×2 | 37.92/0.9603 | 33.43/0.9167 | 31.10/0.8988 | 31.86/0.9257 |
Convolution Structure | Parameters (K) | MACs (G) |
---|---|---|
Large-Kernel Parallel Depthwise Separable Conv. Group | 674 | 570.07 |
7 × 7 Depthwise Separable Convolution | 732 | 624.22 |
Convolution Structure | Scale Factor | Set5 | Set14 | B100 | Urban100 |
---|---|---|---|---|---|
PSNR/SSIM | PSNR/SSIM | PSNR/SSIM | PSNR/SSIM | ||
Dynamic Convolution | 37.90/0.9602 | 33.51/0.9174 | 32.11/0.8990 | 31.93/0.9262 | |
Conv | 37.88/0.9601 | 33.46/0.9167 | 32.10/0.8988 | 31.82/0.9253 |
Convolution Structure | Scale Factor | Set5 | Set14 | B100 | Urban100 |
---|---|---|---|---|---|
PSNR/SSIM | PSNR/SSIM | PSNR/SSIM | PSNR/SSIM | ||
Point Attention Convolution Group | 37.90/0.9602 | 33.51/0.9174 | 32.11/0.8990 | 31.93/0.9262 | |
Standard Convolution Group | 37.90/0.9601 | 33.46/0.9169 | 32.09/0.8986 | 31.86/0.9256 |
Spatial Attention Block | Parameters (K) | MACs (G) |
---|---|---|
Enhanced Spatial Attention Block | 740 | 574.80 |
Efficient Enhanced Spatial Attention Block | 674 | 570.07 |
Large Kernel Parallel | Dynamic | Point Attention | Set5 | Set14 | B100 | Urban100 |
---|---|---|---|---|---|---|
Depthwise Separable Conv Group | Convolution | Convolution Group | PSNR/SSIM | PSNR/SSIM | PSNR/SSIM | PSNR/SSIM |
× | × | × | 37.93/0.9602 | 33.44/0.9166 | 32.11/0.8989 | 31.88/0.9257 |
× | × | ✓ | 37.88/0.9601 | 33.43/0.9159 | 32.10/0.8988 | 31.84/0.9257 |
× | ✓ | × | 37.88/0.9602 | 33.46/0.9168 | 32.10/0.8986 | 31.82/0.9254 |
✓ | × | × | 37.88/0.9601 | 33.46/0.9167 | 32.09/0.8988 | 31.80/0.9253 |
× | ✓ | ✓ | 37.92/0.9603 | 33.43/0.9167 | 31.86/0.9257 | 31.88/0.9258 |
✓ | × | ✓ | 37.87/0.9601 | 33.39/0.9161 | 32.10/0.8986 | 31.82/0.9252 |
✓ | ✓ | × | 37.90/0.9601 | 33.46/0.9169 | 32.11/0.8986 | 31.83/0.9256 |
✓ | ✓ | ✓ | 37.90/0.9602 | 33.51/0.9174 | 32.11/0.8990 | 31.93/0.9262 |
Activation Function | Scaling Factor | Set5 | Set14 | B100 | Urban100 |
---|---|---|---|---|---|
PSNR/SSIM | PSNR/SSIM | PSNR/SSIM | PSNR/SSIM | ||
SiLU | 37.91/0.9602 | 33.45/0.9168 | 32.10/0.8989 | 31.88/0.9259 | |
GELU | 37.90/0.9602 | 33.51/0.9174 | 32.11/0.8990 | 31.93/0.9262 | |
ReLU | 37.91/0.9602 | 33.42/0.9167 | 32.12/0.8991 | 31.94/0.9263 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Long, Y.; Ruan, H.; Zhao, H.; Liu, Y.; Zhu, L.; Zhang, C.; Zhu, X. Adaptive Dynamic Shuffle Convolutional Parallel Network for Image Super-Resolution. Electronics 2024, 13, 4613. https://doi.org/10.3390/electronics13234613
Long Y, Ruan H, Zhao H, Liu Y, Zhu L, Zhang C, Zhu X. Adaptive Dynamic Shuffle Convolutional Parallel Network for Image Super-Resolution. Electronics. 2024; 13(23):4613. https://doi.org/10.3390/electronics13234613
Chicago/Turabian StyleLong, Yiting, Haoyu Ruan, Hui Zhao, Yi Liu, Lei Zhu, Chengyuan Zhang, and Xinghui Zhu. 2024. "Adaptive Dynamic Shuffle Convolutional Parallel Network for Image Super-Resolution" Electronics 13, no. 23: 4613. https://doi.org/10.3390/electronics13234613
APA StyleLong, Y., Ruan, H., Zhao, H., Liu, Y., Zhu, L., Zhang, C., & Zhu, X. (2024). Adaptive Dynamic Shuffle Convolutional Parallel Network for Image Super-Resolution. Electronics, 13(23), 4613. https://doi.org/10.3390/electronics13234613