Next Article in Journal
A Receiver-Driven Named Data Networking (NDN) Congestion Control Method Based on Reinforcement Learning
Previous Article in Journal
Multitasking Feature Selection Using a Clonal Selection Algorithm for High-Dimensional Microarray Data
Previous Article in Special Issue
NRPerson: A Non-Registered Multi-Modal Benchmark for Tiny Person Detection and Localization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Dynamic Shuffle Convolutional Parallel Network for Image Super-Resolution

1
College of Information and Intelligence, Hunan Agricultural University, Changsha 410128, China
2
College of Computer Science and Electronic Engineering, Hunan University, Changsha 410082, China
*
Authors to whom correspondence should be addressed.
Electronics 2024, 13(23), 4613; https://doi.org/10.3390/electronics13234613
Submission received: 29 September 2024 / Revised: 8 November 2024 / Accepted: 11 November 2024 / Published: 22 November 2024
(This article belongs to the Special Issue Big Model Techniques for Image Processing)

Abstract

:
Image super-resolution has experienced significant advancements with the emergence of deep learning technology. However, deploying highly complex super-resolution networks on resource-constrained devices poses a challenge due to their substantial computational requirements. This paper presents the Adaptive Dynamic Shuffle Convolutional Parallel Network (ADSCPN), a novel lightweight super-resolution model designed to achieve an optimal balance between computational efficiency and image reconstruction quality. The ADSCPN framework employs large-kernel parallel depthwise separable convolutions, dynamic convolutions, and an enhanced attention mechanism to optimize feature extraction and improve detail preservation. Extensive evaluations on standard benchmark datasets demonstrate that ADSCPN achieves state-of-the-art performance while significantly reducing computational complexity, making it well-suited for practical applications on devices with limited computational resources.

1. Introduction

Image super-resolution (SR) [1,2,3] refers to the process of reconstructing or enhancing a low-resolution image to obtain a higher-resolution version, with more detailed and finer visual features. It aims to increase the spatial resolution of an image, which typically involves predicting or estimating the missing high-frequency details from the available low-frequency data. SR is crucial for applications where high-quality images are necessary but only low-resolution versions are available, such as in medical imaging [4,5], satellite photography [6,7], image recognition [8,9,10] and video streaming [11,12,13]. Conventional SR techniques, such as interpolation or statistical model-based approaches, exhibit limitations in reconstructing high-frequency details of images. In contrast, deep learning technologies [14,15,16,17,18], particularly through models like convolutional neural networks (CNNs) [19,20,21,22], have significantly enhanced the performance of SR. By employing an end-to-end training framework, deep learning models are capable of automatically learning the complex mapping between low-resolution and high-resolution images from large-scale datasets, thereby circumventing the constraints associated with manually designed feature extraction.
Most current deep learning-based methods [1,23,24,25], such as FSRCNN [26], directly utilize low-resolution images as input, modifying the nonlinear mapping approach of SRCNN [27] by initially reducing the dimensionality and subsequently upscaling it, with convolutional layers employed for feature extraction in between. Following this, DRCN [28] introduced a structure comprising three components: an embedding network, an inference network with recursive layers that increase network depth without significantly increasing the number of parameters, and a reconstruction network. DRRN [29] builds upon these algorithms by incorporating concepts from DRCN and utilizing local skip connections inspired by ResNet [19]. It further integrates global skip connections and recursive blocks to deepen the network while controlling the parameter count. EDSR [30], through an analysis of the residual blocks in SRResNet [31], observed that batch normalization layers tend to distort image color and contrast, negatively affecting image quality. Consequently, they removed the batch normalization layers, simplifying the residual blocks and stacking more of them to increase the model’s capacity. In addition, ESRGCNN [23] introduced an enhanced group convolutional strategy to improve feature extraction, while DSRNet [32] utilized dynamic networks to adaptively adjust its internal structure based on input characteristics. Moreover, HDSRNet [33] proposed a heterogeneous dynamic convolutional network to balance computational cost and performance in SR tasks.
While these deep models substantially enhance the reconstruction performance of super-resolution images, they generally achieve this by increasing network depth through direct block stacking or recursive structures. This approach often results in challenging training processes, high computational complexity, slow processing speeds, and a large number of parameters. Furthermore, these networks do not adequately account for the effects of convolutional structures and components on super-resolution performance, thereby imposing a considerable computational burden. Notably, very deep networks such as VDSR [34] incur significant computational costs due to their depth and large parameter scale, rendering them impractical for mobile devices and real-time applications.
To address the aforementioned challenges, it is crucial to design lightweight SR models to balance model complexity, inference time, and high-resolution reconstruction quality. Recent research efforts have focused on various strategies to reduce the model scale and enhance their efficiency. For instance, Ahn et al. [35] reduced the depth of a new designed Cascading Residual Network (CARN), resulting in the CARN-M model, though with a slight performance trade-off. Similarly, IDN [36] is a lightweight information distillation model constructed by a channel-splitting strategy. Building on this, Information Multi-Distillation Network (IMDN) [37] was designed, which adopts a residual architecture, utilizing multi-distillation through channel-splitting, feature concatenation and contrast-based channel attention, thereby further reducing dimensionality. Liu et al. [38] developed the Residual Feature Distillation Network (RFDN) by replacing the channel-splitting operation of IMDN with multiple feature distillation connections, thereby learning more discriminative feature representations. Kong et al. [39] refined RFDN by replacing feature distillation connections with stacked convolutional layers and activation functions for feature extraction. They also refined the activation functions within the feature extractor, combined with contrastive loss and a warm-start learning rate decay strategy, leading to improved feature extraction capability and enhanced model compactness, while maintaining a balance between accuracy and inference speed. Recent architectures such as Transformers and MLPs have also brought new directions for lightweight SR research. Zhang et al. [40] introduced the concept of superpixel, clustering similar local pixels to form similar feature regions, and proposed a Superpixel Token Interaction Network with intra- and cross-attention blocks, significantly enhancing model interpretability, which resulted in a more compact model with faster inference speeds.
Motivation. Despite the advancements made by existing methods, the trade-off between super-resolution (SR) performance and model compactness remains a persistent challenge. To address this issue, a promising emerging approach involves utilizing large-kernel convolutions for image feature extraction. Notably, AlexNet [41], a pioneering model employing large-kernel convolutions, demonstrated remarkable performance, inspiring subsequent research in the image SR domain. Specifically, large-kernel strided convolutions are often employed in attention mechanisms to capture salient local features from images. While depthwise separable convolutions offer advantages in terms of reduced parameter count and computational complexity, they convolve input features channel-wise, limiting the model’s ability to fully capture spatial information across different channels. ShuffleMixer [42], for instance, leverages large-kernel depthwise separable convolutions to construct a lightweight SR model by incorporating channel projection and shuffling to activate features along the channel dimension, facilitating information exchange between grouped features. However, this approach fails to enhance spatial feature representation or improve feature interactions between neighboring regions in the image, thereby affecting the SR reconstruction quality. Consequently, how to effectively apply large-kernel depthwise separable convolutions to lightweight SR tasks is still a challenge for further research.
Our Method. To address the challenges outlined, this paper introduces a novel Adaptive Dynamic Shuffle Convolutional Parallel Network (ADSCPN). First, the proposed model incorporates a large-kernel parallel depthwise separable convolution technique, which enhances the network’s ability to capture local features from various orientations, thereby expanding the receptive field and improving the extraction of fine details. This approach effectively mitigates the risk of feature loss, leading to more accurate image reconstruction. Second, the method integrates a dynamic convolution mechanism, which leverages an attention mechanism to selectively emphasize key deep features, striking a balance between computational efficiency and robust feature extraction. Moreover, ADSCPN employs a point-attention convolution group (PACG), which refines deep feature representations by applying point-wise attention, thereby minimizing the risk of feature distortion during training. Additionally, the proposed model introduces an Efficient Enhanced Spatial Attention (EESA) block, which streamlines the traditional spatial attention mechanism to preserve performance while reducing computational complexity. Collectively, these techniques enable the SR model to reconstruct high-quality images with lower computational costs, making it highly suitable for deployment on resource-constrained devices.
Contributions. The main contributions of this paper can be summarized as follows:
  • We propose a novel lightweight network architecture, named ADSCPN, designed to enhance feature learning across both channel and spatial dimensions. This innovative approach markedly improves the performance of image SR reconstruction.
  • We introduce an innovative feature processing module, incorporating multi-layer pointwise perceptrons and shuffling attention, which enhances feature extraction through grouped convolutions and channel shuffling. Additionally, the framework integrates large-kernel convolution groups with dynamic convolutions, effectively reducing model size while maintaining high computational efficiency. These advancements enable the model to capture deep semantic features with low complexity, leading to a significant improvement in the quality of reconstructed images.
  • We conduct extensive experiments on multiple benchmark datasets to evaluate the performance of our method. The results demonstrate that our model achieves superior performance compared to several state-of-the-art approaches, while maintaining low computational complexity.
Roadmap. The remainder of this paper is organized as follows: Section 2 provides an overview of related work on lightweight image SR techniques. Section 3 details the proposed methodology, including problem formulation and technical implementation. Section 4 presents the experimental results, while Section 5 concludes the paper with a summary of the key findings and contributions.

2. Related Work

Image super-resolution technology has advanced significantly in recent years, largely driven by deep learning innovations [42,43,44]. Numerous methods have been proposed to enhance image resolution, offering diverse solutions to this challenge. Building upon these foundations, our study introduces several extensions and innovations. This section reviews related work, covering traditional super-resolution techniques, deep learning-based approaches, lightweight SR methods, channel and spatial attention mechanisms, and the application of large-kernel convolutions.

2.1. Traditional Super-Resolution Methods

Early image SR techniques predominantly relied on interpolation algorithms, such as bilinear and bicubic interpolation. These approaches generate high-resolution images by spatially interpolating pixel values, but they often fail to accurately recover fine details and edge structures [45,46]. In response, methods based on sparse representation and dictionary learning were introduced, which reconstruct SR images by learning sparse representations of low-resolution and high-resolution patches, thereby achieving improved preservation of image details [47]. Despite their advantages, these methods are computationally demanding and exhibit limited effectiveness when applied to complex scenes.

2.2. Deep Learning-Based Super-Resolution Methods

With the advent of deep learning, SR methods based on CNNs have garnered significant attention. SRCNN [27] was among the first CNN-based models, achieving notable performance improvements by learning a direct mapping from low-resolution to high-resolution images. Building on this, VDSR [34] introduced a deeper network architecture, further improving SR performance through residual learning. EDSR [30] enhanced reconstruction quality by eliminating redundant batch normalization layers. Although these deep learning-based approaches substantially improve image reconstruction, particularly in restoring high-frequency details, their high computational demands limit their feasibility on resource-constrained devices.

2.3. Lightweight Super-Resolution Methods

To balance reconstruction quality with computational complexity, researchers have developed various lightweight SR models. For instance, MobileNet [48] and ShuffleNet [49] significantly reduce model parameters and computational overhead by leveraging depthwise separable convolutions and group convolutions, making them well-suited for resource-constrained environments such as mobile devices. Additionally, some approaches have incorporated channel attention mechanisms, such as SENet [50], to further enhance model efficiency and performance. While these lightweight models exhibit strong performance under resource limitations, they still face challenges in accurately reconstructing complex scenes.

2.4. Channel and Spatial Attention Mechanisms

The attention mechanism is instrumental in enhancing the feature extraction capabilities of SR networks. Channel attention mechanisms improve reconstruction quality by assigning importance to feature maps across channels, allowing the network to focus on critical features [50]. In contrast, spatial attention mechanisms emphasize distinct spatial regions of the image, thereby further enhancing the network’s ability to capture localized features [51]. The integration of these mechanisms into lightweight SR models significantly enhances both their efficiency and performance.

2.5. Application of Large-Kernel Convolutions

Large-kernel convolutions have gained increasing attention in computer vision tasks in recent years. By expanding the receptive field, they can more effectively capture global information of images. For lightweight super-resolution tasks, researchers have proposed decomposing large-kernel convolutions into multiple small-kernel convolutions, combining them with dynamic convolutions and group convolutions to achieve efficient feature extraction [52,53]. These methods significantly improve the recovery of image details while maintaining low computational complexity, but further optimization of the application of large-kernel convolutions in super-resolution tasks remains an open question.

2.6. Summary

Although considerable advancements have been made in enhancing image super-resolution quality and reducing computational complexity, further performance improvements are needed for more complex scenarios. The ADSCPN model introduced in this paper establishes a novel balance between image reconstruction quality and computational efficiency by integrating large-kernel parallel convolutions, dynamic convolutions, and both channel and spatial attention mechanisms. Compared to existing approaches, ADSCPN excels across multiple benchmark tests, showcasing its substantial potential for application, particularly in resource-constrained environments.

3. Methodology

This section provides an in-depth examination of the proposed ADSCPN model. We commence by defining the relevant notations and problem. Subsequently, we present a detailed description of each model component. The section concludes with a discussion of the model optimization algorithm.

3.1. Notations and Problem Definition

Notations. Without loss of generality, the lightweight image SR is aiming to reconstruct a high-resolution (HR) image I H R from its low-resolution (LR) counterpart I L R . We denote the input low-resolution image as I L R R H × W × C , where H, W, and C represent the height, width, and number of channels of the image, respectively. The high-resolution image is denoted as I H R R r H × r W × C , where r is the scaling factor. Let F e n c be the encoded shallow features extracted from the input LR image I L R using a shallow feature extraction function f e n c ( · ) , and  F K represents the deep features processed through K feature processing blocks. The output SR image I S R is obtained after passing through a series of upsampling and reconstruction layers. Additionally, F p r o j and F s p a t represent the features processed by the channel projection block and the spatial attention block, respectively. The attention weights computed for the spatial features are denoted by A s p a t . For readability, the notations frequently used in this paper are summarized in Table 1.
Problem Definition. Given a training dataset D = { ( I L R ( i ) , I H R ( i ) ) } i = 1 N , where N is the total number of training samples, our goal is to learn a mapping f θ ( · ; θ ) that transforms the low-resolution image I L R into a high-resolution image I S R such that the difference between I S R and the ground-truth high-resolution image I H R is minimized. The objective function is the average loss over D :
min θ 1 N i = 1 N L ( I H R ( i ) , f θ ( I L R ( i ) ; θ ) ) .
This optimization process ensures that the learned model f θ ( · ; θ ) is capable of generating high-quality super-resolution images with low computational complexity, making it suitable for deployment on resource-constrained devices.

3.2. Architecture of ADSCPN

The architecture of ADSCPN is illustrated in Figure 1, comprising three primary components. Given an input low-resolution image I L R , the model first processes it through a shallow feature encoder, f e n c , which consists of a single 3 × 3 convolutional layer. The resulting feature map, denoted as X e n c = f e n c ( I L R ) is subsequently passed through K Feature Processing Blocks, denoted as f F E , which incorporate an innovative combination of large-kernel parallel depthwise separable convolutions, attention mechanisms, and grouped convolutions. Following this, the features are processed by a single 3 × 3 convolutional layer, f 3 × 3 , integrated with a residual connection. The final stage involves the upsampling module, f U P , which is consistent with the structure outlined in Section 3. This module utilizes pixel shuffle operations and 3 × 3 convolutional layers to reconstruct the super-resolution image I S R with the desired scaling factor. The entire process is mathematically formalized as follows:
I S R = f 3 × 3 f U P X e n c + f 3 × 3 f F E K f F E K 1 f F E 1 X e n c .

3.3. Feature Processing Block

The Feature Processing Block, depicted in Figure 2, is composed of several novel modules: the Channel Processing Block (CPB), Feature Extraction Block (FEB), Point-Attention Convolution Group (PACG), and Efficiently Enhanced Spatial Attention (EESA) module. These modules collaboratively perform channel projection and feature shuffle attention learning, effectively enhancing feature extraction, calibration, and spatial augmentation. A residual structure is integrated before and after the Feature Extraction Block to stabilize the training process and facilitate the recovery of low-frequency textures.
The input to the i-th Feature Processing Block is denoted as X i . The channel processing block (CPB) is designed to project and shuffle spatial features, thereby enhancing deep feature extraction while maintaining computational efficiency. The input is subsequently passed through a 3 × 3 convolutional layer and a residual structure, further refining deep feature representations and supporting shallow feature processing. The output is then fused with the original input, followed by parallel operations involving a 1 × 1 convolution and the Efficiently Enhanced Spatial Attention (EESA) module, which calibrates the spatial features by focusing on the most salient information. The computational flow can be formalized as follows:
X i + 1 = X i + f E E S A f 1 × 1 f C P B ( X i ) + f R f F E X i + f C P ( X i ) .

3.3.1. Channel Processing Block

The channel processing block (CPB) is the first component of the feature processing block. It is composed of a point-wise multi-layer perceptron (PW-MLP), a 3 × 3 convolution, and Shuffle Attention, focusing on extracting information from feature channels and spatial dimensions, as shown in Figure 3. The multi-layer point-wise perceptron projects and activates important low-frequency feature information from the input channels through expansion and contraction operations, thereby enhancing the channel features. The  3 × 3 convolution further activates the channel features and extracts deep local spatial information. The learned features are then fed into the Shuffle Attention module, where grouped channel features are processed. The module calculates attention within and across channels and spatial dimensions, enhancing the efficiency of the attention computation while promoting information fusion and interaction among grouped features. Since the Shuffle Attention block can easily lose part of the feature information, residual connections are added around the channel processing block to supplement shallow features.
As the first component of the channel processing block, the PW-MLP facilitates the encoding of spatial information within channels, activates channel features, and lays the foundation for subsequent local feature extraction and channel-wise grouped shuffle attention learning. As shown in Figure 4, it is composed of a 1 × 1 convolution, a GeLU activation function, and another 1 × 1 convolution in sequence. Given an output feature map of size H × W × C , the channel expansion factor is denoted as ϕ . The channel dimension undergoes expansion and contraction processing to activate the channel features. After processing by the first 1 × 1 convolution, the feature map size is expanded as H × W × C × ϕ , which is then reduced by the subsequent 1 × 1 convolution to restore the original size. In our solution, the expansion factor ϕ is set to 2.

3.3.2. Feature Extraction Block

The structure of the newly designed feature extraction block is shown in Figure 5. The first component is a large-kernel parallel convolution processing unit, consisting of a 1 × 7 depthwise separable convolution, a  7 × 1 depthwise separable convolution, and a 1 × 1 group convolution. The input to this block, X , is the low-frequency feature map generated by the CPB module. The  1 × 7 and 7 × 1 convolutions process the same input to capture local features in both horizontal and vertical directions, enhancing the extracted image features while filling in nearby spatial details. This design improves the network’s ability to capture local features and strengthens relationships between surrounding pixels. Compared to a single 7 × 7 depthwise separable convolution, this unit delivers a more robust performance.
Following the large-kernel parallel convolution processing, the two branches are fused into a single output, denoted as F B P . This output is activated by the GeLU function and passed through a 1 × 1 group convolution to further enhance feature aggregation and learning efficiency. The processed feature map, F B P , is then fed into a dynamic convolution with four adjustable convolutional kernels. By modulating the convolutional kernels and adaptively adjusting parameters, this dynamic convolution balances computational cost while significantly improving representational capacity. The output, denoted as X F E , is subsequently refined through a 1 × 1 group convolution, activated by GeLU, and followed by a channel shuffle for final feature aggregation. This output is then fused with the original input and processed by f s h u f f l e to facilitate grouped feature interaction and aggregation. This process effectively refines the image’s high-frequency features, promoting stable gradient flow and enhancing overall model performance. The entire process of the feature extraction block (FEB) can be expressed as follows:
F B P = f 1 × 7 , D w c o n v ( X ) + f 1 × 1 , G c o n v ( X ) + f 7 × 1 , D w c o n v ( X ) ,
X B P = f 1 × 1 , G c o n v ( G e L U ( F B P ) ) ,
X F E = X + f s h u f f l e f 1 × 1 , G c o n v G e L U f D y n a m i c ( X B P ) .

3.3.3. Point-Attention Convolution Group

Similar to EDSR [30], where the model is primarily composed of residual blocks interconnected by 3 × 3 convolutions and activation functions, we introduce a novel point-attention convolution group placed at the end of the feature processing block to further deepen the network. This component performs weighted learning on the deep features processed by the convolution group, enhancing and refining them after convolutional processing. To mitigate potential feature loss during deep model training, these refined features are merged with the output of the convolution group. The convolution group comprises two 3 × 3 convolutional layers connected in series, as depicted in Figure 6. The original input X F E passes through the convolution group, producing an intermediate output X R B . This is followed by a point-attention block, consisting of a 1 × 1 convolution and a Sigmoid activation function, which learns attention weights for the deep features, extracting the most significant ones. The attention-weighted deep features, denoted as X i , are multiplied with X R B to calibrate the convolutional output and minimize feature loss, resulting in the enhanced deep semantic features X R B . Finally, the enhanced features are fused with the original X R B to produce the final output, X R B . The entire process is mathematically expressed as follows:
X R B = f 3 × 3 f R e L U f 3 × 3 ( X F E ) ,
X i = f S i g m o i d f 1 × 1 ( X R B ) ,
X R B = X R B × X i ,
X R B = X R B + X R B .

3.3.4. Efficiently Enhanced Spatial Attention

The lightweight Enhanced Spatial Attention (ESA) module [38], frequently employed in various lightweight super-resolution models, enhances feature representation capabilities while maintaining efficiency and ease of integration. As depicted in Figure 7a, the ESA architecture includes a 1 × 1 convolution for channel dimension reduction, followed by a 3 × 3 convolution and a max pooling layer with a large receptive field, before upsampling back to the original resolution. Subsequently, a  1 × 1 convolution is applied to restore the original channel dimensions, and the output is processed through a Sigmoid activation function to generate the spatial attention map. The ESA has been further optimized into the Efficiently Enhanced Spatial Attention (EESA) module [54], as shown in Figure 7b. In this modified version, the  3 × 3 convolution and upsampling operations are replaced with a single 3 × 3 convolution, effectively reducing and recovering the channel dimensions without the need for additional convolutions in the residual connection. This adjustment reduces computational complexity while preserving essential feature details. Additionally, a pooling layer with a 7 × 7 window size and a stride equivalent to the original max pooling layer is introduced to further expand the receptive field. Despite its simplified design, this lightweight version maintains a comparable performance level to the original ESA module with minimal impact on accuracy.

3.4. Optimization

The optimization procedure of the ADSCPN is outlined in Algorithm 1. Initially, the training dataset undergoes preprocessing, followed by data augmentation to enhance the training scale. Key hyperparameters, including training epochs, batch size, number of iterations, and learning rate, are configured, along with a reduction in the total number of epochs. Subsequently, the model parameters are initialized. As training progresses, learning rate decay is applied based on the current epoch number and iteration count.
For each batch, a fixed number of high-resolution and low-resolution image pairs are input into the model for feature extraction. During this phase, shallow features are encoded, and deep semantic features are extracted. The features are then fused to generate comprehensive semantic representations of the images. The upsampling module reconstructs the corresponding super-resolution images. The  L 1 loss is computed between the reconstructed super-resolution images and their corresponding high-resolution ground truth. Model parameters are iteratively updated using the Adam optimizer through backpropagation. This process continues until the optimal model is achieved and is validated through performance metrics.
Algorithm 1 The pseudo-code of ADSCN Algorithm
Input: 
Training dataset D ; training period T; batch size N; image crop size P; number of iterations I; learning rate η ; model channel C; number of groups G; number of feature processing blocks N f ; batch size N, with each batch size as [ N , channels , height , width ] ;
Output: 
Super-resolution image I S R ; model parameters θ
 1:
Initialize the training environment, crop high-resolution images from the training dataset into low-resolution images based on the crop size P, and adjust the channels.
 2:
Calculate the number of image blocks M required according to the iteration I and batch size N.
 3:
Construct a list of corresponding high-resolution and low-resolution image blocks [ [ I L R 1 , I S R 1 ] , , [ I L R m , I S R m ] ] .
 4:
Build the ADSCPN model, set the convolution kernel size, input channels as C, build the upsampling module based on the scale factor, stack N f feature processing blocks, and initialize the model parameters θ .
 5:
for each t = 1 to I do
 6:
   Adjust the learning rate;
 7:
   for each pair of high-resolution and low-resolution image blocks t = 1 to N do
 8:
     Feed the low-resolution image I L R into ADSCPN, perform feature extraction and multiple feature processing blocks, and obtain the corresponding super-resolution image I S R ;
 9:
     Calculate the loss function L between the generated super-resolution image I S R and the corresponding high-resolution image I H R ;
10:
   end for
11:
   Update model parameters θ using the Adam optimizer and backpropagation;
12:
   if convergence is reached then
13:
     break;
14:
   end if
15:
end for

4. Experiment

This section outlines a comprehensive experimental evaluation conducted on several standard benchmark datasets. We begin by detailing the experimental setup, encompassing the datasets utilized, evaluation metrics, baseline models, and implementation specifics. Following this, we present a performance analysis derived from multiple experimental groups.

4.1. Dataset

To comprehensively evaluate the performance of the proposed super-resolution model, a series of experiments were conducted using several widely-recognized datasets. The DIV2K dataset [55] was employed for model training, while evaluations were performed on standard benchmark datasets, including Set5 [56], Set14 [57], B100 [58], and Urban100 [59]. Furthermore, to assess the model’s robustness under diverse real-world conditions, we incorporated blind super-resolution datasets such as DIV2KRK [60], DRealSR [61], and RealSR [62]. These datasets provide varying levels of real-world complexity, with DRealSR and RealSR capturing real-world noise and degradation processes, making them more challenging than the original benchmark datasets like Set5 [56] and B100 [58], which primarily consist of controlled synthetic degradations. We provide a brief introduction to these datasets below:
  • DIV2K. The DIV2K dataset comprises 1000 high-resolution images, characterized by a rich variety of features and diverse scene categories, including environments, humans, and other objects. For experimental purposes, the first 800 images are designated for training, images 801–900 are allocated to the validation set, and the remaining 100 images are reserved for performance evaluation. Each image in the dataset has a native 2 K resolution, accompanied by corresponding low-resolution images generated through the application of various degradation kernels, such as Gaussian white noise. This dataset is widely utilized for training and benchmarking super-resolution models due to its diversity and high-quality annotations.
  • Set5 and Set14. The Set5 and Set14 datasets are among the earliest small-scale benchmarks introduced in the field of image processing. Set5 contains 5 images, while Set14 consists of 14 images, both encompassing a mix of human subjects and natural scenes. Despite their limited size, these datasets remain widely used for testing and validation purposes, offering a convenient and efficient means of assessing model performance, particularly in super-resolution tasks.
  • B100. The B100 dataset comprises 100 images depicting a variety of subjects, including animals, plants, and real-world scenes. While offering a diverse range of content, the images in this dataset are relatively small in resolution and exhibit less detailed textures compared to other high-resolution benchmarks, making it a useful resource for evaluating super-resolution models under more constrained conditions.
  • Urban100. The Urban100 dataset contains 100 high-resolution images, specifically focused on capturing intricate details of urban architecture. These images provide complex structural patterns and fine-grained textures, making the dataset particularly challenging and suitable for evaluating the performance of super-resolution models on architectural and man-made scenes.
  • DRealSR and RealSR. The RealSR and DRealSR datasets are created using DSLR cameras with multiple zoom levels, varying aperture settings, and different lens focal lengths to simulate signal noise introduced during the degradation process. An image registration algorithm is employed to precisely align high- and low-resolution image pairs. These datasets encompass a wide range of scenes, including natural landscapes, architectural structures, as well as human and animal subjects. Both datasets are widely used for real-world image super-resolution reconstruction tasks, providing realistic and challenging conditions for model evaluation.

4.2. Evaluation Metrics

Quantitative evaluation of the model’s performance was carried out using the Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index (SSIM) metrics [63], ensuring a thorough comparison of the model’s accuracy and perceptual quality across different datasets.
PSNR is a widely used metric for evaluating the quality of reconstructed images at the pixel level by measuring the pixel-wise differences between two images. A higher PSNR value indicates that the reconstructed super-resolution image is closer to the original high-resolution image, suggesting improved reconstruction accuracy. However, PSNR does not account for structural differences or human visual perception, which can lead to high PSNR values even when the reconstructed image visually deviates from the original, particularly in local regions. To compute PSNR, the Mean Squared Error (MSE) between two images, I 1 and I 2 , must first be calculated, where W and H represent the width and height of the image, respectively. The mathematical expression for MSE is given by:
MSE = 1 H × W i = 1 H j = 1 W I 1 ( i , j ) I 2 ( i , j ) 2 .
The PSNR metric is then calculated as follows:
PSNR = 10 log 10 ( 2 n 1 ) 2 MSE ,
where 2 n 1 represents the maximum possible pixel value in the image, with n = 8 for typical grayscale images. For RGB images, the calculation is performed only on the Y channel after converting the image to the YCbCr color space.
SSIM evaluates the overall structural similarity between the reconstructed super-resolution image I S R and the original high-resolution image I H R . SSIM is computed by considering three key components: luminance similarity, contrast similarity, and structural similarity. The SSIM value ranges from 0 to 1, with values closer to 1 indicating that the reconstructed image exhibits greater structural similarity to the original high-resolution image. Thus, a higher SSIM value reflects a more accurate representation of the original image’s structural content. Assume that both I S R and I H R images contain N pixels. First, the average luminance μ and standard deviation σ of each image I are calculated to quantify the luminance and contrast. Then, the normalized image I , which quantifies the structural information, is computed using the following formulas:
μ = 1 N i = 1 N I ( i ) ,
σ = 1 N 1 i = 1 N ( I ( i ) μ ) 2 ,
I ( i ) = I ( i ) μ σ , i [ 1 , . . , N ] .
Next, the luminance similarity l, contrast similarity c, and structural similarity s are calculated. These are expressed as follows:
l = 2 μ H R μ S R + c 1 μ H R 2 + μ S R 2 + c 1 ,
c = 2 σ H R σ S R + c 2 σ H R 2 + σ S R 2 + c 2 ,
s = σ H R , S R + c 3 σ H R σ S R + c 3 ,
where μ H R and μ S R represent the mean luminance of the original high-resolution image and the super-resolution result, respectively. σ H R and σ S R represent the standard deviations, while σ H R , S R represents the covariance between the original high-resolution image and the super-resolution result. c 1 , c 2 , and c 3 are constants introduced to avoid division by zero. Finally, the SSIM value is calculated as:
SSIM = l α c β s γ ,
where α , β , and γ are parameters that adjust the importance of luminance, contrast, and structural details in the calculation. These parameters are typically set to 1.

4.3. Baselines and Implementation Details

Baseline. We compared ADSCPN with several baselines, including Bicubic [46], SRCNN [27], FSRCNN [26], VDSR [34], DRCN [28], DRRN [29], CARN-M [35], CARN [35], MemNet [64], LapSRN [65], IDN [36], MADNet [66], ECBSR (Edge-oriented Convolution Block for Real-time Super-Resolution) [67], and FALSR-A (Fast, Accurate, and Lightweight Super-Resolution) [68]. These methods were chosen as they represent state-of-the-art lightweight and efficient super-resolution techniques, covering a diverse range of architectures from early CNN-based methods (e.g., SRCNN [27], VDSR [34]) to more recent lightweight models (e.g., CARN [35], MADNet [66]), providing a comprehensive benchmark for comparison.
Implementation Details. During the training process, the Adam optimizer was used to update the network parameters, with hyperparameters configured as follows: β 1 = 0.9 , β 2 = 0.999 , ϵ = 10 8 , and the initial learning rate of 5 × 10 4 . The batch size was set to 16, and the L 1 loss function was employed as the model training loss criterion. All experiments were conducted on a workstation equipped with an Intel(R) Core i9-12900K 3.9 GHz CPU, 128 GB of RAM, 1 TB SSD storage, 2 TB HDD storage, and dual NVIDIA GeForce RTX 3090 GPUs, operating under Ubuntu 18.04.1. The implementation of all techniques was carried out using Python 3.6 and PyTorch 1.8.1.

4.4. Performance Evaluation

4.4.1. Model Parameter Analysis

This section examines the influence of hyperparameter configurations, such as learning rate, number of layers, and batch size, on the performance of the ADSCPN model. We conducted extensive experiments to understand how these hyperparameters impact the convergence speed, stability, and final quality of the reconstructed images.
The learning rate plays a critical role in training deep models effectively. In our approach, we employed a staged-halving learning rate schedule to guide the training process. During the initial stages, a higher learning rate was used to accelerate the convergence, enabling the model to quickly capture general patterns in the data. As training proceeded, the learning rate was halved at specific intervals to gradually slow down the learning process. This staged reduction helped stabilize the training, allowing the model to refine its parameters and converge to better local optima. This strategy proved effective, as it balanced fast convergence at the beginning with careful parameter tuning towards the end, leading to consistent improvements in PSNR and SSIM values across all test datasets.
We conducted an ablation study to determine the optimal number of Feature Process Blocks (also referred to as “layers”) in the model. As illustrated in Figure 8a, increasing the number of Feature Process Blocks initially led to significant improvements in PSNR on the B100 test set, establishing a positive correlation between model depth and reconstruction quality. However, when the number of blocks exceeded 6, the performance gains began to plateau, indicating diminishing returns. Based on these findings, we selected 6 blocks as the optimal configuration, which strikes a balance between model complexity and performance enhancement. Increasing the number of blocks beyond 6 resulted in increased computational costs without notable improvements, thus establishing 6 blocks as the best trade-off.
Figure 8b presents PSNR results for the Set5 test set, which demonstrates that increasing the number of channels improves reconstruction quality. However, as the number of channels surpasses 64, the performance gains become less pronounced, indicating a performance plateau. Given the increased complexity and computational burden of using more channels, we determined that 48 channels provide an optimal balance between reconstruction quality and model efficiency, as summarized in Table 2. This selection reduces computational requirements while maintaining high-quality output.
The batch size used during training affects both the stability and efficiency of the optimization process. Inspired by hyperparameter settings from well-established super-resolution models like EDSR and RCAN, we experimented with different batch sizes and concluded that a batch size of 16 was ideal for ADSCPN. This value provided a good trade-off between convergence stability and GPU memory utilization, allowing efficient training without compromising on model quality. While larger batch sizes provided more stable gradient estimates, the performance gains diminished, and smaller batch sizes resulted in higher variance in gradients, which led to less stable training dynamics.
Furthermore, we analyzed the effect of the number of groups within the architecture through experiments involving the division of convolutional groups within the ADSCPN network. As reported, the model achieves optimal performance on the Set5 test set when configured with 8 groups. Increasing the number of groups further leads to reduced model parameters and a smaller memory footprint but with diminishing performance gains. Thus, we established 8 groups as the optimal configuration to balance between model complexity and overall performance.
These findings provide clear guidance for the optimal hyperparameter configuration of the ADSCPN model. Specifically, the staged-halving learning rate schedule allows for both rapid initial convergence and stable parameter refinement, while the chosen number of Feature Process Blocks, channels, and groups ensures that the model remains computationally efficient without sacrificing performance. A batch size of 16 further guarantees a smooth training process, optimizing GPU utilization and stability. These hyperparameter choices effectively balance learning efficiency, generalization, and practical deployment feasibility, making ADSCPN suitable for high-quality image reconstruction tasks.

4.4.2. Quantitative Experimental Results Comparison

In this section, we compare the performance of the proposed model with several popular and effective methods on standard single-image super-resolution benchmark datasets across various scaling factors. The specific results are summarized in Table 3, Table 4 and Table 5, where the best and second-best average PSNR and SSIM values are highlighted in bold and underlined, respectively. The comparative methods include Bicubic [46], SRCNN [27], FSRCNN [26], VDSR [34], LapSRN [65], DRCN [28], DRRN [29], CARN-M [35], CARN [35], MemNet [64], IDN [36], MADNet [66], ECB-SR [67], LINF [69], FPL [70] and FALSR-A [68]. The results are averaged from the data reported in the respective studies.
When the scaling factor is set to 2, the ADSCPN-plus model consistently achieves either the best or second-best performance across all datasets. In particular, for scaling factors of 3 × and 4 × , the ADSCPN model surpasses the ECB-SR and FALSR-A methods, especially on the Urban100 dataset, which features more complex scenes, where it attains the highest performance. Notably, the PSNR and SSIM values for ADSCPN-plus are significantly higher than those of competing methods at scaling factors of 3 × and 4 × , with an average PSNR increase exceeding 0.02 dB for ADSCPN-plus at a scaling factor of 3. Furthermore, both ADSCPN and ADSCPN-plus outperform other methods on the Urban100 and Set14 datasets in terms of PSNR at a scaling factor of 3 × . It is important to highlight that the model is updated with only 1600 batches per epoch, resulting in a remarkably short training time while achieving these performance metrics. This demonstrates the effectiveness, as well as the theoretical and practical value of the model as a lightweight solution for super-resolution tasks.
Table 6 presents the quantitative evaluation results for scaling factors of 2 × and 4 × on the DIV2KRK dataset [60]. The best performances are highlighted for clarity. Notably, despite the lightweight complexity of the proposed model, it achieves super-resolution performance comparable to that of more complex architectures, particularly at a scaling factor of 4 × . The PSNR and SSIM values for ADSCPN-plus with a patch size of 192 are 25.65 dB and 0.6940, respectively, which are slightly superior to the results of ADRBN under the same training conditions, showing a difference of 0.01 dB and 0.0004. Additionally, Table 7 provides the quantitative results for the model trained on the DIV2K dataset [71] and tested across various scaling factors on the DRealSR [61] real-world scenario test set. Furthermore, Table 8 displays the quantitative evaluation results across different scaling factors following retraining and testing on the RealSR dataset [62].

4.4.3. Qualitative Results Comparison

To evaluate the visual quality of reconstructed images generated by different models, localized regions from the test set results were enlarged to facilitate direct visual inspection of the image detail reconstruction. The comparison methods include Bicubic [46], LapSRN [65], CARN-M [35], IDN [36], and the proposed ADSCPN-plus model, with visual results sourced from official publications for each method.
As illustrated in Figure 9, for the 2 × scaling factor, the region highlighted within the red box of the high-resolution image from the Urban100 test set (“image092”) reveals processed horizontal lines. In contrast, the areas reconstructed by the three interpolation methods exhibit significant blurring, with LapSRN [65] displaying the most distorted slanted lines. In comparison, the proposed model achieves superior visual and quantitative results. For the 3 × scaling factor, an enlarged region from the high-resolution image in the B100 test set (“img063”) demonstrates that the model effectively retains the brightness in the eye region, exhibiting sharper edges compared to the high-resolution images. At a 4 × scaling factor, the enlarged image region in the Urban100 test set (“image044”) reveals that the building edges are processed more accurately, resembling high-resolution images, while other methods produce blurred edges. The proposed model successfully captures texture patterns closer to those of the high-resolution images, underscoring its effectiveness in handling complex scenes. Figure 10 presents the results from the DIV2KRK, DRealSR, and RealSR datasets, further demonstrating the model’s ability to generalize across different datasets and maintain high-quality reconstructions.

4.4.4. Model Complexity

In this section, we analyze the model complexity of ADSCPN in terms of FLOPs (Floating Point Operations) and MACs (Multiply-Accumulate Operations). FLOPs represent the total number of floating-point calculations needed by the model, which directly reflects the computational power required. MACs are a measure of the model’s efficiency, representing the number of times two numbers are multiplied and added, a critical metric for evaluating computational load in deep learning models. These metrics are particularly important for understanding the model’s suitability for resource-constrained environments.
When the scaling factor is set to 2 and the input image size is 3 × 1280 × 720 , the model’s complexity was calculated using the ‘thop’ library and compared against several established models, including SRCNN [27], VDSR [34], LapSRN [65], DRCN [28], DRRN [29], MemNet [64], CARN [35], MADNet [66], and FALSR-A [68]. The comparison focuses on model parameters (Flops) and multiply-accumulate operations (MACs), which estimate the memory requirements and computational costs of the models. As demonstrated in Table 9, the proposed model exhibits significantly fewer parameters compared to MADNet, FALSR-A, and CARN, while maintaining a comparable level of performance. Although the model prioritizes convolution operations and consequently incurs a higher number of multiply-accumulate operations, the overall computational load remains relatively manageable for hardware devices. This characteristic renders the model well-suited for real-time applications on embedded mobile devices.

4.5. Ablation Analysis

To substantiate the rationale behind the design of each component within the feature processing block and to assess the effectiveness of these components in achieving high-quality reconstruction, we will conduct ablation experiments. These experiments will validate each component individually and investigate the impact of their combined presence on quantitative performance.

4.5.1. Effectiveness of Large-Kernel Parallel Convolution Groups

The large-kernel parallel depthwise separable convolution group plays a crucial role in enhancing the receptive field without significantly increasing the computational complexity. It is composed of a 1 × 7 depthwise separable convolution, a 1 × 1 group convolution, and a 7 × 1 depthwise separable convolution, which together enable efficient spatial feature learning. Specifically, the 1 × 7 and 7 × 1 convolutions separately learn features along the horizontal and vertical axes, which enhances the 1 × 1 convolution’s ability to integrate these directional features effectively.
Compared to traditional 3 × 3 convolutions often employed in lightweight models, the large-kernel parallel structure significantly expands the receptive field, enabling the model to capture more global contextual information while maintaining a lightweight architecture. This structure provides a richer and more detailed spatial feature representation, leading to improved feature extraction capability.
Unlike a single 7 × 7 depthwise separable convolution, which may introduce more computational burden and inefficiencies, the large-kernel parallel group processes the input using separate directional convolutions. This parallel approach not only maintains computational efficiency but also improves the model’s ability to capture intricate details, which is reflected in the enhanced PSNR and SSIM values seen across multiple test datasets.
The experimental results, summarized in Table 10, demonstrate that the large-kernel parallel convolution group offers superior performance in comparison to using a single 7 × 7 convolution, especially for the Set5 dataset. This effectiveness can be attributed to the improved ability to capture directional features and integrate them efficiently. Table 11 further highlights that this convolution strategy achieves a better trade-off between model complexity and computational cost, which is crucial for achieving high-quality super-resolution while ensuring the model remains practical for deployment.

4.5.2. Effectiveness of Dynamic Convolution

After applying the large-kernel parallel depthwise separable convolution group, dynamic convolution is used to further enhance the model’s ability to extract meaningful features. Dynamic convolution introduces an adaptive mechanism where the convolutional filters are adjusted based on the importance of the features, as determined by an embedded attention mechanism. This attention-driven adaptation ensures that critical regions of the input receive more focus during processing, thereby significantly enhancing the model’s capacity to represent and learn from complex image features.
In contrast to standard convolution, which applies static filters across all regions of the input, dynamic convolution provides a context-dependent filter adaptation that aligns with the content’s specific characteristics. This capability not only enhances feature extraction but also allows the model to better capture the nuances in more challenging datasets, such as Urban100, which contain complex structures and textures.
For this study, the dynamic convolution used a kernel size of 3 × 3 . We conducted an ablation experiment by replacing dynamic convolution with standard 3 × 3 convolution to evaluate its impact. As shown in Table 12, dynamic convolution consistently outperformed the standard convolution, particularly for a scale factor of 2, across all test datasets. Notably, the improvement was most significant on the Urban100 dataset, indicating the model’s enhanced ability to generalize to complex, real-world scenes.
The embedded attention mechanism within dynamic convolution plays a key role in this performance gain by dynamically highlighting the most informative features while suppressing less important ones. This selective focus directly contributes to improved PSNR and SSIM values, thereby enhancing overall reconstruction quality. The adaptive nature of dynamic convolution, driven by attention, makes it particularly effective for scenarios where image content varies widely in terms of texture and detail complexity. This explains its superior performance compared to standard convolution, as evidenced by the experimental results.

4.5.3. Effectiveness of the Point Attention Convolution Group

The point attention convolution group introduces an attention mechanism aimed at enhancing the model’s ability to focus on the most critical features of an image, which significantly impacts overall performance. It improves upon the traditional convolution group—typically consisting of two 3 × 3 convolution layers with ReLU activation—by adding an additional branch consisting of a 1 × 1 convolution followed by a Sigmoid activation. This branch computes attention coefficients for each feature channel, which are then used to recalibrate the input features dynamically.
The recalibration process ensures that more informative features are amplified while less significant features are suppressed, allowing the model to allocate its resources more effectively toward reconstructing key image details. This type of attention mechanism enables better feature calibration, particularly in deeper layers where important image features might otherwise become diluted or distorted.
In our experiments, replacing the point attention convolution group with a traditional convolution group resulted in a noticeable decrease in both PSNR and SSIM values across all test sets, as shown in Table 13. The most significant performance gains were observed in datasets with high structural complexity, such as Urban100, demonstrating the point attention group’s effectiveness in capturing intricate textures and structures.
By dynamically computing and applying attention coefficients, the point attention convolution group helps the model selectively focus on salient regions of the image, thereby improving the quality of feature representation. This selective focus, facilitated by the embedded attention mechanism, directly leads to better feature extraction, less feature distortion, and improved reconstruction quality. The attention-driven recalibration of features also ensures that the network learns to prioritize the most critical information, which is crucial for achieving higher PSNR and SSIM metrics.

4.5.4. Effectiveness of the Efficient Enhanced Spatial Attention Block

The Efficient Enhanced Spatial Attention (EESA) block is a crucial component designed to improve the model’s ability to focus on spatially important regions while maintaining computational efficiency. The EESA block is an optimized version of the Enhanced Spatial Attention (ESA) block, achieved by removing the external 1 × 1 convolution and simplifying the convolution group to a single 3 × 3 convolution. This optimization reduces both model complexity and computational burden, making it more suitable for lightweight deployment without sacrificing performance.
The core mechanism of the EESA block is its ability to dynamically highlight spatial regions that are most important for the reconstruction task. By learning spatial attention maps that emphasize high-frequency details such as edges and textures, the EESA block directs the model’s focus to areas with significant structural information. This selective focus is essential for achieving high-quality reconstruction, particularly in regions with complex textures or fine details.
In our ablation studies, we compared the EESA block to the original ESA block to evaluate its effectiveness. As shown in Table 14, the EESA block not only reduces the number of parameters and multiply-accumulate (MAC) operations but also maintains comparable or better performance in terms of PSNR and SSIM across multiple test datasets. This improvement is largely due to the effective attention mechanism that allows the model to prioritize spatial regions that contribute the most to image quality.
The simplified architecture of the EESA block enhances computational efficiency while ensuring that the model retains the ability to focus on crucial spatial features. This efficiency is particularly beneficial for real-world applications where computational resources are limited. The attention mechanism within the EESA block plays a key role in guiding the model to learn meaningful spatial relationships, thereby enhancing the quality of feature extraction and overall image reconstruction.

4.5.5. Effectiveness of Combined Components

To assess the impact of the combined components within the feature processing block on the model’s quantitative performance and to validate the rationale and superiority of the overall structure presented in this section, we conducted an ablation study focusing on the feature extraction block and the point attention convolution group. Following the previous ablation experiments that modified individual components, we replaced the large-kernel parallel depthwise separable convolution group in the feature extraction block with a single 7 × 7 large-kernel depthwise separable convolution, substituted dynamic convolution with a standard 3 × 3 convolution, and replaced the point attention convolution group with a standard convolution group.
The specific combinations of components are detailed in Table 15, where “✓” indicates the inclusion of the component, and “×” signifies replacement with a standard component. Optimal performance is highlighted in bold. As shown, except for the Set5 test set, the model proposed in this section achieves the best quantitative performance overall, with PSNR and SSIM values on Set14 and Urban100 improving by 0.05 dB and 0.0005, respectively, compared to the second-best performance, and by 0.01 dB and 0.0001 on B100. Notably, when all components were replaced with standard components, the performance on the Urban100 dataset was superior to configurations where only one or two components were replaced with standard components. This finding demonstrates that the feature processing block can effectively learn image features at multiple levels and that the designed combination of the three components is both indispensable and complementary, resulting in robust reconstruction performance with reduced model complexity.

4.5.6. Effectiveness of Activation Functions

Incorporating appropriate activation functions can significantly enhance the nonlinear feature learning capabilities of the model, thereby improving the fidelity of image reconstruction. This section evaluates the performance of three activation functions: SiLU, GELU, and ReLU. With the exception of the Point Attention Convolution Group and the Efficient Enhanced Spatial Attention Block, all other components of the ADSCPN were subjected to an ablation study to compare the effects of these different activation functions. The quantitative results are summarized in Table 16, which identifies the optimal activation function. Notably, when the scaling factor is set to 2, the GELU activation function substantially outperforms the other options on the Set14 dataset, while exhibiting comparable performance across the remaining datasets. Consequently, GELU is selected as the activation function for the entire model in this study.

5. Conclusions

Through comprehensive research and analysis in the realm of lightweight image super-resolution, we propose a novel lightweight super-resolution network founded on a large-kernel parallel convolution group. The network architecture employs a multi-layer point-wise perceptron for channel projection and utilizes shuffle attention for effective channel grouping, thereby enhancing the extraction of deep channel and spatial information while improving feature learning efficiency. To facilitate deeper feature learning, we introduce a large-kernel parallel depthwise separable convolution group alongside grouped convolution. Additionally, the point attention convolution group and efficient enhanced spatial attention block are integrated to optimize spatial feature learning, supplement low-frequency information, and enhance deep feature extraction. This design approach mitigates the distortion of deep features often encountered in deeper networks, significantly improving the quality of super-resolution images while maintaining low model complexity. Both quantitative and qualitative experiments, in comparison with existing methods, demonstrate the superior performance of our proposed approach, while extensive ablation studies further validate the rationale behind the component configurations.
Despite its promising performance, the ADSCPN model has some limitations, including sensitivity to noisy and low-quality inputs, high computational complexity, and limited generalization in diverse scenarios. These limitations may hinder its deployment in resource-constrained environments and challenging real-world applications. Future work will focus on enhancing robustness to noise, optimizing computational efficiency for edge devices, and expanding training with more diverse datasets to improve generalization. Such improvements will make ADSCPN more viable for practical applications like mobile photography, medical imaging, surveillance, and specialized domains.

Author Contributions

Y.L. (Yiting Long) and H.Z. designed the methodology, wrote the original draft, and designed and prepared all figures. H.R. contributed to conceptualization and project administration. C.Z. and L.Z. acquired funding and reviewed and edited the manuscript. Y.L. (Yiting Long) and H.R. conceived the experiments. Y.L. (Yiting Long), H.R. and H.Z. conducted the experiments and acquired the experimental results. Y.L. (Yi Liu), L.Z., and X.Z. analyzed the experimental results. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 62202163, 62072166 and 62372150, the Natural Science Foundation of Hunan Province under Grant 2022JJ40190, 2022JJ30231 and 2023JJ30169, the Scientific Research Project of Hunan Provincial Department of Education under Grant 22A0145 and 22B0559.

Data Availability Statement

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
SRSuper-resolution
HRHigh-resolution
LRLow-resolution
ADSCPNAdaptive dynamic shuffle convolutional parallel network
CNNsConvolutional neural networks
LDFeature processing block
CPBChannel processing block
FEBFeature extraction block
PW-MLPPoint-wise multi-layer perceptron
PACGPoint-attention convolution group
EESAEfficiently enhanced spatial attention
ESAEnhanced spatial attention
PSNRPeak signal-to-noise ratio
SSIMStructural similarity index

References

  1. Chen, Z.; Zhang, Y.; Gu, J.; Kong, L.; Yang, X.; Yu, F. Dual aggregation transformer for image super-resolution. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 12312–12321. [Google Scholar]
  2. Chen, H.; He, X.; Qing, L.; Wu, Y.; Ren, C.; Sheriff, R.E.; Zhu, C. Real-world single image super-resolution: A brief review. Inf. Fusion 2022, 79, 124–145. [Google Scholar] [CrossRef]
  3. Zhang, Q.; Xiao, J.; Tian, C.; Chun-Wei Lin, J.; Zhang, S. A robust deformed convolutional neural network (CNN) for image denoising. CAAI Trans. Intell. Technol. 2023, 8, 331–342. [Google Scholar] [CrossRef]
  4. Wang, C.; Lv, X.; Shao, M.; Qian, Y.; Zhang, Y. A novel fuzzy hierarchical fusion attention convolution neural network for medical image super-resolution reconstruction. Inf. Sci. 2023, 622, 424–436. [Google Scholar] [CrossRef]
  5. Georgescu, M.I.; Ionescu, R.T.; Miron, A.I.; Savencu, O.; Ristea, N.C.; Verga, N.; Khan, F.S. Multimodal multi-head convolutional attention with various kernel sizes for medical image super-resolution. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–7 January 2023; pp. 2195–2205. [Google Scholar]
  6. Cornebise, J.; Oršolić, I.; Kalaitzis, F. Open high-resolution satellite imagery: The worldstrat dataset–with application to super-resolution. Adv. Neural Inf. Process. Syst. 2022, 35, 25979–25991. [Google Scholar]
  7. Arefin, M.R.; Michalski, V.; St-Charles, P.L.; Kalaitzis, A.; Kim, S.; Kahou, S.E.; Bengio, Y. Multi-image super-resolution for remote sensing using deep recurrent networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 206–207. [Google Scholar]
  8. Zhu, L.; Yu, W.; Zhu, X.; Zhang, C.; Li, Y.; Zhang, S. MvHAAN: Multi-view hierarchical attention adversarial network for person re-identification. World Wide Web 2024, 27, 59. [Google Scholar] [CrossRef]
  9. Zhang, C.; Xie, F.; Yu, H.; Zhang, J.; Zhu, L.; Li, Y. PPIS-JOIN: A novel privacy-preserving image similarity join method. Neural Process. Lett. 2022, 54, 2783–2801. [Google Scholar] [CrossRef]
  10. Zheng, M.; Xu, J.; Shen, Y.; Tian, C.; Li, J.; Fei, L.; Zong, M.; Liu, X. Attention-based CNNs for image classification: A survey. In Proceedings of the Journal of Physics: Conference Series; IOP Publishing: Bristol, UK, 2022; Volume 2171, p. 012068. [Google Scholar]
  11. Li, Q.; Chen, Y.; Zhang, A.; Jiang, Y.; Zou, L.; Xu, Z.; Muntean, G.M. A super-resolution flexible video coding solution for improving live streaming quality. IEEE Trans. Multimed. 2022, 25, 6341–6355. [Google Scholar] [CrossRef]
  12. Huang, K.; Tian, C.; Xu, Z.; Li, N.; Lin, J.C.W. Motion Context guided Edge-preserving network for video salient object detection. Expert Syst. Appl. 2023, 233, 120739. [Google Scholar] [CrossRef]
  13. Zhang, Y.; Zhang, Y.; Wu, Y.; Tao, Y.; Bian, K.; Zhou, P.; Song, L.; Tuo, H. Improving quality of experience by adaptive video streaming with super-resolution. In Proceedings of the IEEE INFOCOM 2020-IEEE Conference on Computer Communications, Virtual, 6–9 July 2020; pp. 1957–1966. [Google Scholar]
  14. Qiao, C.; Li, D.; Guo, Y.; Liu, C.; Jiang, T.; Dai, Q.; Li, D. Evaluation and development of deep neural networks for image super-resolution in optical microscopy. Nat. Methods 2021, 18, 194–202. [Google Scholar] [CrossRef]
  15. Gendy, G.; He, G.; Sabor, N. Lightweight image super-resolution based on deep learning: State-of-the-art and future directions. Inf. Fusion 2023, 94, 284–310. [Google Scholar] [CrossRef]
  16. Tian, C.; Fei, L.; Zheng, W.; Xu, Y.; Zuo, W.; Lin, C.W. Deep learning on image denoising: An overview. Neural Netw. 2020, 131, 251–275. [Google Scholar] [CrossRef] [PubMed]
  17. Tian, C.; Xu, Y.; Fei, L.; Yan, K. Deep learning for image denoising: A survey. In Proceedings of the Genetic and Evolutionary Computing: Proceedings of the Twelfth International Conference on Genetic and Evolutionary Computing, Changzhou, China, 14–17 December 2018; Springer: Berlin/Heidelberg, Germany, 2019; pp. 563–572. [Google Scholar]
  18. Zhu, L.; Zhang, C.; Song, J.; Liu, L.; Zhang, S.; Li, Y. Multi-graph based hierarchical semantic fusion for cross-modal representation. In Proceedings of the 2021 IEEE International Conference on Multimedia and Expo (ICME), Virtual, 5–9 July 2021; pp. 1–6. [Google Scholar]
  19. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  20. Tian, C.; Xu, Y.; Li, Z.; Zuo, W.; Fei, L.; Liu, H. Attention-guided CNN for image denoising. Neural Netw. 2020, 124, 117–129. [Google Scholar] [CrossRef] [PubMed]
  21. Zhu, L.; Song, J.; Wei, X.; Yu, H.; Long, J. CAESAR: Concept augmentation based semantic representation for cross-modal retrieval. Multimed. Tools Appl. 2022, 81, 34213–34243. [Google Scholar] [CrossRef]
  22. Tian, C.; Xu, Y.; Zuo, W. Image denoising using deep CNN with batch renormalization. Neural Netw. 2020, 121, 461–473. [Google Scholar] [CrossRef]
  23. Tian, C.; Yuan, Y.; Zhang, S.; Lin, C.W.; Zuo, W.; Zhang, D. Image super-resolution with an enhanced group convolutional neural network. Neural Netw. 2022, 153, 373–385. [Google Scholar] [CrossRef]
  24. Wang, Z.; Chen, J.; Hoi, S.C. Deep learning for image super-resolution: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 3365–3387. [Google Scholar] [CrossRef] [PubMed]
  25. Tian, C.; Zhang, X.; Lin, J.C.W.; Zuo, W.; Zhang, Y.; Lin, C.W. Generative adversarial networks for image super-resolution: A survey. arXiv 2022, arXiv:2204.13620. [Google Scholar]
  26. Dong, C.; Loy, C.C.; Tang, X. Accelerating the super-resolution convolutional neural network. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part II 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 391–407. [Google Scholar]
  27. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image super-resolution using deep convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 295–307. [Google Scholar] [CrossRef]
  28. Kim, J.; Lee, J.K.; Lee, K.M. Deeply-recursive convolutional network for image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1637–1645. [Google Scholar]
  29. Tai, Y.; Yang, J.; Liu, X. Image super-resolution via deep recursive residual network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3147–3155. [Google Scholar]
  30. Lim, B.; Son, S.; Kim, H.; Nah, S.; Mu Lee, K. Enhanced deep residual networks for single image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 136–144. [Google Scholar]
  31. Ledig, C.; Theis, L.; Huszar, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; Volume 2017, pp. 4681–4690. [Google Scholar]
  32. Tian, C.; Zhang, X.; Zhang, Q.; Yang, M.; Ju, Z. Image super-resolution via dynamic network. CAAI Trans. Intell. Technol. 2024, 9, 837–849. [Google Scholar] [CrossRef]
  33. Tian, C.; Zhang, X.; Ren, J.; Zuo, W.; Zhang, Y.; Lin, C.W. A Heterogeneous Dynamic Convolutional Neural Network for Image Super-resolution. arXiv 2024, arXiv:2402.15704. [Google Scholar]
  34. Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
  35. Ahn, N.; Kang, B.; Sohn, K.A. Fast, accurate, and lightweight super-resolution with cascading residual network. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 252–268. [Google Scholar]
  36. Hui, Z.; Wang, X.; Gao, X. Fast and accurate single image super-resolution via information distillation network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 723–731. [Google Scholar]
  37. Hui, Z.; Gao, X.; Yang, Y.; Wang, X. Lightweight image super-resolution with information multi-distillation network. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 2024–2032. [Google Scholar]
  38. Liu, J.; Tang, J.; Wu, G. Residual feature distillation network for lightweight image super-resolution. In Proceedings of the Computer Vision–ECCV 2020 Workshops, Glasgow, UK, 23–28 August 2020; Proceedings, Part III 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 41–55. [Google Scholar]
  39. Kong, F.; Li, M.; Liu, S.; Liu, D.; He, J.; Bai, Y.; Chen, F.; Fu, L. Residual local feature network for efficient super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022; pp. 766–776. [Google Scholar]
  40. Zhang, A.; Ren, W.; Liu, Y.; Cao, X. Lightweight image super-resolution with superpixel token interaction. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–6 October 2023; pp. 12728–12737. [Google Scholar]
  41. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  42. Sun, L.; Pan, J.; Tang, J. Shufflemixer: An efficient convnet for image super-resolution. Adv. Neural Inf. Process. Syst. 2022, 35, 17314–17326. [Google Scholar]
  43. Tian, C.; Zhuge, R.; Wu, Z.; Xu, Y.; Zuo, W.; Chen, C.; Lin, C.W. Lightweight image super-resolution with enhanced CNN. Knowl. Based Syst. 2020, 205, 106235. [Google Scholar] [CrossRef]
  44. Conde, M.V.; Choi, U.J.; Burchi, M.; Timofte, R. Swin2sr: Swinv2 transformer for compressed image super-resolution and restoration. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; pp. 669–687. [Google Scholar]
  45. Hou, H.; Andrews, H. Cubic splines for image interpolation and digital filtering. IEEE Trans. Acoust. Speech Signal Process. 1978, 26, 508–517. [Google Scholar]
  46. Keys, R. Cubic convolution interpolation for digital image processing. IEEE Trans. Acoust. Speech Signal Process. 1981, 29, 1153–1160. [Google Scholar] [CrossRef]
  47. Li, J.; Wu, J.; Deng, H.; Liu, J. A self-learning image super-resolution method via sparse representation and non-local similarity. Neurocomputing 2016, 184, 196–206. [Google Scholar] [CrossRef]
  48. Howard, A.G. MobileNets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  49. Ma, N.; Zhang, X.; Zheng, H.T.; Sun, J. Shufflenet v2: Practical guidelines for efficient cnn architecture design. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 116–131. [Google Scholar]
  50. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  51. Hu, Y.; Li, J.; Huang, Y.; Gao, X. Channel-wise and spatial feature modulation network for single image super-resolution. IEEE Trans. Circuits Syst. Video Technol. 2019, 30, 3911–3927. [Google Scholar] [CrossRef]
  52. Li, Z.; Yang, J.; Liu, Z.; Yang, X.; Jeon, G.; Wu, W. Feedback network for image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 3867–3876. [Google Scholar]
  53. Tian, C.; Zhang, Y.; Zuo, W.; Lin, C.W.; Zhang, D.; Yuan, Y. A heterogeneous group CNN for image super-resolution. In Proceedings of the IEEE Transactions on Neural Networks and Learning Systems, Virtual Event, 23–28 May 2022. [Google Scholar]
  54. Luo, Z.; Huang, H.; Yu, L.; Li, Y.; Fan, H.; Liu, S. Deep constrained least squares for blind image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 21–24 June 2022; pp. 17642–17652. [Google Scholar]
  55. Ma, C.; Yang, C.Y.; Yang, X.; Yang, M.H. Learning a no-reference quality metric for single-image super-resolution. Comput. Vis. Image Underst. 2017, 158, 1–16. [Google Scholar] [CrossRef]
  56. Bevilacqua, M.; Roumy, A.; Guillemot, C.; Alberi-Morel, M.L. Low-complexity single-image super-resolution based on nonnegative neighbor embedding. In Proceedings of the 23rd British Machine Vision Conference (BMVC), Surrey, UK, 3–7 September 2012; BMVA Press: Durham, UK, 2012; pp. 135.1–135.10, ISBN 1-901725-46-4. [Google Scholar] [CrossRef]
  57. Zeyde, R.; Elad, M.; Protter, M. On single image scale-up using sparse-representations. In Proceedings of the Curves and Surfaces: 7th International Conference, Avignon, France, 24–30 June 2010; Revised Selected Papers 7. Springer: Berlin/Heidelberg, Germany, 2012; pp. 711–730. [Google Scholar]
  58. Martin, D.; Fowlkes, C.; Tal, D.; Malik, J. A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In Proceedings of the Eighth IEEE international Conference on Computer Vision (ICCV), Vancouver, BC, Canada, 7–14 July 2001; Volume 2, pp. 416–423. [Google Scholar]
  59. Huang, J.B.; Singh, A.; Ahuja, N. Single image super-resolution from transformed self-exemplars. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, , 7–12 June 2015; pp. 5197–5206. [Google Scholar]
  60. Chen, X.; Wang, X.; Zhou, J.; Qiao, Y.; Dong, C. Activating more pixels in image super-resolution transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–23 June 2023; pp. 22367–22377. [Google Scholar]
  61. Huang, Y.; Li, S.; Wang, L.; Tan, T.; Luo, Z. Unfolding the alternating optimization for blind super resolution. Adv. Neural Inf. Process. Syst. 2020, 33, 5632–5643. [Google Scholar]
  62. Fujimoto, A.; Ogawa, T.; Yamamoto, K.; Matsui, Y.; Yamasaki, T.; Aizawa, K. Manga109 dataset and creation of metadata. In Proceedings of the 1st International Workshop on Comics Analysis, Processing and Understanding, Cancun, Mexico, 4–7 December 2016; pp. 1–5. [Google Scholar]
  63. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  64. Tai, Y.; Yang, J.; Liu, X.; Xu, C. Memnet: A persistent memory network for image restoration. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. pp. 4539–4547.
  65. Lai, W.S.; Huang, J.B.; Ahuja, N.; Yang, M.H. Deep laplacian pyramid networks for fast and accurate super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 624–632. [Google Scholar]
  66. Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A convnet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 11976–11986. [Google Scholar]
  67. Ding, X.; Zhang, X.; Han, J.; Ding, G. Scaling up your kernels to 31x31: Revisiting large kernel design in cnns. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 11963–11975. [Google Scholar]
  68. Chu, X.; Zhang, B.; Ma, H.; Xu, R.; Li, Q. Fast, accurate and lightweight super-resolution with neural architecture search. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 59–64. [Google Scholar]
  69. Gou, Y.; Hu, P.; Lv, J.; Zhu, H.; Peng, X. Rethinking image super resolution from long-tailed distribution learning perspective. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 14327–14336. [Google Scholar]
  70. Yao, J.E.; Tsao, L.Y.; Lo, Y.C.; Tseng, R.; Chang, C.C.; Lee, C.Y. Local implicit normalizing flow for arbitrary-scale image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 1776–1785. [Google Scholar]
  71. Agustsson, E.; Timofte, R.N. Challenge on single image super-resolution: Dataset and study. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2016; pp. 21–26. [Google Scholar]
Figure 1. The architecture of ADSCPN.
Figure 1. The architecture of ADSCPN.
Electronics 13 04613 g001
Figure 2. The structure of the Feature Process Block (FPB).
Figure 2. The structure of the Feature Process Block (FPB).
Electronics 13 04613 g002
Figure 3. The structure of the channel processing block (CPB).
Figure 3. The structure of the channel processing block (CPB).
Electronics 13 04613 g003
Figure 4. The structure of point-wise multi-layer perceptron (PW-MLP).
Figure 4. The structure of point-wise multi-layer perceptron (PW-MLP).
Electronics 13 04613 g004
Figure 5. The structure of the feature extraction block (FEB).
Figure 5. The structure of the feature extraction block (FEB).
Electronics 13 04613 g005
Figure 6. The structure of the point-attention convolution group (PACG).
Figure 6. The structure of the point-attention convolution group (PACG).
Electronics 13 04613 g006
Figure 7. ESA and EESA module structure comparison. (a) Original ESA block. (b) EESA block.
Figure 7. ESA and EESA module structure comparison. (a) Original ESA block. (b) EESA block.
Electronics 13 04613 g007
Figure 8. The impact of different settings on quantitative PSNR performance. (a) The number of Feature Process Blocks. (b) The number of model channels.
Figure 8. The impact of different settings on quantitative PSNR performance. (a) The number of Feature Process Blocks. (b) The number of model channels.
Electronics 13 04613 g008
Figure 9. Visual comparison of the results for different methods.
Figure 9. Visual comparison of the results for different methods.
Electronics 13 04613 g009
Figure 10. Visual comparison of different super-resolution methods on the DRealSR, DIV2KRK, and RealSR datasets.
Figure 10. Visual comparison of different super-resolution methods on the DRealSR, DIV2KRK, and RealSR datasets.
Electronics 13 04613 g010
Table 1. The summary of frequently used notations.
Table 1. The summary of frequently used notations.
NotationDefinition
I L R Low-resolution input image
I H R High-resolution target image
I S R Super-resolution output image
rScaling factor
H , W , C Height, width, and channels of the image
F e n c Shallow feature maps extracted from I L R
F K Deep features after processing through K blocks
F p r o j Features after channel projection
F s p a t Features after spatial attention block
A s p a t Attention weights for spatial features
Table 2. Quantitative comparison of ADSCPN with different group numbers.
Table 2. Quantitative comparison of ADSCPN with different group numbers.
Group NumberParametersMACsSet5Set14B100
(K)(G)PSNR/SSIMPSNR/SSIMPSNR/SSIM
276265137.92/0.960333.45/0.916532.10/0.8987
470359737.88/0.960233.44/0.917032.10/0.8989
668457937.90/0.960233.44/0.916932.10/0.8988
867457037.90/0.960233.51/0.917432.11/0.8990
Table 3. Quantitative comparison of methods with scale factor 2.
Table 3. Quantitative comparison of methods with scale factor 2.
MethodScale FactorSet5Set14B100Urban100
PSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIM
Bicubic × 2 33.66/0.929930.24/0.868829.56/0.843126.88/0.8403
SRCNN × 2 36.66/0.954232.42/0.906331.36/0.896429.50/0.8946
FSRCNN × 2 37.00/0.955832.63/0.908831.53/0.892029.88/0.9009
VDSR × 2 37.53/0.959033.05/0.913031.90/0.896030.77/0.9140
LapSRN × 2 37.52/0.959132.99/0.913431.80/0.895230.41/0.9100
DRCN × 2 37.63/0.958832.98/0.913031.85/0.894230.75/0.9133
DRRN × 2 37.74/0.959133.23/0.914532.05/0.897331.23/0.9188
CARN × 2 37.76/0.959033.52/0.916632.09/0.897831.92/0.9256
MemNet × 2 37.78/0.959733.28/0.914232.08/0.897831.31/0.9195
IDN × 2 37.83/0.960033.30/0.914832.08/0.898531.27/0.9196
FALSR-A × 2 37.82/0.959533.55/0.916832.12/0.898731.93/0.9256
LINF × 2 37.49/—33.38/—32.16/—31.22/—
FPL × 2 35.73/ 0.950931.59/0.905930.75/0.902230.46/0.9257
ADSCPN × 2 37.84/0.960033.42/0.916532.06/0.898331.67/0.9238
ADSCPN-plus × 2 37.90/0.960233.51/0.917432.11/0.899031.93/0.9262
Table 4. Quantitative comparison of methods with scale factor 3.
Table 4. Quantitative comparison of methods with scale factor 3.
MethodScale FactorSet5Set14B100Urban100
PSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIM
Bicubic × 3 30.39/0.868227.55/0.774227.21/0.738524.46/0.7349
SRCNN × 3 32.75/0.909029.30/0.820328.41/0.786326.24/0.8090
FSRCNN × 3 33.06/0.914029.43/0.824228.53/0.791026.43/0.8080
VDSR × 3 33.66/0.921329.77/0.831428.82/0.797627.14/0.8279
DRCN × 3 33.82/0.922629.76/0.831128.80/0.796327.15/0.8276
DRRN × 3 34.03/0.924429.96/0.834928.95/0.800427.53/0.8378
IDN × 3 34.11/0.925330.13/0.836029.01/0.801327.40/0.8359
CARN-M × 3 33.99/0.924530.08/0.835228.91/0.800027.47/0.8371
MemNet × 3 34.09/0.924830.00/0.835028.96/0.800127.56/0.8376
MADNet × 3 34.16/0.925330.21/0.839829.08/0.802327.82/0.8423
LINF × 3 33.94/—29.84/—28.55/—28.39/—
ADSCPN × 3 34.21/0.925230.22/0.839828.99/0.802427.81/0.8444
ADSCPN-plus × 3 34.23/0.925830.24/0.839929.10/0.803328.78/0.8463
Table 5. Quantitative comparison of methods with scale factor 4.
Table 5. Quantitative comparison of methods with scale factor 4.
MethodScale FactorSet5Set14B100Urban100
PSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIM
Bicubic × 4 28.42/0.810426.00/0.702725.96/0.667523.14/0.6577
SRCNN × 4 30.48/0.862827.50/0.751326.90/0.710124.52/0.7221
FSRCNN × 4 30.72/0.866027.61/0.755026.98/0.715024.62/0.7280
VDSR × 4 31.35/0.883828.01/0.767427.29/0.725125.18/0.7524
LapSRN × 4 31.54/0.885228.09/0.770027.32/0.727525.21/0.7562
DRCN × 4 31.53/0.885428.02/0.767027.23/0.723325.14/0.7510
DRRN × 4 31.68/0.888828.21/0.772027.38/0.728425.44/0.7638
IDN × 4 31.82/0.890328.25/0.773027.41/0.729725.41/0.7632
CARN-M × 4 31.92/0.890328.42/0.776227.44/0.730425.62/0.7694
MemNet × 4 31.74/0.889328.26/0.772327.40/0.728125.50/0.7630
MADNet × 4 31.93/0.891728.44/0.778027.47/0.732725.76/0.7746
ECBSR × 4 31.92/0.894628.34/0.781727.48/0.739325.53/0.7773
LINF × 4 31.70/—27.54/—26.62/—25.15/—
FPL × 4 30.39/0.880526.91/0.768826.29/0.739324.78/0.7880
ADSCPN × 4 32.04/0.892928.50/0.779327.52/0.733925.95/0.7812
ADSCPN-plus × 4 32.12/0.894028.49/0.779427.53/0.734525.98/0.7823
Table 6. Quantitative comparison of methods on the DIV2KRK test set.
Table 6. Quantitative comparison of methods on the DIV2KRK test set.
Method × 2 × 4
PSNR/SSIMPSNR/SSIM
Bicubic28.69/0.805825.38/0.6822
ADSCPN-plus29.19/0.822725.65/0.6940
EDSR(192)29.21/0.823425.66/0.6945
ADRBN-plus29.21/0.823325.66/0.6944
Table 7. Quantitative results on the DRealSR test set.
Table 7. Quantitative results on the DRealSR test set.
Method×2×3×4
PSNR/SSIMPSNR/SSIMPSNR/SSIM
Bicubic32.67/0.906031.51/0.870030.56/0.8595
ADSCPN-plus32.82/0.908931.60/0.872730.62/0.8611
Table 8. Quantitative comparison of methods on the RealSR test set.
Table 8. Quantitative comparison of methods on the RealSR test set.
Method×2×3×4
PSNR/SSIMPSNR/SSIMPSNR/SSIM
Bicubic31.67/0.886528.63/0.808527.23/0.7637
ADSCPN-plus34.01/0.922530.88/0.846529.27/0.8255
Table 9. Comparison of model parameters and MACs.
Table 9. Comparison of model parameters and MACs.
MethodParameters (K)MACs (G)
SRCNN5753
VDSR665613
LapSRN81330
DRCN177417,974
DRRN2976797
MemNet6772662
CARN1592223
MADNet878187
FALSR-A1021235
Ours674570
Table 10. Comparison of replacing large kernel parallel depthwise convolution with 7 × 7 depthwise separable convolution.
Table 10. Comparison of replacing large kernel parallel depthwise convolution with 7 × 7 depthwise separable convolution.
ConvolutionScale FactorSet5Set14B100Urban100
StructurePSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIM
Large Kernel Parallel Depthwise Separable Conv Group×237.90/0.960233.51/0.917432.11/0.899031.93/0.9262
7 × 7 Depthwise Separable Conv×237.92/0.960333.43/0.916731.10/0.898831.86/0.9257
Table 11. Comparison of model complexity between large-kernel parallel depthwise separable convolution groups and a single 7 × 7 depthwise separable convolution.
Table 11. Comparison of model complexity between large-kernel parallel depthwise separable convolution groups and a single 7 × 7 depthwise separable convolution.
Convolution StructureParameters (K)MACs (G)
Large-Kernel Parallel Depthwise Separable Conv. Group674570.07
7 × 7 Depthwise Separable Convolution732624.22
Table 12. Quantitative comparison of replacing dynamic convolution with 3 × 3 standard convolution.
Table 12. Quantitative comparison of replacing dynamic convolution with 3 × 3 standard convolution.
Convolution StructureScale FactorSet5Set14B100Urban100
PSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIM
Dynamic Convolution × 2 37.90/0.960233.51/0.917432.11/0.899031.93/0.9262
3 × 3 Conv × 2 37.88/0.960133.46/0.916732.10/0.898831.82/0.9253
Table 13. Comparison between point attention convolution group and standard convolution group.
Table 13. Comparison between point attention convolution group and standard convolution group.
Convolution StructureScale FactorSet5Set14B100Urban100
PSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIM
Point Attention Convolution Group × 2 37.90/0.960233.51/0.917432.11/0.899031.93/0.9262
Standard Convolution Group × 2 37.90/0.960133.46/0.916932.09/0.898631.86/0.9256
Table 14. Comparison of model complexity between enhanced spatial attention block and efficient enhanced spatial attention block.
Table 14. Comparison of model complexity between enhanced spatial attention block and efficient enhanced spatial attention block.
Spatial Attention BlockParameters (K)MACs (G)
Enhanced Spatial Attention Block740574.80
Efficient Enhanced Spatial Attention Block674570.07
Table 15. Quantitative ablation study of different model component combinations.
Table 15. Quantitative ablation study of different model component combinations.
Large Kernel ParallelDynamicPoint AttentionSet5Set14B100Urban100
Depthwise Separable Conv GroupConvolutionConvolution GroupPSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIM
×××37.93/0.960233.44/0.916632.11/0.898931.88/0.9257
××37.88/0.960133.43/0.915932.10/0.898831.84/0.9257
××37.88/0.960233.46/0.916832.10/0.898631.82/0.9254
××37.88/0.960133.46/0.916732.09/0.898831.80/0.9253
×37.92/0.960333.43/0.916731.86/0.925731.88/0.9258
×37.87/0.960133.39/0.916132.10/0.898631.82/0.9252
×37.90/0.960133.46/0.916932.11/0.898631.83/0.9256
37.90/0.960233.51/0.917432.11/0.899031.93/0.9262
Table 16. Ablation study on activation functions.
Table 16. Ablation study on activation functions.
Activation FunctionScaling FactorSet5Set14B100Urban100
PSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIM
SiLU × 2 37.91/0.960233.45/0.916832.10/0.898931.88/0.9259
GELU × 2 37.90/0.960233.51/0.917432.11/0.899031.93/0.9262
ReLU × 2 37.91/0.960233.42/0.916732.12/0.899131.94/0.9263
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Long, Y.; Ruan, H.; Zhao, H.; Liu, Y.; Zhu, L.; Zhang, C.; Zhu, X. Adaptive Dynamic Shuffle Convolutional Parallel Network for Image Super-Resolution. Electronics 2024, 13, 4613. https://doi.org/10.3390/electronics13234613

AMA Style

Long Y, Ruan H, Zhao H, Liu Y, Zhu L, Zhang C, Zhu X. Adaptive Dynamic Shuffle Convolutional Parallel Network for Image Super-Resolution. Electronics. 2024; 13(23):4613. https://doi.org/10.3390/electronics13234613

Chicago/Turabian Style

Long, Yiting, Haoyu Ruan, Hui Zhao, Yi Liu, Lei Zhu, Chengyuan Zhang, and Xinghui Zhu. 2024. "Adaptive Dynamic Shuffle Convolutional Parallel Network for Image Super-Resolution" Electronics 13, no. 23: 4613. https://doi.org/10.3390/electronics13234613

APA Style

Long, Y., Ruan, H., Zhao, H., Liu, Y., Zhu, L., Zhang, C., & Zhu, X. (2024). Adaptive Dynamic Shuffle Convolutional Parallel Network for Image Super-Resolution. Electronics, 13(23), 4613. https://doi.org/10.3390/electronics13234613

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop