Next Article in Journal
Two Lot-Sizing Algorithms for Minimizing Inventory Cost and Their Software Implementation
Next Article in Special Issue
Application of Attention-Enhanced 1D-CNN Algorithm in Hyperspectral Image and Spectral Fusion Detection of Moisture Content in Orah Mandarin (Citrus reticulata Blanco)
Previous Article in Journal
Exploring Community Awareness of Mangrove Ecosystem Preservation through Sentence-BERT and K-Means Clustering
Previous Article in Special Issue
Identifying Smartphone Users Based on Activities in Daily Living Using Deep Neural Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Level Attention Split Network: A Novel Malaria Cell Detection Algorithm

School of Information Science and Engineering, Zhejiang Sci-Tech University, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
Information 2024, 15(3), 166; https://doi.org/10.3390/info15030166
Submission received: 18 January 2024 / Revised: 7 February 2024 / Accepted: 11 March 2024 / Published: 15 March 2024

Abstract

:
Malaria is one of the major global health threats. Microscopic examination has been designated as the “gold standard” for malaria detection by the World Health Organization. However, it heavily relies on the experience of doctors, resulting in long diagnosis time, low efficiency, and a high risk of missed or misdiagnosed cases. To alleviate the pressure on healthcare workers and achieve automated malaria detection, numerous target detection models have been applied to the blood smear examination for malaria cells. This paper introduces the multi-level attention split network (MAS-Net) that improves the overall detection performance by addressing the issues of information loss for small targets and mismatch between the detection receptive field and target size. Therefore, we propose the split contextual attention structure (SPCot), which fully utilizes contextual information and avoids excessive channel compression operations, reducing information loss and improving the overall detection performance of malaria cells. In the shallow detection layer, we introduce the multi-scale receptive field detection head (MRFH), which better matches targets of different scales and provides a better detection receptive field, thus enhancing the performance of malaria cell detection. On the NLM—Malaria Dataset provided by the National Institutes of Health, the improved model achieves an average accuracy of 75.9% in the public dataset of Plasmodium vivax (malaria)-infected human blood smear. Considering the practical application of the model, we introduce the Performance-aware Approximation of Global Channel Pruning (PAGCP) to compress the model size while sacrificing a small amount of accuracy. Compared to other state-of-the-art (SOTA) methods, the proposed MAS-Net achieves competitive results.

1. Introduction

Malaria, as a prominent global public health issue, poses a significant threat to human well-being. Despite notable progress in malaria prevention since 2000, the annual incidence of over 200 million new cases remains alarmingly high, leading to a persistently elevated death toll. The malaria parasites that infect humans include Plasmodium falciparum, Plasmodium vivax, Plasmodium ovale, and Plasmodium malariae [1]. Among these, P. falciparum presents the most severe risk to human health. The life cycle of malaria parasites is intricate, encompassing various stages such as the ring stage, trophozoite stage, schizont stage, and gametocyte stage. Microscopic examination of Giemsa-stained blood smears continues to serve as the preferred method for definitively diagnosing both malaria parasites and malaria cases [2]. Nonetheless, the dependence on experienced physicians for conducting microscopic examination of blood smears remains unchanged. Following successful endeavors in malaria prevention and control, there has been a decline in the incidence of malaria cases. As a result, numerous primary healthcare facilities and infectious disease control institutions have redirected their focus away from conducting microscopic examinations for malaria. Consequently, there has been an enduring deterioration in the proficiency of malaria microscopy [3,4,5], resulting in an insufficient capacity to promptly identify cases of malaria [6].
The field of computer vision has experienced significant advancements in recent years, with deep learning-based neural network models achieving remarkable success in object detection. These advanced models have been widely applied in the field of biological image analysis. Utilizing computer-aided diagnosis can substantially alleviate the workload of healthcare professionals by facilitating the automatic identification of cells and their respective disease types. The proposed method effectively reduces false negatives and false positives resulting from visual fatigue, thereby providing fast and accurate quantitative results. These findings have significant implications for the prevention and control of malaria. EfficientNet, GoogleNet, and VGG19 are mainly used for the recognition and localization of malaria cells in malaria detection. If accurate identification of the types of malaria and judgment of their developmental stages are required, the Faster-RCNN and YOLO series algorithms are more commonly employed. In the past year, the RT-DETR method has been applied in related fields. However, the microscopic examination of malaria blood smears poses several challenges. Malaria parasites reside within red blood cells and undergo minimal changes in shape during different developmental stages. As a result, differentiation can only be achieved by observing the texture of the cytoplasmic staining. Furthermore, individual cells occupy a minuscule portion of the image, and a single blood smear typically contains hundreds of cells, with cells often adhering to and obstructing one another. These factors pose challenges for feature extraction. To address these challenges, it becomes essential to enhance the model’s feature extraction capabilities and effectively utilize information. Hence, we propose MAS-Net, a context-aware feature extraction model. Computer-aided diagnosis can greatly alleviate the workload of healthcare professionals and facilitate the automatic identification of cells and their corresponding disease types. Additionally, it effectively reduces false negatives and false positives caused by visual fatigue while providing rapid and accurate quantitative results. This holds immense significance for malaria prevention and control. The specific approach is described as follows:
  • We introduce the Split Contextual Attention Structure (SPCot) module to incorporate contextual information into our feature extraction network. This module enhances the capability of feature extraction and also reduces the complexity of the model via the introduction of a redundant split convolutional (RSConv).
  • An MRFH detection head is proposed, introducing the adaptive matching receptive field to resolve the mismatch issue between the receptive field and the target size.
  • By using Performance-aware Approximation of Global Channel Pruning (PAGCP) [7] for pruning our model, we significantly reduces the parameter count of our model at the cost of sacrificing minimal accuracy.

2. Related Work

With the development of image processing and computer science technology, the detection of malaria-infected cells has gradually shifted from purely manual to semi-manual or even fully automated. A large number of computer vision techniques have been applied during the analysis of medical data. F. Boray Tek et al. [8] studied the self-detection and identification method of malaria parasites in thin blood smear images and found that Giemsa staining could highlight malaria parasites, white blood cells, platelets, and artificially introduced noise. Therefore, an improved K-nearest neighbor classifier-based binary parasite detection method was proposed. At this stage, malaria detection has transitioned from purely manual to semi-manual, while cell detection is still limited to manually designed features. In 2015, LeCun et al. [9] first proposed the theory of deep learning, which has great significance in the development history of neural networks. In recent years, deep convolutional neural networks have achieved good results in various medical image analysis and processing tasks, and now various machine learning and deep learning algorithms have been applied to the detection of malaria parasites. In 2017, Liang et al. [10] proposed a new machine learning model based on Convolutional Neural Networks (CNNs) to classify uninfected and parasitic cells. Vijayalakshmi [11] applied the Support Vector Machine (SVM) and VGG architecture for malaria detection. In [12,13,14], pre-trained CNNs such as LeNet, AlexNet, GoogleNet, ResNet, and VGGnet were used with transfer learning techniques to detect malaria parasites, achieving high detection accuracy. Hung et al. [15] introduced a malaria detection method based on Faster R-CNN, which employed an object detection model previously used for natural images to identify malaria parasites in blood smear images. M. K. Dath [16] and S. S [17] used VGG19 for malaria detection and compared it with Restnet and SVM, proving the effectiveness of VGG19. Y. Jusman et al. [18] applied VGG19 and GoogLeNet for malaria parasite classification, and demonstrated that GoogLeNet is more suitable for the classification task compared to VGG19. F. A. S. Araujo et al. [19] employed EfficientNet for malaria cell detection and achieved good results. F. Yang [20] proposed the use of Cascade YOLO and YOLOv2. In [21], F. Abdurahman introduced the improved YOLOv3 and YOLOv4, which have higher accuracy and lower computational requirements. Due to efficiency issues with two-stage detection models, research on one-stage paradigm-based detection models has become increasingly popular. D. D. Acula et al. [22] compared the performance of DackNet53 and VGG16 models in malaria detection, and DackNet53 achieved good results. Liu Z et al. [23] used YOLOv5 for detecting malaria cells in blood smears and achieved excellent results. Guemas E et al. [24] introduced RT_DETR for malaria detection, which has lower accuracy compared to mainstream models, but has extremely low parameters and computational complexity. In recent years, both CNN and Transformer have made new research achievements. J. Li proposed SCConv [25], a plug-and-play module with less system redundancy. Y. Li et al. proposed a new self-attention structure called Transform, enhancing the feature extraction capability for small targets [26]. Li C et al. proposed ODConv [27], which is based on multi-channel attention and dynamic convolution. Regarding feature extraction networks, S.-H. Gao introduced Res2Net at a granularity level, representing multi-scale features and increasing receptive fields for each network layer [28]. S. Woo et al. proposed ConvNeXt V2, which approximates the performance of Transformer models [29]. Yu W. et al. introduced InceptionNeXt based on large convolution kernels, and there have also been lightweight works such as GhostNet V2, FasterNet, and EdgeNeXt [30,31,32,33]. Transformer-based models, such as SwinTransformer, EfficientViT, and RepVit [34,35,36], have shown good performance in complex scenes and small object detection. Among them, YOLOv5, proposed in 2021, has achieved a good balance between accuracy and efficiency, surpassing most of the two-stage detection models known for high accuracy. Thus, it can serve as a benchmark for current detection models.

3. MAS-Net

In this section, we will provide a detailed introduction to the proposed MAS-Net. The overall network architecture is shown in Figure 1.
MAS-Net follows the general paradigm of backbone, neck, and head. Firstly, SPCot is used as the unit to construct the feature extraction network, which outputs features maps of three scales: 1/8, 1/16, and 1/32. The specific backbone structure is shown in Figure 2. For the neck, we adopt the Feature Pyramid Network (FPN) structure from YOLOv5. Finally, the features are sent into the MRFH detection head for optimal receptive field adaptation.

3.1. RSConv

To minimize information loss and facilitate feature extraction in multiple channels, our model avoids extensive channel reduction and expansion operations. However, this approach leads to an exponential growth of parameters. To improve the efficiency of our model, we have explored techniques such as MobileNets [37,38,39], ShuffleNets [40,41], and GhostNet [42]. These methods utilize depthwise convolution (DWConv) [43] and group convolution (GConv) [44] to extract spatial features. While this approach helps reduce the number of floating-point operations (FLOPs), it does not effectively address the issue of reducing memory access [32]. Previous studies [32,45] have suggested minimizing redundancy in the feature maps. Inspired by this observation, we designed the segmentation-based RSConv method, illustrated in Figure 3.
The input feature map is denoted as X R c × h × w and the output feature is Y R m × h × w , with c representing the input channel and m representing the output channel. We define W R c × k × k × m as the convolution weights with a kernel size of k × k . The mathematical expression of the convolution operation, given by Y = W X + b , can be expressed as follows:
y 1 y 2 y m = W 11 W 12 W 1 , c W 21 W 12 W 2 , c W m , 1 W m , 2 W m , c x 1 x 2 x c
x i , i 1 , c represents one of the channels of the input feature map, and x i , i , j [ 1 ; c , m ] represents one of the m convolution kernels. After the convolution, we obtain the output y j , j [ 1 , m ] . RSConv divides the input channels into two main components: one component applies a k × k depth-wise convolution (DWConv) to capture intrinsic information, and the other component redundantly applies a cost-effective 1 × 1 convolution to enhance subtle hidden details. The left side of Figure 3 illustrates this configuration. Therefore, our initial RSConv can be represented as:
y 1 y 2 y m = W 11 W 12 W 1 , α c W 21 W 12 W 2 , α c W m , 1 W m , 2 W m , α c x 1 x 2 x α c + W 1 , α c + 1 W 1 , α c + 2 W 1 c W 2 , α c + 1 W 1 , α c + 2 W 2 c W m , α c + 1 W m , α c + 2 W m c x α c + 1 x α c + 2 x c
where W i j , j [ 1 , α c ] , represents the parameters of a 3 × 3 kernel for convolution on channel α c . W i j , j [ α c + 1 , c ] , represents the parameters of the inexpensive 1 × 1 kernel, which operates through point convolution on the remaining ( 1 α ) c redundant features.
Combining feature maps with different receptive fields directly may lead to the loss of crucial geometric details and introduce conflicts in information. In the end, we choose to connect the feature maps from different branches along the channel to preserve both spatial and channel information in the resulting feature maps. Consequently, the preserved feature maps can be utilized to compute the relative importance of each channel. We obtain channel statistics data S c by applying global average pooling, thus incorporating global information.
S c = F g a p U c = 1 H × W i = 1 H j = 1 W U c ( i , j )
Using a fully connected layer ( f c ) to obtain feature z, which guides precise and adaptive selection:
z = F f c ( S c )
We combine these two results, z a and z b , by stacking their vectors together, and then perform cross-channel soft attention operations. a c and b c represent the soft attention vectors of U a and U b .
a c = e z a e z a + e z b , b c = 1 a c ,
The final output Y can be obtained by fusing features U a and U b :
Y = a c · U a + b c · U b , a c + b c = 1
Complexity analysis:
The parameters of a regular convolution can be calculated as:
P C o n v = k × k × c × m
The parameters of RSConv can be calculated as:
P R S C o n v = k × k × α c + 1 × 1 × α c × m + 1 × 1 × 1 α c × m
After simplification, we obtain:
P R S C o n v = k × k × α c + c × m
When we set α = 0.5 and use a 3 × 3 kernel, the parameters can be reduced to 1 / 9 of the original.

3.2. Split Contextual Attention Structure

The YOLOv5 baseline network employs Darknet53 for extracting features. Malaria cells are found in blood smears in a small proportion and frequently overlap and adhere to one another. These cells are mainly characterized by subtle texture features. Additionally, neighboring cells create significant interference. Furthermore, scaling or reducing the channel of the feature map inevitably results in the loss of information about smaller targets. Drawing inspiration from [26,46], we introduce the self-attention structure, SPCot, which integrates contextual information. The structure is shown in Figure 4.
SPCot utilizes RSConv to encode the contextual information of all adjacent keys within a spatially oriented 3 × 3 grid, resulting in the contextual information K 1 R H × W × C . Subsequently, the local static contextual information K 1 is concatenated with the original input feature map X. Two consecutive 1 convolutions are then used to obtain the attention matrix A. In this case, X represents the original input feature map, while W α and W β represent the two convolution layers. The calculation formula is expressed as follows:
A = [ K 1 , X ] W α W β
In this context, X represents the input feature information, while V is defined as V = X W φ , where W φ represents a 1 × 1 convolution weight matrix. Afterwards, a matrix multiplication is performed to obtain the global dynamic contextual information K 2 . The calculation formula is given as follows:
K 2 = V · A
By fusing the local static contextual information K 1 and the global dynamic contextual information K 2 , the output is obtained.
K = K 1 + K 2
To mitigate potential information loss resulting from convolutions, residual structures are employed to compensate for the features.
Y = K + X
The detailed process can be described as Algorithm 1.
Algorithm 1 Algorithm for self-attention module based on contextual information
Input: 
X [batch, channel, height, weight]
Output: 
The final output Y
 1:
K 1 = R S C o n v ( X ) . The feature map X employs the RSConv convolution operation to acquire the local static contextual information K 1 .
 2:
V = X W φ . The feature map X undergoes a 1 × 1 convolution process to derive the output V.
 3:
A = [ K 1 , X ] W α W β . A is the attention matrixn.
 4:
K 2 = V · A . The global dynamic contextual information K 2 is obtained through element-wise multiplication of V and the attention matrixA.
 5:
K = K 1 + K 2 . The global contextual information K is obtained by adding K 1 and K 2 .
 6:
Y = X + K . The final output Y is obtained by adding the global contextual information K to the input features X.
We utilize the SPCot structure to construct our feature extraction network. The main construction parameters of the backbone are presented in the Table 1. The detailed network structure is shown in Figure 3.

3.3. Multi-Scale Receptive Field Detection Head

The detection head in the Dynamic Scale-Aware Head YOLO series typically comprises a 3 × 3 convolutional layer followed by a 1 × 1 convolutional layer. Due to the specific nature of the detection targets, a single type of detection head may exhibit a discrepancy between the receptive field and the target size. When the receptive field is too small, it can only capture local features, rendering it insufficient to capture comprehensive information about the target. Conversely, an excessively large receptive field introduces excessive noise and irrelevant information, leading to suboptimal detector performance. In their study, ref. [47] presented a dynamic selection mechanism in CNN that enables each neuron to adaptively adjust its receptive field size based on the multiple scales of input information. Leveraging this mechanism, we developed the MRFH detection head, as depicted in Figure 5.
After passing the input into MRFH, a series of convolutions are applied to the feature map. Specifically, a 3 × 3 convolution and a 5 × 5 convolution (achieved by two 3 × 3 convolutions with a dilation rate of 2) are used. These convolutions utilize diverse kernels to establish distinct receptive fields. Subsequently, softmax attention is employed to combine the information from these branches, allowing for the fusion of branches with different receptive fields. It is well known that simply summing feature maps with varying receptive fields can result in the dominance of specific geometric details and introduce conflicting information. Additionally, the reduction in channel of the obtained feature maps hampers their ability to effectively guide the learning process of both branches simultaneously. In contrast to SKNet’s [47] approach of pixel-wise summation followed by average pooling, we adopt an alternative strategy. Firstly, separate pooling operations are performed on each branch, and then their results are concatenated instead of using pixel-wise addition. We have decided to connect the feature maps from different branches in the channel in order to obtain feature maps that preserve both spatial and channel information. This enables the computation of the relative importance of each channel. This approach not only allows for the independent determination of channel importance for a specific receptive field, but also identifies the importance of feature maps relative to other branches with distinct receptive fields. In a previous study [48], it was demonstrated that a 3 × 3 convolutional layer fails to provide adequate receptive fields for each branch, leading to a preference for branches with larger receptive fields. Shallow detection heads, in particular, necessitate larger receptive fields to incorporate more global contextual information, thereby enhancing the recognition of small targets. Meanwhile, deep detection heads have already achieved significant receptive fields through the stacking of multiple residual modules and the introduction of SPPF modules. However, integrating MRFH into the deep detection heads quickly reaches a saturation point and introduces a higher number of parameters, resulting in inefficiency. Taking into consideration the balance between detection accuracy and model efficiency, we have made the decision to exclusively employ the MRFH detection head in the shallow detection layers.

3.4. PAGCP

Taking into consideration the limited computing resources commonly available in malaria-endemic areas, we employed the PAGCP [7] method to prune and compress the model, resulting in a more suitable MAS-Net-Tiny model for practical deployment. PAGCP adjusts the pruning optimization objective to:
min θ Ω E x D x S x ; θ s . t . g θ α
S · represents the significance standard used to evaluate the importance of filters. x represents the input image, where x follows the distribution of D . Additionally, g · represents a constraint vector determined by several factors, such as the reduction rate of FLOPs and parameters, while α is a threshold vector in which each channel corresponds to each of the aforementioned factors. Ω is the set of all pruning selections in the original model. l = 1 , 2 , , L and k = 1 , 2 , 3 , , K l , where K is the number of filters in the l-th layer. Ω represents the total set of parameters. The constraint function g · is parameterized by θ , where θ = 0 indicates the deletion of the convolutional kernel, while θ = 1 preserves it. The performance constraint is designed as:
g θ = Δ L 1 θ , Δ L 2 θ , , Δ L T θ
= max 1 t T Δ L t
The layer-wise constrained boundary is designed in a cascading transformation form:
i = 1 L 1 + d 1 λ i 1 = α
Among them, d 1 represents the constraint boundary during the initial layer pruning, which is the value of g θ . λ is the constraint boundary for each optimization step, equivalent to the scaling factor of the constraint boundary in the previous step. α is the global constraint boundary.
PAGCP adjusts the optimization objective to perform a constrained minimization search only on saliency indicators while keeping the model parameter weights unchanged. Among them, the constraint terms are adjusted to include performance loss constraints after model pruning and computational constraints. This ensures that the model parameters are in the optimal state and enables more accurate importance evaluation.

4. Model Training and Result Analysis

4.1. Dataset Introduction

We employed the Plasmodium vivax (malaria) infected human blood smear dataset [49] in this study. It consists of about 1364 images in PNG format. The dataset includes approximately 80,000 cells stained with the Giemsa stain. It covers two categories of uninfected cells: red blood cells and white blood cells. Moreover, it consists of four categories of infected cells: gametocytes, schizonts, merozoites, and trophozoites. The dataset’s annotation covers all these classes, including challenging-to-identify cells. Challenging cells are labeled as difficult types. Each image has a resolution of 1600 × 1200 pixels. In the experiment, we excluded cells labeled as difficult, white blood cells, and uninfected red blood cells. We retained only infected cells for further analysis.

4.2. Experimental Environment

The experimental operating system of this article is Windows 10, completed based on the GPU, PyTorch, and CUDA framework. The CPU of the experimental environment is an AMD Ryzen 7 5800, the graphics card is an NVIDIA RTX 3090 with 24 GB of memory, and the operating memory is 16 GB. Under the above experimental environment, the image dataset is being trained. The learning rate decay strategy is based on the cosine learning rate decay [50], with SGD chosen as the optimizer, and momentum and weight decay values are set to 0.937 and 0.0005, respectively. The batch size is set to 16, and no pre-trained weights are being loaded. The training consists of 300 epochs. The image augmentation strategies include Mixup, Mosaic, HSV transformation, and affine transformation, as well as image concatenation. The loss function combines CIOU and applies NWD [51]. In the PAGCP pruning process, an initial rate of 0.05 is set, an initial threshold of 5 is set, and other parameters are set to default values. It runs for 100 epochs.

4.3. Results and Analysis

This experiment evaluates precision ( P r e ), recall rate ( R e c ), mean Average Precision ( m A P ), floating-point operations ( F L O P s ), F 1 score, and frames per second ( F P S ).

4.4. Grad-CAM

Figure 6 shows the Grad-CAM of MAS-Net. In Figure 6a, it is difficult for our naked eye to detect the infected target at the center of the image, while our network can accurately focus on it. Through Figure 6b, we can clearly see the region of interest that the network is attentive to, which corresponds to the distribution of heme pigment. Additionally, we can observe that the network also pays attention to the swelling of infected cells, which is consistent with the characteristics of manually inspecting malaria cells.

4.5. Experimental Results

We compared the MAS-Net model with other YOLO series models. The results showed that compared to other one-stage models, MAS-Net performed better in terms of the PR curve, as shown in Figure 7. Additionally, there was a significant improvement in mAP value, as detailed in Figure 8.
According to Figure 7, it is evident that our MAS-Net model exhibits a superior precision–recall curve. Furthermore, Figure 8 highlights a slight reduction in the convergence speed of MAS-Net compared to other models. This can be attributed to the heightened depth caused by the complex structure of SPCot, which subsequently affects the training speed. Notably, MAS-Net achieves a remarkable m A P of 75.9% at 280 epochs, surpassing the performance of other SOTA models.
We can see from Table 2 that MAS-Net outperforms other models in terms of precision, recall, mAP, and F 1 score. Although MAS-Net has higher parameter and computational complexity compared to other models, it is lower than the v8 model with similar accuracy. However, MAS-Net has a slower image processing speed, which limits its practical applicability, especially in resource-constrained environments. To address this issue, we compressed the MAS-Net model and obtained MAS-Net-Tiny through PAGCP channel pruning.
From Table 3, it can be seen that our model has made significant progress in simplifying complexity and has also improved in processing speed. This helps improve the applicability of the algorithm in practical applications, but it has also led to a decrease of 5.3% in mAP.
Compared to the EfficientNet method in [19], MAS-Net has a higher computational complexity but slight advantages in mAP. Additionally, MAS-Net also has a faster image processing speed. Our models have shown improvements in performance compared to the models introduced in [23], which incorporate CA and BiFPN. The RT-DETR model in [24] has the advantage of low computational complexity and a high detection speed, as shown in the table. By comparison, our MAS-Net-Tiny model achieves a higher mAP and FPS with lower parameters and computational complexity.
To validate the effectiveness of our model, we compared it with other detection networks. The results are shown in Table 4.
We compared our model with other SOTA models using the PAN structure and trained it for 300 epochs. The results are shown in Table 5. Compared to other models, MAS-Net has higher accuracy but also higher complexity. On the other hand, MAS-Net-Tiny achieves a good balance between computational complexity and accuracy, which is advantageous for the model’s practical applicability.
According to the results presented in Table 2, the accuracy of detecting malaria cells with MAS-Net reaches 75.9%, which demonstrates a significant 6.9% improvement in m A P compared to the baseline YOLOv5 model. Moreover, precision has seen a notable increase of 10.5%, whereas recall has shown a substantial improvement of 4.4%. Furthermore, the computational complexity, measured in terms of G F L O P s , has increased by 15.7%. Moreover, considering the findings outlined in Table 5, it becomes clear that the increase in parameters can be primarily attributed to our MRFH detection head, whereas the deeper layers of MRFH contribute marginally to the improvement in m A P . Therefore, we chose not to employ MRFH in the deeper layers. After adopting PAGCP, the model’s mAP value decreased by 5.3%. However, there was a significant simplification in terms of parameters and computational complexity. The experimental results demonstrate that the MAS-Net model achieved high recognition accuracy in the detection of malaria-infected cells after Giemsa staining. Furthermore, the model’s parameter rationality and performance are superior to other state-of-the-art (SOTA) models.
Figure 9 illustrates the detection results of this model on the dataset, comprising six cell categories: four types of malaria-infected cells, healthy red blood cells, and white blood cells. The model exclusively detects the malaria-infected cells, yielding results that include predicted bounding boxes, class labels (ri: ring, tr: trophozoite, sc: schizont, ga: gametocyte), and confidence scores. Upon comparing the results, it becomes evident that the original YOLOv5 model exhibits a lack of confidence in accurately predicting the bounding boxes containing the targets, resulting in subpar detection performance with numerous missed instances. In contrast, our model effectively resolves the aforementioned issues. The confidence scores are generally higher in comparison to the original model, resulting in fewer missed detections, particularly in densely populated regions where infected cells are small and cellular adhesion is prevalent. This improved performance is attributed to the SPCot structure, which extracts contextual information, enhancing the capabilities of the MAS-Net model compared to the original YOLOv5 model.
The process of blood smear preparation and examination may result in low-quality images due to factors such as lighting conditions and microscope quality. We attempted to simulate this scenario by adding various types of noise and motion blur in the images. The detection results are shown in Figure 10.
Our model has a certain robustness to noise and can effectively identify infected cells. However, the confidence scores are relatively lower compared to the original images. On the other hand, our model exhibits high robustness to motion blur, with confidence scores similar to the original images. It may be because our model is more sensitive to colors in the images. It is difficult for the model to detect overlapping cells in the presence of noise, making it challenging to distinguish them as separate cells.

5. Conclusions

Applying computer science and technology to the field of biomedical research holds great potential. We have proposed the MAS-Net algorithm for the detection and recognition of malaria parasites in blood smear cells. Our SPCot structure effectively leverages contextual information to extract more detailed information. During the network construction process, we reduce the loss of feature information by minimizing channel compression operations and employing RSConv to reduce computational parameters, avoiding parameter explosion. The design of the MRFH detection head provides a larger receptive field for shallow networks. By introducing PAGCP channel pruning, we compress the model to improve its practical applicability. Experimental results have demonstrated the effectiveness and robustness of our model. Using the methods described in this paper, we can significantly improve the detection rate and reduce the workload of relevant personnel, even in the absence of specialized microscopic technicians. This has significant implications, particularly in medical diagnosis within the field of healthcare.

6. Future Work

The training dataset suffers from class imbalance, which is the reason why our model’s detection performance is only 75.9%. In our future work, we will attempt to address this issue by expanding the dataset through unsupervised learning, aiming to improve the recognition ability of our network and validate its generalization on diverse datasets. We also plan to extend the detection scope to malignant malaria parasites and other species of malaria parasites, as well as other bloodborne diseases that require microscopic examination. Furthermore, we will actively seek cooperation to promote the practical application of our detection network in diagnosis.

Author Contributions

Conceptualization, Z.X. and J.W.; Methodology, J.W.; Software, Z.X.; Verification, Z.X.; Resources, J.W.; Data Management, J.W.; Writing—First draft preparation, Z.X.; Writing—Review and Edit, J.W.; Project Management, Z.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Zhejiang 14th five-year graduate education reform project under Grant No. syjsjg2023144.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset used in this study can be found at the link below https://lhncbc.nlm.nih.gov/LHC-research/LHC-projects/image-processing/malaria-datasheet.html (accessed on 17 January 2024). The MAS-Net code can be found at the link below https://github.com/glassxiong/MAS_NET (accessed on 17 January 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MAS-Netmulti-level attention split network
SPCotsplit contextual attention structure
MRFHmulti-scale receptive field detection head
RSConvredundant split convolutional
PAGCPPerformance-aware Approximation of Global Channel Pruning
NWDNormalized Wasser-stein Distance

References

  1. Rougemont, M.; Van Saanen, M.; Sahli, R.; Hinrikson, H.P.; Bille, J.; Jaton, K. Detection of four Plasmodium species in blood from humans by 18S rRNA gene subunit-based and species-specific real-time PCR assays. J. Clin. Microbiol. 2004, 42, 5636–5643. [Google Scholar] [CrossRef]
  2. World Health Organization. Basic Malaria Microscopy Part I. Learner’s Guide; World Health Organization: Geneva, Switzerland, 2010. [Google Scholar]
  3. Bloch, M. The past and the present in the present. Man 1977, 12, 278–292. [Google Scholar] [CrossRef]
  4. Das, D.; Vongpromek, R.; Assawariyathipat, T.; Srinamon, K.; Kennon, K.; Stepniewska, K.; Ghose, A.; Sayeed, A.A.; Faiz, M.A.; Netto, R.L.A.; et al. Field evaluation of the diagnostic performance of EasyScan GO: A digital malaria microscopy device based on machine-learning. Malar. J. 2022, 21, 122. [Google Scholar] [CrossRef]
  5. Ayalew, F.; Tilahun, B.; Taye, B. Performance evaluation of laboratory professionals on malaria microscopy in Hawassa Town, Southern Ethiopia. BMC Res. Notes 2014, 7, 839. [Google Scholar] [CrossRef]
  6. Feng, X.; Levens, J.; Zhou, X.N. Protecting the gains of malaria elimination in China, 2020. Infect. Dis. Poverty 2020, 9, 43. [Google Scholar] [CrossRef]
  7. Ye, H.; Zhang, B.; Chen, T.; Fan, J.; Wang, B. Performance-aware Approximation of Global Channel Pruning for Multitask CNNs. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 10267–10284. [Google Scholar] [CrossRef] [PubMed]
  8. Tek, F.B.; Dempster, A.G.; Kale, I. Parasite detection and identification for automated thin blood film malaria diagnosis. Comput. Vis. Image Underst. 2010, 114, 21–32. [Google Scholar] [CrossRef]
  9. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  10. Liang, Z.; Powell, A.; Ersoy, I.; Poostchi, M.; Silamut, K.; Palaniappan, K.; Guo, P.; Hossain, M.A.; Sameer, A.; Maude, R.J.; et al. CNN-based image analysis for malaria diagnosis. In Proceedings of the 2016 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Shenzhen, China, 15–18 December 2016; pp. 493–496. [Google Scholar] [CrossRef]
  11. Vijayalakshmi, A. Deep learning approach to detect malaria from microscopic images. Multimed. Tools Appl. 2020, 79, 15297–15317. [Google Scholar] [CrossRef]
  12. Rajaraman, S.; Antani, S.K.; Poostchi, M.; Silamut, K.; Hossain, M.A.; Maude, R.J.; Jaeger, S.; Thoma, G.R. Pre-trained convolutional neural networks as feature extractors toward improved malaria parasite detection in thin blood smear images. PeerJ 2018, 6, e4568. [Google Scholar] [CrossRef]
  13. Dong, Y.; Jiang, Z.; Shen, H.; Pan, W.D.; Williams, L.A.; Reddy, V.V.; Benjamin, W.H.; Bryan, A.W. Evaluations of deep convolutional neural networks for automatic identification of malaria infected cells. In Proceedings of the 2017 IEEE EMBS International Conference on Biomedical & Health Informatics (BHI), Orlando, FL, USA, 16–19 February 2017; pp. 101–104. [Google Scholar] [CrossRef]
  14. Fuhad, K.; Tuba, J.F.; Sarker, M.R.A.; Momen, S.; Mohammed, N.; Rahman, T. Deep learning based automatic malaria parasite detection from blood smear and its smartphone based application. Diagnostics 2020, 10, 329. [Google Scholar] [CrossRef]
  15. Hung, J.; Carpenter, A. Applying faster R-CNN for object detection on malaria images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 808–813. [Google Scholar] [CrossRef]
  16. Dath, M.K.; Nazir, N. Diagnosing malaria with AI and image processing. In Proceedings of the 2023 3rd International Conference on Innovative Practices in Technology and Management (ICIPTM), Uttar Pradesh, India, 22–24 February 2023; pp. 1–6. [Google Scholar] [CrossRef]
  17. Suraksha, S.; Santhosh, C.; Vishwa, B. Classification of Malaria cell images using Deep Learning Approach. In Proceedings of the 2023 Third International Conference on Advances in Electrical, Computing, Communication and Sustainable Technologies (ICAECT), Bhilai, India, 5–6 January 2023; pp. 1–5. [Google Scholar] [CrossRef]
  18. Jusman, Y.; Aftal, A.A.; Tyassari, W.; Kanafiah, S.N.A.M.; Hayati, N.; Mohamed, Z. Classification of Parasite Malaria in Schizon Stage with GoogleNet and VGG-19 Pre-Trained Models. In Proceedings of the 2023 10th International Conference on Information Technology, Computer, and Electrical Engineering (ICITACEE), Semarang, Indonesia, 31 August–1 September 2023; pp. 219–223. [Google Scholar] [CrossRef]
  19. Araujo, F.; Colares, N.; Carvalho, U.; Costa Filho, C.; Costa, M. Plasmodium Life Cycle-Stage Classification on Thick Blood Smear Microscopy Images using Deep Learning: A Contribution to Malaria Diagnosis. In Proceedings of the 2023 19th International Symposium on Medical Information Processing and Analysis (SIPAIM), Mexico City, Mexico, 5–17 November 2023; pp. 1–4. [Google Scholar] [CrossRef]
  20. Yang, F.; Quizon, N.; Yu, H.; Silamut, K.; Maude, R.J.; Jaeger, S.; Antani, S. Cascading yolo: Automated malaria parasite detection for plasmodium vivax in thin blood smears. In Proceedings of the Medical Imaging 2020: Computer-Aided Diagnosis, Houston, TX, USA, 16–19 February 2020; Volume 11314, pp. 404–410. [Google Scholar]
  21. Abdurahman, F.; Fante, K.A.; Aliy, M. Malaria parasite detection in thick blood smear microscopic images using modified YOLOV3 and YOLOV4 models. BMC Bioinform. 2021, 22, 112. [Google Scholar] [CrossRef]
  22. Acula, D.D.; Carlos, J.A.P.; Lumacad, M.M.; Minano, J.C.L.O.; Reodica, J.K.R. Detection and classification of plasmodium parasites in human blood smear images using Darknet with YOLO. In Proceedings of the International Conference on Green Energy, Computing and Intelligent Technology (GEn-CITy 2023), Johor, Malaysia, 10–12 July 2023; pp. 24–31. [Google Scholar] [CrossRef]
  23. Liu, Z.; Liu, H.; Sun, Y. Detection and Classification of Malaria Parasite Based on Improved YOLOv5 Model. In Proceedings of the 2023 16th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Taizhou, China, 28–30 October 2023; pp. 1–5. [Google Scholar] [CrossRef]
  24. Guemas, E.; Routier, B.; Ghelfenstein-Ferreira, T.; Cordier, C.; Hartuis, S.; Marion, B.; Bertout, S.; Varlet-Marie, E.; Costa, D.; Pasquier, G. Automatic patient-level recognition of four Plasmodium species on thin blood smear by a real-time detection transformer (RT-DETR) object detection algorithm: A proof-of-concept and evaluation. Microbiol. Spectr. 2024, 12, e01440-23. [Google Scholar] [CrossRef]
  25. Li, J.; Wen, Y.; He, L. SCConv: Spatial and Channel Reconstruction Convolution for Feature Redundancy. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 6153–6162. [Google Scholar] [CrossRef]
  26. Li, Y.; Yao, T.; Pan, Y.; Mei, T. Contextual transformer networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 1489–1500. [Google Scholar] [CrossRef]
  27. Li, C.; Zhou, A.; Yao, A. Omni-dimensional dynamic convolution. arXiv 2022, arXiv:2209.07947. [Google Scholar]
  28. Gao, S.H.; Cheng, M.M.; Zhao, K.; Zhang, X.Y.; Yang, M.H.; Torr, P. Res2net: A new multi-scale backbone architecture. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 652–662. [Google Scholar] [CrossRef]
  29. Woo, S.; Debnath, S.; Hu, R.; Chen, X.; Liu, Z.; Kweon, I.S.; Xie, S. Convnext v2: Co-designing and scaling convnets with masked autoencoders. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 16133–16142. [Google Scholar] [CrossRef]
  30. Yu, W.; Zhou, P.; Yan, S.; Wang, X. Inceptionnext: When inception meets convnext. arXiv 2023, arXiv:2303.16900. [Google Scholar]
  31. Tang, Y.; Han, K.; Guo, J.; Xu, C.; Xu, C.; Wang, Y. GhostNetv2: Enhance cheap operation with long-range attention. Adv. Neural Inf. Process. Syst. 2022, 35, 9969–9982. [Google Scholar]
  32. Chen, J.; Kao, S.h.; He, H.; Zhuo, W.; Wen, S.; Lee, C.H.; Chan, S.H.G. Run, Don’t Walk: Chasing Higher FLOPS for Faster Neural Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 12021–12031. [Google Scholar] [CrossRef]
  33. Maaz, M.; Shaker, A.; Cholakkal, H.; Khan, S.; Zamir, S.W.; Anwer, R.M.; Shahbaz Khan, F. Edgenext: Efficiently amalgamated cnn-transformer architecture for mobile vision applications. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Cham, Switzerland; pp. 3–20. [Google Scholar]
  34. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 9992–10002. [Google Scholar] [CrossRef]
  35. Liu, X.; Peng, H.; Zheng, N.; Yang, Y.; Hu, H.; Yuan, Y. EfficientViT: Memory Efficient Vision Transformer with Cascaded Group Attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 14420–14430. [Google Scholar] [CrossRef]
  36. Wang, A.; Chen, H.; Lin, Z.; Pu, H.; Ding, G. Repvit: Revisiting mobile cnn from vit perspective. arXiv 2023, arXiv:2307.09283. [Google Scholar]
  37. Howard, A.; Sandler, M.; Chu, G.; Chen, L.C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. Searching for mobilenetv3. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1314–1324. [Google Scholar] [CrossRef]
  38. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  39. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar] [CrossRef]
  40. Ma, N.; Zhang, X.; Zheng, H.T.; Sun, J. Shufflenet v2: Practical guidelines for efficient cnn architecture design. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 116–131. [Google Scholar]
  41. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. Shufflenet: An extremely efficient convolutional neural network for mobile devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6848–6856. [Google Scholar] [CrossRef]
  42. Han, K.; Wang, Y.; Tian, Q.; Guo, J.; Xu, C.; Xu, C. Ghostnet: More features from cheap operations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1577–1586. [Google Scholar] [CrossRef]
  43. Sifre, L.; Mallat, S. Rigid-motion scattering for texture classification. arXiv 2014, arXiv:1403.1687. [Google Scholar]
  44. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25. Available online: https://proceedings.neurips.cc/paper_files/paper/2012/file/c399862d3b9d6b76c8436e924a68c45b-Paper.pdf (accessed on 17 January 2024). [CrossRef]
  45. Zhang, Q.; Jiang, Z.; Lu, Q.; Han, J.; Zeng, Z.; Gao, S.H.; Men, A. Split to be slim: An overlooked redundancy in vanilla convolution. arXiv 2020, arXiv:2006.12085. [Google Scholar]
  46. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30. Available online: https://proceedings.neurips.cc/paper_files/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf (accessed on 17 January 2024).
  47. Li, X.; Wang, W.; Hu, X.; Yang, J. Selective kernel networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 510–519. [Google Scholar]
  48. Yu, X.; Lyu, W.; Zhou, D.; Wang, C.; Xu, W. ES-Net: Efficient scale-aware network for tiny defect detection. IEEE Trans. Instrum. Meas. 2022, 71, 1–14. [Google Scholar] [CrossRef]
  49. Ljosa, V.; Sokolnicki, K.L.; Carpenter, A.E. Annotated high-throughput microscopy image sets for validation. Nat. Methods 2012, 9, 637. [Google Scholar] [CrossRef]
  50. He, T.; Zhang, Z.; Zhang, H.; Zhang, Z.; Xie, J.; Li, M. Bag of tricks for image classification with convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 558–567. [Google Scholar] [CrossRef]
  51. Wang, J.; Xu, C.; Yang, W.; Yu, L. A normalized Gaussian Wasserstein distance for tiny object detection. arXiv 2021, arXiv:2110.13389. [Google Scholar]
  52. Tan, M.; Le, Q. Efficientnetv2: Smaller models and faster training. In Proceedings of the International conference on machine learning. PMLR, Virtual, 18–24 July 2021; pp. 10096–10106. [Google Scholar]
  53. He, K.; Zhang, X.; Ren, S.; Sun, J. Identity mappings in deep residual networks. In Proceedings of the Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part IV 14. Springer: Cham, Switzerland, 2016; pp. 630–645. [Google Scholar]
  54. Lv, W.; Xu, S.; Zhao, Y.; Wang, G.; Wei, J.; Cui, C.; Du, Y.; Dang, Q.; Liu, Y. Detrs beat yolos on real-time object detection. arXiv 2023, arXiv:2304.08069. [Google Scholar]
Figure 1. Network structure of MAS-Net. Following the general paradigm, the overall architecture of MAS-Net consists of three components: backbone, neck, and head.
Figure 1. Network structure of MAS-Net. Following the general paradigm, the overall architecture of MAS-Net consists of three components: backbone, neck, and head.
Information 15 00166 g001
Figure 2. Backbone of MAS-Net. The “SPPF” in the diagram is from the YOLOv5 network. The outputs of layers 5, 7, and 10 are represented as M3, M4, and M5.
Figure 2. Backbone of MAS-Net. The “SPPF” in the diagram is from the YOLOv5 network. The outputs of layers 5, 7, and 10 are represented as M3, M4, and M5.
Information 15 00166 g002
Figure 3. RSConv structure diagram, where “Split” denotes channel splitting operation, “GAP” represents global average pooling, and “FC” stands for fully connected layer.
Figure 3. RSConv structure diagram, where “Split” denotes channel splitting operation, “GAP” represents global average pooling, and “FC” stands for fully connected layer.
Information 15 00166 g003
Figure 4. SPCot structure diagram. SPCot enhances detection performance by combining local and global attention. α , β , φ represent 1 × 1 convolutions, while RSConv is used to extract features and reduce the number of parameters. SPCot builds the backbone as a feature extraction module, which you can find in Figure 2.
Figure 4. SPCot structure diagram. SPCot enhances detection performance by combining local and global attention. α , β , φ represent 1 × 1 convolutions, while RSConv is used to extract features and reduce the number of parameters. SPCot builds the backbone as a feature extraction module, which you can find in Figure 2.
Information 15 00166 g004
Figure 5. MRFH structure diagram. the detection head MRFH utilizes a 5 × 5 convolution achieved by a 3 × 3 convolution with a dilation rate of 2.
Figure 5. MRFH structure diagram. the detection head MRFH utilizes a 5 × 5 convolution achieved by a 3 × 3 convolution with a dilation rate of 2.
Information 15 00166 g005
Figure 6. Grad-CAM of MAS-Net. We output the heat map of the shallow layers of the network, where the red part represents the network’s region of interest. The deeper the color, the more interested the network is in that area..
Figure 6. Grad-CAM of MAS-Net. We output the heat map of the shallow layers of the network, where the red part represents the network’s region of interest. The deeper the color, the more interested the network is in that area..
Information 15 00166 g006
Figure 7. The P R curve graph. It shows the P R curve of mainstream one-stage models.
Figure 7. The P R curve graph. It shows the P R curve of mainstream one-stage models.
Information 15 00166 g007
Figure 8. The m A P variation curve. The figure shows the m A P curve of the main SOTA models on our dataset.
Figure 8. The m A P variation curve. The figure shows the m A P curve of the main SOTA models on our dataset.
Information 15 00166 g008
Figure 9. Detection results example.The figure shows a comparison of the detection performance between our model and the baseline model. The results include predicted bounding boxes, class names (ri: ring, tr: trophozoite, sc: schizont, ga: gametocyte), and confidence scores. MAS-Net can detect some hard-to-find targets, and it also achieves higher confidence scores.
Figure 9. Detection results example.The figure shows a comparison of the detection performance between our model and the baseline model. The results include predicted bounding boxes, class names (ri: ring, tr: trophozoite, sc: schizont, ga: gametocyte), and confidence scores. MAS-Net can detect some hard-to-find targets, and it also achieves higher confidence scores.
Information 15 00166 g009
Figure 10. Robustness test results visualization. Gaussian noise, salt-and-pepper noise, Rayleigh noise, uniform noise, and motion blur were added and tested accordingly. The test results consist of predicted bounding boxes, class names (ri: ring, tr: trophozoite, sc: schizont, ga: gametocyte), and confidence scores.
Figure 10. Robustness test results visualization. Gaussian noise, salt-and-pepper noise, Rayleigh noise, uniform noise, and motion blur were added and tested accordingly. The test results consist of predicted bounding boxes, class names (ri: ring, tr: trophozoite, sc: schizont, ga: gametocyte), and confidence scores.
Information 15 00166 g010
Table 1. The backbone network for feature extraction uses the following parameters.
Table 1. The backbone network for feature extraction uses the following parameters.
Output ChannelConvolution SizeNumber of Repetitions
64Conv(kernel: 3 × 3 ,stride 2)1
128Conv(kernel: 3 × 3 ,stride 2)1
128SPCot3
256Conv(kernel: 3 × 3 ,stride 2)1
256SPCot6
512Conv(kernel: 3 × 3 ,stride 2)1
512SPCot9
1024Conv(kernel: 3 × 3 ,stride 2)1
1024SPCot3
1024SPPF1
Table 2. Comparison of MAS-Net with other mainstream one-stage detection algorithms.
Table 2. Comparison of MAS-Net with other mainstream one-stage detection algorithms.
Algorithm Pre % Rec % mAP % Para GFLOPs F 1 FPS
YOLOv563.467.869.946.3 m108.365.546
YOLOv762.966.967.839.2 m105.264.850
YOLOv868.865.371.743.6 m164.866.062
MAS-Net73.972.275.948.7 m126.273.042
MAS-Net-Tiny62.669.870.69.6 m36.866.0125
Table 3. Parameter comparison before and after PAGCP.
Table 3. Parameter comparison before and after PAGCP.
Algorithm mAP % Para % GFLOPs % FPS mAP Para ( % ) GFLOPs ( % )
MAS-Net75.948.7 m126.242---
MAS-Net-Tiny70.69.6 m36.81255.380.270.7
The downward arrow (↓) indicates a decrease.
Table 4. Comparison of MAS-Net with other mainstream one-stage detection algorithms.
Table 4. Comparison of MAS-Net with other mainstream one-stage detection algorithms.
Algorithm Pre % Rec % mAP % Para GFLOPs F 1 FPS
Efficientnet [19]62.374.571.849.2 m80.867.910
EfficientnetV2 [52]64.765.172.274.1 m134.064.913
ConvnextV2 [29]61.360.364.350.6 m113.259.829
MobilenetV3 [37]57.361.364.623.2 m43.159.228
Fasternet [32]63.865.768.123.7 m44.459.228
ResnetV2 [53]61.565.064.851.2 m113.063.239
Edgenext [33]58.864.863.740.1 m87.661.730
EfficientViT [35]68.763.368.533.8 m68.065.917
RT-DETR [54]67.965.461.932.0 m87.266.6100
MAS-Net73.972.275.948.7 m126.273.042
MAS-Net-Tiny62.669.870.69.6 m36.866.0125
Table 5. Ablation experiments. Baseline network is YOLOv5, and all experiments are conducted using mosaic data augmentation method.
Table 5. Ablation experiments. Baseline network is YOLOv5, and all experiments are conducted using mosaic data augmentation method.
MosaicSPCotNetMRFHPAGCPmAP%GFLOPsPara (Million)
---69.0108.347.9
--75.5116.246.6
-75.9126.248.7
*-75.3156.690.6
70.636.89.6
The asterisk (*) indicates the use of MRFH detection head in all detection layers. The checkmark (√) represents adopting the structure, while the dash (-) represents not adopting the structure. Using MRFHdetection heads in deep neural networks comes at a high cost, so MAS-Net only employs MRFH detection heads in shallow neural networks.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xiong, Z.; Wu, J. Multi-Level Attention Split Network: A Novel Malaria Cell Detection Algorithm. Information 2024, 15, 166. https://doi.org/10.3390/info15030166

AMA Style

Xiong Z, Wu J. Multi-Level Attention Split Network: A Novel Malaria Cell Detection Algorithm. Information. 2024; 15(3):166. https://doi.org/10.3390/info15030166

Chicago/Turabian Style

Xiong, Zhao, and Jiang Wu. 2024. "Multi-Level Attention Split Network: A Novel Malaria Cell Detection Algorithm" Information 15, no. 3: 166. https://doi.org/10.3390/info15030166

APA Style

Xiong, Z., & Wu, J. (2024). Multi-Level Attention Split Network: A Novel Malaria Cell Detection Algorithm. Information, 15(3), 166. https://doi.org/10.3390/info15030166

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop