Next Article in Journal
Sentiment Analysis of Comment Data Based on BERT-ETextCNN-ELSTM
Previous Article in Journal
Quality of Life Prediction on Walking Scenes Using Deep Neural Networks and Performance Improvement Using Knowledge Distillation
Previous Article in Special Issue
A Novel Coherent Integration Algorithm for Maneuvering Target Detection Based on Symmetric Instantaneous Autocorrelation Function
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Domain Adaptive Few-Shot Learning for ISAR Aircraft Recognition with Transferred Attention and Weighting Importance

School of Electrical Engineering, Henan University of Technology, Zhengzhou 450001, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(13), 2909; https://doi.org/10.3390/electronics12132909
Submission received: 1 June 2023 / Revised: 28 June 2023 / Accepted: 29 June 2023 / Published: 2 July 2023

Abstract

:
With the enhancement of air-based and space-based perception capabilities, space-aeronautics incorporation and integration is growing in importance. Full domain awareness is crucial for integrated perception systems, in which domain adaptation is one of the key problems in improving the performance of cross-domain perception. Deep learning is currently an advanced technique for complex inverse synthetic aperture radar (ISAR) object recognition. However, the training procedure needs many annotated samples, which is insufficient for certain targets, such as aircraft. Few-shot learning provides a new approach to solving the above problem by transferring useful knowledge from other domains, such as optical images from satellites. Nevertheless, it fails to fully consider the domain shift between the source and target domains, generally neglecting the transferability of training samples in the learning process. Consequently, it produces suboptimal recognition accuracy. To address the composite problems mentioned above, we propose a domain adaptive few-shot learning method from satellites to an ISAR called S2I-DAFSL for aircraft recognition tasks. Furthermore, unlike conventional domain adaptation methods that directly align the distributions, the attention transferred importance-weighting network (ATIN) is proposed to improve the transferability in the domain adaptation procedure. Compared with state-of-the-art methods, it shows that the proposed method achieves better performance, increasing the accuracy and effectiveness of classification, which is more suitable for cross-domain few-shot ISAR aircraft recognition tasks.

1. Introduction

Aircraft recognition is one of the important tasks in ISAR image identification [1,2,3,4,5,6]. In recent years, deep convolution neural networks have achieved better performance in ISAR image classification [7,8,9,10,11,12,13,14], which can be applied to aircraft recognition tasks. However, the lack of labeled samples in the training process often results in over-fitting under the condition of small training sets. Recently, a fast space target ISAR image recognition based on local and global structural feature fusion has been proposed [15]. Further, transfer learning also provides a solution by taking the deep CNN model as a feature extractor or for fine-tuning the weight in the new task to address the problem that the samples in the target classification task are too small to effectively train the deep CNN model [16]. Additionally, since humans can quickly learn transferable knowledge based on their experiences and a few observations [17,18,19], few-shot learning (FSL) approaches are proposed to deal with the problems mentioned above. Including meta-learning, also known as learning to learn, a meta-learner-based stacking network is proposed to realize the classification of space targets by ISAR images with small samples [20]. FSL methods often assume that the samples from the source domain and target domain are from the same distribution without fully considering the domain discrepancy. Domain adaptation (DA) can reduce the domain shift by aligning the distributions between the source and target domains [21,22,23,24,25,26]. It is generally assumed that there are enough training samples when the categories from the source domain and target domain are consistent with each other, known as closed-set DA, which is different from the problem setting of FSL being able to identify new unseen categories.
The above scenarios are described in Figure 1.
From the perspective of DA methods, there exists a categorization of DA settings known as closed-set, partial, and open-set. This categorization is primarily based on the relationship between label spaces. Closed-set DA occurs when both domains possess identical label space, and the primary challenge lies in bridging the domain gap. In scenarios where the label set relationship is often unknown along with a wide domain gap, partial DA techniques are considered appropriate when the label space of the source domain is expansive enough to encompass the label space of the target domain. Conversely, open-set DA methods are more suitable when the label space of the source domain is restricted within the label space of the target domain.
Open-set DA can be seen as a kind of open-set problem, but focusing on the distribution shift between source and target domains. In open-set problems, it is assumed that there are known classes and unknown classes. It often learns with known classes and tests in all classes. The objective is to thoroughly investigate the capabilities of the model in distinguishing between images of known and unknown classes, with the aim of assessing its potential for open-set recognition based on the principle of “Openness”. The complexity of recognition can be determined by the number of target categories, as well as the training and testing categories utilized [27]. For example, an innovative approach based on proportional similarity is proposed in order to improve the accuracy and robustness of open-set recognition (OSR) in synthetic aperture radar (SAR) images [28].
As mentioned above, the open-set DA methodology offers a more practical approach to DA as it can incorporate novel classes that may emerge in the target domain. This approach involves dealing with unfamiliar categories in the target domain that did not appear in the source domain. To effectively overcome the challenges of open-set DA, it is crucial to reduce the domain shift in identical categories while simultaneously enhancing the discrimination among different categories.
Similar to open-set DA, FSL aims to identify novel target categories by leveraging samples from their source domain while also utilizing limited labeled samples from the target categories. In open-set DA, there are private categories within the target domain, in addition to the common categories between source and target domains. Open-set DA often assumes that the label space of the source domain is included within the label space of the target domain. Nevertheless, FSL often focuses on identifying novel categories when there is no intersection between source classes and target classes.
The different methodologies are further illustrated and presented in Figure 2.
As shown in Figure 2, FSL transfers knowledge from the source domain to the target domain: (1) The data distributions of the source domain are consistent with the target domain; (2) The intersection of the categories from the source and target domain is absent; and (3) The labeled samples for each category in the target domain are few.
Closed-set DA also involves transferring knowledge across domains: (1) There is a domain gap between the source and target domains; (2) The categories across domains are the same; and (3) The available samples for each category in the target domain are sufficient.
Open-set DA is similar to closed-set DA except that there are private categories in the target domain, and the source categories are included in the target categories.
DAFSL addresses FSL and DA simultaneously: (1) There is also a domain gap across domains; (2) The intersection of the categories from the source and target domain is absent; and (3) The labeled samples for each category in the target domain is few.
In general, due to the limited availability of labeled data, the field of FSL has recently garnered increased attention. The majority of visual recognition models rely on deep CNN, which necessitates the collection and annotation of sufficient samples per class to enable effective training. However, this requirement can be impractical or unfeasible for infrequent categories. FSL is often treated as a transfer learning issue, where knowledge is transferred from the source to target classes. The primary focus has been on developing a classifier with minimal samples. A fundamental challenge has been neglected, as the target classes often suffer from insufficient representation due to few training samples and may belong to a different domain compared to the source classes. Thus, it is critical to accommodate both novel categories and domain discrepancies using minimal samples from the target categories.
In other words, while the classification of labeled categories from some datasets can be easily accessed, the identification of novel categories presents a twofold challenge. Firstly, imaging modalities utilized in the target domain may differ significantly from those employed in the source training data, resulting in domain discrepancy. Secondly, new categories may not be adequately represented in the training data, causing a serious category gap.
Directly combining the FSL and DA is a straightforward approach but not appropriate since the distribution alignment in DA is harmful to the distinctiveness of new unseen categories in the target domain. Therefore, aligning the distributions of the source and target domains while maintaining certain independence is fundamentally important.
To address the above problems, in this paper, the attention transferred importance-weighting network (ATIN) is proposed, and FSL is implemented simultaneously in both the source and target categories, allowing it to build a discriminative embedding model as well as explore the transferable information. The major contributions of our work are summarized as follows:
(1)
A cross-domain FSL method from satellites to ISAR called S2I-DAFSL is proposed to implement the aircraft recognition task, addressing the DA and FSL simultaneously.
(2)
We propose the ATIN to further improve the transferability and effectiveness in the DA procedure, in which the attention-transferred module focuses on more informative regions, and the importance-weighting module helps choose more appropriate training samples.
(3)
Extensive experimental results demonstrate that the proposed method can improve the accuracy of ISAR aircraft image classification with more efficient implementation.
The remainder of this paper is organized as follows. Section 2 introduces the overall architecture of the proposed method. Section 3 provides the details about each component of the proposed method. In Section 4, we analyze the experiments, and Section 5 is the paper’s conclusion.

2. Overall Architecture

In the conventional problem settings of DA, the data distributions of the source and target domains are different. Formally, the target domain is indicated by Dt, and the source domain is denoted by Ds. ns and nt are the numbers of samples in the source and target domains, respectively. Due to the domain changes, PsPt in the DA settings. Therefore, deep DA approaches are utilized to minimize the distribution discrepancy across domains. As described above, the problem settings of domain adaptive few-shot learning are more challenging than DA or FSL. The proposed method leverages a comprehensive dataset Ds that includes samples from categories Cs in the source domain and few-shot samples Dd from classes Cd in the target domain. In addition, we have access to a separate test set T that represents a different target domain with classes Ct, ensuring the absence of overlaps between these classes. The primary objective of our approach is to teach an effectively generalized model to T by utilizing Ds and Dd. Obviously, the dissimilarities between the source and target domains mean that Ps(x) of source classes Cs differs from that of the target classes CtCd. The overall architecture of the proposed method is shown in Figure 3. The prototypical network is utilized to implement FSL while the ATIN performs the DA procedure, containing the attention transferred module and importance-weighting module.

3. The ATIN Approach

The ATIN is thoroughly proposed in this section. The majority of adversarial-based DA techniques have three parts: a feature extractor Gf, a classifier Gy, and a domain discriminator Gd [29]. The objective functions are shown as
min G f , G y L c l s ( G f , G y ) = 1 n s i = 1 n s L ( G y ( G f ( x i s ) ) , y i s )
min G f max G d L a d v ( G f , G d ) = 1 n s i = 1 n s log ( G d ( G f ( x i s ) ) ) 1 n t j = 1 n t log ( 1 G d ( G f ( x j t ) ) )
where Lcls is s the cross-entropy loss for the classification task of the source-domain; Ladv is the domain adversarial loss.
However, the current adversarial DA techniques may not take into account complicated multi-modal structures in domain distributions [30]. If inappropriate samples and untransferable features are forced to match, this may result in a negative transfer. Therefore, the ATIN is proposed to address the above problems, which include the attention transferred module and importance-weighting module.

3.1. The Attention Transferred Module

Learning transferable representations across domains is crucial in adversarial DA. However, it is difficult to learn features that are actually both transferable and discriminative. To overcome this problem, a self-attention mechanism for conditional adversarial learning is added. We initially input features from Gf into a convolution layer before distilling it into two separate abstract feature maps: M1 and M2, with the dimensions C × H × W . Next, we rearrange them into C × N and conduct a matrix multiplication operation to transpose them. The self-attention map MN×N may be calculated using the softmax operation.
α j i = exp ( M 1 i M 2 j ) i = 1 N exp ( M 1 i M 2 j )
where α j i calculates the influence of location i on position j. The similarities between the two places and their correlation increase as they become closer. Further, a convolution layer is also used to obtain M3 with the dimensions C × H × W and then reshaped into C × N . Then, it is reshaped to C × H × W using the feature map M3 and MN×N.
Next, the weighted feature representations are obtained, which are expressed as
M j = μ i = 1 N α j i M 3 i + M j
The parameter μ with an initial value of 0 is progressively increased in the training process [31]. Singular value decomposition (SVD) is employed to investigate the spectrum characteristics of the attention feature M in batches, further enhancing the transferability and discriminability [32]. Moreover, batch spectral penalization (BSP) is used to improve the discriminability of the feature representation [33]. Thus, using the greatest K singular values as a regularization term, the final calculation for self-attention transfer loss is obtained by
L B S P ( M ) = i = 1 K ( ρ s , i 2 + ρ t , i 2 )
where ρ s , i and ρ t , i are the i-th largest singular values in matrix Σs and Σt.

3.2. The Importance-Weighting Module

In order to determine the significance of each source sample to the target domain, we utilize the output of Gd and the classifier predictions g. In this approach, the proposed network may lessen the impact of outliers in the source domain.
Let h = (f, g) represent the joint variable of the classifier prediction g and feature representation f. The conditioning strategy can be expressed as
T ( h ) = { T ( f , g ) , d f × d g d f e a t T ( f , g ) ,   otherwise  
where and stand for a multi-linear map and an explicit randomized multi-linear map. df and dg are the dimensions of f and g, and dfeat denotes the dimension of the output.
As seen in Equation (7), the domain adversarial loss may be expressed as follows:
min G f G y max G d L a d ν ( G f , G y , G d ) = 1 n s i = 1 n s log ( G d ( T ( h ) i s ) )                          1 n t j = 1 n t log ( 1 G d ( T ( h ) j t ) )
Equation (7) assigns all cases the same weight, which might be problematic since certain samples with ambiguous predictions could make the adversarial adaptation process worse. In most cases, the entropy criteria can be described as
H ( g ) = i = 1 C g i log g i
where gi is the probability of correct prediction of category i and may be used to quantify the uncertainty of classifier predictions. C represents the total classes in the source domain. The prediction is more definite when the H(g) is smaller. As a result, it is possible to condition the domain discriminator by weight w ( H ( g ) ) = e H ( g ) , and the domain adversarial loss is defined as follows:
min G f , G y max G d L a d v ( G f , G y , G d ) = 1 n s i = 1 n s w ( H ( g i s ) ) log ( G d ( T ( h ) i s ) )                        1 n t j = 1 n t w ( H ( g j t ) ) log ( 1 G d ( T ( h ) j t ) )
However, obviously, in the minimax game represented by Equation (9), w ( H ( g i s ) ) soon approaches 1 as training goes on [34]. As each source sample is, thus, equally significant, it will result in inappropriate adaptation.
Intuitively, the transferability of samples can be directly reflected in the output result di of the domain discrimination model. The more uncertain the sample is, the better the transferability is. Thus, the entropy of domain prediction di can be utilized to calculate the transferability of the source domain samples, which is expressed as w ^ = H ( d i ) . The former Gd is only used to generate the transferability w ^ of the source domain samples, which is described as
min G d L d ( G d ) = 1 n s i = 1 n s w ( H ( g i s ) ) log ( G d ( T ( h ) i s ) )            1 n t j = 1 n t w ( H ( g j t ) ) log ( 1 G d ( T ( h ) j t ) )
Meanwhile, another adversarial domain discriminator G ^ d is utilized, which is shown as
min G f , G y max G ^ d L a d v ( G f , G y , G ^ d ) = 1 n s i = 1 n s w ^ w ( H ( g i s ) ) log ( G ^ d ( T ( h ) i s ) )            1 n t j = 1 n t w ( H ( g j t ) ) log ( 1 G ^ d ( T ( h ) j t ) )
Therefore, the objective of the importance-weighting module consists of Equations (10) and (11).

4. The FSL Procedure

During the episode training phase, we aim to emulate the few-shot testing process by generating small episodic training sets from both the source dataset (Ds) and the few-shot dataset (Dd). To achieve this, we randomly choose a subset of N classes from Ds and divide them into support and query sets, creating training episodes. Nevertheless, owing to the restricted number of samples in Dd, we augment the data with standard techniques such as horizontal flips and random crops. To develop our FSL framework, we have leveraged the widely acclaimed prototypical network. This choice is motivated by its remarkable performance and simplicity, making it an ideal cornerstone for our study. The network operates by encoding a prototype and subsequently utilizing a nearest neighbor classifier to classify samples in the query set. It projects samples from visual space into feature space where samples belonging to the same class are closely clustered, while those from different classes are distinctly separated.
During the training phase, the network parameters undergo an updating process utilizing a nonparametric softmax technique that draws upon the outcomes of a distance metric. This approach effectively determines the distribution of class membership for a query sample x residing within the query set Q.
P ( y = k x Q ) = exp ( d i s ( f φ ( x ) , c k ) ) k = 1 C exp ( d i s ( f φ ( x ) , c k ) )
where dis(,) represents the Euclidean distance function. Next, the FSL loss is obtained using support set S and query set Q, where Lfss for the source domain and Lfst for the target domain can be described as
L f s s = E S s , Q s [ ( x , y ) Q s log p φ ( y = k x ) ]
L f s t = E S t , Q t [ ( x , y ) Q t log p φ ( y = k x ) ]
Therefore, the integrated objective functions of the proposed method are Equations (5), (10), (11), (13) and (14).

5. Experiments

As shown in Figure 3, there are 15 classes of optical aircraft in the source domain [35,36]. To increase the diversity of the tasks and verify the robustness of the proposed method, 10 classes of ISAR images are selected in the target domain, in which 5 classes are synthesized and others are obtained from an anechoic chamber [5]. Next, 5-way 1-shot and 5-way 5-shot settings are utilized to evaluate the test set. The study reports the mean accuracy of 600 testing episodes, which were generated at random, coupled with their corresponding 95% confidence intervals.
ADDA [24], CDAN [22], RelationNet [19], MatchingNet [37], ProtoNet [18], and CDAN + ProtoNet (directly combining CDAN and ProtoNet) are selected as baselines. We utilize the ResNet50 model [38] as the backbone architecture. The backbone is meticulously pre-trained from scratch on the training datasets, and subsequently fine-tuned to address the challenging domain adaptation few-shot learning task. The aim of this study is to investigate the efficacy of a novel approach to address the challenges posed by FSL and DA in cross-domain ISAR aircraft recognition. The results are presented in Table 1, highlighting the following noteworthy findings.
Our proposed method outperforms FSL or DA methods, demonstrating its potential as a promising solution to address the composite problems. Furthermore, our comparative analysis with DA + FSL baselines emphasizes the need for more sophisticated models as the naive combination of strategies has limitations.
The proposed method addresses the challenges associated with bridging the gap between different domains by combining both FSL and DA effectively. To evaluate the robustness of the proposed method, we change the number of n and k in the n-way k-shot settings.
As shown in Figure 4, when the number of n decreases or the number of k increases, the accuracy of recognition is improved. When n is 5, as k increases, the recognition accuracy improves significantly. When k is 9, it does not matter if n equals 3 or 5, the recognition accuracy is more than 95%. As n equals 3, it reaches 99.1%. The above results prove the robustness and effectiveness of the proposed method.
In addition, we implemented an ATIN approach to enhance our full model. To evaluate the efficacy of each module, we conducted an ablation study by comparing the performance of our complete model against four simplified versions, i.e., FSL, ATIN, ATIN-I (using only the importance-weighting module), and ATIN-A (using only the attention transferred module) in 5-way 1-shot and 5-way 5-shot settings.
The results, as shown in Figure 5, demonstrate that the integration of each module improves the performance, proving the contribution of each module in resolving the ISAR image classification. Our study further emphasizes the crucial role of the ATIN, as demonstrated by the performance improvements achieved by ATIN over classical FSL. In conclusion, the proposed method presents a promising solution for addressing the challenges associated with DA in FSL for ISAR aircraft recognition.
Furthermore, to analyze the recognition performance of ISAR images that are synthesized or from an anechoic chamber, we only take synthesized images, those from an anechoic chamber and their mixing as a test set, respectively. As shown in Figure 6, Figure 7 and Figure 8, the confusion matrix is obtained under the 5-way 5-shot setting, and the per class indicators for all confusion matrices are presented in Table 2, Table 3 and Table 4.
As shown in Figure 6 and Figure 7, the recognition accuracy of synthesized samples is affected by the appearance of the aircraft. When the samples are only selected from the anechoic chamber, the confusion of each category is increased.
It also can be seen in Figure 8 that there is no confusion between synthesized samples and those from the anechoic chamber, proving the effectiveness of the proposed method.
Further, the t-SNE [39] is used to visualize the tested target features with corresponding category labels according to the above settings, where each category contains 31 samples in the 5-way 1-shot setting and 35 samples in the 5-way 5-shot setting.
As shown in Figure 9, the distributions of features from the anechoic chamber are different compared with those from the synthesized samples. The recognition accuracy under the 5-way 5-shot setting is higher than the 5-way 1-shot setting. Each category can be generally distinguished in each case. When the categories are mixed, the features of the synthesized samples and those from the anechoic chamber (viewed in green and yellow in Figure 9c) can also be well separated, proving the robustness of the proposed method, which is suitable for ISAR aircraft recognition.

6. Conclusions

We present a novel domain adaptive few-shot learning method from satellites to an ISAR called S2I-DAFSL for aircraft recognition. Rather than aligning distributions directly, our approach leverages an ATIN to further enhance transferability during the domain adaptation process. Compared with the state-of-the-art approaches, the proposed method achieves better classification accuracy and effectiveness, making it more suitable for ISAR aircraft recognition tasks. It is worth noting that although our work suggests a promising solution for ISAR aircraft recognition, there is still room for improvements. For example, the proposed method is evaluated on a specific task, and the availability of labeled data in the source domain is assumed, which could be further extended to other tasks and domains. In the future, we can also consider extending the framework to implement open-set recognition tasks as well as zero-shot learning under the unsupervised learning paradigm.

Author Contributions

Conceptualization, B.L., Y.Y. and Q.W.; methodology, B.L., Y.Y. and Q.W.; validation, B.L.; data curation, Q.W.; investigation, Y.Y.; resources, Q.W.; writing—original draft preparation, B.L. and Y.Y.; writing—review and editing, Q.W.; supervision, B.L. and Y.Y.; project administration, B.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Henan Provincial Science and Technology Project (grant numbers 202102210136, 222102220009, 222102220080, and 222103810083), the Henan Province Education Department Natural Science Project (grant number 20A413003), the R&D Project of Zhengzhou (grant number 22ZZRDZX06), and the Natural Science Project of Henan University of Technology (grant numbers 2019BS018 and 2019BS055).

Data Availability Statement

The data used in this study are collected according to references [5,35,36].

Acknowledgments

All authors would sincerely thank the reviewers and editors for their suggestions and opinions for improving this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Benedetto, F.; Riganti, F.; Laudani, A.; Albanese, G. Automatic aircraft target recognition by ISAR image processing based on neural classifier. Int. J. Adv. Comput. Sci. Appl. 2012, 3, 96–103. [Google Scholar] [CrossRef] [Green Version]
  2. Kondaveeti, H.K.; Vatsavayi, V.K. Abridged shape matrix representation for the recognition of aircraft targets from 2D ISAR imagery. Adv. Comput. Sci. Technol. 2017, 10, 1103–1122. [Google Scholar] [CrossRef]
  3. Vatsavayi, V.K.; Kondaveeti, H.K. Efficient ISAR image classification using MECSM representation. J. King Saud Univ. Comput. 2018, 30, 356–372. [Google Scholar] [CrossRef] [Green Version]
  4. Slavyanov, K.; Nikolov, L. An algorithm for ISAR image classification procedure. Industry 2017, 2, 76–79. [Google Scholar]
  5. Kondaveeti, H.K.; Vatsavayi, V.K. Robust ISAR image classification using Abridged Shape Matrices. In Proceedings of the 1st International Conference on Emerging Trends in Engineering, Technology and Science, Pudukkottai, India, 24–26 February 2016; pp. 1–6. [Google Scholar]
  6. Slavyanov, K.O. Neural network classification method for aircraft in ISAR images. In Proceedings of the 12th International Scientific and Practical Conference on Environment, Technology, Resources, Rezekne, Latvia, 20–22 June 2019; pp. 141–145. [Google Scholar]
  7. Xue, R.H.; Bai, X.R.; Zhou, F. SAISAR-Net: A robust sequential adjustment ISAR image classification network. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5214715. [Google Scholar] [CrossRef]
  8. Xue, B.; Yi, W.; Jing, F.; Wu, S. Complex ISAR target recognition using deep adaptive learning. Eng. Appl. Artif. Intell. 2021, 97, 104025. [Google Scholar] [CrossRef]
  9. Liu, C.; Wang, Z. Efficient complex ISAR object recognition using adaptive deep relation learning. IET Comput. Vis. 2020, 14, 185–191. [Google Scholar] [CrossRef]
  10. Lu, W.; Zhang, Y.S.; Yin, C.B.; Lin, C.Y.; Xu, C.; Zhang, X. A deformation robust ISAR image satellite target recognition method based on PT-CCNN. IEEE Access 2021, 9, 23432–23453. [Google Scholar] [CrossRef]
  11. Xue, B.; Tong, N.N. Real-world ISAR object recognition using deep multimodal relation learning. IEEE Trans. Cybern. 2020, 50, 4256–4267. [Google Scholar] [CrossRef]
  12. Xue, B.; Tong, N.N.; Xu, X. DIOD: Fast, semi-supervised deep ISAR object detection. IEEE Sens. J. 2019, 19, 1073–1081. [Google Scholar] [CrossRef]
  13. Xue, B.; Tong, N.N. Real-world ISAR object recognition and relation discovery using deep relation graph learning. IEEE Access 2019, 7, 43906–43914. [Google Scholar] [CrossRef]
  14. Bai, X.; Zhou, X.; Zhang, F.; Wang, L.; Xue, R.H.; Zhou, F. Robust Pol-ISAR target recognition based on ST-MC-DCNN. IEEE Trans. Geosci. Remote Sens. 2019, 57, 9912–9927. [Google Scholar] [CrossRef]
  15. Yang, H.; Zhang, Y.S.; Ding, W.Z. A fast recognition method for space targets in ISAR images based on local and global structural fusion features with lower dimensions. Int. J. Aerosp. Eng. 2020, 2020, 3412582. [Google Scholar] [CrossRef]
  16. Yang, H.; Zhang, Y.; Ding, W. Multiple heterogeneous P-DCNNs ensemble with stacking algorithm: A novel recognition method of space target ISAR images under the condition of small sample set. IEEE Access 2020, 8, 75543–75570. [Google Scholar] [CrossRef]
  17. Choi, J.; Krishnamurthy, J.; Kembhavi, A.; Farhadi, A. Structured set matching networks for one-shot part labeling. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake, UT, USA, 18–22 June 2018; pp. 3627–3636. [Google Scholar]
  18. Snell, J.; Swersky, K.; Zemel, R. Prototypical networks for few-shot learning. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 4080–4090. [Google Scholar]
  19. Sung, F.; Yang, Y.; Zhang, L.; Xiang, T.; Torr, H.S.; Hospedales, M. Learning to compare: Relation network for few-shot learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1199–1208. [Google Scholar]
  20. Zhang, Y.; Yuan, H.X.; Li, H.B.; Chen, J.Y.; Niu, M.Q. Meta-learner-based stacking network on space target recognition for ISAR images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 12132–12148. [Google Scholar] [CrossRef]
  21. Long, M.; Cao, Y.; Wang, J.; Jordan, M.I. Learning transferable features with deep adaptation networks. In Proceedings of the 32nd International Conference on Machine Learning, Lile, France, 6–11 July 2015; pp. 97–105. [Google Scholar]
  22. Long, M.; Cao, Z.; Wang, J.; Jordan, M.I. Conditional adversarial domain adaptation. In Proceedings of the 32nd Conference on Neural Information Processing Systems, Montreal, QC, Canada, 2–8 December 2018; pp. 1640–1650. [Google Scholar]
  23. Long, M.; Zhu, H.; Wang, J.; Jordan, M.I. Unsupervised domain adaptation with residual transfer networks. In Proceedings of the 30th Annual Conference on Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 136–144. [Google Scholar]
  24. Tzeng, E.; Hoffman, J.; Saenko, K.; Darrell, T. Adversarial discriminative domain adaptation. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2962–2971. [Google Scholar]
  25. Tzeng, E.; Hoffman, J.; Zhang, N.; Saenko, K.; Darrell, T. Deep domain confusion: Maximizing for domain invariance. arXiv 2014, arXiv:1412.3474. [Google Scholar]
  26. Venkateswara, H.; Eusebio, J.; Chakraborty, S.; Panchanathan, S. Deep hashing network for unsupervised domain adaptation. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5018–5027. [Google Scholar]
  27. Scheirer, W.J.; Rocha, A.d.; Sapkota, A.; Boult, T.E. Toward open set recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1757–1772. [Google Scholar] [CrossRef]
  28. Giusti, E.; Ghio, S.; Oveis, A.H.; Martorella, M. Proportional Similarity-Based Openmax Classifier for Open Set Recognition in SAR Images. Remote Sens. 2022, 14, 4665. [Google Scholar] [CrossRef]
  29. Ganin, Y.; Lempitsky, V. Unsupervised domain adaptation by backpropagation. arXiv 2014, arXiv:1409.7495. [Google Scholar]
  30. Zhao, A.; Ding, M.; Lu, Z.; Xiang, T.; Niu, Y.L.; Guan, J.C.; Wen, J.R. Domain-adaptive few-shot learning. In Proceedings of the 2021 IEEE Winter Conference on Applications of Computer Vision, Virtual, 5–9 January 2021; pp. 1389–1398. [Google Scholar]
  31. Zhang, H.; Goodfellow, I.; Metaxas, D.N.; Odena, A. Self-attention generative adversarial networks. arXiv 2018, arXiv:1805.08318. [Google Scholar]
  32. Chen, X.; Wang, S.; Long, M.; Wang, J. Transferability vs. discriminability: Batch spectral penalization for adversarial domain adaptation. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 1859–1868. [Google Scholar]
  33. Zhang, C.C.; Zhao, Q.J.; Wang, Y. Transferable attention networks for adversarial domain adaptation. Inf. Sci. 2020, 539, 422–433. [Google Scholar] [CrossRef]
  34. Liu, P.; Xiao, T.; Fan, C.N.; Zhao, W.; Tang, X.L.; Liu, H.W. Importance weighted conditional adversarial network for unsupervised domain adaptation. Expert Syst. Appl. 2020, 155, 113404. [Google Scholar] [CrossRef]
  35. MTARSI 2. Available online: https://doi.org/10.5281/zenodo.5044949 (accessed on 30 June 2021).
  36. MTARSI. Available online: https://doi.org/10.5281/zenodo.2888016 (accessed on 18 May 2019).
  37. Vinyals, O.; Blundell, C.; Lillicrap, T.; Kavukcuoglu, K.; Wierstra, D. Matching networks for one shot learning. In Proceedings of the 30th Annual Conference on Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 3637–3645. [Google Scholar]
  38. He, K.M.; Zhang, X.; Ren, S.Q.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 29th IEEE conference on computer vision and pattern recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  39. Van der Maaten, L.; Hinton, G. Visualizing data using t-sne. J. Mach. Learn. Res. 2008, 9, 2579–2625. [Google Scholar]
Figure 1. Different scenarios of visual recognition problems.
Figure 1. Different scenarios of visual recognition problems.
Electronics 12 02909 g001
Figure 2. Illustration of different methodologies.
Figure 2. Illustration of different methodologies.
Electronics 12 02909 g002
Figure 3. The overall architecture of the proposed method.
Figure 3. The overall architecture of the proposed method.
Electronics 12 02909 g003
Figure 4. Recognition accuracy under varied n-way k-shot settings.
Figure 4. Recognition accuracy under varied n-way k-shot settings.
Electronics 12 02909 g004
Figure 5. Recognition accuracy using each module.
Figure 5. Recognition accuracy using each module.
Electronics 12 02909 g005
Figure 6. Confusion matrix of synthesized images.
Figure 6. Confusion matrix of synthesized images.
Electronics 12 02909 g006
Figure 7. Confusion matrix of images obtained from anechoic chamber.
Figure 7. Confusion matrix of images obtained from anechoic chamber.
Electronics 12 02909 g007
Figure 8. Confusion matrix of images in mixed categories.
Figure 8. Confusion matrix of images in mixed categories.
Electronics 12 02909 g008
Figure 9. Feature visualization in different combinations of categories in 5-way 1-shot and 5-way 5-shot settings (best viewed in color): (a) Features of synthesized samples; (b) Features of samples from anechoic chamber; (c) Features of samples from mixed categories.
Figure 9. Feature visualization in different combinations of categories in 5-way 1-shot and 5-way 5-shot settings (best viewed in color): (a) Features of synthesized samples; (b) Features of samples from anechoic chamber; (c) Features of samples from mixed categories.
Electronics 12 02909 g009aElectronics 12 02909 g009b
Table 1. Comparative accuracies with 95% confidence intervals under the 5-way 1-shot and 5-way 5-shot settings.
Table 1. Comparative accuracies with 95% confidence intervals under the 5-way 1-shot and 5-way 5-shot settings.
Model5-Way 1-Shot5-Way 5-Shot
ADDA [24]73.25 ± 0.5583.98 ± 0.39
CDAN [22]73.72 ± 0.4384.11 ± 0.51
RelationNet [19]72.58 ± 0.4682.47 ± 0.28
MatchingNet [37]73.06 ± 0.5383.62 ± 0.40
ProtoNet [18]71.79 ± 0.2681.88 ± 0.31
CDAN + ProtoNet74.02 ± 0.4184.35 ± 0.29
S2I-DAFSL 75.97 ± 0.3786.93 ± 0.26
Table 2. Per class indicators for confusion matrix of synthesized images.
Table 2. Per class indicators for confusion matrix of synthesized images.
ClassPrecisionRecallF1-Score
10.800.700.75
20.910.670.77
30.711.000.83
41.001.001.00
51.001.001.00
Table 3. Per class indicators for confusion matrix of images obtained from anechoic chamber.
Table 3. Per class indicators for confusion matrix of images obtained from anechoic chamber.
ClassPrecisionRecallF1-Score
10.950.600.74
20.700.470.56
30.530.800.64
40.780.970.86
50.930.900.91
Table 4. Per class indicators for confusion matrix of images in mixed categories.
Table 4. Per class indicators for confusion matrix of images in mixed categories.
ClassPrecisionRecallF1-Score
11.001.001.00
20.800.800.80
30.831.000.91
40.790.730.76
50.750.810.78
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, B.; Yao, Y.; Wang, Q. Domain Adaptive Few-Shot Learning for ISAR Aircraft Recognition with Transferred Attention and Weighting Importance. Electronics 2023, 12, 2909. https://doi.org/10.3390/electronics12132909

AMA Style

Li B, Yao Y, Wang Q. Domain Adaptive Few-Shot Learning for ISAR Aircraft Recognition with Transferred Attention and Weighting Importance. Electronics. 2023; 12(13):2909. https://doi.org/10.3390/electronics12132909

Chicago/Turabian Style

Li, Binquan, Yuan Yao, and Qiao Wang. 2023. "Domain Adaptive Few-Shot Learning for ISAR Aircraft Recognition with Transferred Attention and Weighting Importance" Electronics 12, no. 13: 2909. https://doi.org/10.3390/electronics12132909

APA Style

Li, B., Yao, Y., & Wang, Q. (2023). Domain Adaptive Few-Shot Learning for ISAR Aircraft Recognition with Transferred Attention and Weighting Importance. Electronics, 12(13), 2909. https://doi.org/10.3390/electronics12132909

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop