Few-Shot Image Classification Based on Swin Transformer + CSAM + EMD
Abstract
:1. Introduction
- (1)
- The Swin Transformer is employed for feature extraction, enabling the acquisition of both local and global details from the image are captured and perform CSAM attention mechanism feature weighting on the output feature map.
- (2)
- The EMD measurement module is employed to measure the distance. The main concept involves utilizing block level measurement and incorporating a cross reference weight mechanism to effectively mitigate the influence of significant variations within the same category and cluttered background.
- (3)
- Numerous experiments were conducted on three widely utilized benchmark datasets for FSIC, and the findings demonstrate the significant enhancement achieved by the proposed model. The SOTA classification accuracy for few-shot images has been achieved.
2. Related Work
3. Methodology
3.1. Problem Description
3.2. Feature Extraction Module Based on Swin Transformer Network
3.3. CSAM MODULE
3.4. Measurement Module
3.4.1. Problem Description
3.4.2. Application of EMD in FSIC
3.4.3. End-to-End Training
3.4.4. Weight Generation
3.4.5. How to Set K-Shot?
4. Experimental Section
4.1. Dataset Description
4.2. Experimental Environment
4.3. Implementation Details
4.4. Analysis of Ablation Experiment
4.5. Analysis of Experimental Results
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Tao, H. Smoke Recognition in Satellite Imagery via an Attention Pyramid Network With Bidirectional Multilevel Multigranularity Feature Aggregation and Gated Fusion. IEEE Internet Things J. 2024, 11, 14047–14057. [Google Scholar] [CrossRef]
- Tao, H.; Duan, Q. Learning Discriminative Feature Representation for Estimating Smoke Density of Smoky Vehicle Rear. IEEE Trans. Intell. Transp. Syst. 2022, 23, 23136–23147. [Google Scholar] [CrossRef]
- Liu, M.; Huang, X.; Mallya, A.; Karras, T.; Aila, T.; Lehtinen, J.; Kautz, J. Few-shot unsupervised image-to-image translation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 10551–10560. [Google Scholar]
- Zhang, H.; Zhang, J.; Koniusz, P. Few-shot learning via saliency-guided hallucination of samples. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2770–2779. [Google Scholar]
- Chen, Z.; Fu, Y.; Wang, Y.; Ma, L.; Liu, W.; Hebert, M. Image deformation meta-networks for one-shot learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 8680–8689. [Google Scholar]
- Wang, Y.; Girshick, R.; Hebert, M.; Hariharan, B. Low-shot learning from imaginary data. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7278–7286. [Google Scholar]
- Li, A.; Luo, T.; Lu, Z.; Xiang, T.; Wang, L. Large-scale few-shot learning: Knowledge transfer with class hierarchy. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7212–7220. [Google Scholar]
- Peng, Z.; Li, Z.; Zhang, J.; Li, Y.; Qi, G.; Tang, J. Few-shot image recognition with knowledge transfer. In Proceedings of the IEEE/CVF international Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 441–449. [Google Scholar]
- Mangla, P.; Singh, M.; Sinha, A.; Kumari, N.; Balasubramanian, V.; Krishnamurthy, B. Charting the right manifold: Manifold mixup for few-shot learning. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Snowmass Village, CO, USA, 1–5 March 2020; pp. 2218–2227. [Google Scholar]
- Luo, T.; Li, A.; Xiang, T.; Huang, W.; Wang, L. Few-shot learning with global class representations. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9715–9724. [Google Scholar]
- Zhang, C.; Cai, Y.; Lin, G.; Shen, C. Deepemd: Few-shot image classification with differentiable earth mover’s distance and structured classifiers. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
- Snell, J.; Swersky, K.; Zemel, R. Prototypical networks for few-shot learning. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017; pp. 4077–4087. [Google Scholar]
- Sung, F.; Yang, Y.; Zhang, L.; Xiang, T.; Torr, P.; Hospedales, T. Learning to compare: Relation network for few-shot learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1199–1208. [Google Scholar]
- Li, W.; Wang, L.; Xu, J.; Huo, J.; Gao, Y.; Luo, J. Revisiting local descriptor based image-to-class measure for few-shot learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7260–7268. [Google Scholar]
- Rizve, M.; Khan, S.; Khan, F.; Shah, M. Exploring complementary strengths of invariant and equivariant representations for few-shot learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 10836–10846. [Google Scholar]
- Wu, J.; Zhang, T.; Zhang, Y.; Wu, F. Task-aware part mining network for few-shot learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 8433–8442. [Google Scholar]
- Thrun, S.; Pratt, L. Learning to Learn; Springer Science & Business Media: New York, NY, USA, 2012; pp. 3–17. [Google Scholar]
- Ravi, S.; Larochelle, H. Optimization as a model for few-shot learning. In Proceedings of the International Conference on Learning Representations (ICLR), San Juan, Puerto Rico, 2–4 May 2016. [Google Scholar]
- Gori, M.; Monfardini, G.; Scarselli, F. A new model for learning in graph domains. In Proceedings of the IEEE International Joint Conference on Neural Networks, Montreal, QC, Canada, 31 July–4 August 2005; pp. 729–734. [Google Scholar]
- Kim, J.; Kim, T.; Kim, S.; Yoo, C. Edge-labeling graph neural network for few-shot learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 11–20. [Google Scholar]
- Chen, X.; Wang, G. Few-shot learning by integrating spatial and frequency representation. In Proceedings of the 18th Conference on Robots and Vision (CRV), Burnaby, BC, Canada, 26–28 May 2021; pp. 49–56. [Google Scholar]
- Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021. [Google Scholar]
- Barratt, S. On the differentiability of the solution to convex optimization problems. arXiv 2018, arXiv:1804.05098. [Google Scholar]
- Dontchev, A.; Tyrrell Rockafellar, R. Implicit functions and solution mappings. IEEE Control. Syst. Mag. 2011, 31, 74–77. [Google Scholar]
- Krantz, S.; Parks, H. The Implicit Function Theorem: History, Theory, and Applications; Springer Science & Business Media: New York, NY, USA, 2012; pp. 1–12. [Google Scholar]
- Vinyals, O.; Blundell, C.; Lillicrap, T.; Kavukcuoglu, K.; Wierstra, D. Matching networks for one shot learning. arXiv 2016, arXiv:1606.04080. [Google Scholar]
- Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Andrej, K.; Aditya, K.; Bernstein Alexander, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef]
- Bertinetto, L.; Henriques, J.; Torr, P.; Vedaldi, A. Meta-learning with differentiable closed-form solvers. In Proceedings of the International Conference on Learning Representations (ICLR), New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
- Krizhevsky, A. Learning Multiple Layers of Features from Tiny Images; University of Toronto: Toronto, ON, Canada, 2009. [Google Scholar]
- Wah, C.; Branson, S.; Welinder, P.; Pietro, P.; Belongie, S. The Caltech-ucsd Birds-200–2011 Dataset; California Institute of Technology: Pasadena, CA, USA, 2011. [Google Scholar]
- Ren, M.; Triantafillou, E.; Ravi, S.; Snell, J.; Swersky, K.; Tenenbaum, J.; Larochelle, H.; Zemel, R. Meta-learning for semi-supervised few-shot classification. In Proceedings of the International Conference on Learning Representations (ICLR), Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
- Oreshkin, B.; Rodriguez, P.; Lacoste, A. TADAM: Task dependent adaptive metric for improved few-shot learning. In Proceedings of the Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018, Montréal, QC, Canada, 3–8 December 2018. [Google Scholar]
- Hu, S.; Li, D.; Stühmer, J.; Kim, M.; Hospedales, T. Pushing the Limits of Simple Pipelines for Few-Shot Learning: External Data and Fine-Tuning Make a Difference. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
- Yang, L.; Li, L.; Zhang, Z.; Zhou, X.; Zhou, E.; Liu, Y. DPGN: Distribution Propagation Graph Network for Few-Shot Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 13387–13396. [Google Scholar]
- Zhang, B.; Li, X.; Ye, Y.; Huang, Z.; Zhang, L. Prototype Completion With Primitive Knowledge for Few-Shot Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 3754–3762. [Google Scholar]
- Chen, W.; Liu, Y.; Kira, Z.; Wang, Y.; Huang, J. A Closer Look at Few-shot Classification. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
- Wang, Y.; Chao, W.; Weinberger, K.; Maaten, L. SimpleShot: Revisiting Nearest-Neighbor Classification for Few-Shot Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
- Xu, W.; Xu, Y.; Wang, H.; Tu, Z. Attentional constellation nets for few-shot learning. In Proceedings of the International Conference on Learning Representations, Virtual, 3–7 May 2021. [Google Scholar]
- Ma, J.; Xie, H.; Han, G.; Chang, S.; Galstyan, A.; Wael, A. Partner-assisted learning for few-shot image classification. In Proceedings of the IEEE/CVF international conference on computer vision, Montreal, QC, Canada, 10–17 October 2021; pp. 10573–10582. [Google Scholar]
- Boudiaf, M.; Masud, Z.; Rony, J.; Dolz, J.; Piantanida, P.; Ayed, I. Information Maximization for Few-Shot Learning. In Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, Virtual, 6–12 December 2020. [Google Scholar]
- Ziko, I.; Dolz, J.; Granger, E.; Ayed, I. Laplacian Regularized Few-Shot Learning. In Proceedings of the 37th International Conference on Machine Learning, Virtual, 13–18 July 2020. [Google Scholar]
- Liu, J.; Song, L.; Qin, Y. Prototype rectification for few-shot learning. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020. [Google Scholar]
- Hu, Y.; Gripon, V.; Pateux, S. Leveraging the feature distribution in transfer-based few-shot learning. In Proceedings of the 30th International Conference on Artificial Neural Networks, Bratislava, Slovakia, 14–17 September 2021; pp. 487–499. [Google Scholar]
- Hu, Y.; Gripon, V.; Pateux, S. Squeezing backbone feature distributions to the max for efficient few-shot learning. arXiv 2021, arXiv:2110.09446. [Google Scholar] [CrossRef]
- Bateni, P.; Barber, J.; Meent, J.; Wood, F. Enhancing Few-Shot Image Classification With Unlabelled Examples. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2021; pp. 2796–2805. [Google Scholar]
- Singh, A.; Hadi, J. Transductive decoupled variational inference for few-shot classification. arXiv 2022, arXiv:2208.10559. [Google Scholar]
- Zhou, Z.; Qiu, X.; Xie, J.; Wu, J.; Zhang, C. Binocular mutual learning for improving few-shot classification. In Proceedings of the IEEE/CVF international conference on computer vision, Montreal, QC, Canada, 10–17 October 2021; pp. 8402–8411. [Google Scholar]
- Hu, Y.; Pateux, S.; Gripon, V. Adaptive dimension reduction and variational inference for transductive few-shot classification. arXiv 2022, arXiv:2209.08527. [Google Scholar]
- Finn, C.; Abbeel, P.; Levine, S. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 1126–1135. [Google Scholar]
- Kang, D.; Kwon, H.; Min, J.; Cho, M. Relational embedding for few-shot classification. In Proceedings of the IEEE/CVF international Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021. [Google Scholar]
- Shalam, D.; Korman, S. The self-optimal- transport feature transform. arXiv 2022, arXiv:2204.03065. [Google Scholar]
- Lee, D.; Chung, S. Unsupervised embedding adaptation via early-stage feature reconstruction for few-shot classification. In Proceedings of the 38th International Conference on Machine Learning, Virtual, 18–24 July 2021. [Google Scholar]
- Chen, D.; Chen, Y.; Li, Y.; Mao, F.; He, Y.; Xue, H. Self-supervised learning for few-shot image classification. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021. [Google Scholar]
- Ye, H.; Hu, H.; Zhan, D.; Sha, F. Few-shot learning via embedding adaptation with set-to-set functions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8808–8817. [Google Scholar]
- Schwartz, E.; Karlinsky, L.; Shtok, J.; Harary, S.; Marder, M.; Feris, R.; Kumar, A.; Giryes, R.; Bronstein, A. Delta-encoder: An effective sample synthesis method for few-shot object recognition. In Proceedings of the Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018, Montréal, QC, Canada, 3–8 December 2018. [Google Scholar]
- Liu, Z.; Hu, H.; Lin, Y.; Yao, Z.; Xie, Z.; Wei, Y.; Ning, J.; Cao, Y.; Zhang, Z.; Dong, L.; et al. Swin transformer v2: Scaling up capacity and resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
FSCbenchmark | Mini-Imagenet | Tiered-Imagenet | |||
---|---|---|---|---|---|
Methods | Backbone | Five-Way One-Shot | Five-Way Five-Shot | Five-Way One-Shot | Five-Way Five-Shot |
DPGN [34] | R12 | 67.77 ± 0.32% | 84.6 ± 0.43% | 72.45 ± 0.51% | 87.24 ± 0.39% |
S2M2R [9] | WRN | 64.93 ± 0.18% | 83.18 ± 0.11% | 73.71 ± 0.22% | 88.59 ± 0.14% |
ProtoCompletion [35] | R12 | 73.13 ± 0.85% | 82.06 ± 0.54% | 81.04 ± 0.89% | 87.42 ± 0.57% |
Baseline++ [36] | R18 | 51.87 ± 0.77% | 75.68 ± 0.63% | 51.87 ± 0.77% | 75.68 ± 0.63% |
SimpleShot [37] | WRN | 63.5 ± 0.20% | 80.10 ± 0.15% | 69.75 ± 0.2% | 85.31 ± 0.15% |
ConstellationNet [38] | R12 | 64.89 ± 0.23% | 79.95 ± 0.17% | - | - |
PAL [39] | R12 | 69.37 ± 0.64% | 84.40 ± 0.44% | 72.25 ± 0.72% | 86.95 ± 0.47% |
TIM-GD [40] | WRN | 77.8% | 87.4% | 82.1% | 89.8% |
LaplacianShot [41] | WRN | 70.27 ± 0.19% | 84.07% | 79.13 ± 0.21% | 86.75 ± 0.15% |
BD-CSPN [42] | WRN-28-10 | 70.31 ± 0.93% | 81.89 ± 0.60% | 78.74 ± 0.95 | 86.92 ± 0.63 |
PT + MAP [43] | WRN | 82.92 ± 0.26% | 88.82 ± 0.13% | 85.67 ± 0.26% | 90.45 ± 0.14% |
PEMnE-BMS [44] | WRN | 83.35 ± 0.25% | 89.53 ± 0.13% | 86.07 ± 0.25% | 91.09 ± 0.14% |
TransCNAPS + FETI [45] | R18 | 79.9 ± 0.8% | 91.50 ± 0.4% | 73.8 ± 1.0% | 87.7 ± 0.6% |
TRIDENT [46] | Conv4 | 86.11 ± 0.59% | 95.95 ± 0.28% | 86.97 ± 0.50% | 96.57 ± 0.17% |
P > M > F [33] | ViT-base | 95.3% | 98.4% | - | - |
STCE (Ours) | Swin Transformer | 98.65 ± 0.1% | 99.6 ± 0.2% | 91.6 ± 0.1% | 96.55 ± 0.27% |
FSCbenchmark | FC100 | ||
---|---|---|---|
Methods | Backbone | Five-Way One-Shot | Five-Way Five-Shot |
BML [47] | R12 | 45.00 ± 0.41% | 63.03 ± 0.41% |
IE [15] | R12 | 47.76 ± 0.77% | 65.30 ± 0.76% |
DeepEMD [11] | R12 | 46.47 ± 0.78% | 63.22 ± 0.71% |
TPMN [16] | R12 | 46.93 ± 0.71% | 63.26 ± 0.74% |
PAL [39] | R12 | 47.20 ± 0.60% | 64.00 ± 0.60% |
ConstellationNet [38] | R12 | 43.80 ± 0.20% | 59.70 ± 0.20% |
BAVARDAGE [48] | R12 | 57.27 ± 0.29% | 70.60 ± 0.21% |
STCE (Ours) | Swin Transformer | 64.1 ± 0.3% | 79.8 ± 0.69% |
FSCbenchmark | CIFAR-FS | ||
---|---|---|---|
Methods | Backbone | Five-Way One-Shot | Five-Way Five-Shot |
ProtoNet [12] | ConvNet-64 | 55.50 ± 0.70% | 72.00 ± 0.60% |
DPGN [34] | R12 | 77.9 ± 0.5% | 90.2 ± 0.4% |
IE [15] | R12 | 77.87 ± 0.85% | 89.74 ± 0.57% |
MAML [49] | ConvNet-32 | 58.90 ± 1.90% | 71.50 ± 1.00% |
PAL [39] | R12 | 77.1 ± 0.70% | 88.00 ± 0.50% |
RENet [50] | R12 | 74.51 ± 0.46% | 86.60 ± 0.32% |
ConstellationNet [38] | R12 | 75.4 ± 0.20% | 86.8 ± 0.20% |
BD-CSPN [42] | WRN-28-10 | 72.13 ± 1.01% | 82.28 ± 0.69% |
S2M2R [9] | WRN | 74.81 ± 0.19% | 87.47 ± 0.13% |
BML [47] | R12 | 73.45 ± 0.47% | 88.04 ± 0.33% |
PEMbE-NCM [44] | WRN | 74.84 ± 0.21% | 87.73 ± 0.15% |
PT + MAP + SF + SOT [51] | WRN-28-10 | 89.94 ± 0.6% | 92.83 ± 0.8% |
STCE (Ours) | Swin Transformer | 86.95 ± 0.2% | 94 ± 0.4% |
FSCbenchmark | CUB | ||
---|---|---|---|
Methods | Backbone | Five-Way One-Shot | Five-Way Five-Shot |
MAML [49] | R10 | 70.32 ± 0.99% | 80.93 ± 0.71% |
BD-CSPN + ESFR [52] | R18 | 82.68% | 88.65% |
ProtoNet [12] | R18 | 72.99 ± 0.88% | 86.64 ± 0.51% |
AmdimNet [24] | AmdimNet | 77.09 ± 0.21% | 89.18 ± 0.13% |
MatchingNetwork [53] | R18 | 73.49 ± 0.89% | 84.45 ± 0.58% |
S2M2R [9] | WRN | 80.68 ± 0.81% | 90.85 ± 0.44% |
FEAT [54] | R12 | 73.27 ± 0.22% | 85.77 ± 0.14% |
BML [47] | R12 | 76.21 ± 0.63% | 90.45 ± 0.36% |
DPGN [34] | R12 | 75.71 ± 0.47% | 91.48 ± 0.33% |
PT + NCM [43] | WRN | 80.57 ± 0.20% | 91.15 ± 0.10% |
DeepEMD [11] | R12 | 75.65 ± 0.83% | 88.69 ± 0.50% |
RENet [50] | R12 | 79.49 ± 0.44% | 91.11 ± 0.24% |
Delta-encoder [55] | VGG16 | 69.8% | 82.6% |
LaplacianShot [41] | R18 | 80.96% | 88.68% |
PEMbE-NCM [44] | WRN | 80.82 ± 0.19% | 91.46 ± 0.10% |
STCE (Ours) | Swin Transformer | 83.1 ± 0.4% | 92.88 ± 0.4% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Sun, H.; Zhang, P.; Zhang, X.; Han, X. Few-Shot Image Classification Based on Swin Transformer + CSAM + EMD. Electronics 2024, 13, 2121. https://doi.org/10.3390/electronics13112121
Sun H, Zhang P, Zhang X, Han X. Few-Shot Image Classification Based on Swin Transformer + CSAM + EMD. Electronics. 2024; 13(11):2121. https://doi.org/10.3390/electronics13112121
Chicago/Turabian StyleSun, Huadong, Pengyi Zhang, Xu Zhang, and Xiaowei Han. 2024. "Few-Shot Image Classification Based on Swin Transformer + CSAM + EMD" Electronics 13, no. 11: 2121. https://doi.org/10.3390/electronics13112121
APA StyleSun, H., Zhang, P., Zhang, X., & Han, X. (2024). Few-Shot Image Classification Based on Swin Transformer + CSAM + EMD. Electronics, 13(11), 2121. https://doi.org/10.3390/electronics13112121