Transfer Learning for Radio Frequency Machine Learning: A Taxonomy and Survey
Abstract
:1. Introduction
2. Definitions
- —the source and target tasks have different conditional probability distributions. This most commonly manifests in the form of unbalanced datasets, where a subset of classes has more examples in the source dataset than the target dataset or vice versa. A simple example might be transferring an AMC model between two datasets, both of which only contain BPSK and QPSK signal types. However, the source dataset contains 70% BPSK signals and 30% QPSK signals, while the target dataset contains 30% BPSK signals and 70% QPSK signals.
- —the source and target tasks have different label spaces. For example, the target task contains an additional output class (i.e., for an AMC algorithm, and the source task is a binary BPSK/QPSK output set, while the target task includes a third noise-only class). Alternatively, the target task may be completely unrelated and disjoint from the source task (i.e., the target task is to perform SEI, while the source task was to perform AMC); therefore, the label spaces are also disjoint.
- —the source and target domains have different data distributions. An example of such a scenario includes a transfer of models from one channel environment to another, as described further in Section 4.1.1.
- —the source and target feature spaces differ. An example includes performing SEI using the same set of known emitters but using different modulation schemes in the source and target domain.
3. Related TL Taxonomies
- Self-taught methods that address settings where no labeled data are available in the source domain;
- Multitask learning that assumes the availability of labeled data in both the source and target domains and in which the source and target tasks are learned simultaneously;
- Sequential learning, which also assumes the availability of labeled data in both the source and target domains; however, the source task/domain is learned first and the target task/domain is learned second.
- Domain adaptation, under which the source and target domains differ;
- Sample selection bias, also known as covariance shift, which refers to when both source and target domains and tasks are the same, but the source and/or target training dataset may be incomplete or small.
4. An RFML-Specific Taxonomy
4.1. Domain Adaptation
4.1.1. Environment Adaptation
4.1.2. Platform Adaptation
4.1.3. Environment Platform Co-Adaptation
4.2. Multitask Learning
4.3. Sequential Learning
5. Future Work
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Rondeau, T. Radio Frequency Machine Learning Systems (RFMLS). Available online: https://www.darpa.mil/program/radio-frequency-machine-learning-systems (accessed on 19 January 2022).
- West, N.E.; O’Shea, T. Deep architectures for modulation recognition. In Proceedings of the 2017 IEEE International Symposium on Dynamic Spectrum Access Networks (DySPAN), Baltimore, MD, USA, 6–9 March 2017; pp. 1–6. [Google Scholar]
- Mohammed, S.; El Abdessamad, R.; Saadane, R.; Hatim, K.A. Performance Evaluation of Spectrum Sensing Implementation using Artificial Neural Networks and Energy Detection Method. In Proceedings of the 2018 International Conference on Electronics, Control, Optimization and CS (ICECOCS), Kenitra, Morocco, 5–6 December 2018; pp. 1–6. [Google Scholar]
- Wong, L.J.; Headley, W.C.; Andrews, S.; Gerdes, R.M.; Michaels, A.J. Clustering Learned CNN Features from Raw I/Q Data for Emitter Identification. In Proceedings of the 2018 IEEE Military Comm. Conference (MILCOM), Los Angeles, CA, USA, 29–31 October 2018; pp. 26–33. [Google Scholar]
- Hauser, S.C. Real-World Considerations for Deep Learning in Spectrum Sensing. Master’s Thesis, Virginia Tech, Blacksburg, VA, USA, 2018. [Google Scholar]
- Sankhe, K.; Belgiovine, M.; Zhou, F.; Riyaz, S.; Ioannidis, S.; Chowdhury, K. ORACLE: Optimized radio classification through convolutional neural networks. In Proceedings of the IEEE INFOCOM 2019-IEEE Conference on Computer Communications, Paris, France, 29 April–2 May 2019; pp. 370–378. [Google Scholar]
- Ruder, S. Neural Transfer Learning for Natural Language Processing. Ph.D. Thesis, NUI Galway, Galway, Ireland, 2019. [Google Scholar]
- Clark IV, W.H.; Hauser, S.; Headley, W.C.; Michaels, A.J. Training Data Augmentation for Deep Learning Radio Frequency Systems. J. Def. Model. Simul. 2021, 18, 217–237. [Google Scholar] [CrossRef]
- Pan, S.J.; Yang, Q. A Survey on Transfer Learning. IEEE Trans. Knowl. Data Eng. 2010, 22, 1345–1359. [Google Scholar] [CrossRef]
- Zhuang, F.; Qi, Z.; Duan, K.; Xi, D.; Zhu, Y.; Zhu, H.; Xiong, H.; He, Q. A comprehensive survey on transfer learning. Proc. IEEE 2020, 109, 43–76. [Google Scholar] [CrossRef]
- Csurka, G. Domain adaptation for visual applications: A comprehensive survey. arXiv 2017, arXiv:1702.05374. [Google Scholar]
- Wang, M.; Deng, W. Deep visual domain adaptation: A survey. Neurocomputing 2018, 312, 135–153. [Google Scholar] [CrossRef] [Green Version]
- Lazaric, A. Transfer in reinforcement learning: A framework and a survey. In Reinforcement Learning; Springer: Berlin/Heidelberg, Germany, 2012; pp. 143–173. [Google Scholar]
- Wong, L.J.; Clark IV, W.H.; Flowers, B.; Buehrer, R.M.; Michaels, A.J.; Headley, W.C. The RFML Ecosystem: A Look at the Unique Challenges of Applying Deep Learning to Radio Frequency Applications. arXiv 2020, arXiv:2010.00432. [Google Scholar]
- Tong, Z.; Shi, D.; Yan, B.; Wei, J. A Review of Indoor-Outdoor Scene Classification. In Proceedings of the 2017 2nd International Conference on Control, Automation and Artificial Intelligence (CAAI 2017), Sanya, China, 25–26 June 2017; Atlantis Press: Zhengzhou, China, 2017; pp. 469–474. [Google Scholar] [CrossRef] [Green Version]
- Pei, Y.; Huang, Y.; Zou, Q.; Zhang, X.; Wang, S. Effects of Image Degradation and Degradation Removal to CNN-Based Image Classification. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 1239–1253. [Google Scholar] [CrossRef]
- Chen, S.; Zheng, S.; Yang, L.; Yang, X. Deep Learning for Large-Scale Real-World ACARS and ADS-B Radio Signal Classification. IEEE Access 2019, 7, 89256–89264. [Google Scholar] [CrossRef]
- Pati, B.M.; Kaneko, M.; Taparugssanagorn, A. A Deep Convolutional Neural Network Based Transfer Learning Method for Non-Cooperative Spectrum Sensing. IEEE Access 2020, 8, 164529–164545. [Google Scholar] [CrossRef]
- Wang, Z.; Dai, Z.; Poczos, B.; Carbonell, J. Characterizing and Avoiding Negative Transfer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
- Zhang, W.; Deng, L.; Wu, D. Overcoming Negative Transfer: A Survey. arXiv 2020, arXiv:2009.00909. [Google Scholar]
- Liu, Z.; Lian, T.; Farrell, J.; Wandell, B.A. Neural network generalization: The impact of camera parameters. IEEE Access 2020, 8, 10443–10454. [Google Scholar] [CrossRef]
- Merchant, K. Deep Neural Networks for Radio Frequency Fingerprinting. Ph.D. Thesis, University of Maryland, College Park, MD, USA, 2019. [Google Scholar]
- Venkateswara, H.; Eusebio, J.; Chakraborty, S.; Panchanathan, S. Deep hashing network for unsupervised domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5018–5027. [Google Scholar]
- O’Shea, T.J.; Roy, T.; Clancy, T.C. Over-the-Air Deep Learning Based Radio Signal Classification. IEEE J. Sel. Top. Signal Process. 2018, 12, 168–179. [Google Scholar] [CrossRef] [Green Version]
- Dörner, S.; Cammerer, S.; Hoydis, J.; Ten Brink, S. Deep Learning Based Communication Over the Air. IEEE J. Sel. Top. Signal Process. 2018, 12, 132–143. [Google Scholar] [CrossRef] [Green Version]
- Peters, M.E.; Ruder, S.; Smith, N.A. To tune or not to tune? adapting pretrained representations to diverse tasks. arXiv 2019, arXiv:1903.05987. [Google Scholar]
- Zheng, S.; Chen, S.; Qi, P.; Zhou, H.; Yang, X. Spectrum sensing based on deep learning classification for cognitive radios. China Comm. 2020, 17, 138–148. [Google Scholar] [CrossRef]
- Clark, B.; Leffke, Z.; Headley, C.; Michaels, A. Cyborg Phase II Final Report; Technical Report; Ted and Karyn Hume Center for National Security and Technology: Blacksburg, VA, USA, 2019. [Google Scholar]
- Ju, Y.; Lei, M.; Zhao, M.; Li, M.; Zhao, M. A Joint Jamming Detection and Link Scheduling Method Based on Deep Neural Networks in Dense Wireless Networks. In Proceedings of the 2019 IEEE 90th Vehicular Technology Conference (VTC2019), Honolulu, HI, USA, 22–25 September 2019; pp. 1–5. [Google Scholar] [CrossRef]
- Huang, C.; Chiang, C.; Li, Q. A study of deep learning networks on mobile traffic forecasting. In Proceedings of the 2017 IEEE 28th Annual International Symposium on Personal, Indoor, and Mobile Radio Communications (PIMRC), Montreal, QC, Canada, 8–13 October 2017; pp. 1–6. [Google Scholar] [CrossRef]
- Qiu, C.; Zhang, Y.; Feng, Z.; Zhang, P.; Cui, S. Spatio-Temporal Wireless Traffic Prediction with Recurrent Neural Network. IEEE Wirel. Comm. Lett. 2018, 7, 554–557. [Google Scholar] [CrossRef]
- Lin, W.; Huang, C.; Duc, N.; Manh, H. Wi-Fi Indoor Localization based on Multi-Task Deep Learning. In Proceedings of the 2018 IEEE 23rd International Conference on Digital Signal Processing (DSP), Shanghai, China, 19–21 November 2018; pp. 1–5. [Google Scholar] [CrossRef]
- Ye, N.; Li, X.; Yu, H.; Zhao, L.; Liu, W.; Hou, X. DeepNOMA: A Unified Framework for NOMA Using Deep Multi-Task Learning. IEEE Trans. Wirel. Commun. 2020, 19, 2208–2225. [Google Scholar] [CrossRef]
- Clark, W.H.; Arndorfer, V.; Tamir, B.; Kim, D.; Vives, C.; Morris, H.; Wong, L.; Headley, W.C. Developing RFML Intuition: An Automatic Modulation Classification Architecture Case Study. In Proceedings of the 2019 IEEE Military Communications Conference (MILCOM), Norfolk, VA, USA, 12–14 November 2019; pp. 292–298. [Google Scholar] [CrossRef]
- Mossad, O.S.; ElNainay, M.; Torki, M. Deep Convolutional Neural Network with Multi-Task Learning Scheme for Modulations Recognition. In Proceedings of the 2019 15th International Wireless Communications Mobile Computing Conference (IWCMC), Tangier, Morocco, 24–28 June 2019; pp. 1644–1649. [Google Scholar] [CrossRef]
- Wong, L.J.; McPherson, S. Explainable Neural Network-based Modulation Classification via Concept Bottleneck Models. In Proceedings of the 2021 IEEE Computing and Communications Workshop and Conference (CCWC), Las Vegas, NV, USA, 27–30 January 2021. [Google Scholar]
- Peng, Q.; Gilman, A.; Vasconcelos, N.; Cosman, P.C.; Milstein, L.B. Robust Deep Sensing Through Transfer Learning in Cognitive Radio. IEEE Wirel. Comm. Lett. 2020, 9, 38–41. [Google Scholar] [CrossRef]
- Kuzdeba, S.; Robinson, J.; Carmack, J. Transfer Learning with Radio Frequency Signals. In Proceedings of the 2021 IEEE 18th Annual Consumer Communications Networking Conference (CCNC), Las Vegas, NV, USA, 9–12 January 2021; pp. 1–9. [Google Scholar] [CrossRef]
- Robinson, J.; Kuzdeba, S. RiftNet: Radio Frequency Classification for Large Populations. In Proceedings of the 2021 IEEE 18th Annual Consumer Communications Networking Conference (CCNC), Las Vegas, NV, USA, 9–12 January 2021; pp. 1–6. [Google Scholar] [CrossRef]
- Le, D.V.; Meratnia, N.; Havinga, P.J.M. Unsupervised Deep Feature Learning to Reduce the Collection of Fingerprints for Indoor Localization Using Deep Belief Networks. In Proceedings of the 2018 International Conference on Indoor Positioning and Indoor Navigation (IPIN), Nantes, France, 24–27 September 2018; pp. 1–7. [Google Scholar] [CrossRef]
- Ali, A.; Yangyu, F. Unsupervised feature learning and automatic modulation classification using deep learning model. Phys. Commun. 2017, 25, 75–84. [Google Scholar] [CrossRef]
- O’Shea, T.J.; Corgan, J.; Clancy, T.C. Unsupervised representation learning of structured radio communication signals. In Proceedings of the 2016 First International Workshop on Sensing, Processing and Learning for Intelligent Machines (SPLINE), Aalborg, Denmark, 6–8 July 2016; pp. 1–5. [Google Scholar] [CrossRef] [Green Version]
- Bengio, Y. Deep learning of representations for unsupervised and transfer learning. In Proceedings of the ICML Workshop on Unsupervised and Transfer Learning, JMLR Workshop and Conference Proceedings, Bellevue, WA, USA, 28 June–2 July 2012; pp. 17–36. [Google Scholar]
- Neyshabur, B.; Sedghi, H.; Zhang, C. What is being transferred in transfer learning? In Advances in Neural Information Processing Systems; Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2020; Volume 33, pp. 512–523. [Google Scholar]
- Nguyen, C.; Hassner, T.; Seeger, M.; Archambeau, C. LEEP: A new measure to evaluate transferability of learned representations. In Proceedings of the International Conference on Machine Learning, Virtual, 13–18 July 2020; pp. 7294–7305. [Google Scholar]
- You, K.; Liu, Y.; Long, M.; Wang, J. LogME: Practical Assessment of Pre-trained Models for Transfer Learning. arXiv 2021, arXiv:2102.11005. [Google Scholar]
- Ben-David, S.; Blitzer, J.; Crammer, K.; Pereira, F. Analysis of Representations for Domain Adaptation. In Advances in Neural Information Processing Systems; Schölkopf, B., Platt, J., Hoffman, T., Eds.; MIT Press: Cambridge, MA, USA, 2007; Volume 19. [Google Scholar]
- Yosinski, J.; Clune, J.; Nguyen, A.; Fuchs, T.; Lipson, H. Understanding neural networks through deep visualization. arXiv 2015, arXiv:1506.06579. [Google Scholar]
- Li, Y.; Yosinski, J.; Clune, J.; Lipson, H.; Hopcroft, J. Convergent Learning: Do different neural networks learn the same representations? In Proceedings of the Feature Extraction: Modern Questions and Challenges, Montreal, QC, Canada, 11 December 2015; pp. 196–212. [Google Scholar]
- Wang, L.; Hu, L.; Gu, J.; Wu, Y.; Hu, Z.; He, K.; Hopcroft, J. Towards understanding learning representations: To what extent do different neural networks learn the same representation. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; pp. 9607–9616. [Google Scholar]
- Garriga-Alonso, A.; Rasmussen, C.E.; Aitchison, L. Deep Convolutional Networks as shallow Gaussian Processes. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
- Montavon, G.; Samek, W.; Müller, K.R. Methods for interpreting and understanding deep neural networks. Digit. Signal Process. 2018, 73, 1–15. [Google Scholar] [CrossRef]
- Turek, M. Explainable Artificial Intelligence (XAI). Available online: https://www.darpa.mil/program/explainable-artificial-intelligence (accessed on 19 January 2022).
- Raghu, M.; Gilmer, J.; Yosinski, J.; Sohl-Dickstein, J. SVCCA: Singular vector canonical correlation analysis for deep learning dynamics and interpretability. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 6078–6087. [Google Scholar]
- Morcos, A.S.; Raghu, M.; Bengio, S. Insights on representational similarity in neural networks with canonical correlation. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018. [Google Scholar]
- Kornblith, S.; Norouzi, M.; Lee, H.; Hinton, G. Similarity of neural network representations revisited. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 3519–3529. [Google Scholar]
Domain Elements | Tasks |
---|---|
|
|
Setting | Description |
---|---|
(a) | The traditional ML setting where the source and target domains and tasks are the same. |
(b) | The TL setting in which learned features from one domain are used to support performing the same task in a second domain. For example, using features learned to perform AMC in an AWGN channel to support performing AMC in a fading channel. |
(c) | The setting in which source and target domains are so dissimilar that TL is unsuccessful, despite the source and target tasks being the same. |
(d) | The TL setting in which learned features from one task are used to support a second task, while the source and target domains are the same. For example, using features learned to perform AMC to support SEI with the source and target domains being the same. |
(e) | Likely the most challenging TL setting in which learned features from one domain and task are used to support performing a second task in a new domain. For example, using features learned to perform AMC in an AWGN channel to support performing SEI in a fading channel. |
(f) | The setting in which source and target domains are so dissimilar that TL is unsuccessful, although the source and target tasks are somewhat similar. |
(g) | The setting in which source and target tasks are so dissimilar that TL is unsuccessful, despite the source and target domains being the same. |
(h) | The setting in which source and target tasks are so dissimilar that TL is unsuccessful, despite the source and target domains being somewhat similar. |
(i) | The setting in which both source and target tasks and domains are dissimilar, preventing the use of successful TL. |
TL Setting | Use Case | Source Domain | Source Task | Target Domain | Target Task |
---|---|---|---|---|---|
Environment Adaptation | Move a Tx/Rx pair equipped with an AMC model from an empty field to a city center | Single Tx/Rx pair, AWGN channel | Binary AMC (BPSK/QPSK) | Same Tx/Rx pair, Multipath channel | Binary AMC (BPSK/QPSK) |
Platform Adaptation | Transfer an AMC model between UAV | Single Rx, Many Tx, Fading channel w/ Doppler | Binary AMC (BPSK/QPSK) | Different Rx, Same Tx set, Fading channel w/ Doppler | Binary AMC (BPSK/QPSK) |
Environment Platform Co-Adaptation | Transfer an AMC model between a ground-station and UAV | Single Rx, Many Tx, Multipath channel | Binary AMC (BPSK/QPSK) | Different Rx, Same Tx set, Fading channel w/ Doppler | Binary AMC (BPSK/QPSK) |
Multitask Learning | Simultaneous signal detection and AMC | Single Tx/Rx pair, AWGN channel | Binary AMC (BPSK/QPSK) | Same Tx/Rx pair, AWGN channel | SNR Estimation |
Sequential Learning | Addition of an output class(es) to an | Single Tx/Rx pair, AWGN channel | Binary AMC (BPSK/QPSK) | Same Tx/Rx pair, AWGN channel | Four-class AMC (BPSK/QPSK/ 16QAM/64QAM) |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Wong, L.J.; Michaels, A.J. Transfer Learning for Radio Frequency Machine Learning: A Taxonomy and Survey. Sensors 2022, 22, 1416. https://doi.org/10.3390/s22041416
Wong LJ, Michaels AJ. Transfer Learning for Radio Frequency Machine Learning: A Taxonomy and Survey. Sensors. 2022; 22(4):1416. https://doi.org/10.3390/s22041416
Chicago/Turabian StyleWong, Lauren J., and Alan J. Michaels. 2022. "Transfer Learning for Radio Frequency Machine Learning: A Taxonomy and Survey" Sensors 22, no. 4: 1416. https://doi.org/10.3390/s22041416
APA StyleWong, L. J., & Michaels, A. J. (2022). Transfer Learning for Radio Frequency Machine Learning: A Taxonomy and Survey. Sensors, 22(4), 1416. https://doi.org/10.3390/s22041416