Hyperspectral Imaging Zero-Shot Learning for Remote Marine Litter Detection and Classification
Abstract
:1. Introduction
- A ZSL approach for hyperspectral data, using purely spectral information, that is able to classify pixel by pixel regardless of its shape, which is usually affected by sun exposure and erosion caused by water and rocks;
- This ZSL approach is based on a generalised methodology, which means that it is trained with the known classes, and the classification is performed in all classes, known and unknown. The algorithm is adaptable, meaning that it can classify the samples from unknown classes into different classes;
- The algorithm is tested and evaluated using a real hyperspectral imaging dataset, holding marine litter data acquired during field experiments in Porto Pim bay, Faial island, Azores.
2. Hyperspectral Zero-Shot Learning Method
- Feature Extractor (F)—Convolutional Neural Network (CNN), with channel attention modules after each convolutional block [48];
- Classifier (C)—Fully Connected Neural Network;
- Decoder (D)—Autoencoder;
- Discriminator (P).
2.1. Feature Extractor
2.2. Classifier
2.3. Decoder
2.4. Discriminator
3. Marine Litter Dataset Acquisition and Experimental Setup
- Class imbalance: As shown in Table 1, the number of pixels for each class is different. To minimise the probability of overfitting during training, we use the lower number of pixels as a reference to randomly select samples from the other classes;
- Feature normalisation: The hyperspectral data acquisition is affected by the variability of atmospheric conditions, mainly direct sunlight. To minimise this effect, there is the need to perform the data normalisation in order to normalise the feature data to unit variance.
- Due to the flight altitude, some of the classes will have small (resolution) in the datasets. In addition to this, there are classes with artefacts in the middle due to their physical characteristics and construction. This happens in class 1, orange target, and class 3, rope target. The first, given the physical design of the multiple components (used oyster spat collectors) and their orientation, can accumulate more water, while the second (rope target) has more gaps in the middle of the target. Additionally, this target also uses two different rope targets with variable flutuability.
- The water contains points with high exposure due to the wave effects and the camera adjustment.
- The concrete pier also contains rocks.
- It is necessary to take into account that in class 6, boats have different hulls, and there are artefacts inside the boat that also have different materials. However, the flight altitude combined with the cameras spatial resolution does not allow a better distinction to facilitate the manual labelling process.
4. Results
5. Discussion
- Random Forest (RF), Support Vector Machine (SVM) and the Convolutional Neural Network 3D (CNN3D) were all implemented using only four classes:
- –
- Class 0: For the RF and SVM algorithm, it represents the water and land (houses, trees, streets, cars and other materials) to train the algorithms to learn the characteristics of the non-marine litter pixels. In the case of the CNN3D algorithm, it represents the water;
- –
- Class 1: orange target;
- –
- Class 2: white target;
- –
- Class 3: rope target.
- RF, SVM, and ZSL approaches use the F-BUMA data, while the CNN3D resorts to the drone data, using the same artificial targets and in the same test site. Due to the lack of resolution of the F-BUMA data, the results of the CNN3D were unsatisfactory due to the lack of distinct spatial features. Although the drone data have much more spatial information, as can be observed in [6], it does not contain data from any other class besides the artificial targets and the water, which was not enough for the ZSL algorithm. It would be possible to use RF and SVM with drone data. However, the algorithms will not benefit from the spatial information, and therefore, no novelty will be added to the results obtained with the F-BUMA data;
- The datasets used for RF, SVM and CNN3D algorithms were organised by flyby over the artificial targets, giving each flyby its results. In the case of the SVM and RF, Table 3 only presents the results for a single flybys (there were six flybys in total), while in the case of the CNN3D, all the results are presented.
6. Conclusions
Author Contributions
Funding
Acknowledgments
Conflicts of Interest
References
- Ciappa, A. Marine Litter Detection by Sentinel-2: A Case Study in North Adriatic (Summer 2020). Remote Sens. 2022, 14, 2409. [Google Scholar] [CrossRef]
- Maximenko, N.; Corradi, P.; Law, K.L.; Van Sebille, E.; Garaba, S.P.; Lampitt, R.S.; Galgani, F.E.A. Toward the Integrated Marine Debris Observing System. Front. Mar. Sci. 2019. [Google Scholar] [CrossRef] [Green Version]
- Topouzelis, K.; Papageorgiou, D.; Suaria, G.; Aliani, S. Floating marine litter detection algorithms and techniques using optical remote sensing data: A review. Mar. Pollut. Bull. 2021, 170, 112675. [Google Scholar] [CrossRef] [PubMed]
- Fulton, M.; Hong, J.; Islam, M.J.; Sattar, J. Robotic Detection of Marine Litter Using Deep Visual Detection Models. arXiv 2019, arXiv:1804.01079. [Google Scholar] [CrossRef]
- Freitas, S.; Silva, H.; Silva, E. Remote Hyperspectral Imaging Acquisition and Characterization for Marine Litter Detection. Remote Sens. 2021, 13, 2536. [Google Scholar] [CrossRef]
- Freitas, S.; Silva, H.; Almeida, C.; Viegas, D.; Amaral, A.; Santos, T.; Dias, A.; Jorge, P.A.S.; Pham, C.K.; Moutinho, J.; et al. Hyperspectral Imaging System for Marine Litter Detection. In OCEANS 2021: San Diego—Porto; IEEE: New York, NY, USA, 2021; pp. 1–6. [Google Scholar] [CrossRef]
- Martínez-Vicente, V.; Clark, J.R.; Corradi, P.; Aliani, S.; Arias, M.; Bochow, M.; Bonnery, G.; Cole, M.; Cózar, A.; Donnelly, R.; et al. Measuring Marine Plastic Debris from Space: Initial Assessment of Observation Requirements. Remote Sens. 2019, 11, 2443. [Google Scholar] [CrossRef] [Green Version]
- Kremezi, M.; Kristollari, V.; Karathanassi, V.; Topouzelis, K.; Kolokoussis, P.; Taggio, N.; Aiello, A.; Ceriola, G.; Barbone, E.; Corradi, P. Pansharpening PRISMA Data for Marine Plastic Litter Detection Using Plastic Indexes. IEEE Access 2021, 9, 61955–61971. [Google Scholar] [CrossRef]
- Topouzelis, K.; Papageorgiou, D.; Karagaitanakis, A.; Papakonstantinou, A.; Arias Ballesteros, M. Remote Sensing of Sea Surface Artificial Floating Plastic Targets with Sentinel-2 and Unmanned Aerial Systems (Plastic Litter Project 2019). Remote Sens. 2020, 12, 2013. [Google Scholar] [CrossRef]
- Wolf, M.; van den Berg, K.; Garaba, S.P.; Gnann, N.; Sattler, K.; Stahl, F.; Zielinski, O. Machine learning for aquatic plastic litter detection, classification and quantification (APLASTIC-Q). Environ. Res. Lett. 2020, 15, 114042. [Google Scholar] [CrossRef]
- Park, Y.J.; Garaba, S.P.; Sainte-Rose, B. Detecting the Great Pacific Garbage Patch floating plastic litter using WorldView-3 satellite imagery. Opt. Express 2021, 29, 35288–35298. [Google Scholar] [CrossRef]
- Kumar, N.; Berg, A.C.; Belhumeur, P.N.; Nayar, S.K. Attribute and Simile Classifiers for Face Verification. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Kyoto, Japan, 29 September–2 October 2009. [Google Scholar]
- Lampert, C.H.; Nickisch, H.; Harmeling, S. Attribute-Based Classification for Zero-Shot Visual Object Categorization. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 453–465. [Google Scholar] [CrossRef] [PubMed]
- Xu, B.; Fu, Y.; Jiang, Y.; Li, B.; Sigal, L. Video Emotion Recognition with Transferred Deep Feature Encodings. In Proceedings of the 2016 ACM on International Conference on Multimedia Retrieval, New York, NY, USA, 6–9 June 2016. [Google Scholar]
- Fu, Y.; Hospedales, T.M.; Xiang, T.; Gong, S. Attribute Learning for Understanding Unstructured Social Activity. In Computer Vision—ECCV 2012. ECCV 2012. Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2012; Volume 7575. [Google Scholar] [CrossRef] [Green Version]
- Liu, J.; Kuipers, B.; Savarese, S. Recognizing Human Actions by Attributes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Colorado Springs, CO, USA, 20–25 June 2011; pp. 3337–3344. [Google Scholar]
- Fu, Y.; Hospedales, T.M.; Xiang, T.; Gong, S. Learning Multimodal Latent Attributes. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 303–316. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Blitzer, J.; Foster, D.P.; Kakade, S.M. Zero-Shot Domain Adaptation: A Multi-View Approach. Technical Report TTI-TR-2009-1. 2009. Available online: https://www.ttic.edu/technical_reports/ttic-tr-2009-1.pdf (accessed on 29 July 2022).
- Long, Y.; Liu, L.; Shao, L. Attribute embedding with visual-semantic ambiguity removal for zero-shot learning. In Proceedings of the British Machine Vision Conference 2016, York, UK, 19–22 September 2016. [Google Scholar]
- Zhang, Z.; Saligrama, V. Zero-shot learning via joint latent similarity embedding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 6034–6042. [Google Scholar]
- Xian, Y.; Akata, Z.; Sharma, G.; Nguyen, Q.; Hein, M.; Schiele, B. Latent embeddings for zero-shot classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 69–77. [Google Scholar]
- Norouzi, M.; Mikolov, T.; Bengio, S.; Singer, Y.; Shlens, J.; Frome, A.; Corrado, G.; Dean, J. Zero-Shot Learning by Convex Combination of Semantic Embeddings. arXiv 2013, arXiv:1312.5650. [Google Scholar]
- Zhang, Z.; Saligrama, V. Zero-shot learning via semantic similarity embedding. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 4166–4174. [Google Scholar]
- Akata, Z.; Perronnin, F.; Harchaoui, Z.; Schmid, C. Label-Embedding for Image Classification. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 1425–1438. [Google Scholar] [CrossRef] [Green Version]
- Frome, A.; Corrado, G.; Shlens, J.; Bengio, S.; Dean, J.; Ranzato, M.; Mikolov, T. DeViSE: A Deep Visual-Semantic Embedding Model. In Proceedings of the Advances in Neural Information Processing Systems 26 (NIPS 2013), Lake Tahoe, NV, USA, 5–10 December 2013; pp. 2121–2129. [Google Scholar]
- Paredes, B.R.; Torr, P.H.S. An Embarrassingly Simple Approach to Zero-shot Learning. In Proceedings of the 32nd International Conference on International Conference on Machine Learning—Volume 37. JMLR.org, Lille, France, 7–9 July 2015; pp. 2152–2161. [Google Scholar]
- Changpinyo, S.; Chao, W.; Gong, B.; Sha, F. Synthesized Classifiers for Zero-Shot Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
- Zhang, Z.; Saligrama, V. Learning Joint Feature Adaptation for Zero-Shot Recognition. arXiv 2016, arXiv:1611.07593. [Google Scholar]
- Kodirov, E.; Xiang, T.; Fu, Z.; Gong, S. Unsupervised domain adaptation for zero-shot learning. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 2452–2460. [Google Scholar]
- Zhang, Z.; Saligrama, V. Zero-shot recognition via structured prediction. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016; pp. 533–548. [Google Scholar]
- Fu, Y.; Hospedales, T.M.; Xiang, T.; Fu, Z.; Gong, S. Transductive multi-view embedding for zero-shot recognition and annotation. In Proceedings of the European Conference on Computer Vision (ECCV), Zurich, Switzerland, 6–12 September 2014; pp. 584–599. [Google Scholar]
- Guo, Y.; Ding, G.; Jin, X.; Wang, J. Transductive Zero-shot Recognition via Shared Model Space Learning. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; pp. 3494–3500. [Google Scholar]
- Song, J.; Shen, C.; Yang, Y.; Liu, Y.; Song, M. Transductive Unbiased Embedding for Zero-Shot Learning. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
- Salakhutdinov, R.; Tenenbaum, J.B.; Torralba, A. Learning with Hierarchical-Deep Models. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1958–1971. [Google Scholar] [CrossRef] [Green Version]
- Mensink, T.; Gavves, E.; Snoek, C.G.M. COSTA: Co-Occurrence Statistics for Zero-Shot Classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014; pp. 2441–2448. [Google Scholar] [CrossRef] [Green Version]
- Chao, W.; Changpinyo, S.; Gong, B.; Sha, F. An Empirical Study and Analysis of Generalized Zero-Shot Learning for Object Recognition in the Wild. In Proceedings of the Computer Vision–ECCV 2016, Amsterdam, The Netherlands, 11–14 October 2016; pp. 52–68. [Google Scholar]
- Felix, R.; Kumar, B.G.V.; Reid, I.; Carneiro, G. Multi-modal Cycle-consistent Generalized Zero-Shot Learning. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
- Verma, V.K.; Arora, G.; Mishra, A.; Rai, P. Generalized Zero-Shot Learning via Synthesized Examples. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
- Qu, Y.; Baghbaderani, R.K.; Qi, H. Few-Shot Hyperspectral Image Classification Through Multitask Transfer Learning. In Proceedings of the 10th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Amsterdam, The Netherlands, 24–26 September 2019; pp. 1–5. [Google Scholar] [CrossRef]
- Liu, S.; Shi, Q.; Zhang, L. Few-Shot Hyperspectral Image Classification With Unknown Classes Using Multitask Deep Learning. IEEE Trans. Geosci. Remote Sens. 2021, 59, 5085–5102. [Google Scholar] [CrossRef]
- Wang, Y.; Liu, M.; Yang, Y.; Li, Z.; Du, Q.; Chen, Y.; Li, F.; Yang, H. Heterogeneous Few-Shot Learning for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
- Li, Z.; Liu, M.; Chen, Y.; Xu, Y.; Li, W.; Du, Q. Deep Cross-Domain Few-Shot Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–18. [Google Scholar] [CrossRef]
- Tong, X.; Yin, J.; Han, B.; Qv, H. Few-Shot Learning With Attention-Weighted Graph Convolutional Networks For Hyperspectral Image Classification. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, 25–28 October 2020; pp. 1686–1690. [Google Scholar] [CrossRef]
- Ding, C.; Li, Y.; Wen, Y.; Zheng, M.; Zhang, L.; Wei, W.; Zhang, Y. Boosting Few-Shot Hyperspectral Image Classification Using Pseudo-Label Learning. Remote Sens. 2021, 13, 3539. [Google Scholar] [CrossRef]
- Pal, D.; Bundele, V.; Sharma, R.; Banerjee, B.; Jeppu, Y. Few-Shot Open-Set Recognition of Hyperspectral Images with Outlier Calibration Network. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2022; pp. 2091–2100. [Google Scholar] [CrossRef]
- Dong, Y.; Jiang, X.; Zhou, H.; Lin, Y.; Shi, Q. SR2CNN: Zero-Shot Learning for Signal Recognition. arXiv 2020, arXiv:2004.04892. [Google Scholar] [CrossRef]
- Rodarmel, C.; Shan, J. Principal Component Analysis for Hyperspectral Image Classification. Surv. Land Inf. Syst. 2002, 62, 115–122. [Google Scholar]
- Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. arXiv 2018, arXiv:1807.06521. [Google Scholar] [CrossRef]
- Wen, Y.; Zhang, K.; Li, Z.; Qiao, Y. A Discriminative Feature Learning Approach for Deep Face Recognition. In Computer Vision—ECCV 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 499–515. [Google Scholar]
ID | Description | Example | Number of Pixels |
---|---|---|---|
Class 0 (Blue) Known class | Water | 1,279,132 | |
Class 1 (Red) Unknown class | Orange target (low density polyethylene) | 8118 | |
Class 2 (Green) Known class | White target (White plastic film) | 7913 | |
Class 3 (Magenta) Unknown class | Rope target | 8375 | |
Class 4 (White) Known class | Concrete pier | 16,999 | |
Class 5 (Grey) Known class | Trees | 972,160 | |
Class 6 (Yellow) Known class | Boats | 5023 |
Class 0 Water Known Class | Class 1 Orange Target Unknown Class | Class 2 White Target Known Class | Class 3 Rope Target Unknown Class | Class 4 Concrete Pier Known Class | Class 5 Trees Known Class | Class 6 Boats Known Class | |
---|---|---|---|---|---|---|---|
Precision | 0.9853 | 0.6659 | 0.7508 | 0.5632 | 0.9784 | 0.9927 | 0.5874 |
Recall | 0.9872 | 0.6946 | 0.9960 | 0.5766 | 0.9885 | 0.9927 | 0.9962 |
F1-Score | 0.9862 | 0.6800 | 0.8562 | 0.5698 | 0.9834 | 0.9927 | 0.7390 |
OA | 98.71% |
F-BUMA Flight—Flyby Over the Artificial Targets | ||||||||
Flyby 5 | ||||||||
Random Forest (RF) | Support Vector Machine (SVM) | |||||||
Class | Precision | Recall | F1-Score | Accuracy | Precision | Recall | F1-Score | Accuracy |
0 Water/Land | 0.99 | 1 | 0.99 | 98.09% | 0.99 | 0.99 | 0.99 | 97.06 % |
1 Orangetarget | 0.71 | 0.25 | 0.37 | 0.77 | 0.46 | 0.58 | ||
2 Whitetarget | 0.74 | 0.62 | 0.67 | 0.57 | 0.59 | 0.58 | ||
3 Ropetarget | 0.38 | 0.56 | 0.45 | 0.21 | 0.47 | 0.29 | ||
F-BUMA Flight—Zero-Shot Learning | ||||||||
Class 0 Water Known Class | Class 1 Orange Target Unknown Class | Class 2 White Target Known Class | Class 3 Rope Target Unknown Class | Class 4 Concrete Pier Known Class | Class 5 Trees Known Class | Class 6 Boats Known Class | Accuracy | |
Precision | 0.9853 | 0.6659 | 0.7508 | 0.5632 | 0.9784 | 0.9927 | 0.5874 | 98.71 % |
Recall | 0.9872 | 0.6946 | 0.9960 | 0.5766 | 0.9885 | 0.9927 | 0.9962 | |
F1-Score | 0.9862 | 0.6800 | 0.8562 | 0.5698 | 0.9834 | 0.9927 | 0.7390 | |
Drone Flight—CNN3D | ||||||||
Flyby 4 | ||||||||
Class | Precision | Recall | F1-Score | Accuracy | ||||
0—Water | 0.98 | 0.93 | 0.95 | 91.67 % | ||||
1—Orange target | 0.75 | 0.78 | 0.77 | |||||
2—White target | 0.94 | 0.95 | 0.94 | |||||
3—Rope target | 0.69 | 0.96 | 0.80 | |||||
Flyby 5 | ||||||||
Class | Precision | Recall | F1-Score | Accuracy | ||||
0—Water | 0.83 | 0.97 | 0.90 | 84.84 % | ||||
1—Orange target | 0.62 | 0.50 | 0.55 | |||||
2—White target | 0.98 | 0.84 | 0.91 | |||||
3—Rope target | 0.88 | 0.54 | 0.67 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Freitas, S.; Silva, H.; Silva, E. Hyperspectral Imaging Zero-Shot Learning for Remote Marine Litter Detection and Classification. Remote Sens. 2022, 14, 5516. https://doi.org/10.3390/rs14215516
Freitas S, Silva H, Silva E. Hyperspectral Imaging Zero-Shot Learning for Remote Marine Litter Detection and Classification. Remote Sensing. 2022; 14(21):5516. https://doi.org/10.3390/rs14215516
Chicago/Turabian StyleFreitas, Sara, Hugo Silva, and Eduardo Silva. 2022. "Hyperspectral Imaging Zero-Shot Learning for Remote Marine Litter Detection and Classification" Remote Sensing 14, no. 21: 5516. https://doi.org/10.3390/rs14215516
APA StyleFreitas, S., Silva, H., & Silva, E. (2022). Hyperspectral Imaging Zero-Shot Learning for Remote Marine Litter Detection and Classification. Remote Sensing, 14(21), 5516. https://doi.org/10.3390/rs14215516