Multi-Label Classification of Complaint Texts: Civil Aviation Service Quality Case Study
Abstract
:1. Introduction
2. Related Work
2.1. Classification of Complaint Texts
2.2. Multi-Label Text Classification
3. Study Design
3.1. Modeling Framework
3.2. Text Vectorization Representation Layer
3.2.1. BERT Word Embedding
3.2.2. Label Feature Word Embedding
- Feature word set construction for civil aviation service quality complaint text labeling:First, the corpus corresponding to each tag was segmented and lexically labeled sentence by sentence. The noun component of the sentence was taken as the basic transaction, and single-word nouns were filtered to create the basic transaction set. Then, the frequent 1-item sets were obtained using the association rule Apriori algorithm, and the frequent 2-item sets were obtained from .The Apriori algorithm for association rules is a classic algorithm widely used in the field of data mining, aimed at discovering interesting association relationships between item sets in a dataset. This algorithm adopts an iterative strategy of a layer-by-layer search, first identifying all frequent item sets, then constructing new candidate sets based on these frequent item sets and calculating their support to filter out the next order of frequent item sets. This process continues until no more frequent item sets containing more items can be found.Since the frequent item set mining algorithm of association rules does not consider the positional relationships of nouns in the complaint text, the two nouns of may not form a phrase. So, when filtering according to the dependency relation of non-phrase features, only the noun phrases whose relation was a noun phrase (NP) were considered to obtain a new set of frequent 2-item sets, . Then, from the candidate feature set , we introduced domain dictionaries such as “civil aviation” and “civil aviation thesaurus” to supplement the mining of infrequent domain features. This enabled us to collect more comprehensive and accurate domain features, which we used to obtain .The redundant features were then filtered by the minimum independent support. The minimum independent support is the support of the frequent 1-item set containing the feature word minus the absolute support of the frequent 2-item set containing the word. If the minimum independent support of a feature word was less than a threshold value, T, the word was filtered as a redundant feature, and the final set of frequent 1-item sets satisfying the independent support constraint was obtained, referred to as .Finally, the semi-automatic updating of the civil aviation domain feature dictionary was realized by manually updating the dictionary with feedback on the mined frequent features, which saves human effort. After candidate feature mining, pruning, filtering, and infrequent item set supplementation, the domain feature set was finally obtained as .
- Feature word embedding:(1) Training the Word2Vec model:First, we chose the model type. Word2Vec has two main model types, Skip-gram and CBOW (continuous bag of words). Skip-gram usually performs better on small datasets, while CBOW is faster for large datasets.Next, we set the hyperparameters of the model, including the vector dimension, window size, number of training iterations, and others.Then, the model was trained. The Word2Vec model was trained using preprocessed text data. The model learns word vectors by predicting the context of a word (Skip-gram) or predicting a word using the context (CBOW).(2) Obtaining word vectors for the topic feature words:For a feature word set consisting of t words, , we obtained the word vectors of label feature words from the trained Word2Vec model as .
3.2.3. Text Vector Representation Based on Attention Weighting of Label Features
3.3. Multi-Channel Feature Extraction Layer
3.3.1. Text-CNN-Based Localized Feature Extraction
3.3.2. BiLSTM-Based Global Feature Extraction
3.4. Feature Interaction Layer
3.5. Multi-Label Classification Layer
3.5.1. Binary Relevance Model
3.5.2. GBDT Integration Algorithm
3.6. Output Layer
4. Experimental Results and Analysis
4.1. Experimental Data
4.2. Measurement Indicators
4.3. Result Analysis
4.3.1. Model Performance Comparison
4.3.2. Analysis of Ablation Experiments
4.3.3. Label Classification Effect Analysis
- (1)
- High-frequency incidents: Complaints in the categories of ticketing services and irregular flight services are usually closely related to passengers’ daily travel. Problems in these two categories are common and therefore have a higher chance of giving rise to complaints.
- (2)
- Direct impact on the passenger experience: Ticketing and flight services have a direct impact on the overall passenger experience and travel plans. Any problems related to ticketing errors, itinerary changes, flight delays, cancellations, etc., may lead to traveler complaints, as they can significantly affect passenger travel.
- (3)
- Measurability: Problems in the ticket service and irregular flight service categories are usually easily measurable and recordable and therefore more likely to be recorded in complaint data. In contrast, other types of problems may be less easily captured and recorded.
- (4)
- Cumulative effect: Problems with ticketing or flight services that occur during a single trip may result in multiple complaints, as they may affect multiple stages and aspects of the trip. This can lead to an increase in the percentage of complaints in both categories.
5. Conclusions and Future Work
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Liang, X.; Li, M. Text categorization of complain in telecommunication industry and its applied research. Chin. J. Manag. Sci. 2015, 23, 188–192. [Google Scholar]
- Wang, J.; Yang, Y.; Yu, B. Customer complaint text classification model based on multi-head collaborative attention mechanism. Data Anal. Knowl. Discov. 2023, 7, 10. [Google Scholar]
- Joachims, T. Text categorization with support vector machines: Learning with many relevant features. In Proceedings of the 10th European Conference on Machine Learning (ECML), Chemnitz, Germany, 21–23 April 1998; pp. 137–142. [Google Scholar]
- Xu, B.; Zhang, Y. A new svm chinese text of classification algorithm based on the semantic kernel. In Proceedings of the 2011 International Conference on Multimedia Technology (ICMT), Hangzhou, China, 26–28 July 2011; pp. 2857–2860. [Google Scholar]
- Hussain, M.G.; Rashidul Hasan, M.; Rahman, M.; Protim, J.; Al Hasan, S. Detection of bangla fake news using mnb and svm classifier. In Proceedings of the 2020 International Conference on Computing, Electronics Communications Engineering (ICCECE), Southend, UK, 17–18 August 2020; pp. 81–85. [Google Scholar]
- Rongyan, L. A new algorithm of chinese text classification. J. Beijing Norm. Univ. (Nat. Sci.) 2006, 42, 501–505. [Google Scholar]
- Tan, Y. An improved knn text classification algorithm based on k-medoids and rough set. In Proceedings of the 2018 10th International Conference on Intelligent Human-Machine Systems and Cybernetics (IHMSC), Hangzhou, China, 25–26 August 2018; pp. 109–113. [Google Scholar]
- Moldagulova, A.; Sulaiman, R.B. Using knn algorithm for classification of textual documents. In Proceedings of the 2017 8th International Conference on Information Technology (ICIT), Amman, Jordan, 17–18 May 2017; pp. 665–671. [Google Scholar]
- Chen, Z.; Shi, G.; Wang, X. Text classification based on naive bayes algorithm with feature selection. Information 2012, 15, 4255–4260. [Google Scholar]
- El Hindi, K.M.; Aljulaidan, R.R.; AlSalman, H. Lazy fine-tuning algorithms for naive bayesian text classification. Appl. Soft Comput. 2020, 96, 106652. [Google Scholar] [CrossRef]
- Salton, G.; Wong, A.; Yang, C.-S. A vector space model for automatic indexing. Commun. ACM 1975, 18, 613–620. [Google Scholar] [CrossRef]
- Zhu, Y. Research on news text classification based on deep learning convolutional neural network. Wireless Commun. Mob. Comput. 2021, 2021, 1508150. [Google Scholar] [CrossRef]
- Blei, D.M.; Ng, A.Y.; Jordan, M.I. Latent Dirichlet Allocation. J. Mach. Learn. Res. 2003, 3, 993–1022. [Google Scholar]
- Pavlinek, M.; Podgorelec, V. Text classification method based on self-training and LDA topic models. Expert Syst. Appl. 2017, 80, 83–93. [Google Scholar] [CrossRef]
- Mikolov, T. Efficient estimation of word representations in vector space. arXiv 2013, arXiv:1301.3781. [Google Scholar]
- Rui, Z.; Yutai, H. Research on short text classification based on word2vec microblog. In Proceedings of the 2020 International Conference on Computer Science and Management Technology (ICCSMT), Shanghai, China, 20–22 November 2020; pp. 178–182. [Google Scholar]
- Hu, H.; Liao, M.; Zhang, C.; Jing, Y. Text classification based recurrent neural network. In Proceedings of the 2020 IEEE 5th Information Technology and Mechatronics Engineering Conference (ITOEC), Chongqing, China, 12–14 June 2020; pp. 652–655. [Google Scholar]
- Chen, G.; Ye, D.; Xing, Z.; Chen, J.; Cambria, E. Ensemble application of convolutional and recurrent neural networks for multi-label text categorization. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 2377–2383. [Google Scholar]
- Tian, Q.; Kong, W.; Teng, J.; Wang, Z. Text sentiment analysis model based on parallel hybrid network and attention mechanism. Comput. Eng. 2022, 48, 266–273. [Google Scholar]
- Han, Y.; Chen, C.; Su, H.; Liang, Y. A hybrid neural network text classification model integrating channel features. J. Chin. Inf. Process. 2021, 35, 78–88. [Google Scholar]
- Ma, Y.; Huang, J.; Wang, F.; Rui, X. Multi-label classification of science and technology policy content using BERT and multi-scale CNN. J. Inf. 2022, 41, 157–163. [Google Scholar]
- Gu, Y.; Shi, J. The social public issues analysis model based on deep learning. Sci. Program. 2022, 2022, 8676124. [Google Scholar] [CrossRef]
- Rhanoui, M.; Mikram, M.; Yousfi, S.; Barzali, S. A CNN-BiLSTM model for document-level sentiment analysis. Mach. Learn. Knowl. Extract. 2019, 1, 832–847. [Google Scholar] [CrossRef]
- Kavianpour, P.; Kavianpour, M.; Jahani, E.; Ramezani, A. A CNN-BiLSTM model with attention mechanism for earthquake prediction. J. Supercomput. 2023, 79, 19194–19226. [Google Scholar] [CrossRef]
- Xu, X.; Wang, B. Research on multi-label patent classification algorithm based on text and historical data. Comput. Sci. 2024, 51, 172–178. [Google Scholar]
- Bahdanau, D. Neural machine translation by jointly learning to align and translate. arXiv 2014, arXiv:1409.0473. [Google Scholar]
- Yan, Y.; Liu, F.; Zhuang, X.; Ju, J. An R-transformer-BiLSTM model based on attention for multi-label text classification. Neural Process. Lett. 2023, 55, 1293–1316. [Google Scholar] [CrossRef]
- Zhao, H.; Li, X.; Wang, F.; Zeng, Q.; Diao, X. Incorporating keyword extraction and attention for multi-label text classification. J. Intell. Fuzzy Syst. 2023, 45, 2083–2093. [Google Scholar] [CrossRef]
- Wang, Z.; Dong, K.; Huang, J.; Wang, B. SEMFA: A large-scale multi-label text classification model based on semantic features and associative attention. Comput. Sci. 2023, 50, 270–278. [Google Scholar]
- Lu, J.; Yang, J.; Batra, D.; Parikh, D. Hierarchical question-image co-attention for visual question answering. In Proceedings of the 30th International Conference on Neural Information Processing Systems (NIPS), Barcelona, Spain, 5–10 December 2016; pp. 289–297. [Google Scholar]
- Hu, L.; Zhao, Z.; Qi, W.; Song, X.; Nie, L. Multimodal matching-aware co-attention networks with mutual knowledge distillation for fake news detection. Inf. Sci. 2024, 664, 120310. [Google Scholar] [CrossRef]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems (NIPS), Long Beach, CA, USA, 4–9 December 2017; pp. 6000–6010. [Google Scholar]
- Khan, M.; Gueaieb, W.; El Saddik, A.; Kwon, S. MSER: Multimodal speech emotion recognition using cross-attention with deep fusion. Expert Syst. Appl. 2024, 245, 122946. [Google Scholar] [CrossRef]
- Boutell, M.R.; Luo, J.; Shen, X.; Brown, C.M. Learning multi-label scene classification. Pattern Recognit. 2004, 37, 1757–1771. [Google Scholar] [CrossRef]
- Zhang, M.-L.; Li, Y.-K.; Liu, X.-Y.; Geng, X. Binary relevance for multi-label learning: An overview. Frontiers Comput. Sci. 2018, 12, 191–202. [Google Scholar] [CrossRef]
- Yang, Z.; Emmert-Streib, F. Optimal performance of binary relevance CNN in targeted multi-label text classification. Knowl.-Based Syst. 2024, 284, 111286. [Google Scholar] [CrossRef]
- Tsoumakas, G.; Katakis, I. Multi-label classification: An overview. Int. J. Data Warehous. Mining (IJDWM) 2007, 3, 1–13. [Google Scholar] [CrossRef]
- Li, R.; Liu, W.; Lin, Y.; Zhao, H.; Zhang, C. An ensemble multi-label classification for disease risk prediction. J. Healthcare Eng. 2017, 2017, 8051673. [Google Scholar] [CrossRef]
- Read, J.; Pfahringer, B.; Holmes, G.; Frank, E. Classifier chains for multi-label classification. Mach. Learn. 2011, 85, 333–359. [Google Scholar] [CrossRef]
- Trajdos, P.; Kurzynski, M. Dynamic classifier chains for multi-label learning. In Proceedings of the 2019 DAGM German Conference on Pattern Recognition (GCPR), Dortmund, Germany, 10–13 September 2019; pp. 567–580. [Google Scholar]
- Komatsu, T.; Watanabe, S.; Miyazaki, K.; Hayashi, T. Acoustic event detection with classifier chains. In Proceedings of the Interspeech 2021, Brno, Czech Republic, 30 August–3 September 2021; pp. 601–605. [Google Scholar]
- Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef]
- Asadi, S.; Roshan, S.E. A bi-objective optimization method to produce a near-optimal number of classifiers and increase diversity in bagging. Knowl.-Based Syst. 2021, 213, 106656. [Google Scholar] [CrossRef]
- Schapire, R.E.; Freund, Y. Boosting: Foundations and algorithms. Kybernetes 2013, 42, 164–166. [Google Scholar] [CrossRef]
- Aravkin, A.Y.; Bottegal, G.; Pillonetto, G. Boosting as a kernel-based method. Mach. Learn. 2019, 108, 1951–1974. [Google Scholar] [CrossRef]
- Grønlund, A.; Kamma, L.; Larsen, K.G.; Mathiasen, A.; Nelson, J. Margin-based generalization lower bounds for boosted classifiers. In Proceedings of the 33rd International Conference on Neural Information Processing Systems (NeurIPS), Vancouver, BC, Canada, 8–14 December 2019; pp. 11963–11972. [Google Scholar]
- Freund, Y.; Schapire, R.E. A decision-theoretic generalization of on-line learning and an application to boosting. J. Comput. Syst. Sci. 1997, 55, 119–139. [Google Scholar] [CrossRef]
- Schapire, R.E.; Singer, Y. Boostexter: A boosting-based system for text categorization. Mach. Learn. 2000, 39, 135–168. [Google Scholar] [CrossRef]
- Winata, G.I.; Khodra, M.L. Handling imbalanced dataset in multi-label text categorization using bagging and adaptive boosting. In Proceedings of the 2015 International Conference on Electrical Engineering and Informatics (ICEEI), Denpasar, Indonesia, 10–11 August 2015; pp. 500–505. [Google Scholar]
- Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Statist. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
- Devlin, J. BERT: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
- Yoo, J.; Cho, Y. ICSA: Intelligent chatbot security assistant using text-CNN and multi-phase real-time defense against SNS phishing attacks. Expert Syst. Appl. 2022, 207, 117893. [Google Scholar] [CrossRef]
- Zhang, W.; Yang, X.; Yang, J.; Wang, Y.; Yu, H.; Zhang, M. Optimal design of digital FIR filters based on text convolutional neural network for aliasing errors reduction in FI-DAC. IEICE Electron. Express 2024, 21, 6–12. [Google Scholar] [CrossRef]
- Li, W.; Qi, F.; Tang, M.; Yu, Z. Bidirectional LSTM with self-attention mechanism and multi-channel features for sentiment classification. Neurocomputing 2020, 387, 63–77. [Google Scholar] [CrossRef]
- Lin, H.; Zhang, S.; Li, Q.; Li, Y.; Li, J.; Yang, Y. A new method for heart rate prediction based on LSTM-BiLSTM-ATT. Measurement 2023, 207, 112384. [Google Scholar] [CrossRef]
- Zhang, X.; Yang, Y.; Liu, J.; Zhang, Y.; Zheng, Y. A CNN-BiLSTM monthly rainfall prediction model based on SCSSA optimization. IEICE Electron. Express 2024, 15, 4862–4876. [Google Scholar] [CrossRef]
- Li, X.; Wei, Z.; Lyu, Z.; Yuan, X.; Xu, J.; Zhang, Z. Federated reinforcement learning based on multi-head attention mechanism for vehicle edge caching. In Proceedings of the 17th International Conference on Wireless Algorithms, Systems, and Applications (WASA), Dalian, China, 24–26 November 2022; pp. 648–656. [Google Scholar]
Form | Number of Texts | Proportion of Samples |
---|---|---|
Ticketing service | 2988 | 53.39% |
Irregular flight service | 1380 | 24.66% |
Check-in and boarding | 468 | 8.36% |
Baggage services | 248 | 4.43% |
Air services | 127 | 2.27% |
Member services | 108 | 1.93% |
First aid | 99 | 1.77% |
Notice of information | 75 | 1.34% |
Special passenger services | 35 | 0.63% |
Terminal basic services | 29 | 0.52% |
Overselling | 25 | 0.45% |
Airport merchant services | 10 | 0.18% |
Ground transportation services | 5 | 0.09% |
Model | F1 Value | Recall | Hamming Loss |
---|---|---|---|
BR-GBDT | 0.7200 | 0.7504 | 0.0856 |
Multi-Head Co-attention | 0.6952 | 0.7086 | 0.0905 |
MAG (Ours) | 0.7260 | 0.7656 | 0.0847 |
Ablation Experiment | F1 Value | Recall | Hamming Loss |
---|---|---|---|
Model 1 | 0.7130 | 0.7549 | 0.0876 |
Model 2 | 0.7164 | 0.7656 | 0.0876 |
Model 3 | 0.7073 | 0.7362 | 0.0876 |
Model 4 | 0.7182 | 0.7656 | 0.0868 |
MAG (Ours) | 0.7260 | 0.7656 | 0.0847 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Cai, H.; Shao, X.; Zhou, P.; Li, H. Multi-Label Classification of Complaint Texts: Civil Aviation Service Quality Case Study. Electronics 2025, 14, 434. https://doi.org/10.3390/electronics14030434
Cai H, Shao X, Zhou P, Li H. Multi-Label Classification of Complaint Texts: Civil Aviation Service Quality Case Study. Electronics. 2025; 14(3):434. https://doi.org/10.3390/electronics14030434
Chicago/Turabian StyleCai, Huali, Xuanya Shao, Pengpeng Zhou, and Hongtao Li. 2025. "Multi-Label Classification of Complaint Texts: Civil Aviation Service Quality Case Study" Electronics 14, no. 3: 434. https://doi.org/10.3390/electronics14030434
APA StyleCai, H., Shao, X., Zhou, P., & Li, H. (2025). Multi-Label Classification of Complaint Texts: Civil Aviation Service Quality Case Study. Electronics, 14(3), 434. https://doi.org/10.3390/electronics14030434