Aspect-Based Sentiment Analysis of Patient Feedback Using Large Language Models
Abstract
:1. Introduction
1.1. Online Patients’ Feedback
1.2. Significance of the Study
2. Related Works
Author(s) | Aspects (Entities) | Training Data | Model | Findings | Weaknesses |
---|---|---|---|---|---|
[53] | Diseases and treatments | Arabic medical journals | BERT-based model (ABioNER) | ABioNER outperformed AraBERT and the multilingual BERT model, achieving a higher F1-score, especially in “Disease or Syndrome” and “Therapeutic or Preventive Procedure” categories, confirming the potential of small-scale, domain-specific pre-training | Limited to two entity types without evaluation on broader NER tasks due to data scarcity; lacks assessment of model performance on additional Arabic biomedical entity types or relation extraction tasks. |
[58] | Nan | COVID-19 Category, Vaccine Sentiment (VC), Maternal Vaccine Stance (MVS), Stanford Sentiment Treebank 2 (SST-2) and Twitter Sentiment SemEval (SE) | BERT-based model CT-BERT | CT-BERT showed superior performance over the BERT-LARGE model with up to 30% improvement in classification tasks, especially in COVID-19 and health-related datasets. It yielded a mean F1 score improvement across datasets, with the most significant gains in COVID-19-specific contexts | Limited to classification tasks; not assessed for other NLP tasks like named entity recognition. Potential for further improvement through hyperparameter optimization. Additionally, only one COVID-19-specific dataset was available for fine-tuning and evaluation. |
[57] | POS tagging, NER and textclassification | WNUT16 NER, WNUT17, oct27.traindev andoct27.test etc. | BERTweet-language model based on BERT | BERTweet outperformed baselines RoBERTa and XLM-R on tweet NLP tasks, achieving new state-of-the-art results, particularly in NER (+14% improvement) and text classification tasks for sentiment and irony detection (+5% and +4% improvement). “Soft” normalization outperformed “hard” lexical normalization for this domain | Although effective on the targeted tasks, BERTweet lacks evaluation on broader NLP tasks outside of Twitter and may not generalize beyond tweet-style text. No “large” version of BERTweet was included, limiting comparisons to larger models like RoBERTa-large |
[61] | medical terms: the Clinical finding, Substance, Body parts, Procedure, and pharmaceutical product | Corpus collected from social media health related posts | COMET-A, a corpus for medical entity linking in social media | Neural methods, combined with dictionary and string-matching baselines, were effective, though COMETA posed significant challenges, especially in zero-shot scenarios. The study observed a 28–46% gap from perfect performance, highlighting the complexity of social media language and the need for multi-view (text and graph) models for improved EL in health contexts. | Performance in zero-shot settings remains challenging due to the highly diverse and informal nature of social media language. The study noted limited performance improvements even with neural baselines, highlighting difficulties in capturing layman terminology in medical contexts. |
[59] | patient’s health conditions and adverse drug reactions | 1.4 million health-related posts online | RuDR-BERT | RuDR-BERT outperformed multilingual and Russian-language BERT baselines, achieving higher F1 scores across both NER and multilabel sentence classification tasks, particularly in detecting ADRs and DI. The model revealed the difficulties of handling diverse regular person language used in patient reviews, particularly in ADRs. | Limited to Russian-language text, and the annotated portion includes only a subset of therapeutic categories, potentially limiting generalizability. Further limitations include the challenge of mapping informal language in user-generated content to formal medical terminology |
[62] | Disease, drugs (chemical), drug (protein), species | NCBI, JNLPBA, BC2GM, LINNAEUS, Species-800, BC5CDR, BC4CHEMD | BioALBERT | BioALBERT outperformed other models, including BioBERT, on eight benchmark BioNER datasets, achieving significant F1 score improvements across all tested categories, especially in Drug/Chem and Disease categories. It also demonstrated faster training speeds and lower memory usage than BERT-based models. | BioALBERT’s training is limited to biomedical corpora. Currently evaluated only on BioNER tasks. |
3. Methodology
3.1. DataSet
3.2. Methods
3.2.1. Content Analysis
3.2.2. Data Pre-Processing
3.2.3. Textual Cohesion and Readability Analysis
3.2.4. Credibility Analysis
3.2.5. Aspect-Based Sentiment Analysis
3.2.6. Architectures of Deep Learning Models
3.2.7. Evaluation of ABSA Model Based on BERT
3.2.8. Overfitting Handling
Regularization Techniques
Cross-Validation Approach
Evaluation on a Held-Out Test Set
3.2.9. ChatGPT Few-Shots
3.2.10. Comparison Between the Models
3.3. Human Validation
4. Results
4.1. Patient Profiles Findings
4.2. Wording Findings
4.3. Aspect-Based Sentiment Analysis Findings
4.4. Deep Learning Models Findings
5. Discussion
5.1. Implications
5.2. Future Directions
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Data Availability Statement
Conflicts of Interest
References
- Dhajate, N.; Joshi, R. Classification of reviews of e-healthcare services to improve patient satisfaction: Insights from an emerging economy. J. Bus. Res. 2023, 164, 114015. [Google Scholar] [CrossRef] [PubMed]
- Baker, P.; Brookes, G.; Evans, C. The language of patient feedback: A corpus linguistic study of online health communication. Engl. Specif. Purp. 2019, 56, 1–2. [Google Scholar] [CrossRef]
- Farnood, A.; Johnston, B.; Mair, F.S. A mixed methods systematic review of the effects of patient online self-diagnosing in the ‘smart-phone society’ on the healthcare professional-patient relationship and medical authority. BMC Med. Inf. Decis. Mak. 2020, 20, 1–14. [Google Scholar] [CrossRef]
- NHS. Involving People in Their Own Health and Care: Statutory Guidance for Clinical Commissioning Groups and NHS England; NHS: London, UK, 2017. [Google Scholar]
- NHS. GP Patient Survey 2015–2016. Available online: https://www.england.nhs.uk/statistics/2016/07/07/gp-patient-survey-2015-16/ (accessed on 4 May 2023).
- Cuteanu, A.; Seguin, M.; Ziebland, S.; Pope, C.; Leydon, G.; Barnes, R.; Murray, E.; Atherton, H.; Stevenson, F. Qualitative study: Patients’ enduring concerns about discussing internet use in general practice consultations. BMJ Open 2021, 11, e047508. [Google Scholar] [CrossRef]
- Hedges, L.; Couey, C. How Patients Use Online Reviews. Available online: https://www.softwareadvice.com/resources/how-patients-use-online-reviews/ (accessed on 12 March 2024).
- Powell, J.; Athertono, H.; Williams, V.; Mazanderani, F.; Dudhwala, F.; Woolgaro, S.; Boylano, A.-M.; Fleming, J.; Kirkpatrick, S.; Martin, A.; et al. Using online patient feedback to improve NHS services: The INQUIRE multimethod study. Health Serv. Deliv. Res. 2019, 7, 1–150. [Google Scholar] [CrossRef]
- Atherton, H.; Fleming, J.; Williams, V.; Powell, J. Online patient feedback: A cross-sectional survey of the attitudes and experiences of United Kingdom health care professionals. J. Health Serv. Res. Policy 2019, 24, 235–244. [Google Scholar] [CrossRef]
- Boylano, A.-M.; Turk, A.; van Velthoven, M.H.; Powell, J. Online patient feedback as a measure of quality in primary care: A multimethod study using correlation and qualitative analysis. BMJ Open 2020, 10, e031820. [Google Scholar] [CrossRef]
- Dudhwala, F.; Boylano, A.-M.; Williams, V.; Powell, J. VIEWPOINT: What counts as online patient feedback, and for whom? Digit. Health 2017, 3, 2055207617728186. [Google Scholar] [CrossRef]
- Entwistle, V.; Firnigl, D.; Ryan, M.; Francis, J.; Kinghorn, P. Which experiences of health care delivery matter to service users and why? A critical interpretive synthesis and conceptual map. J. Health Serv. Res. Policy 2012, 17, 70–80. [Google Scholar] [CrossRef]
- Vahdat, S.; Hamzehgardeshi, L.; Hessam, S.; Hamzehgardeshi, Z. Patient Involvement in Health Care Decision Making: A Review. Iran. Red Crescent Med. J. 2014, 16, e12454. [Google Scholar] [CrossRef]
- NIHR. Improving Care by Using Patient Feedback. Available online: https://content.nihr.ac.uk/nihrdc/themedreview-04327-PE/Patient-Feedback-WEB.pdf (accessed on 2 April 2024).
- Bez, S.M.; Georgescu, I.; Farazi, M.S. TripAdvisor of healthcare:Opportunities for value creation through patient feedback platforms. Technovation 2023, 121, 102625. [Google Scholar] [CrossRef]
- Björklund, K.; Stenfors, T.; Nilsson, G.; Leanderson, C. Learning from patients’ written feedback: Medical students’ experiences. Int. J. Med. Educ. 2022, 31, 19–27. [Google Scholar] [CrossRef] [PubMed]
- Eijkelboom, M.C.L.; de Kleijn, R.A.M.; van Diemen, W.J.M.; Maljaars, C.D.N.; van der Schaaf, P.; Frenkel, P. Patients as Feedback Providers: Exploring Medical Students’ Credibility Judgments. Perspect. Med. Educ. 2023, 12, 129–140. [Google Scholar] [CrossRef] [PubMed]
- Fox, S. The Social Life of Health Information; Pew Research Center: Washington, DC, USA, 2011. [Google Scholar]
- Jia, X.; Pang, Y.; Liu, L. Sally Online Health Information Seeking Behavior: A Systematic Review. Healthcare 2021, 9, 1740. [Google Scholar] [CrossRef]
- Gupta, A.; Tyagi, M.; Sharma, D. Use of social media marketing in healthcare. J. Health Manag. 2013, 15, 293–302. [Google Scholar] [CrossRef]
- Coulson, N.S. How do online patient support communities affect the experience of inflammatory bowel disease? An online survey. JRSM 2013, 4, 2042533313478004. [Google Scholar] [CrossRef]
- Hossain, S.N.; Jaglal, S.B.; Shepherd, J.; Perrier, L.; Tomasone, J.R.; Sweet, S.N.; Luong, D.; Allin, S.; Nelson, M.L.; Guilcher, S.J.; et al. Web-Based Peer Support Interventions for Adults Living with Chronic Conditions: Scoping Review. JMIR Rehabil. Assist. Technol. 2021, 8, e14321. [Google Scholar] [CrossRef]
- Bartlett, Y.K.; Coulson, N.S. An investigation into the empowerment effects of using online support groups and how this affects health professional/patient communication. Patient Educ. Couns. 2011, 83, 113–119. [Google Scholar] [CrossRef]
- Ziebland, S.; Powell, J.; Briggs, P.; Jenkinson, C.; Wyke, S.; Sillence, E.; Harris, P.; Perera, R.; Mazanderani, F.; Martin, A.; et al. Examining the Role of Patients’ Experiences as a Resource for Choice and Decision-Making in Health Care: A Creative, Interdisciplinary Mixed-Method Study in Digital Health; NIHR Journals Library: Southampton, UK, 2016. [Google Scholar]
- Harrington, M.; Heywood, B.; Rura, S.; Wicks, P. PRM95—An open research exchange for online patient feedback in pro development. Value Health 2014, 17, A197. [Google Scholar] [CrossRef]
- Vilendrer, S.; Levoy, E.; Miller-Kuhlmann, R.; Amano, A.; Brown-Johnson, C.; De Borba, L.; Luu, J.H.; Sakamuri, S.; Gold, C.A. Physician Perceptions of Performance Feedback and Impact on Personal Well-Being: A Qualitative Exploration of Patient Satisfaction Feedback in Neurology. Jt. Comm. J. Qual. Patient Saf. 2023, 49, 138–148. [Google Scholar] [CrossRef]
- Schäfer, H.; Idrissi-Yaghir, A.; Bewersdorff, J.; Frihat, S.; Friedrich, C.M.; Zesch, T. Medication event extraction in clinical notes: Contribution of the WisPerMed team to the n2c2 2022 challenge. J. Biomed. Inform. 2023, 143, 104400. [Google Scholar] [CrossRef] [PubMed]
- McMaster, C.; Chan, J.; Liew, D.F.; Su, E.; Frauman, A.G.; Chapman, W.W.; Pires, D.E. Developing a deep learning natural language processing algorithm for automated reporting of adverse drug reactions. J. Biomed. Inform. 2023, 137, 104265. [Google Scholar] [CrossRef] [PubMed]
- Mazanderani, F.; Kirkpatrick, S.F.; Ziebland, S.; Locock, L.; Powell, J. Caring for care: Online feedback in the context of public healthcare services. Soc. Sci. Med. 2021, 285, 114280. [Google Scholar] [CrossRef] [PubMed]
- Liu, W.; Tang, J.; Qin, J.; Xu, L.; Li, Z.; Liang, X. MedDG: A Large-scale Medical Consultation Dataset for Building Medical Dialogue System. arXiv 2020, arXiv:2010.07497. [Google Scholar]
- Zeng, G.; Yang, W.; Ju, Y.; Wang, S.; Zhang, R.; Zhou, M.; Zeng, J.; Dong, X.; Zhang, R.; Fang, H.; et al. Meddialog: Large-scale medical dialogue datasets. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online, 16–20 November 2020; pp. 9241–9250. [Google Scholar]
- Zhang, N.; Chen, M.; Bi, Z.; Liang, X.; Li, L.; Shang, X.; Yin, K.; Tan, C.; Xu, J.; Huang, F.; et al. Cblue: A Chinese biomedical language understanding evaluation benchmark. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Dublin, Ireland, 22–27 May 2022; pp. 7888–7915. [Google Scholar]
- Kim, Y.M.; Lee, T.H.; Na, S.O. Constructing novel datasets for intent detection and Ner in a Korean healthcare advice system: Guidelines and empirical results. Appl. Intell. 2022, 53, 941–961. [Google Scholar] [CrossRef]
- Gu, Y.; Tinn, R.; Cheng, H.; Lucas, M.; Usuyama, N.; Liu, X.; Naumann, T.; Gao, J.; Poon, H. Domain-Specific Language Model Pretraining for Biomedical Natural Language Processing. ACM Trans. Comput. Healthc. 2020, 3, 1–23. [Google Scholar] [CrossRef]
- Jin, Q.; Dhingra, B.; Liu, Z.; Cohen, W.W.; Lu, X. Pubmedqa: A dataset for biomedical research question answering. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP, Hong Kong, China, 7 November 2019; pp. 2567–2577. [Google Scholar]
- Huy, T.D.; Tu, N.A.; Vu, T.H.; Minh, N.P.; Phan, N.; Bui, T.H.; Truong, S.Q. ViMQ: A Vietnamese Medical Question Dataset for Healthcare Dialogue System Development. In Neural Information Processing; Springer: Berlin/Heidelberg, Germany, 2023. [Google Scholar]
- Mondal, I.; Ahuja, K.; Jain, M.; O’Neill, J.; Bali, K.; Choudhury, M. Global Readiness of Language Technology for Healthcare: What Would It Take to Combat the Next Pandemic? In Proceedings of the 29th International Conference on Computational Linguistics, Gyeongju, Republic of Korea, 12–17 October 2022; pp. 4320–4335. [Google Scholar]
- Bai, G.; He, S.; Liu, K.; Zhao, J. Incremental intent detection for medical domain with contrast replay networks. In Proceedings of the Findings of the Association for Computational Linguistics: ACL 2022, Dublin, Ireland, 22–27 May 2022. [Google Scholar]
- Bao, Q.; Ni, L.; Liu, J. Hhh: An online medical chatbot system based on knowledge graph and hierarchical bi-directional attention. In Proceedings of the Australasian Computer Science Week Multiconference, Melbourne, VIC, Australia, 4–6 February 2020; pp. 1–10. [Google Scholar]
- Chen, Q.; Zhuo, Z.; Wang, W. BERT for Joint Intent Classification and Slot Filling. arXiv 2019, arXiv:1902.10909. [Google Scholar]
- Mehta, D.; Santy, S.; Mothilal, R.K.; Srivastava, B.M.L.; Sharma, A.; Shukla, A.; Prasad, V.; Sharma, A.; Bali, K. Learnings from technological interventions in a low resource language: A case-study on Gondi. In Proceedings of the Twelfth Language Resources and Evaluation Conference, Marseille, France, 11–16 May 2020; pp. 2832–2838. [Google Scholar]
- Daniel, J.E.; Brink, W.; Eloff, R.; Copley, C. Towards automating healthcare question answering in a noisy multilingual low-resource setting. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 948–953. [Google Scholar]
- Mullick, A.; Mondal, I.; Ray, S.; Raghav, R.; Chaitanya, G.S.; Goyal, P. Intent Identification and Entity Extraction for Healthcare Queries in Indic Languages. arXiv 2023, arXiv:2302.09685. [Google Scholar]
- Kotei, E.; Thirunavukarasu, R. A Systematic Review of Transformer-Based Pre-Trained Language Models through Self-Supervised Learning. Information 2023, 14, 187. [Google Scholar] [CrossRef]
- Smailhodzic, E.; Hooijsma, W.; Boonstra, A.; Langley, D.J. David Social media use in healthcare: A systematic review of effects on patients and on their relationship with healthcare professionals. BMC Health Serv. Res. 2016, 16, 1–14. [Google Scholar] [CrossRef]
- Broom, A. The eMale: Prostate cancer, masculinity and online support as a challenge to medical expertise. J. Sociol. 2005, 41, 87–104. [Google Scholar] [CrossRef]
- Chiu, Y.C.; Hsieh, Y.L. Communication online with fellow cancer patients: Writing to be remembered, gain strength, and find survivors. J. Health Psychol. 2013, 18, 1572–1581. [Google Scholar] [CrossRef] [PubMed]
- Armstrong, N.; Powell, J. Patient perspectives on health advice posted on Internet discussion boards: A qualitative study. Health Expect. 2009, 12, 313–320. [Google Scholar] [CrossRef] [PubMed]
- Hadert, A.; Rodham, K. The invisible reality of arthritis: A qualitative analysis of an online message board. Invis. Real. Arthritis: A Qual. Anal. Online Message Board 2008, 6, 181–196. [Google Scholar] [CrossRef] [PubMed]
- Shigaki, C.L.; Smarr, K.L.; Gong, Y.; Donovan-Hanson, K.; Siva, C.; Johnson, R.A.; Ge, B.; Musser, D.R. Social interactions in an online self-management program for rheumatoid arthritis. Chronic Illn. 2008, 4, 239–246. [Google Scholar] [CrossRef]
- Lempp, H.; Scott, D.; Kingsley, G. The personal impact of rheumatoid arthritis on patients’ identity: A qualitative study. Chronic Illn. 2006, 2, 109–120. [Google Scholar] [CrossRef] [PubMed]
- Zhao, Y.; Zhang, L.; Zeng, C.; Lu, W.; Chen, Y.; Fan, T. Construction of an aspect-level sentiment analysis model for online medical reviews. Inf. Process. Manag. 2023, 60, 103513. [Google Scholar] [CrossRef]
- Boudjellal, N.; Zhang, H.; Khan, A.; Ahmad, A.; Naseem, R.; Shang, J.; Dai, L. ABioNER: A BERT-based model for Arabic biomedical named-entity recognition. Complexity 2021, 2021, 6633213. [Google Scholar] [CrossRef]
- Kalyan, K.S.; Rajasekharan, A.; Sangeetha, S. AMMU: A survey of transformer-based biomedical pretrained language models. J. Biomed. Inform. 2022, 126, 103982. [Google Scholar] [CrossRef]
- Nerella, S.; Bandyopadhyay, S.; Zhang, J.; Contreras, M.; Siegel, S.; Bumin, A.; Silva, B.; Sena, J.; Shickel, B.; Bihorac, A.; et al. Transformers in Healthcare: A Survey. arXiv 2023, arXiv:2307.00067. [Google Scholar]
- Wang, B.; Xie, Q.; Pei, J.; Chen, Z.; Tiwari, P.; Li, Z.; Fu, J. Pre-trained Language Models in Biomedical Domain: A Systematic Survey. ACM Comput. Surv. 2023, 56, 1–52. [Google Scholar] [CrossRef]
- Nguyen, D.Q.; Vu, T.; Nguyen, A.T. Bertweet: A pre-trained language model for English tweets. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, Online, 16–20 November 2020; pp. 9–14. [Google Scholar]
- Müller, M.; Salathé, M.; Per, E.; Kummervold, P.E. COVID-Twitter-BERT: A natural language processing model to analyse COVID-19 content on Twitter. Front. Artif. Intell. 2023, 6, 1023281. [Google Scholar] [CrossRef] [PubMed]
- Tutubalina, E.; Alimova, I.; Miftahutdinov, Z.; Sakhovskiy, A.; Malykh, V.; Nikolenko, S. The russian drug reaction corpus and neural models for drug reactions and effectiveness detection in user reviews. Bioinformatics 2021, 37, 243–249. [Google Scholar] [CrossRef] [PubMed]
- Madan, M.; Madan, R.; Thakur, P. Analysing the patient sentiments in healthcare domain using Machine learning. Procedia Comput. Sci. 2024, 238, 683–690. [Google Scholar] [CrossRef]
- Basaldella, M.; Liu, F.; Shareghi, E.; Collier, N. Cometa: A corpus for medical entity linking in the social media. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online, 16–20 November 2020; pp. 3122–3137. [Google Scholar]
- Naseem, U.; Khushi, M.; Reddy, V.; Rajendran, S.; Razzak, I.; Kim, J. Bioalbert: A simple and effective pre-trained language model for biomedical named entity recognition. In Proceedings of the 2021 International Joint Conference on Neural Networks (IJCNN), Shenzhen, China, 18–22 July 2021. [Google Scholar]
- Clusmann, J.; Kolbinger, F.R.; Muti, H.S.; Carrero, Z.I.; Eckardt, J.N.; Laleh, N.G.; Löffler, C.M.L.; Schwarzkopf, S.C.; Unger, M.; Veldhuizen, G.P.; et al. The future landscape of large language models in medicine. Commun. Med. 2023, 3, 141. [Google Scholar] [CrossRef]
- Peng, C.; Yang, X.; Chen, A.; Smith, K.E.; PourNejatian, N.; Costa, A.B.; Martin, C.; Flores, M.G.; Zhang, Y.; Magoc, T.; et al. A study of generative large language model for medical research and healthcare. NPJ Digit. Med. 2023, 6, 210. [Google Scholar] [CrossRef]
- Thirunavukarasu, A.J.; Ting, D.S.J.; Elangovan, K.; Gutierrez, L.; Tan, T.F.; Ting, D.S.W. Large language models in medicine. Nat. Med. 2023, 29, 1930–1940. [Google Scholar] [CrossRef]
- Wornow, M.; Xu, Y.; Thapa, R.; Patel, B.; Steinberg, E.; Fleming, S.; Pfeffer, M.A.; Fries, J.; Shah, N.H. The shaky foundations of large language models and foundation models for electronic health records. NPJ Digit. Med. 2023, 6, 135. [Google Scholar] [CrossRef]
- Patient.Info. About-Us. Available online: https://patient.info/about-us (accessed on 15 December 2023).
- Bond, C.S.; Ahmed, O.H. Can I help you? Information sharing in online discussion forums by people living with a long-term condition. J. Innov. Health Inf. 2016, 23, 620–626. [Google Scholar] [CrossRef]
- Hsieh, H.F.; Shannon, S.E. Three Approaches to Qualitative Content Analysis. Qual. Health Res. 2005, 15, 1277–1289. [Google Scholar] [CrossRef]
- Glenn. Analyzing Open-Ended Questions. Available online: http://intelligentmeasurement.net/2007/12/18/analyzing-open-ended-questions/ (accessed on 20 May 2024).
- Akgül, Y. Evaluating the performance of websites from a public value, usability, and readability perspectives: A review of Turkish national government websites. Univers. Access Inf. Soc. 2024, 23, 975–990. [Google Scholar] [CrossRef]
- Flesch, R. A new readability yardstick. J. Appl. Psychol. 1948, 23, 221–233. [Google Scholar] [CrossRef] [PubMed]
- Zellers, R.; Holtzman, A.; Rashkin, H.; Bisk, Y.; Farhadi, A.; Roesner, F.; Choi, Y. Defending against neural fake news. Adv. Neural Inf. Process. Syst. 2019, 32. [Google Scholar] [CrossRef]
- Fogg, B.J. Persuasive Technology: Using Computers to Change What We Think and Do; Morgan Kaufmann: Burlington, MA, USA, 2003. [Google Scholar]
- Jin, W.; Zhao, L.; Liu, C.; Yu, H. Back to common sense: Oxford dictionary descriptive knowledge augmentation for aspect-based sentiment analysis. Inf. Process. Manag. 2023, 60, 103260. [Google Scholar] [CrossRef]
- Yang, H.; Zeng, B.; Xu, M.; Wang, T. Back to Reality: Leveraging Pattern-driven Modeling to Enable Affordable Sentiment Dependency Learning. arXiv 2021, arXiv:2110.08604. [Google Scholar]
- Cohen, S.; Presil, D.; Katz, O.; Arbili, O.; Messica, S.; Rokach, L. Enhancing social network hate detection using back translation and GPT-3 augmentations during training and test-time. Inf. Fusion 2023, 99, 101887. [Google Scholar] [CrossRef]
- Rolczyński, R. Do You Trust in Aspect-Based Sentiment Analysis? Testing and Explaining Model Behaviors. In Rafał Rolczyński; SCALAC SP. Z O. O.: Warsaw, Poland, 2021; Volume 2023. [Google Scholar]
- Wang, H.; Li, J.; Hua, W.; Hovy, E.; Sun, Y. Pre-Trained Language Models and Their Applications. Engineering 2022, 25, 51–65. [Google Scholar] [CrossRef]
- Kamış, S.; Goularas, D. Evaluation of Deep Learning Techniques in Sentiment Analysis from Twitter Data. In Proceedings of the 2019 International Conference on Deep Learning and Machine Learning in Emerging Applications (Deep-ML), Istanbul, Turkey, 26–28 August 2019. [Google Scholar]
- Singhal, K.; Azizi, S.; Tu, T.; Mahdavi, S.S.; Wei, J.; Chung, H.W.; Scales, N.; Tanwani, A.; Cole-Lewis, H.; Pfohl, S.; et al. Large language models encode clinical knowledge. Nature 2023, 620, 172–180. [Google Scholar] [CrossRef]
- Lu, H.Y.; Liu, T.C.; Cong, R.; Yang, J.; Gan, Q.; Fang, W.; Wu, X.J. QAIE: LLM-based Quantity Augmentation and Information Enhancement for few-shot Aspect-Based Sentiment Analysis. Inf. Process. Manag. 2025, 62, 103917. [Google Scholar] [CrossRef]
- Zhang, H.; Zhang, Y.; Zhan, L.M.; Chen, J.; Shi, G.; Lam, A.; Wu, X.M. Effectiveness of Pre-Training for Few-Shot Intent Classification. arXiv 2021, arXiv:2109.05782. [Google Scholar]
- Petrič, G.; Cugmas, M.; Petrič, R.; Atanasova, S. The quality of informational social support in online health communities: A content analysis of cancer-related discussions. Digit. Health 2023, 9, 20552076231155681. [Google Scholar] [CrossRef] [PubMed]
- Srimarut, T.; Techasatian, K. Use of social media in health care by patients and health care professionals: Motives & barriers in Thailand. Utopía Y Prax. Latinoam. 2019, 24, 215–223. [Google Scholar]
- Loo, J.; Greaves, G.; Lewis, P.J. Exploring patients’ pharmacy stories: An analysis of online feedback. Int. J. Clin. Pharm. 2021, 43, 1584–1593. [Google Scholar] [CrossRef] [PubMed]
Topic | Aspects (These Words Were Used by Patients and Bearing Misspelling)-Sample from Feedbacks Under Bone-Related Disease/Disorder |
---|---|
Medicines | cortisone, acupuncture, MSM, Pentosan Polysulfate (PPS), Celebrex, Pazital and Lyrica, icyhot cream/icyhot patches, Biofreeze, Co-codamol, Naproxen and codeine, Bedranol, opiods, opioids, tramadol, tapered steroid, NSAID Etova, vicodin, prednisone, polymyalgia rheumatica, acetaminophen, glucosamine, prescription, zomorph, Flexiseq dihydrocodeine naproxen gabapentin amytriptaline |
Complaints (parts of human body about which patients complain) | collar bone, palm, spinal circles, toes, leg, thumb, neck, low back, hip & pelvis, thumbs, right knee, right hip, right shoulder, SI joints, spine, ankle, fingers, jaw and foot, Joint, joint replacement relief, thumb, wrist, left thumb, cramps in my legs, tingling, finger, pinky, bone spur, hip bursitis |
Medical procedures | ultra-sound, X ray, ACDF surgery, bilateral TKR, knee replacement, adipose stem cell & PRP, MRI and a spinal surgeon, chiropractor, Cac surgery, Pentosan Polysulfate, physical therapy, suture anchor, ankle fusion, kinsiology tape, Brodtens bracelet, X-rays and blood tests |
Diagnoses | arthritis, osteoarthritis, polymyalgia rheumatica, osteoporosis, OE, Fibromyalgia, sciatica, Inflammation, Stiffness, herniated disc, MCAS, osteopenia, rhematic, osteo |
Medical staff | orthopaedist, doctor, surgeon, physios, drs, DOC, rheumatologist, GP, physiotherapist |
Activities | stretching, & walk when I can. Have a heating blanket, sports |
Food and herbs | cider vinegar, fish oils, CBD oil, glucosamine sulphate, Capsaicin cream, turmeric |
Measure | Value |
---|---|
mean | 65.05 |
std | 105.39 |
min | −3939.59 |
max | 206.84 |
Measure | Value |
---|---|
mean | 0.27 |
std | 0.54 |
min | 0.0 |
max | 4.00 |
Architecture | Original Model | Fine-Tuned Model | Fine-Tuned Model with LSTM | Bidirectional LSTM Model |
---|---|---|---|---|
Architecture | Embedding | Embedding | Embedding | Embedding |
LSTM | Bidirectional | |||
Flatten | Flatten | Dropout | LSTM Dropout | |
Dense | Dense | LSTM | Bidirectional | |
Dense | Dense | Dropout | LSTM Dropout | |
Output | Dense | Dense | Dense | |
Output | Output | Output | ||
Activation Functions | ReLU (except for the Output layer) | |||
Loss Function: | Sparse Cross Entropy Loss | |||
Optimizer/Epochs/Batch Size | Adam/20/32 | |||
Regularization | None | None | Dropout (0.2) | Dropout (0.2) |
Model | Output1 (Medicine): (MSE) | Output2 (Pain Locations): (MSE) | Output3 (Medical Procedures): (MSE) | Output4 (Diagnoses): (MSE) | Output5 (Medical Staff): (MSE) | Output6 (Food): (MSE) | Output7 (Activity): (MSE) | Overall MSE |
---|---|---|---|---|---|---|---|---|
Original model | 0.1704 | 0.4423 | 0.2604 | 0.3082 | 0.2800 | 0.0174 | 0.1749 | 0.2362 |
Fine-tuned model | 0.1777 | 0.4451 | 0.2830 | 0.3159 | 0.2752 | 0.0165 | 0.1817 | 0.2421 |
Fine-tuned model using LSTM layers | 0.1737 | 0.5096 | 0.3716 | 0.3824 | 0.4106 | 0.0156 | 0.2374 | 0.3001 |
Bidirectional LSTM model with Dropout layers | 0.1605 | 0.5564 | 0.3580 | 0.3813 | 0.3672 | 0.0142 | 0.1879 | 0.2894 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Alkhnbashi, O.S.; Mohammad, R.; Hammoudeh, M. Aspect-Based Sentiment Analysis of Patient Feedback Using Large Language Models. Big Data Cogn. Comput. 2024, 8, 167. https://doi.org/10.3390/bdcc8120167
Alkhnbashi OS, Mohammad R, Hammoudeh M. Aspect-Based Sentiment Analysis of Patient Feedback Using Large Language Models. Big Data and Cognitive Computing. 2024; 8(12):167. https://doi.org/10.3390/bdcc8120167
Chicago/Turabian StyleAlkhnbashi, Omer S., Rasheed Mohammad, and Mohammad Hammoudeh. 2024. "Aspect-Based Sentiment Analysis of Patient Feedback Using Large Language Models" Big Data and Cognitive Computing 8, no. 12: 167. https://doi.org/10.3390/bdcc8120167
APA StyleAlkhnbashi, O. S., Mohammad, R., & Hammoudeh, M. (2024). Aspect-Based Sentiment Analysis of Patient Feedback Using Large Language Models. Big Data and Cognitive Computing, 8(12), 167. https://doi.org/10.3390/bdcc8120167