Nested Sentiment Analysis for ESG Impact: Leveraging FinBERT to Predict Market Dynamics Based on Eco-Friendly and Non-Eco-Friendly Product Perceptions with Explainable AI
Abstract
:1. Introduction
2. Related Studies
3. Proposed Method
3.1. Materials
Dataset Creation and ESG Score Calculation
3.2. Conceptual Framework
Algorithm 1. Sentiment Analysis for ESG Impact Using FinBERT and Explainable AI. | ||||
Require Dataset D with sentences, products labelled as eco-friendly and non-eco-friendly, related sentiments, and ESG scores. | ||||
1. Input: Sentiment dataset D. | ||||
2. Tokenization: | ||||
Tokenize each sentence s∈D using FinBERT’s tokenizer. | ||||
T(s) = Tokenize(s) | ||||
3. Feature Extraction: | ||||
For | ||||
each sentence s∈D, apply tokenization to generate input features Xs. | ||||
Xs = FeatureExtraction(T(s)) | ||||
4. Dataset Splitting: | ||||
Split dataset D into training set Dtrain and testing set Dtest using stratified sampling. | ||||
Dtrain,Dtest = StratifiedSplit(D) | ||||
5. Model Fine-Tuning (with AdamW Optimizer and L2 Regularization): | ||||
Fine-tune FinBERT on the training set Dtrain using the Trainer API, | ||||
Let the model parameters be denoted by θ. | ||||
θ∗ = arg minθ+ | ||||
Where is the loss function, typically cross-entropy loss. and λ is the regularization strength for L2 regularization. AdamW is used as the optimizer to minimize this loss. | ||||
6. Training Process: | ||||
For | ||||
each epoch e: | ||||
Update the model parameters θ to minimize the loss using AdamW optimizer and L2 regularization: | ||||
θe+1 = θe − η∇θe (fθe(Xs),ys) + | ||||
Where η is the learning rate, and λ represents the L2 regularization term. | ||||
7. Model Evaluation: | ||||
Evaluate the model performance on the validation set Dval using accuracy, where Acc represents accuracy. | ||||
Acc = | ||||
8. Final Model Evaluation: | ||||
Evaluate the final model on the test set Dtest, including accuracy and additional metrics such as precision, recall, and F1-score. | ||||
9. Explainability with LIME: | ||||
Utilize XAI (LIME) to interpret and visualize predictions for specific instances si providing local explanations by highlighting important features that drive predictions. | ||||
LIME(si) = LocalExplanation(si,fθ) |
3.3. LIME (Local Interpretable Model-Agnostic Explanations) Process
- 1.
- Data Perturbation: LIME generates multiple variations of the input sentence by altering features.
- 2.
- Black-Box Predictions: Each perturbed sentence is fed into the original model (e.g., FinBERT), and the model predicts the sentiment;
- 3.
- Simple Model Fitting: LIME fits a simpler model (such as a linear model) to approximate how each word in the sentence contributes to the prediction.
- 4.
- Local Explanation: The simpler model provides local explanations for the prediction, identifying which features were most influential.
- 5.
- Final Interpretation: LIME presents a human-readable explanation, showing how much each feature influenced the prediction.
4. Results and Analysis
4.1. Measurement Metrics
Metrics Applied
4.2. Performance Metrics Summary for ESG Scores
4.2.1. Model Training and Validation Performance
4.2.2. Model Evaluation on Test Data
4.2.3. Comparative Analysis with Existing Models
4.2.4. Classification Performance Metrics
4.3. Result Visualization
4.4. Model Interpretability Using XAI with LIME
5. Discussion
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Chou, J.S.; Chong, P.L.; Liu, C.Y. Deep learning-based chatbot by natural language processing for supportive risk management in river dredging projects. Eng. Appl. Artif. Intell. 2024, 131, 107744. [Google Scholar] [CrossRef]
- Sharifi, A.; Allam, Z.; Bibri, S.E.; Khavarian-Garmsir, A.R. Smart cities and sustainable development goals (SDGs): A systematic literature review of co-benefits and trade-offs. Cities 2024, 146, 104659. [Google Scholar] [CrossRef]
- Roumeliotis, K.I.; Tselikas, N.D.; Nasiopoulos, D.K. LLMs and NLP Models in Cryptocurrency Sentiment Analysis: A Comparative Classification Study. Big Data Cogn. Comput. 2024, 8, 63. [Google Scholar] [CrossRef]
- Krastev, B.; Krasteva-Hristova, R. Challenges and Trends in Green Finance in the Context of Sustainable Development—A Bibliometric Analysis. J. Risk Financ. Manag. 2024, 17, 301. [Google Scholar] [CrossRef]
- Taddeo, S.; Agnese, P.; Busato, F. Rethinking the effect of ESG practices on profitability through cross-dimensional substitutability. J. Environ. Manag. 2024, 352, 120115. [Google Scholar] [CrossRef]
- Martiny, A.; Taglialatela, J.; Testa, F.; Iraldo, F. Determinants of environmental social and governance (ESG) performance: A systematic literature review. J. Clean. Prod. 2024, 456, 142213. [Google Scholar] [CrossRef]
- Gutierrez, A.M.J.; Chiu, A.S.F.; Seva, R. A Proposed Framework on the Affective Design of Eco-Product Labels. Sustainability 2020, 12, 3234. [Google Scholar] [CrossRef]
- Ariyer, G.; Mangla, S.K.; Chowdhury, S.; Sozen, M.E.; Kazancoglu, Y. Predictive and prescriptive analytics for ESG performance evaluation: A case of Fortune 500 companies. J. Bus. Res. 2024, 181, 114742. [Google Scholar]
- Agnese, P.; Carè, R.; Cerciello, M.; Taddeo, S. Reconsidering the impact of environmental, social and governance practices on firm profitability. Manag. Decis. 2024. [Google Scholar] [CrossRef]
- Do Jung, H.; Jang, B. Enhancing Financial Sentiment Analysis Ability of Language Model via Targeted Numerical Change-Related Masking. IEEE Access 2024, 12, 50809–50820. [Google Scholar] [CrossRef]
- Xie, Q.; Zhang, X.; Han, W.; Lai, Y.; Peng, M.; Lopez-Lira, A.; Huang, J. PIXIU: A Large Language Model, Instruction Data and Evaluation Benchmark for Finance. Adv. Neural Inf. Process. Syst. 2023, 36, 1–16. [Google Scholar]
- Sonkiya, P.; Bajpai, V.; Bansal, A. Stock price prediction using BERT and GAN. arXiv 2021, arXiv:2107.09055. [Google Scholar]
- Thomson Reuters. ESG Scores Methodology. 2022. Available online: https://www.refinitiv.com/en/sustainable-finance/esg-scores#methodology (accessed on 20 August 2024).
- Rizinski, M.; Peshov, H.; Mishev, K.; Jovanovik, M.; Trajanov, D. Sentiment Analysis in Finance: From Transformers Back to eXplainable Lexicons (XLex). IEEE Access 2024, 12, 7170–7198. [Google Scholar] [CrossRef]
- Nabila, P.; Setiawan, E.B. Adam and AdamW Optimization Algorithm Application on BERT Model for Hate Speech Detection on Twitter. In Proceedings of the 2024 International Conference on Data Science and Its Applications (ICoDSA), Bali, Indonesia, 10–11 July 2024; pp. 346–351. [Google Scholar] [CrossRef]
- Raghunathan, N.; Saravanakumar, K. Challenges and Issues in Sentiment Analysis: A Comprehensive Survey. IEEE Access 2023, 11, 69626–69642. [Google Scholar] [CrossRef]
- Lee, H.; Kim, J.H.; Jung, H.S. Deep-learning-based stock market prediction incorporating ESG sentiment and technical indicators. Sci. Rep. 2024, 14, 10262. [Google Scholar] [CrossRef]
- Jain, J.K.; Agrawal, R. FB-GAN: A Novel Neural Sentiment-Enhanced Model for Stock Price Prediction. In Proceedings of the Joint Workshop of the 7th Financial Technology and Natural Language Processing, the 5th Knowledge Discovery from Unstructured Data in Financial Services, and the 4th Workshop on Economics and Natural Language Processing, Torino, Italy, 20 May 2024; pp. 85–93. [Google Scholar]
- Mutinda, J.; Mwangi, W.; Okeyo, G. Sentiment Analysis of Text Reviews Using Lexicon-Enhanced Bert Embedding (LeBERT) Model with Convolutional Neural Network. Appl. Sci. 2023, 13, 1445. [Google Scholar] [CrossRef]
- Rahab, H.; Haouassi, H.; Laouid, A. Rule-Based Arabic Sentiment Analysis using Binary Equilibrium Optimization Algorithm. Arab. J. Sci. Eng. 2023, 48, 2359–2374. [Google Scholar] [CrossRef]
- Cam, H.; Cam, A.V.; Demirel, U.; Ahmed, S. Sentiment analysis of financial Twitter posts on Twitter with the machine learning classifiers. Heliyon 2024, 10, e23784. [Google Scholar] [CrossRef]
- Lin, W.; Liao, L.C. Lexicon-based prompt for financial dimensional sentiment analysis. Expert Syst. Appl. 2024, 244, 122936. [Google Scholar] [CrossRef]
- Du, K.; Xing, F.; Mao, R.; Cambria, E. An Evaluation of Reasoning Capabilities of Large Language Models in Financial Sentiment Analysis. In Proceedings of the IEEE Conference on Artificial Intelligence (IEEE CAI), Singapore, 25–27 June 2024. [Google Scholar]
- Xing, F. Designing Heterogeneous LLM Agents for Financial Sentiment Analysis. ACM Trans. Manag. Inf. Syst. 2024. [Google Scholar] [CrossRef]
- Menzio, M.; Paris, D.; Fersini, E. Unveiling Currency Market Dynamics: Leveraging Federal Reserve Communications for Strategic Investment Insights. In Proceedings of the Joint Workshop of the 7th Financial Technology and Natural Language Processing, the 5th Knowledge Discovery from Unstructured Data in Financial Services, and the 4th Workshop on Economics and Natural Language Processing, Torino, Italy, 20 May 2024; pp. 94–102. [Google Scholar]
- Kim, J.; Kim, H.S.; Choi, S.Y. Forecasting the S&P 500 Index Using Mathematical-Based Sentiment Analysis and Deep Learning Models: A FinBERT Transformer Model and LSTM. Axioms 2023, 12, 835. [Google Scholar] [CrossRef]
- Huang, A.H.; Wang, H.; Yang, Y. FinBERT: A Large Language Model for Extracting Information from Financial Text. Contemp. Account. Res. 2023, 40, 806–841. [Google Scholar] [CrossRef]
- Kawamura, K.; Li, Z.; Lin, C.; Mcdanel, B. Revelata at the FinLLM Challenge Task: Improving Financial Text Summarization by Restricted Prompt Engineering and Fine-tuning. In Proceedings of the Eighth Financial Technology and Natural Language Processing and the 1st Agent AI for Scenario Planning, Jeju, Republic of Korea, 3 August 2024; pp. 146–152. [Google Scholar]
- Leippold, M. Sentiment spin: Attacking financial sentiment with GPT-3. Financ. Res. Lett. 2023, 55, 103957. [Google Scholar] [CrossRef]
- Sy, E.; Peng, T.C.; Huang, S.H.; Lin, H.Y.; Chang, Y.C. Fine-Grained Argument Understanding with BERT Ensemble Techniques: A Deep Dive into Financial Sentiment Analysis. In Proceedings of the The 35th Conference on Computational Linguistics and Speech Processing (ROCLING 2023), Taipei City, Taiwan, 20–21 October 2023; pp. 242–249. [Google Scholar]
- Popoola, G.; Abdullah, K.K.; Fuhnwi, G.S.; Agbaje, J. Sentiment Analysis of Financial News Data using TF-IDF and Machine Learning Algorithms. In Proceedings of the 2024 IEEE 3rd International Conference on AI in Cybersecurity (ICAIC), Houston, TX, USA, 7–9 February 2024; pp. 1–6. [Google Scholar] [CrossRef]
- Shahapur, S.S.; Koralli, A.; Chippalakatti, G.; Balikai, M.M.; Mudalagi, D.; Dias, R.; Devali, S.; Wajantari, K. Discovering Untapped Potential in Finance Markets Using NLP-Driven Sentiment Analysis. Indian J. Sci. Technol. 2024, 17, 2240–2249. [Google Scholar] [CrossRef]
- Wu, W.; Xu, M.; Su, R.; Ullah, K. Modeling crude oil volatility using economic sentiment analysis and opinion mining of investors via deep learning and machine learning models. Energy 2024, 289, 130017. [Google Scholar] [CrossRef]
- Memiş, E.; Akarkamçı, H.; Yeniad, M.; Rahebi, J.; Lopez-Guede, J.M. Comparative Study for Sentiment Analysis of Financial Tweets with Deep Learning Methods. Appl. Sci. 2024, 14, 588. [Google Scholar] [CrossRef]
- Peng, B.; Chersoni, E.; Hsu, Y.; Qiu, L.; Huang, C. ren Supervised Cross-Momentum Contrast: Aligning representations with prototypical examples to enhance financial sentiment analysis. Knowl.-Based Syst. 2024, 295, 111683. [Google Scholar] [CrossRef]
- Duan, G.; Yan, S.; Zhang, M. A Hybrid Neural Network Model for Sentiment Analysis of Financial Texts Using Topic Extraction, Pre-Trained Model, and Enhanced Attention Mechanism Methods. IEEE Access 2024, 12, 98207–98224. [Google Scholar] [CrossRef]
- Abdelfattah, B.A.; Darwish, S.M.; Elkaffas, S.M. Enhancing the Prediction of Stock Market Movement Using Neutrosophic-Logic-Based Sentiment Analysis. J. Theor. Appl. Electron. Commer. Res. 2024, 19, 116–134. [Google Scholar] [CrossRef]
- Küçüklerli, K.B.; Ulusoy, V. Sentiment-Driven Exchange Rate Forecasting: Integrating Twitter Analysis with Economic Indicators. J. Appl. Financ. Bank. 2024, 14, 75–96. [Google Scholar] [CrossRef]
- Sakhare, N.N.; Shaik, I.S. Spatial federated learning approach for the sentiment analysis of stock news stored on blockchain. Spat. Inf. Res. 2024, 32, 13–27. [Google Scholar] [CrossRef]
- Saxena, A. ESG Sentiment Dataset. Available online: https://github.com/aradhana298/ESG-Sentiment-Dataset (accessed on 20 August 2024).
- Wijekoon, R.; Sabri, M.F. Determinants that influence green product purchase intention and behavior: A literature review and guiding framework. Sustainability 2021, 13, 6219. [Google Scholar] [CrossRef]
- Camilleri, M.A.; Cricelli, L.; Mauriello, R.; Strazzullo, S. Consumer Perceptions of Sustainable Products: A Systematic Literature Review. Sustainability 2023, 15, 8923. [Google Scholar] [CrossRef]
- Oliver Yébenes, M. Climate Change, ESG Criteria and Recent Regulation: Challenges and Opportunities; Springer International Publishing: Cham, Switzerland, 2024; Volume 14, ISBN 0123456789. [Google Scholar]
- Li, C.; Tang, W.; Liang, F.; Wang, Z. The impact of climate change on corporate ESG performance: The role of resource misallocation in enterprises. J. Clean. Prod. 2024, 445, 141263. [Google Scholar] [CrossRef]
- Fatouros, G.; Soldatos, J.; Kouroumali, K.; Makridis, G.; Kyriazis, D. Transforming sentiment analysis in the financial domain with ChatGPT. Mach. Learn. Appl. 2023, 14, 100508. [Google Scholar] [CrossRef]
- Nousi, C.; Tjortjis, C. A Methodology for Stock Movement Prediction Using Sentiment Analysis on Twitter and StockTwits Data. In Proceedings of the 6th South-East Europe Design Automation, Computer Engineering, Computer Networks and Social Media Conference (SEEDA-CECNSM), Preveza, Greece, 24–26 September 2021; pp. 1–7. [Google Scholar] [CrossRef]
- Sohangir, S.; Wang, D.; Pomeranets, A.; Khoshgoftaar, T.M. Big Data: Deep Learning for financial sentiment analysis. J. Big Data 2018, 5, 3. [Google Scholar] [CrossRef]
Reference | Algorithm and Dataset Used | Accuracy Achieved | Research Gap |
---|---|---|---|
This paper | FinBERT with XAI, ECO and non-eco-friendly labelled datasets used | 90.66%, F score 0.91 | Better accuracy with XAI, but eco/non-eco classification not used, nested sentiment applied |
[17] | Bi-LSTM, S&P 500 index, ESG sentiment data from LexisNexis, technical indicators | MAPE of 3.05% | Lacks a nested sentiment analysis specifically for eco-friendly and non-eco-friendly products and does not use XAI for model interpretability |
[14] | XLex (Explainable Lexicons) on financial texts | 84.30% | Inferior to FinBERT with XAI for financial texts; lacks ESG-specific and eco/non-eco sentiment analysis, less accuracy |
[24] | Heterogeneous LLM Agents framework using six FSA datasets | Achieved an average accuracy of 79.53%. | The study used GPT-3.5, a more resource-intensive model compared to FinBERT, and lacked XAI, eco/non-eco labelled data, and nested sentiment analysis, less accuracy |
[21] | Naive Bayes, Logistic Regression, SVM, KNN, Decision Trees, Multilayer Perceptron, Dataset: Turkish financial tweets | 89% with SVM | Did not use FinBERT, which is better suited for financial data, nor did it employ XAI for interpretability or address eco/non-eco product labelling and nested sentiment analysis, less accuracy |
[31] | Random Forest, Naive Bayes, K-Nearest Neighbour; Dataset: 50,000 tweets related to financial news | 81% by RF and NB | The paper lacks the use of FinBERT, XAI, and does not address eco-friendly vs. non-eco-friendly labelled datasets or perform nested sentiment analysis with less accuracy |
[23] | LLMs (PaLM-2, GPT-3.5, GPT-4) using PhraseBank and Twitter Financial News datasets | 96.39% on PhraseBank-100%) | General LLMs, not FinBERT, were used; XAI and nested sentiment analysis were not employed, and eco/non-eco product categorization was not considered; models were heavier and less specialized for finance. |
[32] | Multinomial Naïve Bayes Dataset: 6000 newspaper articles and social media comments related to finance | 81.39% Multinomial NB | The study did not use finance-specific algorithms like FinBERT, also not use XAI, and lacked eco/non-eco labeling for ESG analysis. It also did not employ nested sentiment analysis, focusing instead on simpler models, less accuracy |
[33] | AR, SVR, MLP, RNN, GRU, LSTM. Dataset: Investor remarks from the Eastmoney forum, INE crude oil futures data | 78.853% by LSTM | FinBERT and XAI were not used; no nested sentiment analysis; no focus on eco/non-eco-friendly labels, less accuracy |
[34] | Neural Network, CNN, Turkish financial tweets dataset | 83.02% by CNN with pre-trained word embedding | The study does not use FinBERT or XAI, lacks eco-friendly labelling, and uses simple binary sentiment analysis without linking sentiment to ESG metrics. Models are heavier and less efficient than FinBERT. |
[22] | BERT, RoBERTa, Electra, T5. Datasets | MAE 0.278 with LP+DANN+ method | The paper lacks FinBERT and XAI, doesn’t use an eco-friendly labelled dataset, and focuses on binary rather than nested sentiment analysis. The models used are heavier than FinBERT. |
[35] | SuCroMoCo, BERT, RoBERTa, Dataset: FinTextSen | 87.75% by FinTextSen: | No finance-specific algorithm was used like FinBERT, no use of XAI, and eco-friendly vs. non-eco-friendly labelling. binary sentiment analysis without linking sentiment to ESG metrics, resulting in less accuracy. |
[36] | Hybrid Neural Network (LFBP) Datasets: FiQATask1 | Accuracy not explicitly mentioned; F1-score improvement of 2.05–7.27% | The paper lacks a nested sentiment analysis with eco-friendly labels, does not analyze sentiment-ESG relationships, and does not use FinBERT or XAI for financial-specific tasks. |
[3] | GPT-4, BERT, FinBERT; Crypto News + dataset | 86.7% by GPT-4: | The research gap includes the absence of FinBERT, no use of XAI, reliance on a non-nested binary sentiment analysis, and no focus on ESG metrics. Additionally, no finance-specific algorithm like FinBERT is used. |
[37] | Neutrosophic Logic, LSTM, StockNet dataset | 78.48% | Omission of FinBERT, absence of XAI, use of a non-nested binary sentiment analysis, and a lack of focus on ESG metrics. Additionally, a heavier model than FinBERT is employed, which is less specialized for financial tasks. |
[38] | LSTM Neural Network, XGBoost, RNN; Twitter data | 65% | This study did not utilize FinBERT or XAI and focused on a non-nested, simple sentiment analysis without eco-friendly labelling. It also did not explore the relationship between sentiment and ESG metrics |
[39] | GAN, Dataset: News data from Economic Times, NIFTY 50 index | 95.80% by GAN | The paper does not use FinBERT, lacks XAI methods, and does not incorporate eco-friendly/non-eco-friendly labels or ESG metric analysis. It also uses simpler binary sentiment analysis instead of nested sentiment analysis. |
Dataset | Unique Feature/Gap |
---|---|
The dataset created during the study | Includes product-level ESG and sentiment analysis with eco vs. non-eco classification. |
Financial PhraseBank [10] | No product names, no eco/non-eco labels, no ESG or environmental scores. |
FiQA [11] | No product names or ESG data focused on financial question-answer sentiment. |
StockTwits [12] | Lacks ESG and product details, and focuses on stock sentiment from social media. |
Refinitiv ES [13] | Lacks sentiment and product data, and focuses on company ESG scores and industry comparison. |
Product Name | Sentence | Sentiment | Environmental Score | ESG Score | Status |
---|---|---|---|---|---|
Non-renewable Cars | Non-renewable Cars are widely recognized for their impact. Unfortunately, it fails to meet expectations and is not recommended. | Negative | 30 | 36 | Non-Eco-friendly |
Styrofoam Shampoo Corp | Styrofoam Shampoo Corp celebrated its Pesticides impressive initiative. | Positive | 40 | 11 | Non-Eco-friendly |
Paper Food | The Paper Food is a disaster in terms of quality and has received a lot of negative reviews. | Negative | 94 | 76 | Eco-friendly |
Natural Cars Corp | Natural Cars Corp receives criticism for its Cruelty-free initiative. | Positive | 49 | 49 | Eco-friendly |
Parameter | Value | Brief Description |
---|---|---|
Model | FinBERT (yiyanghkust/finbert-tone) | A transformer-based model fine-tuned for financial text sentiment analysis. |
Number of Labels | 2 (Positive/Negative) | Binary classification to identify sentiment as Positive or Negative. |
Tokenizer | BertTokenizer (yiyanghkust/finbert-tone) | Converts input text into token IDs that the model can process. |
Number of Training Epochs | 100 (Options: 50, 80, 100) | Number of complete passes through the training dataset. |
Batch Size (Training) | 16 | Number of samples processed before model weights are updated. |
Batch Size (Evaluation) | 16 | Number of samples evaluated at once during validation/testing. |
Learning Rate | AdamW Optimizer, L2 Regularization | AdamW: A variant of the Adam optimizer; L2: Regularizes the model to prevent overfitting. |
Evaluation Strategy | Epoch | Evaluates the model at the end of every epoch (one full pass through the dataset). |
Save Strategy | Epoch | Saves the model after each epoch. |
Metric | Accuracy | Measures how often the model correctly predicts the label. |
Loss Function | Cross-Entropy Loss | A function that calculates the difference between predicted and actual labels. |
Regularization | L2 | Penalizes large weights to reduce overfitting, helping the model generalize better. |
Early Stopping Criteria | Not Applied | Early stopping stops training if the model’s performance does not improve after a set number of epochs. |
Split Ratio (Train/Test) | 80/20 | Percentage of the data used for training and testing (80% for training, 20% for testing). |
Epochs | Training Loss | Validation Loss | Accuracy (%) |
---|---|---|---|
50 | 0.0302 | 1.0495 | 87.91 |
80 | 0.0157 | 0.8751 | 90.66 |
100 | 0.0136 | 0.9218 | 90.66 |
100 (with Adam W& L2) | 0.01 | 0.85 | 91.76 |
Evaluation Metric | Loss | Accuracy | Runtime | Samples per Second | Steps per Second | Epoch |
---|---|---|---|---|---|---|
Value | 0.9218 | 0.9066 | 0.1546 | 1177.049 | 77.608 | 100 |
Year | Paper Title | Algorithm Used | Accuracy | F1-Score | Precision | Recall | AUC |
---|---|---|---|---|---|---|---|
This Paper | FinBERT | 91.76% | 0.91 | 0.94 (+ve), 0.87 (−ve) | 0.88 (+ve), 0.94 (−ve) | 0.98 | |
2023 | [45] | FinBERT | 56% | 0.556 | 0.56 | 0.562 | 0.54 |
GPT-P1 | 73% | 0.725 | 0.76 | 0.73 | 0.3 | ||
GPT-P2 | 79% | 0.79 | 0.797 | 0.79 | 0.227 | ||
GPT-P3 | 74% | 0.737 | 0.78 | 0.735 | 0.282 | ||
GPT-P4 | 78% | 0.789 | 0.804 | 0.784 | 0.221 | ||
2021 | [46] | SVM | 65.80% | 76.30% | Not Specified | Not Specified | 0.67 |
2018 | [47] | Logistic Regression | 71% | 0.7056 | 0.7134 | 0.698 | 0.7088 |
Class | Precision | Recall | F1-Score | Support |
---|---|---|---|---|
Negative | 0.87 | 0.94 | 0.91 | 86 |
Positive | 0.94 | 0.88 | 0.91 | 96 |
Accuracy | 0.91 | 182 | ||
macro avg | 0.91 | 0.91 | 0.91 | 182 |
weighted avg | 0.91 | 0.91 | 0.91 | 182 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Saxena, A.; Santhanavijayan, A.; Shakya, H.K.; Kumar, G.; Balusamy, B.; Benedetto, F. Nested Sentiment Analysis for ESG Impact: Leveraging FinBERT to Predict Market Dynamics Based on Eco-Friendly and Non-Eco-Friendly Product Perceptions with Explainable AI. Mathematics 2024, 12, 3332. https://doi.org/10.3390/math12213332
Saxena A, Santhanavijayan A, Shakya HK, Kumar G, Balusamy B, Benedetto F. Nested Sentiment Analysis for ESG Impact: Leveraging FinBERT to Predict Market Dynamics Based on Eco-Friendly and Non-Eco-Friendly Product Perceptions with Explainable AI. Mathematics. 2024; 12(21):3332. https://doi.org/10.3390/math12213332
Chicago/Turabian StyleSaxena, Aradhana, A. Santhanavijayan, Harish Kumar Shakya, Gyanendra Kumar, Balamurugan Balusamy, and Francesco Benedetto. 2024. "Nested Sentiment Analysis for ESG Impact: Leveraging FinBERT to Predict Market Dynamics Based on Eco-Friendly and Non-Eco-Friendly Product Perceptions with Explainable AI" Mathematics 12, no. 21: 3332. https://doi.org/10.3390/math12213332
APA StyleSaxena, A., Santhanavijayan, A., Shakya, H. K., Kumar, G., Balusamy, B., & Benedetto, F. (2024). Nested Sentiment Analysis for ESG Impact: Leveraging FinBERT to Predict Market Dynamics Based on Eco-Friendly and Non-Eco-Friendly Product Perceptions with Explainable AI. Mathematics, 12(21), 3332. https://doi.org/10.3390/math12213332