Explainable Deep Learning: A Visual Analytics Approach with Transition Matrices
Abstract
:1. Introduction
- A novel method that employs a transition matrix to transform the outcomes of a formal DL model into a format that is more intuitive for human understanding, i.e., a mental model, by leveraging an ML model with features that are more easily interpreted by humans.
- An improved method for obtaining a separating hyperplane for classification tasks using a transition matrix based on VA and the HITL principle.
2. Basic Definitions and Concepts
3. Materials and Methods
3.1. Interpretation of Results Obtained by DL Models through ML Models
3.1.1. Problem Statement and Proposed Solution Method
3.1.2. Illustrative Numerical Example
3.2. Obtaining a Separating Hyperplane Using the Human-in-the-Loop Principle
3.2.1. Problem Statement and Improved Method
3.2.2. Illustrative Numerical Example
3.3. Methodology and Experimental Setting
3.3.1. Datasets
3.3.2. Design of Computational Experiments
- Construction of matrix B. For ML model (or MM) experiments, 10,000 vectors from the MNIST dataset training sample were randomly selected. These vectors were obtained from 28 × 28 MNIST images by “slicing” and concatenating the image rows. In other words, for matrix B, each 28 × 28 handwritten digit image was represented as a vector (one-dimensional array) with 784 features. Consequently, matrix B had a size of 10,000 × 784.
- Construction of matrix A. A DL model (or FM) was created using a classic CNN architecture. The neural network was trained on the full MNIST dataset of 60,000 items. The architecture of the trained CNN is presented in Table A1 in Appendix B. From the penultimate layer of the created neural network model (490 weight coefficients), a vector was formed for each image from the formed sub-sample of 10,000, which was sequentially added to matrix A. As a result of these actions, two experimental matrices were obtained: matrix and matrix .
- Calculation of the Transition Matrix T: Using Formula (6), the transition matrix T was calculated to find new representations for the rows of matrix A through basis B.
- Construction of Matrix B: We randomly selected 10,000 labeled headline and article pairs from the training part of the FNC-1 dataset. Each article was preprocessed by tokenizing the text, removing stop words, and applying stemming. The processed text was then converted into a vector representation using the TF-IDF (Term Frequency-Inverse Document Frequency) encoding, resulting in one-dimensional vectors with a predefined number of features of 1024. As a result, we received matrix , which served as an ML (or MM) model.
- Construction of Matrix A: A RoBERTa model fine-tuned for fake news detection was crafted based on the architecture described in work [40]. The model was trained on the full FNC-1 dataset. The architecture of the trained RoBERTa model is presented in Table A2 in Appendix B. From the penultimate layer of the fine-tuned RoBERTa model, a vector of 768 was formed for each pair “headline/article” from the formed subsample of 10,000, sequentially added to matrix A. Consequently, two experimental matrices were obtained: matrix and matrix , where 768 is the size of the penultimate (fully-connected) layer of the constructed RoBERTa model.
- Using Formula (6), the transition matrix T was calculated to find new representations for the rows of matrix A through basis B. The transition matrix T in this case was of 768 × 1024.
- Four numerical features, sepal length, sepal width, petal length, and petal width, formed matrix .
- Dimensionality reduction of matrix to a two-dimensional space was performed using MDS, resulting in matrix .
- Matrix was visualized on a graph; a separating line was also added to the graph to visually separate one class of the Iris dataset from all others.
- Matrix was used to construct the transition matrix T (6).
- The transition matrix T was used to construct the separating hyperplane using Formula (13) and an evolutionary algorithm with Formula (14).
3.3.3. Quantitative Evaluation Criteria
4. Results and Discussion
4.1. Results of Transitioning from the DL Model to the ML Model Using the Proposed Method
4.1.1. Visual Interpretation based on the MNIST Dataset
4.1.2. Interpreting Stance Detection Based on the FNC-1 Dataset
- Identifying crucial elements (such as specific words or phrases) within the transformed vector that played a significant role in the model’s decision-making process.
- Matching these essential features against the original headline and article body to see if the model’s attention was rightly placed on the main components of the text.
- Conducting a qualitative analysis of how interpretable the model’s decisions were, based on the examination of the transformed features. We evaluated if the features identified as important within the MM space offered a clear and intuitive rationale for the model’s predictions. For instance, a prediction of “Agree” supported by features that drew parallels between the headline and body text would signify that the model’s reasoning was understandable and aligned with human logic.
- Checking for consistency in the model’s focus across similar cases. If the model reliably pinpointed the same meaningful features as relevant in various instances of stance detection, it would indicate that our method, i.e., the use of the transition matrix T, ensured dependable interpretability.
4.2. Results and Discussion for Obtaining a Separating Hyperplane Using the Human-in-the-Loop Principle
- Four numerical features, i.e., sepal length, sepal width, petal length, and petal width, formed matrix .
- Dimensionality reduction of matrix to a two-dimensional space was performed using MDS, resulting in matrix .
- Matrix was visualized on a graph. Also, following the HITL principle, a separating line was added to the graph, visually separating one class of the Iris dataset from all others. The visualization result is demonstrated in Figure 12.
- 4.
- Matrix was used to construct the transition matrix T (6) using the approach proposed in this work.
- 5.
- Finally, the transition matrix T was used to construct the separating hyperplane using Formula (13) and an evolutionary algorithm with Formula (14). The coefficients of the hyperplane are as follows:
4.3. Limitations of the Proposed Methods
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
Abbreviations
AI | Artificial Intelligence |
CAM | Class Activation Mapping |
CNN | Convolutional Neural Network |
DL | Deep Learning |
DNN | Deep Neural Network |
FACTS | Fairness, Accountability, Confidentiality, Transparency, and Safety |
FM | Formal Model |
FNC | Fake News Challenge |
HITL | Human-in-the-Loop |
MAX | Maximum Pixel Value |
MDS | Multidimensional Scaling |
ML | Machine Learning |
MM | Mental Model |
MNIST | Modified National Institute of Standards and Technology |
MSE | Mean Squared Error |
PSNR | Peak Signal-Noise Ratio |
RoBERTa | Robustly Optimized Bidirectional Encoder Representations from Transformers |
SHAP | SHapley Additive exPlanations |
SSIM | Structural Similarity Index Measure |
TF-IDF | Term Frequency-Inverse Document Frequency |
VA | Visual Analytics |
XAI | Explainable Artificial Intelligence |
Appendix A
Appendix B
Component | Layer Type | Input Channels | Output Channels | Kernel Size | Stride | Padding | Additional Details |
---|---|---|---|---|---|---|---|
conv_block_1 | Conv2d | 1 | 10 | 3 × 3 | 1 | 1 | - |
ReLU | - | - | - | - | - | - | |
Conv2d | 10 | 10 | 3 × 3 | 1 | 1 | - | |
ReLU | - | - | - | - | - | - | |
MaxPool2d | - | - | 2 × 2 | 2 | 0 | - | |
conv_block_2 | Conv2d | 10 | 10 | 3 × 3 | 1 | 1 | - |
ReLU | - | - | - | - | - | - | |
Conv2d | 10 | 10 | 3 × 3 | 1 | 1 | - | |
ReLU | - | - | - | - | - | - | |
MaxPool2d | - | - | 2 × 2 | 2 | 0 | - | |
Classifier Block | Flatten | - | - | - | - | - | start_dim = 1, end_dim = −1 |
Linear | - | - | - | - | - | in_features = 490, out_features = 10, bias = True |
Component | Layer Type | Input Features | Output Features | Additional Details |
---|---|---|---|---|
Embedding Block | Embedding Layer | Token IDs | 768 | Positional Embedding, Token Type Embedding |
Encoder Block 1 | Self-Attention | 768 | 768 | 12 attention heads, Layer Normalization |
Feed-Forward | 768 | 768 | GELU Activation, Layer Normalization | |
Encoder Block 2 | Self-Attention | 768 | 768 | 12 attention heads, Layer Normalization |
Feed-Forward | 768 | 768 | GELU Activation, Layer Normalization | |
Encoder Block 3 | Self-Attention | 768 | 768 | 12 attention heads, Layer Normalization |
Feed-Forward | 768 | 768 | GELU Activation, Layer Normalization | |
... | ... | ... | ... | ... |
Encoder Block 12 | Self-Attention | 768 | 768 | 12 attention heads, Layer Normalization |
Feed-Forward | 768 | 768 | GELU Activation, Layer Normalization | |
Classifier Head | Linear | 768 | Number of classes | Dropout, SoftMax Activation |
References
- Belle, V.; Papantonis, I. Principles and practice of explainable machine learning. Front. Big Data 2021, 4, 688969. [Google Scholar] [CrossRef] [PubMed]
- Loyola-González, O. Black-box vs. White-box: Understanding their advantages and weaknesses from a practical point of view. IEEE Access 2019, 7, 154096–154113. [Google Scholar] [CrossRef]
- Phillips, P.J.; Hahn, C.A.; Fontana, P.C.; Yates, A.N.; Greene, K.; Broniatowski, D.A.; Przybocki, M.A. Four Principles of Explainable Artificial Intelligence; NIST Interagency/Internal Report (NISTIR) 8312; National Institute of Standards and Technology: Gaithersburg, MD, USA, 2021; p. 48. [Google Scholar] [CrossRef]
- Casey, B.; Farhangi, A.; Vogl, R. Rethinking explainable machines: The GDPR’s right to explanation debate and the rise of algorithmic audits in enterprise. Berkeley Technol. Law J. 2019, 34, 143–188. [Google Scholar] [CrossRef]
- Greenstein, S. Preserving the rule of law in the era of artificial intelligence (AI). Artif. Intell. Law 2022, 30, 291–323. [Google Scholar] [CrossRef]
- Laux, J.; Wachter, S.; Mittelstadt, B. Trustworthy artificial intelligence and the European Union AI act: On the conflation of trustworthiness and acceptability of risk. Regul. Gov. 2024, 18, 3–32. [Google Scholar] [CrossRef] [PubMed]
- Chamberlain, J. The risk-based approach of the European Union’s proposed artificial intelligence regulation: Some comments from a tort law perspective. Eur. J. Risk Regul. 2023, 14, 1–13. [Google Scholar] [CrossRef]
- Maclure, J. AI, Explainability and public reason: The argument from the limitations of the human mind. Minds Mach. 2021, 31, 421–438. [Google Scholar] [CrossRef]
- Merry, M.; Riddle, P.; Warren, J. A mental models approach for defining explainable artificial intelligence. BMC Med. Inform. Decis. Mak. 2021, 21, 344. [Google Scholar] [CrossRef] [PubMed]
- Manziuk, E.; Barmak, O.; Krak, I.; Mazurets, O.; Skrypnyk, T. Formal model of trustworthy artificial intelligence based on standardization. In Proceedings of the 2nd International Workshop on Intelligent Information Technologies & Systems of Information Security (IntelITSIS-2022), Khmelnytskyi, Ukraine, 24–26 March 2021; Hovorushchenko, T., Savenko, O., Popov, P., Lysenko, S., Eds.; CEUR-WS: Aachen, Germany, 2021; Volume 2853, pp. 190–197. [Google Scholar]
- Hoffman, R.R.; Mueller, S.T.; Klein, G.; Litman, J. Measures for explainable AI: Explanation goodness, user satisfaction, mental models, curiosity, trust, and human-AI performance. Front. Comput. Sci. 2023, 5, 1096257. [Google Scholar] [CrossRef]
- Watson, D.S. Conceptual challenges for interpretable machine learning. Synthese 2022, 200, 65. [Google Scholar] [CrossRef]
- Guidotti, R.; Monreale, A.; Pedreschi, D.; Giannotti, F. Principles of explainable artificial intelligence. In Explainable AI Within the Digital Transformation and Cyber Physical Systems: XAI Methods and Applications; Sayed-Mouchaweh, M., Ed.; Springer International Publishing: Cham, Switzerland, 2021; pp. 9–31. [Google Scholar] [CrossRef]
- Rudin, C.; Chen, C.; Chen, Z.; Huang, H.; Semenova, L.; Zhong, C. Interpretable machine learning: Fundamental principles and 10 grand challenges. Stat. Surv. 2022, 16, 1–85. [Google Scholar] [CrossRef]
- Barredo Arrieta, A.; Díaz-Rodríguez, N.; Del Ser, J.; Bennetot, A.; Tabik, S.; Barbado, A.; Garcia, S.; Gil-Lopez, S.; Molina, D.; Benjamins, R.; et al. Explainable artificial intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Inf. Fusion 2020, 58, 82–115. [Google Scholar] [CrossRef]
- Ras, G.; Xie, N.; van Gerven, M.; Doran, D. Explainable deep learning: A field guide for the uninitiated. J. Artif. Intell. Res. 2022, 73, 329–396. [Google Scholar] [CrossRef]
- Izonin, I.; Tkachenko, R.; Kryvinska, N.; Tkachenko, P.; Greguš ml., M. Multiple linear regression based on coefficients identification using non-iterative SGTM neural-like structure. In Proceedings of the 15th International Work-Conference on Artificial Neural Networks (IWANN-2019), Gran Canaria, Spain, 12–14 June 2019; Rojas, I., Joya, G., Catala, A., Eds.; Springer International Publishing: Cham, Switzerland, 2019; Volume 11506, pp. 467–479. [Google Scholar] [CrossRef]
- Fauvel, K.; Lin, T.; Masson, V.; Fromont, É.; Termier, A. XCM: An explainable convolutional neural network for multivariate time series classification. Mathematics 2021, 9, 3137. [Google Scholar] [CrossRef]
- Bai, X.; Wang, X.; Liu, X.; Liu, Q.; Song, J.; Sebe, N.; Kim, B. Explainable deep learning for efficient and robust pattern recognition: A survey of recent developments. Pattern Recognit. 2021, 120, 108102. [Google Scholar] [CrossRef]
- Olteanu, A.; Garcia-Gathright, J.; de Rijke, M.; Ekstrand, M.D.; Roegiest, A.; Lipani, A.; Beutel, A.; Olteanu, A.; Lucic, A.; Stoica, A.-A.; et al. FACTS-IR: Fairness, accountability, confidentiality, transparency, and safety in information retrieval. SIGIR Forum 2021, 53, 20–43. [Google Scholar] [CrossRef]
- Choo, J.; Liu, S. Visual analytics for explainable deep learning. IEEE Comput. Graph. Appl. 2018, 38, 84–92. [Google Scholar] [CrossRef] [PubMed]
- La Rosa, B.; Blasilli, G.; Bourqui, R.; Auber, D.; Santucci, G.; Capobianco, R.; Bertini, E.; Giot, R.; Angelini, M. State of the art of visual analytics for explainable deep learning. Comput. Graph. Forum 2023, 42, 319–355. [Google Scholar] [CrossRef]
- Krak, I.V.; Kudin, G.I.; Kulyas, A.I. Multidimensional scaling by means of pseudoinverse operations. Cybern. Syst. Anal. 2019, 55, 22–29. [Google Scholar] [CrossRef]
- Mosqueira-Rey, E.; Hernández-Pereira, E.; Alonso-Ríos, D.; Bobes-Bascarán, J.; Fernández-Leal, Á. Human-in-the-loop machine learning: A state of the art. Artif. Intell. Rev. 2023, 56, 3005–3054. [Google Scholar] [CrossRef]
- La Barbera, D.; Roitero, K.; Mizzaro, S. Combining human intelligence and machine learning for fact-checking: Towards a hybrid human-in-the-loop framework. Intell. Artif. 2023, 17, 163–172. [Google Scholar] [CrossRef]
- Retzlaff, C.O.; Das, S.; Wayllace, C.; Mousavi, P.; Afshari, M.; Yang, T.; Saranti, A.; Angerschmid, A.; Taylor, M.E.; Holzinger, A. Human-in-the-loop reinforcement learning: A survey and position on requirements, challenges, and opportunities. J. Artif. Intell. Res. 2024, 79, 359–415. [Google Scholar] [CrossRef]
- Ben-Israel, A.; Greville, T.N.E. Existence and construction of generalized inverses. In Generalized Inverses: Theory and Applications; CMS Books in Mathematics; Springer: New York, NY, USA, 2003; pp. 40–51. [Google Scholar] [CrossRef]
- Akritas, A.G.; Malaschonok, G.I. Applications of singular-value decomposition (SVD). Math. Comput. Simul. 2004, 67, 15–31. [Google Scholar] [CrossRef]
- Kalyta, O.; Barmak, O.; Radiuk, P.; Krak, I. Facial emotion recognition for photo and video surveillance based on machine learning and visual analytics. Appl. Sci. 2023, 13, 9890. [Google Scholar] [CrossRef]
- Krak, I.; Barmak, O.; Manziuk, E.; Kulias, A. Data classification based on the features reduction and piecewise linear separation. In Proceedings of the 2nd International Conference on Intelligent Computing and Optimization (ICO 2019), Koh Samui, Thailand, 3–4 October 2019; Vasant, P., Zelinka, I., Weber, G.-W., Eds.; Springer International Publishing: Cham, Switzerland, 2020; Volume 1072, pp. 282–289. [Google Scholar] [CrossRef]
- Korn, G.A.; Korn, T.M. Mathematical Handbook for Scientists and Engineers: Definitions, Theorems, and Formulas for Reference and Review; Revised edition; Dover Publications: Mineola, NY, USA, 2000; p. 1152. ISBN 978-0-486-41147-7. [Google Scholar]
- Kryvonos, I.G.; Krak, I.V.; Barmak, O.V.; Ternov, A.S.; Kuznetsov, V.O. Information technology for the analysis of mimic expressions of human emotional states. Cybern. Syst. Anal. 2015, 51, 25–33. [Google Scholar] [CrossRef]
- Belkina, A.C.; Ciccolella, C.O.; Anno, R.; Halpert, R.; Spidlen, J.; Snyder-Cappione, J.E. Automated optimized parameters for T-distributed stochastic neighbor embedding improve visualization and analysis of large datasets. Nat. Commun. 2019, 10, 5415. [Google Scholar] [CrossRef] [PubMed]
- Meilă, M.; Zhang, H. Manifold learning: What, how, and why. Annu. Rev. Stat. Its Appl. 2024, 11, 1–25. [Google Scholar] [CrossRef]
- Deng, L. The MNIST database of handwritten digit images for machine learning research [Best of the Web]. IEEE Signal Process. Mag. 2012, 29, 141–142. [Google Scholar] [CrossRef]
- Hanselowski, A.; PVS, A.; Schiller, B.; Caspelherr, F.; Chaudhuri, D.; Meyer, C.M.; Gurevych, I. A retrospective analysis of the fake news challenge stance-detection task. In Proceedings of the 27th International Conference on Computational Linguistics (COLING-2018), Santa Fe, NM, USA, 20–26 August 2018; Bender, E.M., Derczynski, L., Isabelle, P., Eds.; Association for Computational Linguistics: Santa Fe, NM, USA, 2018; pp. 1859–1874. [Google Scholar]
- Unwin, A.; Kleinman, K. The Iris data set: In search of the source of Virginica. Significance 2021, 18, 26–29. [Google Scholar] [CrossRef]
- Alturayeif, N.; Luqman, H.; Ahmed, M. A systematic review of machine learning techniques for stance detection and its applications. Neural Comput. Appl. 2023, 35, 5113–5144. [Google Scholar] [CrossRef] [PubMed]
- Ferreira, W.; Vlachos, A. Emergent: A novel data-set for stance classification. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT’2016), San Diego, CA, USA, 12–17 June 2016; Association for Computational Linguistics: Kerrville, TX, USA, 2016; pp. 1163–1168. [Google Scholar]
- Sepúlveda-Torres, R.; Vicente, M.; Saquete, E.; Lloret, E.; Palomar, M. Exploring summarization to enhance headline stance detection. In Proceedings of the Natural Language Processing and Information Systems (NLDB-2021), Saarbrücken, Germany, 23–25 June 2021; Métais, E., Meziane, F., Horacek, H., Kapetanios, E., Eds.; Springer International Publishing: Cham, Switzerland, 2021; Volume 12801, pp. 243–254. [Google Scholar] [CrossRef]
- Horé, A.; Ziou, D. Image quality metrics: PSNR vs. SSIM. In Proceedings of the 2010 20th International Conference on Pattern Recognition (ICPR-2010), Istanbul, Turkey, 23–26 August 2010; IEEE Inc.: New York, NY, USA, 2010; pp. 2366–2369. [Google Scholar] [CrossRef]
- Galbraith, B.; Rao, D. FNC-1-Baseline: A Baseline Implementation for FNC-1. Available online: https://github.com/FakeNewsChallenge/fnc-1-baseline (accessed on 13 March 2024).
- Baird, S.; Sibley, D.; Pan, Y. Talos Targets Disinformation with Fake News Challenge Victory. Available online: https://blog.talosintelligence.com/talos-fake-news-challenge/ (accessed on 13 March 2024).
- Hanselowski, A. Description of the System Developed by Team Athene in the FNC-1. Available online: https://github.com/hanselowski/athene_system (accessed on 13 March 2024).
- Riedel, B.; Augenstein, I.; Spithourakis, G.P.; Riedel, S. A simple but tough-to-beat baseline for the fake news challenge stance detection task. arXiv 2018, arXiv:1707.03264. [Google Scholar] [CrossRef]
- Zhang, Q.; Liang, S.; Lipani, A.; Ren, Z.; Yilmaz, E. From stances’ imbalance to their hierarchical representation and detection. In Proceedings of the 2019 the World Wide Web Conference (WWW’19), San Francisco, CA, USA, 13–17 May 2019; Liu, L., White, R., Eds.; Association for Computing Machinery: New York, NY, USA, 2019; pp. 2323–2332. [Google Scholar] [CrossRef]
# | Mental Model | Formal Model → Mental Model | Diff | ||||||
---|---|---|---|---|---|---|---|---|---|
1 | −1.97939 | 7.95931 | −1.38111 | −1.72964 | −1.71146 | 1.95563 | −1.71140 | −1.69587 | 0.426 |
2 | −1.97492 | 1.94851 | −1.72661 | −1.76121 | −1.77768 | 1.97691 | −1.81213 | −1.83549 | 0.229 |
3 | −1.84391 | 1.99818 | −1.91285 | −1.97511 | −1.84391 | 1.99818 | −1.91285 | −1.97511 | 0 |
4 | −1.99863 | 1.99967 | −1.99844 | −1.99976 | −1.91013 | 2.01946 | −2.01357 | −2.11472 | 0.147 |
5 | −1.99937 | 1.99889 | −1.99960 | −1.99891 | −1.97636 | 2.04074 | −2.11430 | −2.25434 | 0.284 |
6 | 1.997776 | −1.84400 | 1.660111 | −1.37353 | 2.06518 | −1.96399 | 0.94976 | −1.03481 | 0.798 |
7 | 1.818753 | −1.90968 | 1.20663 | −1.40799 | 1.90491 | −1.86938 | 0.73427 | −1.31160 | 0.491 |
8 | 1.992024 | −1.92380 | 0.70659 | −1.54378 | 1.99202 | −1.9238 | 0.70659 | −1.54378 | 0 |
9 | 1.999174 | −1.99759 | 0.21221 | −1.58697 | 2.04090 | −1.96562 | 0.66440 | −1.74106 | 0.480 |
10 | 1.997854 | −1.99941 | −0.24340 | −1.82758 | 2.15582 | −2.0721 | 0.65076 | −1.97568 | 0.922 |
11 | 0.851626 | 1.57420 | 1.58102 | 1.57393 | 1.00800 | 1.64037 | 1.55452 | 1.55364 | 0.173 |
12 | 1.008513 | 1.57079 | 1.59565 | 1.74176 | 1.07422 | 1.61909 | 1.65524 | 1.69325 | 0.112 |
13 | 1.107744 | 1.61547 | 1.72358 | 1.80761 | 1.10774 | 1.61547 | 1.72358 | 1.80761 | 0 |
14 | 1.089898 | 1.61136 | 1.88253 | 1.87352 | 1.11836 | 1.62421 | 1.76924 | 1.90428 | 0.121 |
15 | 1.290406 | 1.69528 | 1.95350 | 1.94625 | 1.15515 | 1.61882 | 1.84081 | 2.02117 | 0.206 |
# | Formal Model → Mental Model | Diff | |||
---|---|---|---|---|---|
1 | −1.868316882 | 1.917325741 | −1.465195993 | −1.6609 | 0.160916565 |
2 | −1.902561782 | 1.94939568 | −1.615068122 | −1.77742 | 0.133943398 |
3 | −1.936806681 | 1.981465618 | −1.76494025 | −1.89393 | 0.19333596 |
4 | −1.971051581 | 2.013535556 | −1.914812378 | −2.01044 | 0.089781324 |
5 | −2.00529648 | 2.045605494 | −2.064684507 | −2.12696 | 0.151151106 |
6 | 2.045409347 | −1.873238392 | 1.632130028 | −1.33258 | 0.07472559 |
7 | 1.880332443 | −1.782230819 | 1.257434896 | −1.54882 | 0.206033576 |
8 | 1.900664267 | −1.864730953 | 0.750368655 | −1.6372 | 0.149932904 |
9 | 1.925070093 | −1.924359043 | 0.170899595 | −1.66796 | 0.138278882 |
10 | 2.041989021 | −2.033726874 | −0.261215771 | −1.77314 | 0.08004815 |
11 | 0.88397757 | 1.609971462 | 1.532978489 | 1.549243 | 0.072418013 |
12 | 0.91822247 | 1.577901524 | 1.682850618 | 1.665757 | 0.146909333 |
13 | 1.03606919 | 1.601359125 | 1.780350637 | 1.776381 | 0.097646352 |
14 | 1.212437185 | 1.663686004 | 1.841190179 | 1.882882 | 0.139821386 |
15 | 1.321923723 | 1.681590851 | 1.943927409 | 1.994095 | 0.059680911 |
Classes | The Model by Matrix B | ||||||
---|---|---|---|---|---|---|---|
1 | Class 1 | 2.8 | −1.8 | −2.8 | 1.3 | 0.4 | 0.853461 |
2 | 2.9 | −1.9 | −2.9 | 1.4 | 0.5 | 0.870563 | |
3 | 3.0 | −2 | −3.0 | 1.5 | 0.6 | 0.887666 | |
4 | 3.1 | −2.1 | −3.1 | 1.6 | 0.7 | 0.904769 | |
5 | 3.2 | −2.2 | −3.2 | 1.7 | 0.8 | 0.921871 | |
6 | Class 2 | −1.6 | −2.5 | 1.5 | 0.2 | 0.6 | −0.05021 |
7 | −1.3 | −2.7 | 1.3 | 0.4 | 0.8 | −0.0599 | |
8 | −1.0 | −3.0 | 1.5 | 0.6 | 1.0 | −0.37686 | |
9 | −0.7 | −3.2 | 1.7 | 0.8 | 1.2 | −0.74127 | |
10 | −0.5 | −3.5 | 1.9 | 1.0 | 1.4 | −1.01434 | |
11 | Class 3 | 1.2 | −1.2 | 0.7 | −0.3 | −2.8 | −0.40886 |
12 | 1.1 | −1.1 | 0.8 | −0.4 | −2.9 | −0.42597 | |
13 | 1.0 | −1.0 | 0.84 | −0.44 | −3.0 | −0.42784 | |
14 | 0.9 | −0.9 | 0.85 | −0.45 | −3.1 | −0.41905 | |
15 | 0.8 | −0.8 | 0.9 | −0.5 | −3.2 | −0.42244 |
Title 1 | SSIM | PSNR |
---|---|---|
Mean | 0.697 | 17.94 |
Standard Deviation | 0.087 | 1.84 |
Minimum | 0.352 | 11.46 |
25th Percentile | 0.643 | 16.78 |
Median (50th Percentile) | 0.702 | 17.76 |
75th Percentile | 0.758 | 18.94 |
Maximum | 0.898 | 24.08 |
Approach | -Score, % | m, % | Weighted Accuracy, % | |||
---|---|---|---|---|---|---|
Agree | Disagree | Discuss | Unrelated | |||
FNC-1 baseline [42] | 45.62 | 12.46 | 62.32 | 92.40 | 51.60 | 74.67 |
Talos [43] | 50.13 | 3.33 | 69.16 | 93.44 | 55.88 | 84.63 |
Athene [44] | 45.29 | 14.21 | 70.98 | 93.62 | 57.98 | 85.01 |
UCLMR [45] | 44.58 | 10.75 | 67.98 | 92.97 | 55.97 | 84.06 |
Zhang et al. [46] | 62.75 | 66.42 | 76.35 | 92.98 | 77.68 | 89.40 |
Sepulveda-Torres et al. [40] | 69.02 | 60.43 | 78.26 | 93.35 | 77.71 | 89.42 |
Our approach | 68.96 | 60.44 | 78.22 | 93.37 | 77.76 | 89.38 |
Classes | The Model by Matrix B | |||||
---|---|---|---|---|---|---|
Sepal Length | Sepal Width | Petal Length | Petal Width | |||
1 | Class 0 | 5.1 | 3.5 | 1.4 | 0.2 | 0.000036713 |
2 | 4.9 | 3.0 | 1.4 | 0.2 | 0.000020684 | |
3 | 4.7 | 3.2 | 1.3 | 0.2 | 0.000029910 | |
4 | 4.6 | 3.1 | 1.5 | 0.2 | 0.000017786 | |
5 | 5.0 | 3.6 | 1.4 | 0.2 | 0.000039177 | |
… | … | … | … | … | … | |
51 | Class 1 | 7.0 | 3.2 | 4.7 | 1.4 | −0.000125944 |
52 | 6.4 | 3.2 | 4.5 | 1.5 | −0.000122347 | |
53 | 6.9 | 3.1 | 4.9 | 1.5 | −0.000139889 | |
54 | 5.5 | 2.3 | 4.0 | 1.3 | −0.000128922 | |
55 | 6.5 | 2.8 | 4.6 | 1.5 | −0.000138092 | |
… | … | … | … | … | … | |
101 | Class 2 | 6.3 | 3.3 | 6.0 | 2.5 | −0.000202593 |
102 | 5.8 | 2.7 | 5.1 | 1.9 | −0.000173582 | |
103 | 7.1 | 3.0 | 5.9 | 2.1 | −0.000195748 | |
104 | 6.3 | 2.9 | 5.6 | 1.8 | −0.000184621 | |
105 | 6.5 | 3.0 | 5.8 | 2.2 | −0.000196451 | |
… | … | … | … | … | … |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Radiuk, P.; Barmak, O.; Manziuk, E.; Krak, I. Explainable Deep Learning: A Visual Analytics Approach with Transition Matrices. Mathematics 2024, 12, 1024. https://doi.org/10.3390/math12071024
Radiuk P, Barmak O, Manziuk E, Krak I. Explainable Deep Learning: A Visual Analytics Approach with Transition Matrices. Mathematics. 2024; 12(7):1024. https://doi.org/10.3390/math12071024
Chicago/Turabian StyleRadiuk, Pavlo, Olexander Barmak, Eduard Manziuk, and Iurii Krak. 2024. "Explainable Deep Learning: A Visual Analytics Approach with Transition Matrices" Mathematics 12, no. 7: 1024. https://doi.org/10.3390/math12071024
APA StyleRadiuk, P., Barmak, O., Manziuk, E., & Krak, I. (2024). Explainable Deep Learning: A Visual Analytics Approach with Transition Matrices. Mathematics, 12(7), 1024. https://doi.org/10.3390/math12071024