Next Article in Journal
Tailored Direct Oral Anticoagulation in Patients with Atrial Fibrillation: The Future of Oral Anticoagulation?
Next Article in Special Issue
Application of nnU-Net for Automatic Segmentation of Lung Lesions on CT Images and Its Implication for Radiomic Models
Previous Article in Journal
The Gupta Perioperative Risk for Myocardial Infarct or Cardiac Arrest (MICA) Calculator as an Intraoperative Neurologic Deficit Predictor in Carotid Endarterectomy
Previous Article in Special Issue
Detection of Lumbar Spondylolisthesis from X-ray Images Using Deep Learning Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Systematic Review

Artificial Intelligence in the Diagnosis of Hepatocellular Carcinoma: A Systematic Review

by
Alessandro Martinino
1,
Mohammad Aloulou
2,
Surobhi Chatterjee
3,
Juan Pablo Scarano Pereira
4,
Saurabh Singhal
5,
Tapan Patel
6,
Thomas Paul-Emile Kirchgesner
7,
Salvatore Agnes
8,
Salvatore Annunziata
9,
Giorgio Treglia
10,11,12,* and
Francesco Giovinazzo
8,*
1
Department of Surgery, University of Illinois Chicago, Chicago, IL 60607, USA
2
Faculty of Medicine, University of Aleppo, Aleppo 12212, Syria
3
Department of Internal Medicine, King George’s Medical University, Lucknow 226003, Uttar Pradesh, India
4
Faculty of Medicine, Universidad Complutense de Madrid, 28040 Madrid, Spain
5
Department of HPB Surgery and Liver Transplantation, BLK-MAX Superspeciality Hospital, New Delhi 110005, Delhi, India
6
Department of Surgery, Baroda Medical College and SSG Hospital, Vadodara 390001, Gujarat, India
7
Département of Radiology and Medical Imaging, Cliniques Universitaires Saint-Luc, Institut de Recherche Expérimentale et Clinique, Université Catholique de Louvain, 1348 Brussels, Belgium
8
General Surgery and Liver Transplantation Unit, Fondazione Policlinico Universitario Agostino Gemelli IRCCS, 00168 Rome, Italy
9
Unit of Nuclear Medicine, Department of Radiology, Radiotherapy and Hematology, Fondazione Policlinico Universitario A. Gemelli IRCCS, 00168 Rome, Italy
10
Imaging Institute of Southern Switzerland, Ente Ospedaliero Cantonale, 6500 Bellinzona, Switzerland
11
Faculty of Biomedical Sciences, Università della Svizzera Italiana, 6900 Lugano, Switzerland
12
Faculty of Biology and Medicine, University of Lausanne, 1015 Lausanne, Switzerland
*
Authors to whom correspondence should be addressed.
J. Clin. Med. 2022, 11(21), 6368; https://doi.org/10.3390/jcm11216368
Submission received: 13 September 2022 / Revised: 21 October 2022 / Accepted: 26 October 2022 / Published: 28 October 2022
(This article belongs to the Special Issue Artificial Intelligence in Radiology: Present and Future Perspectives)

Abstract

:
Hepatocellular carcinoma ranks fifth amongst the most common malignancies and is the third most common cause of cancer-related death globally. Artificial Intelligence is a rapidly growing field of interest. Following the PRISMA reporting guidelines, we conducted a systematic review to retrieve articles reporting the application of AI in HCC detection and characterization. A total of 27 articles were included and analyzed with our composite score for the evaluation of the quality of the publications. The contingency table reported a statistically significant constant improvement over the years of the total quality score (p = 0.004). Different AI methods have been adopted in the included articles correlated with 19 articles studying CT (41.30%), 20 studying US (43.47%), and 7 studying MRI (15.21%). No article has discussed the use of artificial intelligence in PET and X-ray technology. Our systematic approach has shown that previous works in HCC detection and characterization have assessed the comparability of conventional interpretation with machine learning using US, CT, and MRI. The distribution of the imaging techniques in our analysis reflects the usefulness and evolution of medical imaging for the diagnosis of HCC. Moreover, our results highlight an imminent need for data sharing in collaborative data repositories to minimize unnecessary repetition and wastage of resources.

1. Introduction

Artificial intelligence (AI) is “a field of science and engineering concerned with the computational understanding of what is commonly called intelligent behavior, and with creating artefacts that exhibit such behavior” [1].
Alan Turing first described the use of computers for the simulation of critical thinking and intelligence in 1950. In 1956, John McCarthy coined the definition of AI, the all-encompassing term for computer programs replicating human intelligence. Machine learning is a subset of AI that trains on learning from previous experience and rectifies its functioning sequentially. Deep learning (DL) is a further subset of machine learning that utilizes multi-layered networks between the computing units termed “neurons” that process and validate large training datasets between input and output units, and it leads to meaningful predictions in multiple spheres of medical research (diagnostic, therapeutic, prognostic, etc.) [2].
Hepatocellular carcinoma (HCC) ranks fifth amongst the most common malignancies and is the third most common cause of cancer-related death globally [3]. Though there have been several breakthroughs in the treatment and diagnostic capability, the prognosis of HCC remains dismal due to delayed diagnosis and limited treatment strategies. AI has far reaching potential in the sphere of (a) risk factor stratification, (b) characterization, and (c) improved prognostication in established cases [2]. HCC is a notorious cancer with multiple and overlapping risk factors with the spectrum of its evolving conditions, including NAFLD (Non-Alcoholic Fatty liver disease), NASH (Non-Alcoholic steatohepatitis), and subsequent cirrhosis. Several AI modalities have now been modelled to differentiate and predict the risk of incident HCC [2]. The next challenge lies in classifying indeterminate liver lesions requiring histopathological evidence. The use of computed tomography (CT) and magnetic resonance imaging (MRI) based on DL and radiomics and the success in differentiating between HCC and non-HCC liver nodules with high diagnostic accuracy serve as an essential impetus for creating universal standardized liver tumor segmentation techniques [4]. The following systematic review will expand on the current role of artificial intelligence in HCC detection and characterization, regardless of the instrumental technique.

2. Materials and Methods

Following the Preferred Reporting Items for Systematic Reviews and Meta-analyses (PRISMA) reporting guidelines, we conducted a systematic review. This review reports qualitative data, and because of inconsistent reporting of outcome measures and differences in populations and study design, we did not perform a meta-analysis.

2.1. Searches

PubMed, Scopus, and Cochrane were searched using a combination of the following key words: ((Artificial Intelligence) OR (Machine Learning)) AND ((Hepatocellular Carcinomas) OR (HCC) OR (Liver Cancer)) to retrieve articles reporting the application of AI in detecting or diagnosing HCC. Results were admitted from the time of inception up to and including 5 May 2022. The search terms were modified to fit each database (the terms and their adjustments are found in the Supplementary Materials File S1). Additionally, the reference list of included articles and relevant reviews was checked manually to identify other papers.

2.2. Inclusion and Exclusion Criteria

Only published articles reporting the application of AI in detecting or diagnosing HCC were included, excluding all the studies reporting the application of AI outside the Diagnosis of HCC, such as risk prediction, prognosis, or treatment. Only diagnoses based on CT, MRI, Ultrasound (US), 18F-FDG Positron emission tomography (PET), or X-ray were selected, while other methods like pathology reports or biomarkers were excluded. Reviews, letters, editorials, conference papers, preprints, commentaries, book chapters, or any article in languages other than English were excluded too.

2.3. Quality Assessment

Studies were assessed for quality based on three items:
  • The number of images, estimating the risk of bias and overfitting: fewer than 50 (score 0), 50 to 100 (score 1), and more than 100 (score 2) [5]. This factor was considered the most frequently reported in articles. Where only the number of patients was reported, we considered at least one image per patient.
  • The use of a completely independent cohort for validation: no cohort (score 0), the partition of the cohort between completely separated training and test set (score 1), external validation cohort (score 2).
  • By 2011, the speed of graphics processing units had increased significantly, making it possible to train convolutional neural networks “without” the layer-by-layer pre-training. With the increased computing speed, deep learning had significant advantages in terms of efficiency and speed: no data (score 0), before 2011 (score 1), 2011 or after (score 2).
A simple quality score (QS), consisting of the sum of the 3 previously stated items, was calculated. A maximum possible score of 6 meant a high-quality study design of the article.

2.4. Study Selection & Data Extraction

Duplicates were removed using Endnote X9. Titles and/or abstracts of studies identified using our search criteria were screened independently by 2 authors (A.M. & M.A.) to identify all studies meeting our inclusion criteria. Any disagreement was resolved through discussion with a third reviewer (F.G.). Random included articles were used to generate an extraction sheet. Three authors (A.M., M.A., and J.P.S.P) reviewed the full texts for inclusion and data extraction. Any discrepancies were corrected by consensus. The following parameters were extracted from each article:
  • PMID; First author; Year of publication; Country; Journal.
  • The number of patients; Diagnostic method; AI method.
  • Research question; Key findings.
  • Quality score.
F.G. then reviewed all articles, rechecked data, and analyzed them using an Excel (R) sheet. Statistical calculations were performed with Jamovi (R) software version 2.0.0.0 [6,7].

3. Results

3.1. Searching Results

The study flow diagram is illustrated in Figure 1. Searches identified 3160 records: 1677 from PubMed, 1426 from Scopus, and 57 from Cochrane. A total of 1052 were duplicates and automatically excluded using EndNoteX9. A total of 2108 studies were evaluated by title/abstract screening against the eligibility criteria, and 2032 were excluded. Of these, 1813 were not related to the topic, 5 not including HCC, 5 not including AI, 12 not discussing diagnosis, 80 duplicates not detected by the software, 62 conference papers, 26 reviews, 5 book chapters, and 24 letters. Of the remaining 76 records potentially eligible, after the full-texts screening, 27 articles were included, and 49 were excluded because 6 were not related to the topic, 7 did not include HCC, 3 did not include AI, 11 did not discuss diagnosis, 9 diagnosed based on methods other than CT/PET/MRI/US/X-ray, 5 articles were in a language other than English, 4 were reviews, 3 articles were not available, and 1 was a clinical trial with no published data. After the manual search, 19 articles were further identified. Thus, a total of 46 cited articles were included in this review, published between 1998 and 2022 (Table 1).

3.2. Quality Assessments

The mean of the “Number of Images” score was 1.70, identifying 36 articles (78.3%) where at least 100 images were analyzed. (Table 2) The mean of the “Cohort for Validation” Score was 0.609. Indeed, an external validation cohort was used only in 2 articles (4.3%). (Table 3) The mean of the “Year of Publication” score was 1.87, documenting that most of the works (87.0%) included in this systematic review were published in 2011 or later. (Table 4) On average, the Total Quality Score was 4.17, with a median of 4.00 and SD of 1.04. (Table 5) The contingency table correlating the Total Score with the Year of Publication reports a statistically significant constant improvement over the years of the quality score (p = 0.004). (Table 6) A total of 3 articles (6.52%) scored a QS lower than 3, while 2 (4.34%) received the maximum score. Results from articles with a QS strictly lower than 3 are written in italics in Table 7.

3.3. Results

Different AI methods have been adopted in the included articles, such as CNN (Convolutional Neural Network), SVM (Support-Vector Machine), RF (Random Forest), KNN (K-Nearest Neighbor), PM-DL (pattern matching and deep learning), ANN (Artificial Neural Network), DNN (Deep Neural Network), CDNs (Convolutional Dense Networks), DLS (Deep Learning System), GLM (Generalized Linear Model), DWT (Discrete Wavelet Transform), LSTM (Long Short-Term Memory), NNE (Neural Network Ensemble), and LDA (Linear Discriminant Analysis). A total of 19 articles used CT (41.30%), 20 used US (43.47%), and 7 used MRI (15.21%) in their work. No article has discussed the use of artificial intelligence in PET and X-ray technology. Table 7 lists the total study population, diagnostic method, research question or purpose, AI method, key findings included in this systematic review to summarize how artificial intelligence is used today in diagnosing HCC. Moreover, when the information was available, we reported in Table 7 the background of the images studied, i.e., whether HCC on the cirrhotic or healthy liver and whether other cancerous and benign lesions studied were present.
Table 7. Characteristics of the studies included in the systematic review.
Table 7. Characteristics of the studies included in the systematic review.
StudyCountry/RegionJournalTotal Study PopulationDiagnostic MethodResearch Question/PurposeAI MethodKey Findings
Ziegelmayer et al., 2022 [8]GermanyInvestigative Radiology60 patientsCTTo compare the robustness of CNN features versus radiomics features to technical variations in image acquisition parameters.CNNCNN features were more stable.
Xu et al., 2022 [9]ChinaComputational and Mathematical Methods in Medicine211 patients (122 training set, 89 testing set)CTTo establish an SVM based on radiomic features at non-contrast CT to train a discriminative model for HCC and ICCA at early stage.SVMThe model may facilitate the differential diagnosis of HCC and ICCA in the future.
Turco et al., 2022 [10]USAIEEE Transactions on Ultrasonics, Ferroelectrics, and Frequency control72 patientsUSProposes an interpretable radiomics approach to differentiate between malignant and benign FLLs on CEUS.Logistic regression, SVM, RF, and KNNAspects related to perfusion (peak time and wash-in time), the microvascular architecture (spatiotemporal coherence), and the spatial characteristics of contrast enhancement at wash-in (global kurtosis) and peak (GLCM Energy) are particularly relevant to aid FLLs diagnosis.
Sato et al., 2022 [11]JapanJournal of Gastroenterology and Hepatology972 patients (864 training set, 108 testing set)USTo analyse the diagnostic performance of deep multimodal representation model-based integration of tumour image, patient background, and blood biomarkers for the differentiation of liver tumours observed using B-mode US.CNNThe integration of patient background information and blood biomarkers in addition to US images, multi-modal representation learning outperformed the CNN model that used US images alone.
Rela et al., 2022 [12]IndiaInternational Journal of Advanced Technology and Engineering Exploration68 patients (51 training set, 17 testing set)CTDifferent machine learning algorithms are used to classify the tumour as liver abscess and HCC.SVM, KNN, Decision tree, Ensemble, and Naive BayesSVM classifier gives better performance compared to all other AI methods in the study.
Zheng et al., 2021 [13]ChinaPhysics in Medicine and Biology120 patients (56 training set with 5376 images, 64 testing set with 6144 images)MRITo investigate the feasibility of automatic detection of small HCC (≤2 cm) based on PM-DL model.CNNThe superior performance both in the validation cohort and external test cohort indicated the proposed PM-DL model may be feasible for automatic detection of small HCCs with high accuracy.
Yang et al., 2021 [14]TaiwanPLoS One731 patients (394 training set with 10,130 images, 337 testing set with 22,116 images)CTTo use a previously proposed mask region–based CNN for automatic abnormal liver density detection and segmentation based on HCC CT datasets from a radiological perspective.CNNThe study revealed that this single deep learning model cannot replace the complex and subtle medical evaluations of radiologists, but it can reduce tedious labour.
Stollmayer et al., 2021 [15]HungaryWorld Journal of Gastroenterology69 patients (training set with 186 images, testing set with 30 images)MRITo compare diagnostic efficiency of 2D and 3D-densely connected CNN (DenseNet) for FLLs on multi-sequence MRI.CNNBoth 2D and 3D-DenseNets can differentiate FNH, HCC and MET with good accuracy when trained on hepatocyte-specific contrast-enhanced multi-sequence MRI volumes.
Kim et al., 2021 [16]KoreaEuropean Radiology1320 patients (training set with 568 images, testing set with 589 images, tuning set with 193 images)CTTo develop and evaluate a deep learning-based model capable of detecting primary hepatic malignancies in multiphase CT images of patients at high risk for HCC.CNNThe proposed model exhibited an 84.8% of sensitivity with 4.80 false positives per CT scan in the test set.
Căleanu et al., 2021 [17]RomaniaSensors91 patientsUSTo examine the application of CEUS for automated FLL diagnosis using DNN.DNNThis deep learning-based method provides comparable or better results, for an increased number of FLL types.
Zhou et al., 2020 [18]ChinaFrontiers in Oncology435 patients (616 liver lesions; 462 training set, 154 testing set)CTTo propose a framework based on hierarchical CNNs for automatic detection and classification FLLs in multi-phasic CT.Hierarchical CNNsOverall, this preliminary study demonstrates that the proposed multi-modality and multi-scale CNN structure can locate and classify FLLs accurately in a limited dataset and would help inexperienced physicians to reach a diagnosis in clinical practice.
Kim et al., 2020 [19]South KoreaScientific Reports549 patients, and external validation data set (54 patients)MRITo develop a fully automated deep learning model to detect HCC using hepatobiliary phase MR images and evaluate its performance in detecting HCC on liver MRI compared to human readersFine-tuned CNNThe optimised CNN architecture achieved 94% sensitivity, 99% specificity, and 0.97 area under curve (AUC) for HCC cases in the test dataset and achieved 87% sensitivity and 93% specificity and an AUC of 0.90 for external validation datasets.
Huang et al., 2020 [20]ChinaIEEE journal of biomedical and health informaticsData set 1: 155 patients with FNH and 49 patients with atypical HCC
Data set 2: 102 patients with FNH and only 36 patients with atypical HCC
USTo propose a novel liver tumour CAD approach extracting spatial-temporal semantics for atypical HCC.SVMThe average accuracy reaches 94.40%, recall rate 94.76%, F1-score value 94.62%, specificity 93.62% and sensitivity 94.76%.
Shi et al., 2020 [21]NAAbdominal Radiology342 patients with 449 untreated lesions (194 HCC group; 255 non-HCC group)CTTo evaluate whether a three-phase dynamic contrast-enhanced CT protocol, when combined with a deep learning model, has similar accuracy in differentiating HCC from other FLLs) compared with a four-phase protocol.CDNsWhen combined with a CDN, a three-phase CT protocol without pre-contrast showed similar diagnostic accuracy as a four-phase protocol in differentiating HCC from other FLLs, suggesting that the multiphase CT protocol for HCC diagnosis might be optimised by removing the pre-contrast phase to reduce radiation dose.
Zhen et al., 2020 [22]ChinaFrontiers in Oncology1210 patients (31,608 images), and external validated cohort of 201 patients (6816 images)MRITo develop a DLS to classify liver tumours.CNNDLS that integrated these models could be used as an accurate and timesaving assisted-diagnostic strategy for liver tumours in clinical settings, even in the absence of contrast agents. DLS therefore has the potential to avoid contrast-related side effects and reduce economic costs associated with current standard MRI inspection practices for liver tumour patients.
Krishan et al., 2020 [23]IndiaProceedings of the Institution of Mechanical Engineers, Part H: Journal of Engineering in Medicine794 normal liver images and 844 abnormal liver types (483 MET, 361 HCC)CTTo detect the presence of a tumour region in the liver and classify the different stages of the tumour from CT images.R-part decision tree, AdaBoost, RF, k-SVM, GLM, and NN. A multi-level ensemble model is also developed.The accuracy achieved for different classifiers varies between 98.39% and 100% for tumour identification and between 76.38% and 87.01% for tumour classification. The multi-level ensemble model achieved high accuracy in both the detection and classification of different tumours.
Brehar et al., 2020 [24]RomaniaSensors268 patientsUSTo compare deep-learning and conventional machine-learning methods for the automatic recognition of the HCC areas from US imagesCNNsThe achieved results show that the deep-learning approach overcomes classical machine-learning solutions, by providing a higher classification performance.
Hamm et al., 2019 [25]NAEuropean radiology296 patients; 334 imaging studies; 494 hepatic lesions divided into training (434) and test sets (60) MRITo develop and validate a proof-of-concept CNN–based DLS that classifies common hepatic lesions on multi-phasic MRI.CNNThis preliminary deep learning study demonstrated feasibility for classifying lesions with typical imaging features from six common hepatic lesion types.
Das et al., 2019 [26]IndiaPattern Recognition and Image Analysis123 real-time images (63 HCC, and 60 metastasis carcinoma)CTTo present an automatic approach that integrates the adaptive thresholding and spatial fuzzy clustering approach for detection of cancer region in CT scan images of liver.Multilayer perceptron and C4.5 decision tree classifiersThis result proves that the spatial fuzzy c-means-based segmentation with C4.5 decision tree classifier is an effective approach for automatic recognition of the liver cancer.
Trivizakis et al., 2019 [27]GreeceIEEE Journal of Biomedical and Health Informatics130 images for the training and validation of the networkMRIPropose and evaluate a novel 3D CNN designed for tissue classification in medical imaging and applied for discriminating between primary and metastatic liver tumours from diffusion weighted MRI data.3D CNNThe proposed 3D CNN architecture can bring significant benefit in diffusion weighted MRI liver discrimination and potentially, in numerous other tissue classification problems based on tomographic data, especially in size-limited, disease-specific clinical datasets.
Kutlu et al., 2019 [28]TurkeySensors56 images benign and 56 malignant liver tumoursCTA new liver and brain tumour classification method is proposed by using the power of CNN in feature extraction, the power of DWT in signal processing, and the power of LSTM in signal classification.CNN in feature extraction, DWT in signal processing, and LSTM in signal classificationThe proposed method has a satisfactory accuracy rate at the liver tumour and brain tumour classifying.
Nayak et al., 2019 [29]IndiaInternational Journal of Computer Assisted Radiology and Surgery40 patients (healthy 14, cirrhosis 12, and cirrhosis with HCC 14)CTTo proposes a CAD system for detecting cirrhosis and HCC in a very efficient and less time-consuming approach.SVMThe proposed CAD system showed promising results and can be used as effective screening tool in medical image analysis.
Schmauch et al., 2019 [30]FranceDiagnostic and Interventional Imaging544 patients (367 training set, 177 test set)USTo create an algorithm that simultaneously detects and characterises (benign vs. malignant) FLL using deep learning.ANNThis method could prove to be highly relevant for medical imaging once validated on a larger independent cohort.
Jansen et al., 2019 [31]The NetherlandsPLoS ONE95 patients (213 images)MRIAdditional MR sequences and risk factors are used for automatic classification.Randomised trees classifierThe proposed classification can differentiate five common types of lesions and is a step forward to a clinically useful aid.
Das et al., 2019 [32]IndiaCognitive Systems Research75 patients (225 images)CTTo introduce a new automated technique based on watershed–Gaussian segmentation approach.DNNThe developed system is ready to be tested with huge database and can aid the radiologist in detecting the liver cancer using CT images.
Mokrane et al., 2019 [33]FranceEuropean Radiology178 patients (142 training set, 36 validations set)CTTo enhance clinician’s decision-making by diagnosing HCC in cirrhotic patients with indeterminate liver nodules using quantitative imaging features extracted from triphasic CT scans.KNN, SVM, and RFA proof of concept that machine-learning-based radiomics signature using change in quantitative CT features across the arterial and portal venous phases can allow a non-invasive accurate diagnosis of HCCs in cirrhotic patients with indeterminate nodules.
Lakshmipriya et al., 2019 [34]IndiaJournal of Biomedical and Health Informatics634 images (440 images training set, 194 images validation set)CTAn ensemble FCNet classifier is proposed to classify hepatic lesions from the deep features extracted using GoogleNetLReLU transfer learning approach.CNNResults demonstrate the efficacy of the proposed classifier design in achieving better classification accuracy.
Acharya et al., 2018 [35]MalaysiaComputers in biology and medicine101 patients with 463 images USThis study initiates a CAD system to aid radiologists in an objective and more reliable interpretation of ultrasound images of liver lesions.Radon transform and bi-directional empirical mode decomposition to extract features from the focal liver lesions. The accuracy, sensitivity, and specificity of lesion classification were 92.95%, 90.80%, and 97.44%, respectively.
Ta et al., 2018 [36]USARadiology106 images (54 malignant, 51 benign, and one indeterminate FLL)USTo assess the performance of CAD systems and to determine the dominant US features when classifying benign versus malignant FLLs by using contrast material–enhanced US cine clips.ANN and SVMCAD systems classified benign and malignant FLLs with an accuracy like that of an expert reader. CAD improved the accuracy of both readers. Time-based features of TIC were more discriminating than intensity-based features.
Bharti et al., 2018 [37]IndiaUltrasonic Imaging94 patients (189 images)USTo deal with this difficult visualisation problem, a method has been developed for classifying four liver stages, that is, normal, chronic, cirrhosis, and HCC evolved over cirrhosis.KNN, SVM, rotation forest, CNNsThe experimental results strongly suggest that the proposed ensemble classifier model is beneficial for differentiating liver stages based on US images.
Yasaka et al., 2018 [38]JapanRadiology560 patients (460 patients training set with 55,536 images, 100 patients validation set with 100 images)CTTo investigate diagnostic performance by using a deep learning method with a CNN for the differentiation of liver masses at dynamic contrast agent-enhanced CT.CNNDeep learning with CNN showed high diagnostic performance in differentiation of liver masses at dynamic CT.
Hassan et al., 2017 [39]EgyptArabian Journal for Science and Engineering110 patients (110 images)USA new classification framework is introduced for diagnosing focal liver diseases based on deep learning architecture.Stacked Sparse AutoencoderOur proposed method presented overall accuracy of 97.2% compared with multi-SVM, KNN, and Naïve Bayes.
Guo et al., 2017 [40]ChinaClinical Hemorheology and Microcirculation93 patientsUSTo propose a novel two-stage multi-view learning framework for the CEUS based CAD for liver tumours, which adopted only three typical CEUS images selected from the arterial phase, portal venous phase and late phase.Deep canonical correlation analysis and multiple kernel learningThe experimental results indicate that the proposed achieves best performance for discriminating benign liver tumours from malignant liver cancers.
Kondo et al., 2017 [41]JapanTransactions on Medical Imaging98 patientsUSTo propose an automatic classification method based on machine learning in CEUS of FLLs using the contrast agent Sonazoid.SVMThe results indicated that combining the features from the arterial, portal, and post-vascular phases was important for classification methods based on machine learning for Sonazoid CEUS.
Gatos et al., 2015 [42]NAMedical physics52 patients; (30 benign and 22 malignant)USDetect and classify FLLs from CEUS imaging by means of an automated quantification algorithm.SVMsThe proposed quantification system that employs FLLs detection and classification algorithms may be of value to physicians as a second opinion tool for avoiding unnecessary invasive procedures.
Virmani et al., 2014 [43]IndiaJournal of Digital Imaging108 imagesUSAn NNE-based CAD system to assist radiologists in differential diagnosis between FLLs.NNEThe promising results obtained by the proposed system indicate its usefulness to assist radiologists in differential diagnosis of FLLs.
Wu et al., 2014 [44]ChinaOptik22 patientsUSTo propose a diagnostic system for liver disease classification based on CEUS imaging.DNNQuantitative comparisons demonstrate that the proposed method outperforms the compared classification methods in accuracy, sensitivity, and specificity
Virmani et al., 2013 [45]IndiaDefence Science Journal108 images comprising of 21 NOR images, 12 Cyst, 15 HEM, 28 HCC, and 32 METUSTo investigate the contribution made by texture of regions inside and outside of the lesions in FLLs. SVMThe proposed PCA-SVM based CAD system yielded classification accuracy of 87.2% with the individual class accuracy of 85%, 96%, 90%, 87.5%, and 82.2% for NOR, Cyst, HEM, HCC, and MET cases, respectively. The accuracy for typical, atypical, small HCC and large HCC cases is 87.5%, 86.8%, 88.8%, and 87%, respectively.
Streba et al., 2012 [46]RomaniaWorld Journal of Gastroenterology224 patientsUSTo study the role of time-intensity curve analysis parameters in a complex system of neural networks designed to classify liver tumours.ANNNeural network analysis of CEUS-obtained time-intensity curves seem a promising field of development for future techniques, providing fast and reliable diagnostic aid for the clinician.
Mittal et al., 2011 [47]IndiaComputerized Medical Imaging and Graphics88 patients with 111 images comprising 16 normal liver, 17 Cyst, 15 HCC, 18 HEM and 45 METUSIt proposes a CAD system to assist radiologists in identifying focal liver lesions in B-mode ultrasound images.Two step neural network classifierThe classifier has given correct diagnosis of 90.3% (308/340) in the tested segmented regions-of-interest from typical cases and 77.5% (124/160) in tested segmented regions-of-interest from atypical cases.
Sugimoto et al., 2010 [48]JapanWorld Journal of Radiology137 patients (74 HCCs, 33 liver metastases and 30 liver hemangiomas)USTo introduce CAD aimed at differential Diagnosis of FLLs by use of CEUS.ANNsThe classification accuracies were 84.8% for metastasis, 93.3% for hemangioma, and 98.6% for all HCCs. In addition, the classification accuracies for histologic differentiation types of HCCs were 65.2% for w-HCC, 41.7% for m-HCC, and 80.0% for p-HCC.
Shiraishi et al., 2008 [49]JapanMedical Physics97 patients, (103 images; 26 metastases, 16 hemangiomas, and 61 HCCs)USTo develop a CAD scheme for classifying focal liver lesions as liver metastasis, hemangioma, and three histologic differentiation types of HCC, by use of microflow imaging of CEUS.ANNsThe classification accuracies for the 103 FLLs were 88.5% for metastasis, 93.8% for hemangioma, and 86.9% for all HCCs. In addition, the classification accuracies for histologic differentiation types of HCCs were 79.2% for w-HCC, 50.0% for m-HCC, and 77.8% for p-HCC.
Stoitsis et al., 2006 [50]GreeceNuclear Instruments and Methods in Physics Research147 images (normal liver 76, hepatic cyst 19, hemangioma 28, HCC 24)CTTo classify of four types of hepatic tissue: normal liver, hepatic cyst, hemangioma, and hepatocellular carcinoma, from CT images.Combined use of texture features and classifiersThe achieved classification performance was 100%, 93.75%, and 90.63% in the training, validation, and testing set, respectively.
Matake et al., 2006 [51]NAAcademic radiology120 patientsCTTo apply an ANN for differential diagnosis of certain hepatic masses on CT images and evaluate the effect of ANN output on radiologist diagnostic performance.ANNThe ANN can provide useful output as a second opinion to improve radiologist diagnostic performance in the differential diagnosis of hepatic masses seen on contrast-enhanced CT.
Gletsos et al., 2003 [52]GreeceIEEE transactions on information technology in biomedicine147 patientsCTTo present a CAD system for the classification of hepatic lesions from CT images. Neural-Network ClassifierThe suitability of co-occurrence texture features, the superiority of GAs for feature selection, compared to sequential search methods, and the high performance achieved by the NN classifiers in the testing images set have been demonstrated.
Chen et al., 1998 [53]TaiwanIEEE Transactions on Biomedical Engineering30 patientsCTTo present a CT liver image diagnostic classification system which will automatically find, extract the CT liver boundary, and further classify liver diseases.Modified probabilistic NNThe proposed system was evaluated by 30 liver cases and shown to be efficient and very effective.
AI: Artificial Intelligence; CT: Computerized Tomography; CNN: Convolutional Neural Network; SVM: Support-Vector Machine; HCC: Hepatocellular Carcinoma; ICCA: Intrahepatic Cholangiocarcinoma; CEUS: Contrast Enhanced Ultrasound; FLL: Focal Liver Lesion; RF: Random Forest; KNN: K-Nearest Neighbor; US: Ultrasound; MRI: Magnetic Resonance Imaging; PM-DL: pattern matching and deep learning; 2D: Two-Dimensional; 3D: Three-Dimensional; FNH: Focal Nodular Hyperplasia; MET: Metastatic; ANN: Artificial Neural Network; DNN: Deep Neural Network; CAD: Computer-Aided Diagnosis; NA: Not Applicable; CDNs: Convolutional Dense Network; DLS: Deep Learning System; GLM: Generalized Linear Model; NN: Neural Network; DWT: Discrete Wavelet Transform; LSTM: Long Short-Term Memory; NNE: Neural Network Ensemble; LDA: Linear Discriminant Analysis.

4. Discussion

Artificial Intelligence is a rapidly growing field of interest. It has immense potential to be the standard of care in resource-limited settings where there is lesser availability of expert care and a heavy burden of cancer volume load. However, the use of AI and ML-based algorithms are limited in current practice owing to their limited generalizability. ML algorithms require large training sets of data, processing using GPUs and functions on the GIGO (Garbage in Garbage out) principle, which means that the output is as robust as the input obtained. However, robustness and standardization of large datasets, including follow-up evaluation and patient quality of care, is extremely cumbersome and difficult. The incongruity between modelled datasets versus real-world data is a fundamental challenge that must be overcome in the future [2].
We have grouped systemically the articles using artificial intelligence in HCC detection and characterization in a unique table, helping to plan further research projects. Indeed, for each article, we extracted the scope, the AI method used, and the key findings related to that AI approach, with the idea of having an index of all projects carried out to date. The significant heterogeneity of the studies reflected the difficulty of extrapolating several variables related to the different radiological techniques and pulling them together (e.g., gold standard used for the diagnosis of HCC, patient features, radiologist’s opinion, dose and type of contrast agent, and follow-up imaging).
In this work, 27 articles were analyzed with our composite score for the evaluation of the quality of the publications, with an overall score at 4.17/6. The “Cohort for Validation” score was the lowest, indeed, an external validation cohort was used only in 2 articles. This phenomenon, although explained by the difficulty of collecting data, limits the generalizability of the conclusions. We observe a statistically significant constant improvement over the years on our composite criterion combining the number of images and the presence of a validation cohort. (p = 0.004) This improvement is probably due to the publication of guidelines, dedicated checklists to ensure proper methodology, and technological improvement in the field of AI.
Our results highlight an imminent need for data sharing in collaborative data repositories to minimize unnecessary repetition and wastage of resources. In addition, universal standardized data sharing protocols for sharing datasets from clinical trials are essential to help make the available data robust and fill in the missing data. One such example is the creation of the Human Brain Project and project EBRAINS by the European Union to handle data related to brain research and its broader usage in the development of AI networks [54]. To help make the datasets uniformly accessible and usable, it is also imperative to diversify the data. Most of the work on AI-based algorithms was done on small-scale datasets due to economic and logistic constraints in high-income developed countries with limited to no data from lower middle- and low-income countries, which puts their credibility in ambiguity. Significant work needs to be done to increase the transparency and understanding of AI algorithms so that healthcare professionals gain confidence in using them in clinical settings.
Our systematic approach has shown that previous works in HCC detection and characterization have assessed the comparability of conventional interpretation with machine learning using US, CT, and MRI. The distribution of the imaging techniques in our analysis reflects the usefulness and evolution of medical imaging for the diagnosis of HCC. Ultrasound and CT are overrepresented in our analysis as both are easily available imaging techniques that have long proven their usefulness in the diagnosis of HCC. More recent and limited access to MRI may explain its absence before and low representation since 2019 in our analysis. On the opposite side, no study investigated X-ray or PET techniques. Indeed, even if X-ray has an interesting role in interventional therapeutic procedures, this technique has not been used for diagnostic purposes in this field. Moreover, unlike the other branches of medicine, such as neurology [55], head and neck cancer [56], or lung imaging [57,58,59], artificial intelligence in PET technology has not yet been studied and tested in HCC diagnosis. As PET, in combination with CT scan, is already used in other cancer to define undetermined lesions with high sensitivity and precision, AI and PET technology in HCC have not been explored yet. The most straightforward explanation can be found in the difficulty in analyzing structural and morphological characteristics, the hepatocellular cancer lesions having a variable degree of avidity for the PET tracers such as 18F-FDG. Indeed, the liver is unique in its capacity to maintain glucose homeostasis, thus leading to low 18F-FDG uptakes in low-grade (i.e., relatively metabolically less active) tumors [60]. It has been reported that only up to two-thirds of the tumors are 18F-FDG avid, although higher standardized uptake values (SUV) indicate a more malignant tumor [61,62]. Using other tracers such as 18F-Choline and 11C-Acetate may be a promising approach to increase the accuracy of results and openness to new AI technologies in combination with PET in diagnosing HCC [63,64].
In the future, DL algorithms combining clinical, radiological, pathological, and molecular information can help identify and better prognosticate patients. In addition, algorithms trained on post-chemotherapy patients could help in the early identification of their response and the time to switch between other therapeutic options. This will enable earlier identification of patients with poor treatment response and pre-emptive therapy adjustment based on molecular signature and imaging [2,4]. Anyhow, conducting high quality AI studies with large sets of data remains a real challenge whatever the medical imaging technique. Supervised and moreover unsupervised training-based algorithms need very large sets of data for training but also for validation purpose. High quality methodology requires standardized multi-parametric imaging acquisition protocols and solid diagnostic methods including multiple reader assessment, follow-up imaging, and/or anatomopathological. Multi-center AI studies and pooled imaging data could be an effective solution to spare time and financial resources.

Limitations and Strengths

The most significant limitation of this review is a wide diversity from one article to another in terms of textural parameters and methods used, which meant that even for similar subjects, it was challenging to aggregate and compare the articles between them. Secondly, the scale used to assess the quality of the articles was practical but rather simplistic. This score made it possible to evaluate many articles with high reproducibility at the expense of a thorough analysis of the methods. At the same time, to the best of the authors’ knowledge, this is the first systematic review in the scientific literature focusing on the use of AI in radiological HCC detection and characterization, omitting pathology and prognosis. This allowed for a detailed analysis that described all the scientific techniques and efforts studied in this narrow field, providing an overview that can provide points for reflection and guide future research.

5. Conclusions

Our systematic approach has shown that previous works in HCC detection and characterization have assessed the comparability of conventional interpretation with machine learning using US, CT, and MRI. The distribution of the imaging techniques in our analysis reflects the usefulness and evolution of medical imaging for the diagnosis of HCC. Moreover, our results highlight an imminent need for data sharing in collaborative data repositories to minimize unnecessary repetition and wastage of resources.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/jcm11216368/s1, File S1: Search Strategy.

Author Contributions

Conceptualization: A.M., M.A., S.A. (Salvatore Annunziata), G.T. and F.G.; Methodology: A.M., M.A., S.A. (Salvatore Annunziata), G.T. and F.G.; Software: A.M., M.A. and F.G.; Validation, S.A. (Salvatore Annunziata), S.A. (Salvatore Agnes), G.T. and F.G.; Writing: A.M., M.A., S.C., J.P.S.P., S.S., T.P. and T.P.-E.K.; Supervision: S.A. (Salvatore Annunziata), S.A. (Salvatore Agnes), G.T. and F.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shapiro, S.C. Encyclopaedia of Artificial Intelligence, 2nd ed.; John Wiley & Sons: Nashville, TN, USA, 1992; Volume 1. [Google Scholar]
  2. Calderaro, J.; Seraphin, T.P.; Luedde, T.; Simon, T.G. Artificial intelligence for the prevention and clinical management of hepatocellular carcinoma. J. Hepatol. 2022, 76, 1348–1361. [Google Scholar] [CrossRef] [PubMed]
  3. Sung, H.; Ferlay, J.; Siegel, R.L.; Laversanne, M.; Soerjomataram, I.; Jemal, A.; Bray, F. Global Cancer Statistics 2020: GLOBOCAN Estimates of Incidence and Mortality Worldwide for 36 Cancers in 185 Countries. CA Cancer J. Clin. 2021, 71, 209–249. [Google Scholar] [CrossRef] [PubMed]
  4. Laino, M.E.; Viganò, L.; Ammirabile, A.; Lofino, L.; Generali, E.; Francone, M.; Lleo, A.; Saba, L.; Savevski, V. The Added Value of Artificial Intelligence to LI-RADS Categorization: A Systematic Review. Eur. J. Radiol. 2022, 150, 110251. Available online: https://www.ejradiology.com/article/S0720-048X(22)00101-2/fulltext (accessed on 5 July 2022). [CrossRef] [PubMed]
  5. Morland, D.; Triumbari, E.K.A.; Boldrini, L.; Gatta, R.; Pizzuto, D.; Annunziata, S. Radiomics in Oncological PET Imaging: A Systematic Review—Part 1, Supradiaphragmatic Cancers. Diagnostics 2022, 12, 1329. [Google Scholar] [CrossRef]
  6. The Jamovi Project. Jamovi. (Version 2.2) [Computer Software]. 2021. Available online: https://www.jamovi.org (accessed on 1 August 2022).
  7. R Core Team. R: A Language and Environment for Statistical Computing. (Version 4.0) [Computer Software]. 2021. Available online: https://cran.r-project.org (accessed on 1 August 2022).
  8. Ziegelmayer, S.; Reischl, S.; Harder, F.; Makowski, M.; Braren, R.; Gawlitza, J. Feature Robustness and Diagnostic Capabilities of Convolutional Neural Networks against Radiomics Features in Computed Tomography Imaging. Investig. Radiol. 2022, 57, 171–177. [Google Scholar] [CrossRef]
  9. Xu, X.; Mao, Y.; Tang, Y.; Liu, Y.; Xue, C.; Yue, Q.; Liu, Y.; Wang, J.; Yin, Y. Classification of Hepatocellular Carcinoma and Intrahepatic Cholangiocarcinoma Based on Radiomic Analysis. Comput. Math. Methods Med. 2022, 2022, 5334095. [Google Scholar] [CrossRef]
  10. Turco, S.; Tiyarattanachai, T.; Ebrahimkheil, K.; Eisenbrey, J.; Kamaya, A.; Mischi, M.; Lyshchik, A.; Kaffas, A.E. Interpretable Machine Learning for Characterization of Focal Liver Lesions by Contrast-Enhanced Ultrasound. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 2022, 69, 1670–1681. [Google Scholar] [CrossRef]
  11. Sato, M.; Kobayashi, T.; Soroida, Y.; Tanaka, T.; Nakatsuka, T.; Nakagawa, H.; Nakamura, A.; Kurihara, M.; Endo, M.; Hikita, H.; et al. Development of novel deep multimodal representation learning-based model for the differentiation of liver tumors on B-mode ultrasound images. J. Gastroenterol. Hepatol. 2022, 37, 678–684. [Google Scholar] [CrossRef]
  12. Rela, M.; Rao, S.N.; Patil, R.R. Performance Analysis of Liver Tumor Classification Using Machine Learning Algorithms. IJATEE 2022, 9, 143. Available online: https://www.accentsjournals.org/paperInfo.php?journalPaperId=1393 (accessed on 20 July 2022).
  13. Zheng, R.; Wang, L.; Wang, C.; Yu, X.; Chen, W.; Li, Y.; Li, W.; Yan, F.; Wang, H.; Li, R. Feasibility of automatic detection of small hepatocellular carcinoma (≤2 cm) in cirrhotic liver based on pattern matching and deep learning. Phys. Med. Biol. 2021, 66, 8. [Google Scholar] [CrossRef]
  14. Yang, C.J.; Wang, C.K.; Fang, Y.H.D.; Wang, J.Y.; Su, F.C.; Tsai, H.M.; Lin, Y.J.; Tsai, H.W.; Yee, L.R. Clinical application of mask region-based convolutional neural network for the automatic detection and segmentation of abnormal liver density based on hepatocellular carcinoma computed tomography datasets. PLoS ONE 2021, 16, e0255605. [Google Scholar] [CrossRef] [PubMed]
  15. Stollmayer, R.; Budai, B.K.; Tóth, A.; Kalina, I.; Hartmann, E.; Szoldán, P.; Bérczi, V.; Maurovich-Horvat, P.; Kaposi, P.N. Diagnosis of focal liver lesions with deep learning-based multi-channel analysis of hepatocyte-specific contrast-enhanced magnetic resonance imaging. World J. Gastroenterol. 2021, 27, 5978–5988. [Google Scholar] [CrossRef]
  16. Kim, D.W.; Lee, G.; Kim, S.Y.; Ahn, G.; Lee, J.G.; Lee, S.S.; Kim, K.W.; Park, S.H.; Lee, Y.J.; Kim, N. Deep learning-based algorithm to detect primary hepatic malignancy in multiphase CT of patients at high risk for HCC. Eur. Radiol. 2021, 31, 7047–7057. [Google Scholar] [CrossRef] [PubMed]
  17. Căleanu, C.D.; Sîrbu, C.L.; Simion, G. Deep Neural Architectures for Contrast Enhanced Ultrasound (CEUS) Focal Liver Lesions Automated Diagnosis. Sensors 2021, 21, 4126. [Google Scholar] [CrossRef] [PubMed]
  18. Zhou, J.; Wang, W.; Lei, B.; Ge, W.; Huang, Y.; Zhang, L.; Yan, Y.; Zhou, D.; Ding, Y.; Wu, J.; et al. Automatic Detection and Classification of Focal Liver Lesions Based on Deep Convolutional Neural Networks: A Preliminary Study. Front. Oncol. 2020, 10, 581210. [Google Scholar] [CrossRef]
  19. Kim, J.; Min, J.H.; Kim, S.K.; Shin, S.Y.; Lee, M.W. Detection of Hepatocellular Carcinoma in Contrast-Enhanced Magnetic Resonance Imaging Using Deep Learning Classifier: A Multi-Center Retrospective Study. Sci. Rep. 2020, 10, 9458. [Google Scholar] [CrossRef]
  20. Huang, Q.; Pan, F.; Li, W.; Yuan, F.; Hu, H.; Huang, J.; Yu, J.; Wang, W. Differential Diagnosis of Atypical Hepatocellular Carcinoma in Contrast-Enhanced Ultrasound Using Spatio-Temporal Diagnostic Semantics. IEEE J. Biomed. Health Inform. 2020, 24, 2860–2869. [Google Scholar] [CrossRef]
  21. Shi, W.; Kuang, S.; Cao, S.; Hu, B.; Xie, S.; Chen, S.; Chen, Y.; Gao, D.; Chen, S.; Zhu, Y.; et al. Deep learning assisted differentiation of hepatocellular carcinoma from focal liver lesions: Choice of four-phase and three-phase CT imaging protocol. Abdom. Radiol. 2020, 45, 2688–2697. [Google Scholar] [CrossRef]
  22. Zhen, S.H.; Cheng, M.; Tao, Y.B.; Wang, Y.F.; Juengpanich, S.; Jiang, Z.Y.; Jiang, Y.K.; Yan, Y.Y.; Lu, W.; Lue, J.M.; et al. Deep Learning for Accurate Diagnosis of Liver Tumor Based on Magnetic Resonance Imaging and Clinical Data. Front. Oncol. 2020, 10, 680. [Google Scholar] [CrossRef]
  23. Ensembled Liver Cancer Detection and Classification Using CT Images—Abhay Krishan, Deepti Mittal. 2021. Available online: https://journals.sagepub.com/doi/abs/10.1177/0954411920971888?journalCode=pihb (accessed on 20 July 2022).
  24. Brehar, R.; Mitrea, D.A.; Vancea, F.; Marita, T.; Nedevschi, S.; Lupsor-Platon, M.; Rotaru, M.; Badea, R.I. Comparison of Deep-Learning and Conventional Machine-Learning Methods for the Automatic Recognition of the Hepatocellular Carcinoma Areas from Ultrasound Images. Sensors 2020, 20, 3085. [Google Scholar] [CrossRef]
  25. Hamm, C.A.; Wang, C.J.; Savic, L.J.; Ferrante, M.; Schobert, I.; Schlachter, T.; Lin, M.; Duncan, J.S.; Weinreb, J.C.; Chapiro, J.; et al. Deep learning for liver tumor diagnosis part I: Development of a convolutional neural network classifier for multi-phasic MRI. Eur. Radiol. 2019, 29, 3338–3347. [Google Scholar] [CrossRef] [PubMed]
  26. Das, A.; Das, P.; Panda, S.S.; Sabut, S. Detection of Liver Cancer Using Modified Fuzzy Clustering and Decision Tree Classifier in CT Images. Pattern Recognit. Image Anal. 2019, 29, 201–211. [Google Scholar] [CrossRef]
  27. Trivizakis, E.; Manikis, G.C.; Nikiforaki, K.; Drevelegas, K.; Constantinides, M.; Drevelegas, A.; Constantinides, M.; Drevelegas, A.; Marias, K. Extending 2-D Convolutional Neural Networks to 3-D for Advancing Deep Learning Cancer Classification with Application to MRI Liver Tumor Differentiation. IEEE J. Biomed. Health Inform. 2019, 23, 923–930. [Google Scholar] [CrossRef] [PubMed]
  28. Kutlu, H.; Avcı, E. A Novel Method for Classifying Liver and Brain Tumors Using Convolutional Neural Networks, Discrete Wavelet Transform and Long Short-Term Memory Networks. Sensors 2019, 19, 1992. [Google Scholar] [CrossRef] [Green Version]
  29. Nayak, A.; Baidya Kayal, E.; Arya, M.; Culli, J.; Krishan, S.; Agarwal, S.; Mehndiratta, A. Computer-aided diagnosis of cirrhosis and hepatocellular carcinoma using multi-phase abdomen CT. Int. J. Comput. Assist. Radiol. Surg. 2019, 14, 1341–1352. [Google Scholar] [CrossRef]
  30. Schmauch, B.; Herent, P.; Jehanno, P.; Dehaene, O.; Saillard, C.; Aubé, C.; Luciani, A.; Lassau, N.; Jégou, S. Diagnosis of focal liver lesions from ultrasound using deep learning. Diagn. Interv. Imaging 2019, 100, 227–233. [Google Scholar] [CrossRef]
  31. Automatic Classification of Focal Liver Lesions Based on MRI and Risk Factors. Available online: https://pubmed.ncbi.nlm.nih.gov/31095624/ (accessed on 20 July 2022).
  32. Das, A.; Acharya, U.R.; Panda, S.S.; Sabut, S. Deep learning based liver cancer detection using watershed transform and Gaussian mixture model techniques. Cogn. Syst. Res. 2019, 54, 165–175. [Google Scholar] [CrossRef]
  33. Mokrane, F.Z.; Lu, L.; Vavasseur, A.; Otal, P.; Peron, J.M.; Luk, L.; Yang, H.; Ammari, S.; Saenger, Y.; Rousseau, H. Radiomics machine-learning signature for diagnosis of hepatocellular carcinoma in cirrhotic patients with indeterminate liver nodules. Eur. Radiol. 2020, 30, 558–570. [Google Scholar] [CrossRef]
  34. Balagourouchetty, L.; Pragatheeswaran, J.K.; Pottakkat, B.; Ramkumar, G. GoogLeNet-Based Ensemble FCNet Classifier for Focal Liver Lesion Diagnosis. IEEE J. Biomed. Health Inform. 2020, 24, 1686–1694. [Google Scholar] [CrossRef]
  35. Acharya, U.R.; Koh, J.E.W.; Hagiwara, Y.; Tan, J.H.; Gertych, A.; Vijayananthan, A.; Yaakup, N.A.; Abdullah, B.J.J.; Bin Mohd Fabell, M.K.; Yeong, C.H. Automated diagnosis of focal liver lesions using bidirectional empirical mode decomposition features. Comput. Biol. Med. 2018, 94, 11–18. [Google Scholar] [CrossRef]
  36. Ta, C.N.; Kono, Y.; Eghtedari, M.; Oh, Y.T.; Robbin, M.L.; Barr, R.G.; Kummel, A.C.; Mattrey, R.F. Focal Liver Lesions: Computer-aided Diagnosis by Using Contrast-enhanced US Cine Recordings. Radiology 2018, 286, 1062–1071. [Google Scholar] [CrossRef] [PubMed]
  37. Bharti, P.; Mittal, D.; Ananthasivan, R. Preliminary Study of Chronic Liver Classification on Ultrasound Images Using an Ensemble Model. Ultrason. Imaging 2018, 40, 357–379. [Google Scholar] [CrossRef] [PubMed]
  38. Deep Learning with Convolutional Neural Network for Differentiation of Liver Masses at Dynamic Contrast-Enhanced CT: A Preliminary Study. Available online: https://pubmed.ncbi.nlm.nih.gov/29059036/ (accessed on 20 July 2022).
  39. Hassan, T.M.; Elmogy, M.; Sallam, E.S. Diagnosis of Focal Liver Diseases Based on Deep Learning Technique for Ultrasound Images. Arab. J. Sci. Eng. 2017, 42, 3127–3140. [Google Scholar] [CrossRef]
  40. Guo, L.H.; Wang, D.; Qian, Y.Y.; Zheng, X.; Zhao, C.K.; Li, X.L.; Bo, X.W.; Yue, W.W.; Zhang, Q.; Shi, J.; et al. A two-stage multi-view learning framework based computer-aided diagnosis of liver tumors with contrast enhanced ultrasound images. Clin. Hemorheol. Microcirc. 2018, 69, 343–354. [Google Scholar] [CrossRef]
  41. Kondo, S.; Takagi, K.; Nishida, M.; Iwai, T.; Kudo, Y.; Ogawa, K.; Kamiyama, T.; Shibuya, H.; Kahata, K.; Shimizu, C. Computer-Aided Diagnosis of Focal Liver Lesions Using Contrast-Enhanced Ultrasonography with Perflubutane Microbubbles. IEEE Trans. Med. Imaging 2017, 36, 1427–1437. [Google Scholar] [CrossRef]
  42. Gatos, I.; Tsantis, S.; Spiliopoulos, S.; Skouroliakou, A.; Theotokas, I.; Zoumpoulis, P.; Hazle, J.D.; Kagadis, G.C. A new automated quantification algorithm for the detection and evaluation of focal liver lesions with contrast-enhanced ultrasound. Med. Phys. 2015, 42, 3948–3959. [Google Scholar] [CrossRef]
  43. Virmani, J.; Kumar, V.; Kalra, N.; Khandelwal, N. Neural network ensemble based CAD system for focal liver lesions from B-mode ultrasound. J. Digit. Imaging 2014, 27, 520–537. [Google Scholar] [CrossRef] [Green Version]
  44. Wu, K.; Chen, X.; Ding, M. Deep learning based classification of focal liver lesions with contrast-enhanced ultrasound. Optik 2014, 125, 4057–4063. [Google Scholar] [CrossRef]
  45. Virmani, J.; Kumar, V.; Kalra, N.; Khandelwal, N. Characterization of Primary and Secondary Malignant Liver Lesions from B-Mode Ultrasound. J. Digit. Imaging 2013, 26, 1058–1070. [Google Scholar] [CrossRef]
  46. Streba, C.T.; Ionescu, M.; Gheonea, D.I.; Sandulescu, L.; Ciurea, T.; Saftoiu, A.; Vere, C.C.; Rogoveanu, I. Contrast-enhanced ultrasonography parameters in neural network diagnosis of liver tumors. World J. Gastroenterol. 2012, 18, 4427–4434. [Google Scholar] [CrossRef]
  47. Mittal, D.; Kumar, V.; Saxena, S.C.; Khandelwal, N.; Kalra, N. Neural network based focal liver lesion diagnosis using ultrasound images. Comput. Med. Imaging Graph. 2011, 35, 315–323. [Google Scholar] [CrossRef] [PubMed]
  48. Sugimoto, K.; Shiraishi, J.; Moriyasu, F.; Doi, K. Computer-aided diagnosis for contrast-enhanced ultrasound in the liver. World J. Radiol. 2010, 2, 215–223. [Google Scholar] [CrossRef]
  49. Shiraishi, J.; Sugimoto, K.; Moriyasu, F.; Kamiyama, N.; Doi, K. Computer-aided diagnosis for the classification of focal liver lesions by use of contrast-enhanced ultrasonography. Med. Phys. 2008, 35, 1734–1746. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  50. Stoitsis, J.; Valavanis, I.; Mougiakakou, S.G.; Golemati, S.; Nikita, A.; Nikita, K.S. Computer aided diagnosis based on medical image processing and artificial intelligence methods. Nucl. Instrum. Methods Phys. Res. Sect. A Accel. Spectrometers Detect. Assoc. Equip. 2006, 569, 591–595. [Google Scholar] [CrossRef]
  51. Usefulness of Artificial Neural Network for Differential Diagnosis of Hepatic Masses on CT Images—ScienceDirect. Available online: https://www.sciencedirect.com/science/article/abs/pii/S1076633206002613 (accessed on 20 July 2022).
  52. Gletsos, M.; Mougiakakou, S.G.; Matsopoulos, G.K.; Nikita, K.S.; Nikita, A.S.; Kelekis, D. A computer-aided diagnostic system to characterize CT focal liver lesions: Design and optimization of a neural network classifier. IEEE Trans. Inf. Technol. Biomed. 2003, 7, 153–162. [Google Scholar] [CrossRef]
  53. Chen, E.L.; Chung, P.C.; Chen, C.L.; Tsai, H.M.; Chang, C.I. An automatic diagnostic system for CT liver image classification. IEEE Trans. Biomed. Eng. 1998, 45, 783–794. [Google Scholar] [CrossRef] [Green Version]
  54. Short Overview of the Human Brain Project. Available online: https://www.humanbrainproject.eu/en/about/overview/ (accessed on 2 August 2022).
  55. Data-Efficient and Weakly Supervised Computational Pathology on Whole-Slide Images. Available online: https://pubmed.ncbi.nlm.nih.gov/33649564/ (accessed on 5 July 2022).
  56. Wu, B.; Khong, P.L.; Chan, T. Automatic detection and classification of nasopharyngeal carcinoma on PET/CT with support vector machine. Int. J. Comput. Assist. Radiol. Surg. 2012, 7, 635–646. [Google Scholar] [CrossRef] [Green Version]
  57. Kirienko, M.; Cozzi, L.; Rossi, A.; Voulaz, E.; Antunovic, L.; Fogliata, A.; Chiti, A.; Sollini, M. Ability of FDG PET and CT radiomics features to differentiate between primary and metastatic lung lesions. Eur. J. Nucl. Med. Mol. Imaging 2018, 45, 1649–1660. [Google Scholar] [CrossRef]
  58. Gao, X.; Chu, C.; Li, Y.; Lu, P.; Wang, W.; Liu, W.; Yu, L. The method and efficacy of support vector machine classifiers based on texture features and multi-resolution histogram from (18)F-FDG PET-CT images for the evaluation of mediastinal lymph nodes in patients with lung cancer. Eur. J. Radiol. 2015, 84, 312–317. [Google Scholar] [CrossRef]
  59. Wu, J.; Aguilera, T.; Shultz, D.; Gudur, M.; Rubin, D.L.; Loo, B.W.; Diehn, M.; Li, R. Early-Stage Non-Small Cell Lung Cancer: Quantitative Imaging Characteristics of (18)F Fluorodeoxyglucose PET/CT Allow Prediction of Distant Metastasis. Radiology 2016, 281, 270–278. [Google Scholar] [CrossRef] [Green Version]
  60. Sacks, A.; Peller, P.J.; Surasi, D.S.; Chatburn, L.; Mercier, G.; Subramaniam, R.M. Value of PET/CT in the management of liver metastases, part 1. AJR Am. J. Roentgenol. 2011, 197, W256–W259. [Google Scholar] [CrossRef] [PubMed]
  61. Khan, M.A.; Combs, C.S.; Brunt, E.M.; Lowe, V.J.; Wolverson, M.K.; Solomon, H. Positron emission tomography scanning in the evaluation of hepatocellular carcinoma. J. Hepatol. 2000, 32, 792–797. [Google Scholar] [CrossRef]
  62. Delbeke, D.; Martin, W.H.; Sandler, M.P.; Chapman, W.C.; Wright, J.K.; Pinson, C.W. Evaluation of benign vs malignant hepatic lesions with positron emission tomography. Arch. Surg. 1998, 133, 510–515; discussion 515–516. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  63. Ho, C.; Chen, S.; Yeung, D.W.C.; Cheng, T.K.C. Dual-tracer PET/CT imaging in evaluation of metastatic hepatocellular carcinoma. J. Nucl. Med. 2007, 48, 902–909. [Google Scholar] [CrossRef] [Green Version]
  64. Park, J.W.; Kim, J.H.; Kim, S.K.; Kang, K.W.; Park, K.W.; Choi, J.I.; Lee, W.J.; Kim, C.M.; Nam, B.H. A prospective evaluation of 18F-FDG and 11C-acetate PET/CT for detection of primary and metastatic hepatocellular carcinoma. J. Nucl. Med. 2008, 49, 1912–1921. [Google Scholar] [CrossRef]
Figure 1. PRISMA Flow Diagram.
Figure 1. PRISMA Flow Diagram.
Jcm 11 06368 g001
Table 1. Breakdown of Articles by Year of Publication.
Table 1. Breakdown of Articles by Year of Publication.
LevelsCounts% of TotalCumulative %
199812.2%2.2%
200312.2%4.3%
200624.3%8.7%
200812.2%10.9%
201012.2%13.0%
201112.2%15.2%
201212.2%17.4%
201312.2%19.6%
201424.3%23.9%
201512.2%26.1%
201736.5%32.6%
201848.7%41.3%
20191021.7%63.0%
2020715.2%78.3%
2021510.9%89.1%
2022510.9%100.0%
Median2019
Table 2. “Number of Images” Score.
Table 2. “Number of Images” Score.
LevelsCounts% of TotalCumulative %
048.7%8.7%
1613.0%21.7%
23678.3%100.0%
Mean: 1.70Median: 2.00SD: 0.628
Table 3. “Cohort for Validation” Score.
Table 3. “Cohort for Validation” Score.
LevelsCounts% of TotalCumulative %
02043.5%43.5%
12452.2%95.7%
224.3%100.0%
Mean: 0.609Median: 1.00SD: 0.577
Table 4. “Year of Publication” Score.
Table 4. “Year of Publication” Score.
LevelsCounts% of TotalCumulative %
1613.0%13.0%
24087.0%100.0%
Mean: 1.87Median: 2.00SD: 0.341
Table 5. Total Score.
Table 5. Total Score.
LevelsCounts% of TotalCumulative %
112.2%2.2%
224.3%6.5%
3715.2%21.7%
41634.8%56.5%
51839.1%95.7%
624.3%100.0%
Mean: 4.17Median: 4.00SD: 1.04
Table 6. Contingency Tables; Total Score and Year.
Table 6. Contingency Tables; Total Score and Year.
Year χ2 Tests Year χ2 Tests Year χ2 Tests Year χ2 Tests Year χ2 Tests
Total Score1998200320062008201020112012201320142015201720182019202020212022TotalValuedfp
110000000000000001111750.004
200000000100010002
300111000012010007
4011000100013321316
5000001011001534218
600000000000002002
Total1121111121341075546
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Martinino, A.; Aloulou, M.; Chatterjee, S.; Scarano Pereira, J.P.; Singhal, S.; Patel, T.; Kirchgesner, T.P.-E.; Agnes, S.; Annunziata, S.; Treglia, G.; et al. Artificial Intelligence in the Diagnosis of Hepatocellular Carcinoma: A Systematic Review. J. Clin. Med. 2022, 11, 6368. https://doi.org/10.3390/jcm11216368

AMA Style

Martinino A, Aloulou M, Chatterjee S, Scarano Pereira JP, Singhal S, Patel T, Kirchgesner TP-E, Agnes S, Annunziata S, Treglia G, et al. Artificial Intelligence in the Diagnosis of Hepatocellular Carcinoma: A Systematic Review. Journal of Clinical Medicine. 2022; 11(21):6368. https://doi.org/10.3390/jcm11216368

Chicago/Turabian Style

Martinino, Alessandro, Mohammad Aloulou, Surobhi Chatterjee, Juan Pablo Scarano Pereira, Saurabh Singhal, Tapan Patel, Thomas Paul-Emile Kirchgesner, Salvatore Agnes, Salvatore Annunziata, Giorgio Treglia, and et al. 2022. "Artificial Intelligence in the Diagnosis of Hepatocellular Carcinoma: A Systematic Review" Journal of Clinical Medicine 11, no. 21: 6368. https://doi.org/10.3390/jcm11216368

APA Style

Martinino, A., Aloulou, M., Chatterjee, S., Scarano Pereira, J. P., Singhal, S., Patel, T., Kirchgesner, T. P. -E., Agnes, S., Annunziata, S., Treglia, G., & Giovinazzo, F. (2022). Artificial Intelligence in the Diagnosis of Hepatocellular Carcinoma: A Systematic Review. Journal of Clinical Medicine, 11(21), 6368. https://doi.org/10.3390/jcm11216368

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop