Journal List > Korean J Radiol > v.21(4) > 1144158

Park, Park, and Lee: Radiomics and Deep Learning: Hepatic Applications

Abstract

Radiomics and deep learning have recently gained attention in the imaging assessment of various liver diseases. Recent research has demonstrated the potential utility of radiomics and deep learning in staging liver fibroses, detecting portal hypertension, characterizing focal hepatic lesions, prognosticating malignant hepatic tumors, and segmenting the liver and liver tumors. In this review, we outline the basic technical aspects of radiomics and deep learning and summarize recent investigations of the application of these techniques in liver disease.

INTRODUCTION

Imaging plays a pivotal role in the evaluation of various liver diseases, including screening, surveillance, diagnosis, and prognostication of diffuse liver disorders and hepatic neoplasms. Recent advances in computer science have enabled the clinical application of computer-assisted analysis in imaging examinations; of which radiomics and deep learning are currently the most actively investigated techniques. Although they involve completely different technical processes, both radiomics and deep learning utilize high-dimensional features extracted from images for diagnostic and predictive tasks. Radiomics and deep learning may also expand the role of imaging in the assessment of various liver diseases beyond the domain of traditional visual image analysis, by obtaining additional diagnostic information from images, assessing image features in a comprehensive and objective manner, and facilitating labor-intensive tasks such as liver segmentation. Hence, the goal of our article is to review the basic technical aspects of radiomics and deep learning and to summarize recent investigations on the application of these techniques in assessing liver disorders.

Radiomics

Radiomics refers to a set of techniques for extracting a large number of quantitative features from medical images (1) and subsequently mining these features to retrieve clinically useful diagnostic and prognostic information. Radiomics has gained considerable attention in the field of oncology as a method for supporting clinical decision-making and precision medicine. This methodology is based on the hypothesis that a radiologic phenotype may reflect genetic alterations in carcinogenesis and tumor biology and may thus be predictive of the biologic behavior of the tumor (12). Radiomics is also an effective method for assessing the morphologic and textural changes of the liver that are associated with various disease processes. Unlike visual assessments of clinical images, it may allow for objective and comprehensive assessments of these changes based on quantitative indices.

Radiomics Features

A number of radiomics features can be extracted from a given volume of interest (VOI) drawn on two-dimensional (2D) images or three-dimensional (3D) volume data. Radiomics features can be divided into morphologic features, histogram features, textural features, and higher-order features.

Morphologic Features

Morphologic features describe the size, volume, and shape of the VOI, usually for tumors. Unlike a visual assessment of tumor morphology by radiologists, morphologic features are expressed as statistical values in radiomics (Fig. 1). For example, the circularity on a 2D image describes the ratio of the area to the perimeter of a given VOI, reflecting how close the VOI is to a complete circle (34).

Histogram Features

A histogram is a plot displaying the pixel frequency in accordance with pixel values. Multiple features can be calculated from a histogram, which describe the magnitude (mean), dispersion (standard deviation), asymmetry (skewness), peakedness or flatness (kurtosis), randomness (entropy), uniformity (energy and uniformity), and dispersion relative to the magnitude (coefficient of variation) of gray-level pixel values. These histogram features describe the distribution pattern of gray-level pixel values within a VOI as a whole, but cannot address the spatial relationship among pixels or the textural pattern (456) (Fig. 2).

Textural Features

Textural features are a key component of radiomics features and describe the spatial relationship between each individual pixel and its neighboring pixels. Two commonly used matrices for textural analysis are the gray-level co-occurrence matrix (GLCM) and the gray-level run-length matrix (GLRLM). The GLCM is a matrix describing the frequency of two neighboring pixels with certain gray-level pixel values, while the GLRLM describes the length of a continuous pixel with a certain gray-level pixel value. Both the GLCM and GLRLM are dependent on direction. To improve directional invariance, textural features are calculated by aggregating information from different directional matrices using several 2D or 3D-based methods (457) (Fig. 3).

Higher-Order Features

Higher-order features refer to textural features extracted from filtered images. Various filters have been used to emphasize the characteristics of images. A Gaussian filter is a smoothing filter that reduces the sensitivity to image noise. A Laplacian filter is an edge-enhancing filter. Since the Laplacian filter enhances any rapid intensity changes on an image, it may amplify image noise as well as edges. A Laplacian of Gaussian filter is a combination of both filters (1) and, thus, is frequently used to enhance edges while preventing amplification of image noise. Wavelets transform images using a matrix of complex linear or radial waves, allowing for the separation and emphasis of a high-frequency component (i.e., edge part) or low-frequency component (i.e., smooth part) of the images (578) (Fig. 4).

Process of Radiomics Analysis

The radiomics analysis of medical images involves multiple processes, including image preprocessing, segmentation, feature extraction, feature selection, and classification.
Image preprocessing is an important step for achieving valid and reproducible radiomics features. Image normalization may be required to standardize the gray-scale pixel value, and it can be performed based on the histogram distribution of pixel values or internal reference values (i.e., spleen signal value). Since textural and higher-order features are dependent on pixel dimensions, images with non-isometric pixels or variable resolutions may lead to invalid results. In these cases, the image resolution should be standardized by the resampling the images at a fixed isometric resolution. After image preprocessing, the segmentation of VOIs is performed manually or by using an automatic segmentation algorithm to select the volume or area for which the radiomics features are extracted.
Radiomics features can be extracted in 2D or 3D using in-house software (6910111213) or commercial software (1415). The number of extracted features can be variable, largely depending on the number of textural features and the number of filters used for extracting higher-order features.
A larger number of extracted features do not necessarily indicate better-quality measurements. Since radiomics features are highly correlated with each other (16), the analysis of high-dimensional features may lead to problems of multicollinearity and overfitting. A recent phantom study revealed that the information provided by multiple radiomics features could be summarized using only 10 features because of redundancy (16).
Feature selection is a process performed to reduce the dimensionality of features by selecting informative and reliable features and excluding redundant features among the extracted features. Classification is a process used to build a classifier or prediction model using the selected feature to perform a given classification or prediction task. Feature selection and classification can be performed together as a single process or separately using different algorithms. Unreliable features may be excluded prior to feature selection and classification, based on the results of inter- or intra-observer agreement or test-retest repeatability analyses (6111718192021). To reduce redundancy in the features, informative features showing a high dynamic range may be selected among the correlated features in hierarchical feature clustering (1822). Traditional statistical methods may not be successful in dealing with high-dimensional radiomics features (i.e., too many variables relative to the number of observations). A number of machine learning methods have therefore been used for feature selection and/or classification (10212324). Among the methods for feature selection and classification, regression with Ridge, least absolute shrinkage and selection operator (LASSO), and elastic net regularization have been commonly used (61011121317192526), likely because these algorithms allow for the development of a regression model that is more familiar to radiologists than other machine learning classifiers. These regression analyses incorporate regularization and penalization algorithms for correlated variables; the LASSO regression method is robust for feature selection, but the Ridge regression is more effective in dealing with multicollinearity. Elastic net regression takes advantage of both methods (627). Other commonly used classification methods include regression, support vector machine (SVM), decision tree, and random forest. In machine learning, hyperparameters to control the machine learning process need to be optimized to different data patterns. Following hyperparameter optimization, the machine learning algorithm is trained through the learning process using given training data (28). Figure 5 schematically depicts the development process of a radiomics classification model.

Clinical Application of Radiomics in Liver Disease

Radiomics has been used to evaluate the severity of chronic liver disease and assess the prognosis of malignant liver tumors. The study methodology and the results of some representative reports are presented in Table 1.

Chronic Liver Disease

Chronic liver disease is accompanied by changes in liver volume, morphology, and texture. Several recent studies have shown the potential value of radiomics as a method for comprehensive and objective analysis of such changes in the liver using imaging examinations. Park et al. (6) developed a radiomics fibrosis index based on radiomics features extracted from gadoxetic acid-enhanced hepatobiliary phase magnetic resonance (MR) images. They demonstrated that the radiomics fibrosis index had a high diagnostic performance in staging liver fibrosis (area under the receiver-operating curve [AUROC], 0.89–0.91) and significantly outperformed the normalized liver enhancement and serum fibrosis indices. Liu et al. (12) reported the feasibility of CT-based radiomics analysis for the diagnosis of clinically significant portal hypertension. These authors devised a model based on texture features, morphologic features, and the liver and spleen volumes with the hepatic venous pressure gradient as the reference standard. The performance of this model was significantly better than were those of models using liver stiffness measurements as well as other radiologic and clinical indices (12). Several exploratory studies have indicated the potential of radiomics of multiparametric ultrasound (21) and histogram features of CT images (2930) in staging liver fibrosis and in diagnosing nonalcoholic steatohepatitis. However, the results of these studies were not conclusive because of the small study populations and the lack of proper validation (212930).

Prognostication of Malignant Liver Tumors

Radiomics has been applied to determine the prognosis of hepatocellular carcinoma (HCC) after radiofrequency ablation (3132), surgical resection (1322253133), and liver transplantation (20). Zheng et al. (13) developed nomograms incorporating CT-based radiomics and clinical variables to predict recurrence-free and overall survival outcomes after surgical resection of solitary HCC and reported that these nomograms had better prognostic performance than traditional staging. Kim et al. (22) devised radiomics models for predicting the early and late post-surgical recurrence of HCC using gadoxetic acid-enhanced MRI, incorporating variable extents of peritumor border extension. In that study, a radiomics model with 3-mm or 5-mm peritumoral border extension showed a higher prediction performance than the models without a border extension, indicating that the features of a peri-tumor liver parenchyma are important for predicting early or late recurrence in HCC patients. Since microvascular invasion (MVI) is one of the most important prognostic factors for HCC after surgery (343536), several studies have evaluated the potential of using radiomics to predict it (10242637). Xu et al. (24) developed a prediction model combining a CT-based radiomics score, radiologist image analysis, and laboratory findings and demonstrated a high accuracy (AUROC, 0.889) for predicting MVI in a test dataset. In their study, however, subsequent decision curve analysis failed to demonstrate the incremental value of the radiomics score in comparison with conventional visual image analysis. Two prior studies have reported the incremental value of a CT-based radiomics model in predicting lymph node metastases in patients with cholangiocarcinoma, and noted that incorporating a radiomics signature into the CT-reported lymph node status improved the detection of lymph node metastasis (1117). A recent study (9) has also demonstrated the potential role of radiomics features extracted from gadoxetic acid-enhanced hepatobiliary MR images in assisting with precision immunotherapy of HCC. This study showed that a model combining radiomics and clinical variables accurately predicted the immune-score, which is known to be associated with the therapeutic response to an immune checkpoint blockage (9).

Pitfalls of Radiomics

There are some disadvantages to using radiomics approaches. These methods are labor-intensive and time-consuming as they involve segmentation, feature extraction, and machine learning or modeling processes. Hence, a radiomics study will only produce real clinical value if it generates incremental diagnostic information beyond that obtained with classic visual image interpretation. Radiomics features are also highly dependent on the imaging protocol, VOI selection, and feature extraction methods. All of these factors may be sources of variation in terms of extracted radiomics features (163839). Radiomics models or classifiers thus have inherent limitations in terms of generalization. Optimal image preprocessing, including gray-level normalization and resolution standardization, may partly overcome the imaging protocol dependency of radiomics features. Recently, an algorithm has been proposed that reduces the variation in radiomics features according to different CT protocols, and thus facilitates radiomics analysis using multicenter image data (40). Further research is warranted to develop an optimal method of minimizing the variations in radiomics features. Textural features are also dependent on settings for feature extraction, such as bin size (i.e., the size of gray-level discretization). Research papers on radiomics should therefore clearly state the methods used for radiomics feature extraction so that they can be replicated. The lack of a standardized method for radiomics feature extraction has been an important cause of the poor generalizability of radiomics studies. To overcome this problem, the Image Biomarker Standardization Initiative recently published consensus guidelines to standardize the methods for image processing, the nomenclature and definitions of radiomics features, and the reporting methods (4). A recent review article has further suggested some strategies for reproducible and generalizable radiomics analysis (39). These methodological guidelines may be useful for improving the generalizability of radiomic studies.

Deep Learning

Deep learning is a subset of machine learning, which is based on a neural network structure inspired by the human brain (4142). Unlike radiomics and traditional machine learning, which rely on predefined, hand-engineered features, deep learning is based on representation learning in which the algorithm learns the best features to carry out a given task on its own by navigating the provided data.

Convolutional Neural Network

Convolutional neural network (CNN) is the most popular type of deep learning architecture in medical imaging analysis (4142). A CNN consists of an input layer, hidden layers, and the output layer. The hidden layers may include convolution and pooling layers and fully connected layers. Convolution and pooling layers extract high-dimensional manageable features from given images, which is conceptually similar to the feature extraction process used in radiomics analysis. Convolution operations generate feature maps using a group of filters, followed by activation functions typically using a rectified linear unit. Activation functions add nonlinearity to the outputs of convolutions, allowing the selection of features to pass through to the next layer. Pooling operations reduce the resolution of the feature maps to gain computational performance, obtain spatially invariable features, and reduce the chance of overfitting (4143). The fully connected layers integrate and transform all of the features fed from the convolution and pooling layers into a vector form. The output layer then returns a categorical distribution for class probability through a softmax function. The details of deep learning and CNN can be understood further from previous review articles (4142). Figure 6 schematically presents the architecture and training process of a CNN algorithm.

Training of a Deep Learning Algorithm

The training of a deep learning algorithm is usually performed with supervised learning using labeled training data. A deep learning algorithm typically requires large volumes of high-quality ground truth training data, although the amount of required data may vary for different deep learning algorithm tasks: an algorithm for a segmentation task may require a smaller dataset, while a classification task requires a much larger dataset (44). When a training dataset is not sufficiently large, data augmentation may be used to enlarge it artificially, which is performed through random transformation of original images by adding random noise, flipping, or rotation (41). Data augmentation may also be required to overcome the potential problems of data imbalance. If the size of the training data is imbalanced across different classes, a classification algorithm may have poor classification accuracy for the minority classes (45). This may be prevented by data augmentation for those classes. Datasets for the development and validation of a deep learning algorithm typically consist of training, validation, and test datasets. The data available for the development of the algorithm may be divided into training and validation datasets. The validation dataset is used for monitoring the performance of the algorithm during the training process and/or comparing multiple models based on different CNN architectures or hyperparameters. Once the final model is selected and all its parameters are fixed, its performance is evaluated in the test dataset. The test dataset is used only at the final step of the study to report the final model performance (41). A deep learning algorithm is trained by adjusting network weights. Starting from a random initial configuration, parameters are adjusted to find a set of parameters that perform best on the training dataset. During the training phase, the output of the algorithm is compared with the ground truth by using a loss function that quantitatively measures the error in the prediction in comparison with the ground truth. The error is then back propagated to optimize network weights (Fig. 6). The training phase continues until the loss function reaches a minimum.

Application of Deep Learning in Liver Disease

Deep learning has been widely applied to liver imaging for various tasks, including organ segmentation, staging liver fibrosis, tumor detection, or classification, and improving image quality. The study methodology and the results of some representative studies are summarized in Table 2.

Liver Segmentation

Liver segmentation has direct clinical applications, including liver volume measurement, which is important in pre-operative planning for liver resection (4647), determination of the radiation dose in liver tumor radioembolization, and measurement of quantitative indices such as the proton density fat fraction (PDFF) from the whole liver (48). Notably, however, liver segmentation is labor-intensive and time-consuming, which limits its usage in clinical practice. Thus, deep learning has been applied for automated segmentation of the liver. The U-net architecture is most commonly used for segmentation tasks (49) and consists of a series of contracting and expanding layers that extract and process features from input images and return a pixel-wise probability map. The segmentation performance is typically evaluated using the Dice similarity score (DSS), defined as 2 × true positive pixels / [2 × true positive pixels + false negative pixels + false positive pixels]. Some prior studies have reported the use of deep learning algorithms for automated liver segmentation on CT or MRI (5051525354), and some have utilized a deep learning algorithm combined with image processing methods (5052). All of these studies reported high performance values in liver segmentation, with the reported DSS values ranging from 0.92 to 0.95 (5051525354). Recently, Wang et al. (48) demonstrated the feasibility of generalized CNN, which can be used for liver segmentation on CT scans and various MRI sequences using the transfer learning technique. They reported DSS values ranging from 0.92 to 0.95 for liver segmentation on CT and MR images. Furthermore, these authors demonstrated a close agreement between the PDFF values measured using deep learning-based automatic liver segmentation and those measured by manual liver segmentation, indicating the potential role of deep learning-based liver segmentation for automatic measurement of quantitative indices from the whole liver. Despite these promising results, however, further clinical validation may be required for the actual clinical application of deep learning algorithms for automated liver segmentation. For example, algorithm performance should be evaluated in a healthy liver, fatty liver, and in chronic liver disease and liver cirrhosis. With continued improvements in deep learning-based organ segmentation methods, it is expected that fully automated liver segmentation would become clinically available in the near future.

Liver Fibrosis Staging

A few deep learning algorithms for liver fibrosis staging have been reported, to date. Liu et al. (55) proposed sequential algorithms to diagnose cirrhosis using ultrasound images, which first detect liver capsules on the images by using a sliding window detector, extract features from image patches by using a CNN algorithm, and finally classify an image as indicative of cirrhosis or not by using an SVM. In that report, CNN was used only for feature extraction whereas classification was performed with the SVM because of the small amount of training data. Yasaka et al. (56) developed CNN algorithms for liver fibrosis staging using cropped CT images and cropped gadoxetic acid-enhanced hepatobiliary phase MR images (57). They reported area under the curves (AUCs) of 0.73–0.76 for the CT-based algorithm and 0.84–0.85 for the MRI-based algorithm in staging liver fibrosis. However, the use of a small test dataset (100 patients) and lack of any external validation limited the generalizability of their study results. Choi et al. (45) reported the use of a deep learning algorithm for fully automated liver fibrosis staging using portal venous phase CT images. Using a large training dataset (7491 patients) and internal and external test data (891 patients), these authors reported a high accuracy (AUCs, 0.95–0.97) of the deep learning algorithm in liver fibrosis staging, surpassing that of the serum fibrosis indices and visual image analyses by radiologists. A recent multicenter prospective study reported a higher accuracy (AUCs, 0.97–0.98) with a deep learning algorithm using cropped 2D shear wave elastographic images in staging liver fibrosis in comparison with liver stiffness measurement results (8).

Diagnosis of Fatty Liver Disease

The feasibility of using deep learning for the diagnosis and grading of fatty liver disease using ultrasound images has been evaluated in several previous reports (585960). Although these prior studies demonstrated the technical feasibility of deep learning, its clinical applicability has not been well proven because of the small size of the test data, lack of external validation, and the use of a less reliable reference standard (i.e., ultrasound determined fatty liver grade).

Detection, Segmentation, and Classification of Liver Tumors

Vorontsov et al. (61) have reported the use of a deep learning algorithm for the automatic detection and segmentation of malignant liver tumors on CT images. In a small test dataset (26 CT examinations) in that study, the algorithm showed high accuracy in detecting liver lesions larger than 2 cm with a sensitivity of 85% and positive predictive value of 94%, whereas it was not accurate in the detection of small lesions (sensitivity, 10% for lesions < 1 cm) or in automatic tumor segmentation (DSS of 0.14–0.68). Schmauch et al. (62) also described the technical feasibility of applying deep learning to the detection of focal liver lesions using ultrasound images. The potential utility of deep learning for the classification of focal hepatic lesions has now been evaluated in several studies, all of which devised deep learning algorithms to classify liver lesions into five to six predefined categories based on manually cropped CT or MR images containing these lesions (6364). Yasaka et al. (63) developed an algorithm for classifying liver masses using multi-phasic CT images and reported an accuracy of 84% in the test dataset. Hamm et al. (64) reported the results of algorithms based on multiphasic MRI, describing an accuracy of 90% for lesion diagnosis and 92% for lesion categorization based on the liver imaging reporting and data system. The same researchers (54) also demonstrated the feasibility of deep learning in identifying individual radiologic features of focal hepatic lesions on MR images, reporting a sensitivity of 82.9% and positive predictive value of 76.5% for the algorithm. Despite these promising results, however, all prior studies on the application of deep learning to liver lesion detection and characterization are considered preliminary. These earlier reports focused mainly on the technical feasibility of deep learning, since the algorithms used involved data processes not suitable for a real clinical workflow (e.g., image cropping by radiologists) and were not fully validated using a large-scale external dataset.

Image Quality Improvement and Image Synthesis

Deep learning has now been used for automatic evaluation of image quality (6566). Ma et al. (65) reported a deep learning algorithm to identify technically optimal portal venous phase CT images. Esses et al. (66) described an algorithm to discriminate diagnostic and nondiagnostic T2-weighted MR images. With further improvements, these techniques may be clinically usable for real-time scanning optimization through automatic image quality monitoring. Recent research findings have further suggested the potential utility of deep learning as a method to improve MR image quality (6768). Tamada et al. (68) presented a method to reduce respiratory motion artifacts in gadoxetic acid-enhanced arterial phase MR images using a CNN algorithm. Liu et al. (67) developed a deep-learning-based MR image reconstruction algorithm by adopting generative adversarial networks (GANs). These authors demonstrated that their GAN-based reconstruction algorithm produced superior image quality in comparison with a reconstruction algorithm based on compressed sensing and parallel imaging. This suggested the potential of deep learning-based image reconstruction combined with data under-sampling for fast MRI.

Clinical Validation of Radiomics Models and Deep Learning Algorithms

Radiomics models and deep learning algorithms are subject to the overfitting problem since they are based on numerous image-derived parameters. Overfitting refers to a condition whereby a model customizes itself too much to the training data, to the extent that it explains not only generalizable patterns but also noise and idiosyncratic statistical variations of the training data (6970). An overfitted model performs well on the training data but poorly on other data, reducing the generalizability of the model. Rigorous clinical validation is therefore required for all radiomics and deep learning algorithms. Internal validation methods such as cross-validation, bootstrapping, and split-sample validation (i.e., splitting the entire dataset randomly into training and the validation data) may not sufficiently guarantee the generalizability of radiomics models or deep learning algorithms (7071). External validation using a separate dataset is preferred, which may be conducted using data collected from a different site (i.e., geographic validation) or during a different period from the training data (i.e., temporal validation). In addition, clinical validation needs to be performed in a relevant clinical setting where the radiomics models or deep learning algorithms are actually applied. Further details regarding the clinical validation of artificial intelligence models can be found in previous reviews (7072). Guidelines for transparent reporting of a multivariable prediction model (717374) can also be used as references for choosing proper methods for model development and validation.

CONCLUSION

Radiomics and deep learning are promising techniques for imaging assessment of liver diseases. Recent research findings have demonstrated the potential utility of radiomics and deep learning in staging liver fibrosis, detecting portal hypertension, characterizing focal hepatic lesions, prognosticating malignant hepatic tumors, and segmenting liver and liver tumors. However, as reported in a recent study (75), most previous investigations have focused mainly on the technical feasibility of using radiomics or deep learning algorithms, whereas their applicability and generalizability to actual clinical practice has not been fully evaluated. For radiomics or deep learning algorithms to become a valid clinical tool, their performance should be validated through properly conducted clinical tests. In addition, future research endeavors need to address the clinical impact of radiomics and deep learning and determine how these techniques can be incorporated into real-world clinical practice.

Notes

This research was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF), which is funded by the Ministry of Science, ICT and Future Planning (NRF-2017R1A2B4003114).

Conflicts of Interest: The authors have no potential conflicts of interest to disclose.

References

1. Gillies RJ, Kinahan PE, Hricak H. Radiomics: images are more than pictures, they are data. Radiology. 2016; 278:563–577. PMID: 26579733.
crossref
2. Lee G, Lee HY, Park H, Schiebler ML, van Beek EJR, Ohno Y, et al. Radiomics and its emerging role in lung cancer research, imaging biomarkers and clinical management: state of the art. Eur J Radiol. 2017; 86:297–307. PMID: 27638103.
crossref
3. Legland D, Kiêu K, Devaux MF. Computation of Minkowski measures on 2D and 3D binary images. Image Anal Stereol. 2007; 26:83–92.
crossref
4. Zwanenburg A, Leger S, Vallières M, Löck S. Image biomarker standardisation initiative [updated May 2019]. arXiv:1612.07003 [cs.CV]. 2016. Accessed August 31, 2019. Available at: https://arxiv.org/abs/1612.07003v9.
5. Parekh V, Jacobs MA. Radiomics: a new application from established techniques. Expert Rev Precis Med Drug Dev. 2016; 1:207–226. PMID: 28042608.
crossref
6. Park HJ, Lee SS, Park B, Yun J, Sung YS, Shim WH, et al. Radiomics analysis of gadoxetic acid-enhanced MRI for staging liver fibrosis. Radiology. 2019; 290:380–387. PMID: 30615554.
crossref
7. Scalco E, Rizzo G. Texture analysis of medical images for radiotherapy applications. Br J Radiol. 2017; 90:20160642. PMID: 27885836.
crossref
8. Wang K, Lu X, Zhou H, Gao Y, Zheng J, Tong M, et al. Deep learning radiomics of shear wave elastography significantly improved diagnostic performance for assessing liver fibrosis in chronic hepatitis B: a prospective multicentre study. Gut. 2019; 68:729–741. PMID: 29730602.
crossref
9. Chen S, Feng S, Wei J, Liu F, Li B, Li X, et al. Pretreatment prediction of immunoscore in hepatocellular cancer: a radiomics-based clinical model based on Gd-EOB-DTPA-enhanced MRI imaging. Eur Radiol. 2019; 29:4177–4187. PMID: 30666445.
crossref
10. Feng ST, Jia Y, Liao B, Huang B, Zhou Q, Li X, et al. Preoperative prediction of microvascular invasion in hepatocellular cancer: a radiomics model using Gd-EOB-DTPA-enhanced MRI. Eur Radiol. 2019; 29:4648–4659. PMID: 30689032.
crossref
11. Ji GW, Zhu FP, Zhang YD, Liu XS, Wu FY, Wang K, et al. A radiomics approach to predict lymph node metastasis and clinical outcome of intrahepatic cholangiocarcinoma. Eur Radiol. 2019; 29:3725–3735. PMID: 30915561.
crossref
12. Liu F, Ning Z, Liu Y, Liu D, Tian J, Luo H, et al. Development and validation of a radiomics signature for clinically significant portal hypertension in cirrhosis (CHESS1701): a prospective multicenter study. EBioMedicine. 2018; 36:151–158. PMID: 30268833.
crossref
13. Zheng BH, Liu LZ, Zhang ZZ, Shi JY, Dong LQ, Tian LY, et al. Radiomics score: a potential prognostic imaging feature for postoperative survival of solitary HCC patients. BMC Cancer. 2018; 18:1148. PMID: 30463529.
crossref
14. Szczypiński PM, Strzelecki M, Materka A, Klepaczko A. MaZda--a software package for image texture analysis. Comput Methods Programs Biomed. 2009; 94:66–76. PMID: 18922598.
crossref
15. van Griethuysen JJM, Fedorov A, Parmar C, Hosny A, Aucoin N, Narayan V, et al. Computational radiomics system to decode the radiographic phenotype. Cancer Res. 2017; 77:e104–e107. PMID: 29092951.
crossref
16. Berenguer R, Pastor-Juan MDR, Canales-Vázquez J, Castro-García M, Villas MV, Mansilla Legorburo F, et al. Radiomics of CT features may be nonreproducible and redundant: influence of CT acquisition parameters. Radiology. 2018; 288:407–415. PMID: 29688159.
17. Ji GW, Zhang YD, Zhang H, Zhu FP, Wang K, Xia YX, et al. Biliary tract cancer at CT: a radiomics-based model to predict lymph node metastasis and survival outcomes. Radiology. 2019; 290:90–98. PMID: 30325283.
crossref
18. Kumar V, Gu Y, Basu S, Berglund A, Eschrich SA, Schabath MB, et al. Radiomics: the process and the challenges. Magn Reson Imaging. 2012; 30:1234–1248. PMID: 22898692.
crossref
19. Cozzi L, Dinapoli N, Fogliata A, Hsu WC, Reggiori G, Lobefalo F, et al. Radiomics based analysis to predict local control and survival in hepatocellular carcinoma patients treated with volumetric modulated arc therapy. BMC Cancer. 2017; 17:829. PMID: 29207975.
crossref
20. Guo D, Gu D, Wang H, Wei J, Wang Z, Hao X, et al. Radiomics analysis enables recurrence prediction for hepatocellular carcinoma after liver transplantation. Eur J Radiol. 2019; 117:33–40. PMID: 31307650.
crossref
21. Li W, Huang Y, Zhuang BW, Liu GJ, Hu HT, Li X, et al. Multiparametric ultrasomics of significant liver fibrosis: a machine learning-based analysis. Eur Radiol. 2019; 29:1496–1506. PMID: 30178143.
crossref
22. Kim S, Shin J, Kim DY, Choi GH, Kim MJ, Choi JY. Radiomics on gadoxetic acid-enhanced magnetic resonance imaging for prediction of postoperative early and late recurrence of single hepatocellular carcinoma. Clin Cancer Res. 2019; 25:3847–3855. PMID: 30808773.
crossref
23. Parmar C, Grossmann P, Rietveld D, Rietbergen MM, Lambin P, Aerts HJ. Radiomic machine-learning classifiers for prognostic biomarkers of head and neck cancer. Front Oncol. 2015; 5:272. PMID: 26697407.
crossref
24. Xu X, Zhang HL, Liu QP, Sun SW, Zhang J, Zhu FP, et al. Radiomic analysis of contrast-enhanced CT predicts microvascular invasion and outcome in hepatocellular carcinoma. J Hepatol. 2019; 70:1133–1144. PMID: 30876945.
crossref
25. Zhou Y, He L, Huang Y, Chen S, Wu P, Ye W, et al. CT-based radiomics signature: a potential biomarker for preoperative prediction of early recurrence in hepatocellular carcinoma. Abdom Radiol (NY). 2017; 42:1695–1704. PMID: 28180924.
crossref
26. Hu HT, Wang Z, Huang XW, Chen SL, Zheng X, Ruan SM, et al. Ultrasound-based radiomics score: a potential biomarker for the prediction of microvascular invasion in hepatocellular carcinoma. Eur Radiol. 2019; 29:2890–2901. PMID: 30421015.
crossref
27. Ogutu JO, Schulz-Streeck T, Piepho HP. Genomic selection using regularized linear regression models: ridge regression, lasso, elastic net and their extensions. BMC Proc. 2012; 6 Suppl 2:S10. PMID: 22640436.
crossref
28. Wang S, Summers RM. Machine learning and radiology. Med Image Anal. 2012; 16:933–951. PMID: 22465077.
crossref
29. Lubner MG, Malecki K, Kloke J, Ganeshan B, Pickhardt PJ. Texture analysis of the liver at MDCT for assessing hepatic fibrosis. Abdom Radiol (NY). 2017; 42:2069–2078. PMID: 28314916.
crossref
30. Naganawa S, Enooku K, Tateishi R, Akai H, Yasaka K, Shibahara J, et al. Imaging prediction of nonalcoholic steatohepatitis using computed tomography texture analysis. Eur Radiol. 2018; 28:3050–3058. PMID: 29404772.
crossref
31. Shan QY, Hu HT, Feng ST, Peng ZP, Chen SL, Zhou Q, et al. CT-based peritumoral radiomics signatures to predict early recurrence in hepatocellular carcinoma after curative tumor resection or ablation. Cancer Imaging. 2019; 19:11. PMID: 30813956.
crossref
32. Yuan C, Wang Z, Gu D, Tian J, Zhao P, Wei J, et al. Prediction early recurrence of hepatocellular carcinoma eligible for curative ablation using a radiomics nomogram. Cancer Imaging. 2019; 19:21. PMID: 31027510.
crossref
33. Akai H, Yasaka K, Kunimatsu A, Nojima M, Kokudo T, Kokudo N, et al. Predicting prognosis of resected hepatocellular carcinoma by radiomics analysis with random survival forest. Diagn Interv Imaging. 2018; 99:643–651. PMID: 29910166.
crossref
34. Iwatsuki S, Dvorchik I, Marsh JW, Madariaga JR, Carr B, Fung JJ, et al. Liver transplantation for hepatocellular carcinoma: a proposal of a prognostic scoring system. J Am Coll Surg. 2000; 191:389–394. PMID: 11030244.
35. Lim KC, Chow PK, Allen JC, Chia GS, Lim M, Cheow PC, et al. Microvascular invasion is a better predictor of tumor recurrence and overall survival following surgical resection for hepatocellular carcinoma compared to the Milan criteria. Ann Surg. 2011; 254:108–113. PMID: 21527845.
crossref
36. Iguchi T, Shirabe K, Aishima S, Wang H, Fujita N, Ninomiya M, et al. New pathologic stratification of microvascular invasion in hepatocellular carcinoma: predicting prognosis after living-donor liver transplantation. Transplantation. 2015; 99:1236–1242. PMID: 25427164.
37. Peng J, Zhang J, Zhang Q, Xu Y, Zhou J, Liu L. A radiomics nomogram for preoperative prediction of microvascular invasion risk in hepatitis B virus-related hepatocellular carcinoma. Diagn Interv Radiol. 2018; 24:121–127. PMID: 29770763.
crossref
38. Zhao B, Tan Y, Tsai WY, Qi J, Xie C, Lu L, et al. Reproducibility of radiomics for deciphering tumor phenotype with imaging. Sci Rep. 2016; 6:23428. PMID: 27009765.
crossref
39. Park JE, Park SY, Kim HJ, Kim HS. Reproducibility and generalizability in radiomics modeling: possible strategies in radiologic and statistical perspectives. Korean J Radiol. 2019; 20:1124–1137. PMID: 31270976.
crossref
40. Orlhac F, Frouin F, Nioche C, Ayache N, Buvat I. Validation of a method to compensate multicenter effects affecting CT radiomics. Radiology. 2019; 291:53–59. PMID: 30694160.
crossref
41. Chartrand G, Cheng PM, Vorontsov E, Drozdzal M, Turcotte S, Pal CJ, et al. Deep learning: a primer for radiologists. Radiographics. 2017; 37:2113–2131. PMID: 29131760.
crossref
42. Lee JG, Jun S, Cho YW, Lee H, Kim GB, Seo JB, et al. Deep learning in medical imaging: general overview. Korean J Radiol. 2017; 18:570–584. PMID: 28670152.
crossref
43. Zhou LQ, Wang JY, Yu SY, Wu GG, Wei Q, Deng YB, et al. Artificial intelligence in medical imaging of the liver. World J Gastroenterol. 2019; 25:672–682. PMID: 30783371.
crossref
44. Choy G, Khalilzadeh O, Michalski M, Do S, Samir AE, Pianykh OS, et al. Current applications and future impact of machine learning in radiology. Radiology. 2018; 288:318–328. PMID: 29944078.
crossref
45. Choi KJ, Jang JK, Lee SS, Sung YS, Shim WH, Kim HS, et al. Development and validation of a deep learning system for staging liver fibrosis by using contrast agent-enhanced CT images in the liver. Radiology. 2018; 289:688–697. PMID: 30179104.
46. Iranmanesh P, Vazquez O, Terraz S, Majno P, Spahr L, Poncet A, et al. Accurate computed tomography-based portal pressure assessment in patients with hepatocellular carcinoma. J Hepatol. 2014; 60:969–974. PMID: 24362073.
crossref
47. Nakayama Y, Li Q, Katsuragawa S, Ikeda R, Hiai Y, Awai K, et al. Automated hepatic volumetry for living related liver transplantation at multisection CT. Radiology. 2006; 240:743–748. PMID: 16857979.
crossref
48. Wang K, Mamidipalli A, Retson T, Bahrami N, Hasenstab K, Blansit K, et al. Automated CT and MRI liver segmentation and biometry using a generalized convolutional neural network. Radiology: Artificial Intelligence. 2019; 3. 27. [Epub]. DOI: 10.1148/ryai.2019180022.
49. Ronneberger O, Fischer P, Brox T. U-net: convolutional networks for biomedical image segmentation. arXiv:1505.04597 [cs.CV]. 2015. Accessed August 31, 2019. Available at: https://arxiv.org/abs/1505.04597.
50. Hu P, Wu F, Peng J, Liang P, Kong D. Automatic 3D liver segmentation based on deep learning and globally optimized surface evolution. Phys Med Biol. 2016; 61:8676–8698. PMID: 27880735.
crossref
51. Huo Y, Terry JG, Wang J, Nair S, Lasko TA, Freedman BI, et al. Fully automatic liver attenuation estimation combing CNN segmentation and morphological operations. Med Phys. 2019; 46:3508–3519. PMID: 31228267.
crossref
52. Lu F, Wu F, Hu P, Peng Z, Kong D. Automatic 3D liver location and segmentation via convolutional neural network and graph cut. Int J Comput Assist Radiol Surg. 2017; 12:171–182. PMID: 27604760.
crossref
53. van Gastel MDA, Edwards ME, Torres VE, Erickson BJ, Gansevoort RT, Kline TL. Automatic measurement of kidney and liver volumes from MR images of patients affected by autosomal dominant polycystic kidney disease. J Am Soc Nephrol. 2019; 30:1514–1522. PMID: 31270136.
crossref
54. Wang CJ, Hamm CA, Savic LJ, Ferrante M, Schobert I, Schlachter T, et al. Deep learning for liver tumor diagnosis part II: convolutional neural network interpretation using radiologic imaging features. Eur Radiol. 2019; 29:3348–3357. PMID: 31093705.
crossref
55. Liu X, Song JL, Wang SH, Zhao JW, Chen YQ. Learning to diagnose cirrhosis with liver capsule guided ultrasound image classification. Sensors (Basel). 2017; 17:E149. PMID: 28098774.
crossref
56. Yasaka K, Akai H, Kunimatsu A, Abe O, Kiryu S. Deep learning for staging liver fibrosis on CT: a pilot study. Eur Radiol. 2018; 28:4578–4585. PMID: 29761358.
crossref
57. Yasaka K, Akai H, Kunimatsu A, Abe O, Kiryu S. Liver fibrosis: deep convolutional neural network for staging by using gadoxetic acid-enhanced hepatobiliary phase MR images. Radiology. 2018; 287:146–155. PMID: 29239710.
crossref
58. Biswas M, Kuppili V, Edla DR, Suri HS, Saba L, Marinhoe RT, et al. Symtosis: a liver ultrasound tissue characterization and risk stratification in optimized deep learning paradigm. Comput Methods Programs Biomed. 2018; 155:165–177. PMID: 29512496.
crossref
59. Byra M, Styczynski G, Szmigielski C, Kalinowski P, Michałowski Ł, Paluszkiewicz R, et al. Transfer learning with deep convolutional neural network for liver steatosis assessment in ultrasound images. Int J Comput Assist Radiol Surg. 2018; 13:1895–1903. PMID: 30094778.
crossref
60. Cao W, An X, Cong L, Lyu C, Zhou Q, Guo R. Application of deep learning in quantitative analysis of 2-dimensional ultrasound imaging of nonalcoholic fatty liver disease. J Ultrasound Med. 2020; 39:51–59. PMID: 31222786.
crossref
61. Vorontsov E, Cerny M, Régnier P, Di Jorio L, Pal CJ, Lapointe R, et al. Deep learning for automated segmentation of liver lesions at CT in patients with colorectal cancer liver metastases. Radiology: Artificial Intelligence. 2019; 3. 13. [Epub]. DOI: 10.1148/ryai.2019180014.
62. Schmauch B, Herent P, Jehanno P, Dehaene O, Saillard C, Aubé C, et al. Diagnosis of focal liver lesions from ultrasound using deep learning. Diagn Interv Imaging. 2019; 100:227–233. PMID: 30926443.
crossref
63. Yasaka K, Akai H, Abe O, Kiryu S. Deep learning with convolutional neural network for differentiation of liver masses at dynamic contrast-enhanced CT: a preliminary study. Radiology. 2018; 286:887–896. PMID: 29059036.
64. Hamm CA, Wang CJ, Savic LJ, Ferrante M, Schobert I, Schlachter T, et al. Deep learning for liver tumor diagnosis part I: development of a convolutional neural network classifier for multi-phasic MRI. Eur Radiol. 2019; 29:3338–3347. PMID: 31016442.
crossref
65. Ma J, Dercle L, Lichtenstein P, Wang D, Chen A, Zhu J, et al. Automated identification of optimal portal venous phase timing with convolutional neural networks. Acad Radiol. 2020; 27:e10–e18. PMID: 31151901.
crossref
66. Esses SJ, Lu X, Zhao T, Shanbhogue K, Dane B, Bruno M, et al. Automated image quality evaluation of T2-weighted liver MRI utilizing deep learning architecture. J Magn Reson Imaging. 2018; 47:723–728. PMID: 28577329.
crossref
67. Liu F, Samsonov A, Chen L, Kijowski R, Feng L. SANTIS: Sampling-Augmented Neural neTwork with Incoherent Structure for MR image reconstruction. Magn Reson Med. 2019; 82:1890–1904. PMID: 31166049.
crossref
68. Tamada D, Kromrey ML, Ichikawa S, Onishi H, Motosugi U. Motion artifact reduction using a convolutional neural network for dynamic contrast enhanced MR imaging of the liver. Magn Reson Med Sci. 2020; 19:64–76. PMID: 31061259.
crossref
69. Chaudhary K, Poirion OB, Lu L, Garmire LX. Deep learning-based multi-omics integration robustly predicts survival in liver cancer. Clin Cancer Res. 2018; 24:1248–1259. PMID: 28982688.
crossref
70. Park SH, Han K. Methodologic guide for evaluating clinical performance and effect of artificial intelligence technology for medical diagnosis and prediction. Radiology. 2018; 286:800–809. PMID: 29309734.
71. Moons KG, Altman DG, Reitsma JB, Ioannidis JP, Macaskill P, Steyerberg EW, et al. Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD): explanation and elaboration. Ann Intern Med. 2015; 162:W1–W73. PMID: 25560730.
crossref
72. England JR, Cheng PM. Artificial intelligence for medical image analysis: a guide for authors and reviewers. AJR Am J Roentgenol. 2019; 212:513–519. PMID: 30557049.
crossref
73. Collins GS, Reitsma JB, Altman DG, Moons KGM. Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD). Ann Intern Med. 2015; 162:735–736.
crossref
74. Han K, Song K, Choi BW. How to develop, validate, and compare clinical prediction models involving radiological parameters: study design and statistical methods. Korean J Radiol. 2016; 17:339–350. PMID: 27134523.
crossref
75. Kim DW, Jang HY, Kim KW, Shin Y, Park SH. Design characteristics of studies reporting the performance of artificial intelligence algorithms for diagnostic analysis of medical images: results from recently published papers. Korean J Radiol. 2019; 20:405–410. PMID: 30799571.
crossref
Fig. 1

Schematic description of morphologic features.

Area and perimeter are calculated from ROI drawn on image. Ellipse fitted to given ROI is obtained. Then, major and minor axes of ellipse and convex area are calculated. Based on these values, morphologic features (circularity, roundness, aspect ratio, solidity, compactness, and others) are calculated according to equations shown in Figure 1. Values of morphologic features for angular and rod shapes are compared with those for complete circle, which has value of 1 for all morphologic features. ROI = region of interest
kjr-21-387-g001
Fig. 2

Schematic description of histogram features.

From ROI drawn on image, histogram of gray-scale pixel values is obtained. Then, multiple features are calculated from histogram to describe pattern of distribution of gray-level values within ROI. CV = coefficient of variation, ENT= entropy, SD = standard deviation
kjr-21-387-g002
Fig. 3

Schematic description of textural feature extraction assuming 3 × 3-pixel image with three different gray-scale levels.

GLCM describes frequency of two neighboring pixels having certain gray-level pixel values, while GLRLM describes length of continuous pixel having certain gray-level pixel value. After aggregating different directional matrices, secondary features are calculated from matrices to describe textural pattern of given image, including CON, ENT, CORR, and HOM, and others from GLCM and SRE, LRE, LGRE, and HGRE, and others for GLRLM. CON = contrast, CORR = correlation, ENT= entropy, GLCM = gray-level co-occurrence matrix, GLRLM = gray-level run-length matrix, HGRE = high gray-level run emphasis, HOM = homogeneity, LGRE = low gray-level run emphasis, LRE = long run emphasis, SRE = short run emphasis
kjr-21-387-g003
Fig. 4

Example images depicting effects of image filters.

Portal venous phase CT image was transformed by using Gaussian, Laplacian, and LOG filters and using wavelet transformation of high-frequency and low-frequency parts. Higher-order features are histogram and textural features extracted from these transformed images. LOG = Laplacian of Gaussian
kjr-21-387-g004
Fig. 5

Schematic description of development process for radiomics classification model.

Model for staging liver fibrosis using gadoxetic acid-enhanced hepatobiliary phase magnetic resonance images is assumed for demonstration purposes. Input images undergo preprocessing, including normalization of gray-scale pixel values and image resampling to standardize image resolution. Radiomics features are then extracted, which may include shape, histogram, texture, and high-order features. Feature selection is performed to reduce feature dimension, and classification model is then developed using selected radiomics features. Final radiomics model is used for classification of new input images. LASSO = least absolute shrinkage and selection operator
kjr-21-387-g005
Fig. 6

Schematic depiction of training CNN.

From input images, Conv layer extracts feature maps, and pooling layer downsizes feature maps. ReLU is usually followed by Conv layer as activation function. High-level features are extracted through multiple Conv and pooling layers, and then fed into fully connected layer. Fully connected layers integrate all features to perform classification task. In this schematic diagram, liver fibrosis staging using CT images is presented as classification task for demonstration purposes. Using softmax function, fully connected layer returns probability of each class as output. During training phase, output of CNN is compared with ground truth to calculate errors using loss function. Error is then back propagated, and weights of network are adjusted to decrease loss and thereby maximize accuracy of CNN for given classification task. CNN = convolutional neural network, Conv = convolution, ReLU = rectified linear unit
kjr-21-387-g006
Table 1

Representative Studies on Clinical Application of Radiomics in Liver Disease

kjr-21-387-i001
Reference Task Imaging Training Group Test Group Validation Method* Test Performance
Park et al., 2019 (6) Liver fibrosis staging Gadoxetic acid-enhanced MRI 329 patients 107 patients Internal (split-sample) AUC of radiomics-based model for fibrosis staging, 0.89–0.91
Liu et al., 2018 (12) Detection of portal Contrast-enhanced CT 222 patients 163 patients External (geographic, multi-center) AUC of radiomics-based model for detecting clinically significant portal hypertension, 0.85
Zheng et al., 2018 (13) Prediction of post-operative prognosis in HCC Contrast-enhanced CT 212 patients 107 patients Internal (split-sample) AUC of radiomics-based nomogram for predicting overall survival, 0.71
Kim et al., 2019 (22) Prediction of early and late recurrence of HCC after curative resection Gadoxetic acid-enhanced MRI 128 patients 39 patients External (temporal) AUC of combined clinicopathologic radiomics model, 0.72
Yuan et al., 2019 (32) Prediction of early recurrence of HCC after curative ablation Contrast-enhanced CT 129 patients 55 patients Internal (split-sample) AUC of combined clinicopathologic radiomics model, 0.76
Xu et al., 2019 (24) Prediction of MVI in HCC Contrast-enhanced CT 350 patients 145 patients Internal (split-sample) AUC of combined clinicopathologic radiomics model, 0.889
Hu et al., 2019 (26) Prediction of MVI in HCC Contrast-enhanced US 341 patients 141 patients External (temporal) AUC of combined clinical and radiomics nomogram, 0.73
Ji et al., 2019 (17) Prediction of lymph node metastasis in biliary tract cancers Contrast-enhanced CT 177 patients 70 patients External (temporal) AUC of combined clinical and radiomics nomogram, 0.80 in test group
Ji et al., 2019 (11) Prediction of lymph node metastasis in IHCC Contrast-enhanced CT 103 patients 52 patients External (temporal) AUC of combined clinical and radiomics nomogram, 0.89
Chen et al., 2019 (9) Prediction of immunoscore of HCC Gadoxetic acid-enhanced MRI 150 patients 57 patients Internal (split-sample) AUC of combined clinical and radiomics model for predicting immunoscore, 0.93

*Validation methods were classified as internal (i.e., cross-validation, bootstrapping, and split-sample validation) or external (temporal and geographic validation). AUC = area under curve, HCC = hepatocellular carcinoma, IHCC = intrahepatic cholangiocarcinoma, MVI = microvascular invasion, US = ultrasound

Table 2

Representative Studies on Clinical Application of Deep Learning in Liver Disease

kjr-21-387-i002
Reference Task Imaging Training Group Test Group Validation Method* Test Performance
Wang et al., 2019 (48) Liver segmentation Gadoxetic acidenhanced MRI, contrast-enhanced CT 10 CT scans and 320 MRI scans 50 CT scans and 133 MRI scans Internal and external (geographic, multi-center) DSS for liver segmentation, 0.92–0.95
Choi et al., 2018 (45) Liver fibrosis staging Contrast-enhanced CT 7491 patients 891 patients Internal and external (geographic, multi-center) AUC, 0.95–0.97
Yasaka et al., 2018 (57) Liver fibrosis staging Gadoxetic acidenhanced MRI 534 patients 100 patients Internal (split-sample) AUC, 0.84–0.85
Wang et al., 2019 (8) Liver fibrosis staging US elastography 266 patients 132 patients External (multi-center) AUC, 0.97–0.98
Vorontsov et al., 2019 (61) Detection and segmentation of liver metastases Contrast-enhanced CT 115 scans 26 scans Internal (split-sample) Per-lesion sensitivity for lesions ≥ 20 mm, 0.85; DSS for lesions ≥ 20 mm, 0.68
Yasaka et al., 2018 (63) Classification of liver tumors Contrast-enhanced CT 460 patients 100 patients External (temporal) Mean accuracy for classification, 0.84
Hamm et al., 2019 (64) Classification of liver tumors Contrast-enhanced MRI 434 lesions 60 lesions Internal (split-sample) Accuracy for classification, 0.92
Liu et al., 2019 (67) MR image reconstruction Gadoxetic acidenhanced MRI 77 scans 16 scans Internal (split-sample) Lower errors and higher similarity compared to compressed sensing
Tamada et al., 2020 (68) Motion artifact reduction Gadoxetic acidenhanced MRI, arterial phase 14 patients 20 patients Internal (split-sample) Significant reduction in artifact score

*Validation methods were classified as internal (i.e., cross-validation, bootstrapping, and split-sample validation) or external (temporal and geographic validation). DSS = Dice similarity score

TOOLS
Similar articles