Journal List > Healthc Inform Res > v.31(3) > 1516092148

Yoon and Kim: Machine Learning-Based Age Prediction with Feature Subset Selection from Magnetic Resonance Angiography Data

Abstract

Objectives

The objective of this study was to evaluate the effectiveness of machine learning (ML) models using selected subsets of features to predict age based on intracranial arterial segments’ tortuosity and diameter characteristics derived from magnetic resonance angiography (MRA) data. Additionally, this study aimed to identify key vascular features important for predicting vascular age.

Methods

Three-dimensional time-of-flight MRA image data from 171 subjects were analyzed. After annotating the endpoints for each arterial segment, 169 features—comprising tortuosity metrics and arterial segment diameter statistics—were extracted. Five ML models (random forest, linear regression, AdaBoost, XGBoost, and lightGBM) were trained and validated. Two feature selection methods, correlation-based feature selection (CFS) and Relief-F, were applied to identify optimal feature subsets.

Results

The random forest model utilizing the CFS-based 50% feature subset achieved the best performance, with a root mean square error of 14.0 years, a coefficient of determination (R2) of 0.275, and a Pearson correlation coefficient of 0.560. Tortuosity metrics (e.g., triangular index of the left posterior cerebral artery P1 segment) appeared more frequently than diameter statistics among the top five most important features.

Conclusions

CFS-based feature selection enhanced the performance of ML-based age prediction compared with using the complete feature set. Linear regression consistently demonstrated the poorest performance across all evaluation metrics. ML-based age prediction using segmental tortuosity metrics and diameter statistics is feasible, potentially revealing significant features related to vascular aging.

I. Introduction

Aging is a well-established independent risk factor contributing to vascular structural and functional impairments, ultimately leading to cardiovascular and neurological diseases [13]. Age-related vascular changes include decreased cerebral microvascular density, reduced vessel count, decreased capillary density, increased arteriolar tortuosity, and arterial stiffening. These changes are associated with thickened vessel walls, modest increases in vessel diameters, and outward hypertrophic remodeling. Arterial stiffening can decrease cerebral blood flow with advancing age, resulting in reduced oxygen and glucose supply to brain tissue. This sustained decline in cerebral blood flow may accelerate neurodegenerative processes, such as vascular dementia.
Numerous imaging tools are available to study age-related vascular and functional changes in the brain [4,5]. For instance, three-dimensional (3D) optical coherence tomography angiography has been employed to measure capillary vessel density and vessel tortuosity [3]. Positron emission tomography has demonstrated age-related declines in cerebral blood flow during brain activity [6]. Magnetic resonance angiography (MRA) image data are useful for assessing chronological age since aging is known to influence morphological characteristics of vessels [7]. Deep convolutional neural networks (CNNs) have been proposed to evaluate vascular aging using brain time-of-flight (TOF) MRA data from healthy subjects [8]. Additionally, combining MRA data with T1-weighted imaging has proven effective for predicting brain age with deep CNN models [9].
Arterial tortuosity, a characteristic associated with aging, is also influenced by factors such as female sex, high blood pressure, genetic disorders involving aortic dissection and dilation, and various other cardiovascular risk factors [1012]. Recent studies have explored segmental arterial tortuosity and diameter as inputs for machine learning regression models for predicting age [13]. However, these studies have been limited to specific segments, primarily focusing on the left and right internal carotid arteries (ICAs) and the basilar artery (BA).
In the current study, we expanded this approach by incorporating not only ICA and BA segments but also additional arterial segments surrounding the circle of Willis. This broader inclusion of arterial segments facilitates a more comprehensive analysis of arterial morphological attributes relevant to age prediction. Given the increased number of available features, we further investigated the effectiveness of feature subset selection methods to enhance the performance of machine learning regression models.

II. Methods

Figure 1 illustrates the overall image preprocessing pipeline, including labeling of arterial segments and extraction of arterial tortuosity and diameter statistics features. Figure 2 summarizes the study as a flowchart outlining feature extraction, selection of relevant features, and development and validation of machine learning regression models. Each procedure is described in detail below.

1. Data and Labeling

The 3D MRA dataset was obtained from the IXI datasets (https://brain-development.org/ixi-dataset). After data exclusion, data from 171 subjects were used for analysis. Following manual identification of three seed points in the left and right ICAs and the BA, seeded region-growing was conducted for arterial segmentation [14,15]. The skeletonization function in the scikit-image library [16] was used to extract centerlines of the arteries in 3D. The Plotly Python library (https://plotly.com/python) was used to manually annotate arterial landmarks [14]. The Dijkstra algorithm was then employed to determine the shortest path between two identified endpoints [17]. Resulting segments of the centerlines were color-coded, as shown in Figure 3, and saved as .html files, which were visually inspected to ensure annotation accuracy. Data with incorrect pathfinding results were excluded from analysis. Centerline points underwent smoothing via spline interpolation. Missing centerlines in certain posterior communicating artery (PComm) segments were treated as missing values, imputed using the K-nearest neighbor (KNN) method from scikit-learn [18] with nneighbors = 5.

2. Feature Extraction

The features extracted from each arterial segment included six tortuosity metrics and seven diameter statistics.
1) Tortuosity metrics: The Euclidean length (l) between two endpoints, the curve length (lc), relative length, triangular index, sum of angles metric, and product of angle distance [19].
Relative length (RL):
(1)
RL=llc,
where l is the Euclidean length, and lc is the curve length (Figure 4A).
Triangular index (TI):
(2)
TI=Σi=1n(ai+bi)/cin,
where ai and bi are lengths of the two sides along the arterial segment, and ci is the length of the remaining side in the i-th triangle (Figure 4B).
Sum of angles metric (SOAM):
(3)
SOAM=Σi=1n(180°-φi)lc,
where ϕi is the angle between two adjacent lines connecting the center and two endpoints for the i-th center (Figure 4C). Product of angle distance (PAD):
(4)
PAD=SOAMRL.
2) Diameter statistics: Mean, minimum, maximum, standard deviation, 25th percentile, 50th percentile, and 75th percentile values. Diameters were calculated along the centerline points using the distance_transform_edt() function from the SciPy library [20] on 3D binary vessel masks.
The 13 arterial segments were considered for feature extraction, and their names and anatomical locations are shown in Table 1 and Figure 3, respectively. Hence, the total number of features per subject was 13 segments × (6 + 7) features/segment = 169, which is equal to 13 arterial segments multiplied by 13 features (i.e., 6 tortuosity metrics and 7 diameter statistics).

3. Feature Selection

Two methods, correlation-based feature selection (CFS) and Relief-F, were applied to reduce dataset dimensionality.
CFS is a sophisticated multivariate filter method designed to optimize feature subsets for predictive modeling [21]. It selects subsets that correlate strongly with the target variable while minimizing inter-feature correlations, ensuring each feature uniquely contributes to prediction. By assessing predictive value and redundancy among features, CFS maximizes predictive accuracy. This method refines the predictive capability of models while simplifying feature selection, making it valuable in machine learning applications. It enhances the effectiveness of predictive models by focusing on essential and non-overlapping features.
Relief-F extends the earlier Relief algorithm, specifically tailored for multi-class datasets [22]. It selects features by choosing samples and identifying their closest matches within the same class. By comparing attribute values between these samples and their nearest hits (similar instances) and misses (dissimilar instances), Relief-F continuously refines relevance scores for each feature [23]. Features that consistently show similar values within a class and distinct values across classes are identified as highly relevant, thereby enhancing predictive accuracy.
Feature subsets comprising 70%, 50%, 30%, and 10% of the total 169 features were selected using both methods. Age, the target variable, was categorized into three groups (20–39, 40–59, 60–89 years). For CFS, subsets with the highest merit based on feature-class and feature-feature correlations were selected from all possible subsets. Relief-F utilized the Relief-F library (https://pypi.org/project/ReliefF/) with “n_features_to_keep” set to 118 (70%), 85 (50%), 51 (30%), and 17 (10%), fixing “n_neighbors” at 25. The “n_neighbors” parameter specifies the number of nearest neighbors evaluated when assessing feature significance. These selected subsets were subsequently input into machine learning models.

4. Model Development

Five machine learning regression models were compared for age prediction: random forest, linear regression, AdaBoost, XGBoost, and lightGBM regression. Random forest, linear regression, and AdaBoost models were trained and validated using scikit-learn [18]. XGBoost regression models were trained and validated using XGBoost [24] (https://xgboost.readthedocs.io/en/stable/). LightGBM regression models were trained and validated using LightGBM [25] (https://lightgbm.readthedocs.io/en/stable/). Each model was evaluated using four-fold cross-validation. For each fold, hyper-parameters were tuned via randomized search [26] using candidate parameter sets.

5. Model Evaluation

The machine learning regression models’ prediction performance was evaluated using the following metrics. When xi is the real age of the i-th subject, and yi is the predicted age of the i-th subject, the Pearson correlation coefficient is defined as
(5)
r=Σi=1N(xi-x¯)(yi-y¯)Σi=1N(xi-x¯)2Σi=1N(yi-y¯)2,
where χ̄ is the mean of the xi values, and is the mean of the yi values. The correlation coefficient r ranges from −1 to 1. Root mean square error (RMSE) is defined as
(6)
RMSE=Σi=1N(xi-yi)2N.
Mean absolute percentage error (MAPE) is defined as
(7)
MAPE=1Ni=1N|xi-yixi|.
The coefficient of determination or R2 is defined as
(8)
R2=1-Σi=1N(xi-yi)2Σi=1N(xi-x¯)2.
Feature importance was calculated for all models using the scikit-learn library.

6. Model Interpretation

The Shapley additive explanations (SHAP) method was used to interpret model predictions [27]. SHAP utilizes concepts from cooperative game theory, specifically Shapley values, to quantify the fair contribution of each feature to a model’s predictions. This approach measures each feature’s impact on individual predictions, thereby enhancing the transparency and interpretability of the model’s decisions [28]. SHAP values can be computed through various methods, including Kernel SHAP, Deep SHAP, and Tree SHAP. In this study, Tree SHAP was used, as it is specifically designed for tree-based machine learning models such as decision trees, random forests, and gradient-boosted trees [29]. SHAP values offer significant advantages in terms of both global and local interpretability. Unlike traditional feature importance metrics, SHAP assesses not only the overall importance of each feature but also indicates whether the effect of a feature is positive or negative on the prediction. Furthermore, SHAP values are calculated for each observation, providing personalized insights. Consequently, SHAP facilitates a comprehensive understanding of the model at both global and individual levels.

III. Results

Figure 5 presents a histogram depicting the age distribution grouped in 10-year bins. The largest age group was 60–70 years, comprising 40 subjects, while the smallest group, aged 80–90 years, included only four subjects.
Table 2 summarizes the prediction performance of the five machine learning regression models. The random forest regression model combined with 50% correlation-based feature subset selection exhibited the best overall performance, with the lowest average RMSE (14.001 years), highest average R2 (0.275), and highest average Pearson correlation coefficient (0.560). Conversely, linear regression had the poorest performance, indicated by the highest average RMSE (15.756 years), lowest average R2 (0.065), highest average MAPE (0.301), and lowest Pearson correlation coefficient (0.454). However, linear regression performance improved (lower RMSE) as the feature subset percentage decreased, suggesting that feature selection positively affected its accuracy. Across all models, the correlation-based feature selection method consistently outperformed the Relief-F method. Among the tested models, the LightGBM regression model achieved the lowest average MAPE (0.273).
Table 3 shows the five most important features identified by each regression model. In the random forest and Ada-Boost models, L-P1_TI (triangular index of the left posterior cerebral artery P1 segment) was consistently the most important feature across all degrees of feature subset selection. L-P1_TI also frequently appeared among the top features in the XGBoost and LightGBM models, occurring in 4 out of 5 instances for XGBoost and in 3 out of 5 for LightGBM. In the XGBoost model, L-P2_TI (triangular index of the left posterior cerebral artery P2 segment) was identified as the most important feature in 2 of the 5 feature subset selections. L-P2_TI frequently appeared as an important feature, specifically 4 out of 5 times in random forest, AdaBoost, and XGBoost models. Notably, arterial diameter-related features were identified as important less frequently, appearing in only 32 of the 125 cases analyzed, predominantly within the right ICA segment (25 of 32 cases).
Figure 6 displays box plots for the two most important features, L-P1_TI and L-P2_SOAM, obtained from training the random forest regression model using a 10% feature subset selection (see Table 3). The correlation-based feature selection method was applied to select the top 10% (17) features. Feature importance determined by the random forest regression model was ranked as follows: L-P1_TI, L-P2_SOAM, L-P1_SOAM, R-ICA_d_min, L-P2_RL, and L-P1_PAD, with importance values of 0.152, 0.111, 0.074, 0.073, 0.071, and 0.007, respectively. Three age groups were compared: Group A (20–39 years, n = 56), Group B (40–59 years, n = 57), and Group C (60–89 years, n = 58). Figure 6A demonstrates that the L-P1_TI value increased with age. Mean values for Groups A, B, and C were 1.000271, 1.000317, and 1.000495, respectively. Although the range of TI values was narrow, clear differences in median values are noticeable in the box plots. Similarly, Figure 6B indicates that L-P2_SOAM values increased with age, with mean values of 5.005, 6.052, and 6.865 for Groups A, B, and C, respectively.
Figure 7 displays a Shapley dot plot calculated from a validation dataset. For each feature, individual points on the plot are positioned along the horizontal SHAP value axis, with color indicating the magnitude of the feature value. The SHAP values show that higher feature values correspond to higher predicted age for the L-P1_TI, L-P1_PAD, and L-P2_PAD features. This relationship is expected because larger TI and PAD values indicate increased arterial tortuosity, a feature anticipated in older age groups. Conversely, higher feature values correspond to lower predicted age for the L-P2_RL and R-P2_RL features. This result aligns with expectations because, according to Eq. (1), a smaller RL value indicates greater arterial tortuosity, typically associated with older age.

IV. Discussion

This study evaluated the performance of machine learning-based age prediction using TOF-MRA data with features derived from tortuosity and diameter measurements of 13 arterial segments in the circle of Willis. Multiple regression models were tested using various feature subset selection methods, and their performance was assessed using metrics such as RMSE, Pearson correlation coefficient, R2, and MAPE.
Compared to a previous study [13] that focused solely on the ICA and BA segments, this study, which incorporated 13 arterial segments, demonstrated improved prediction performance. For instance, the random forest regression model in the current study achieved lower average RMSE (14.001 ± 0.444 years) and higher R2 (0.275 ± 0.065) compared to the earlier study’s RMSE (14.867 ± 0.515 years) and R2 (0.186 ± 0.031) [13]. These results highlight the value of incorporating additional arterial segments and employing feature subset selection methods. One notable exception was the linear regression model, which performed worse in the current study, showing higher average RMSE (15.756 ± 1.919 years) and lower R2 (0.065 ± 0.258) compared to the previous study (RMSE 15.000 ± 1.032 years; R2 0.162 ± 0.140). This suggests that linear regression may not effectively handle the increased complexity introduced by additional features.
The difference in performance between linear regression and other machine learning models can be attributed to inherent model characteristics. Linear regression assumes a straightforward linear relationship between features and the target variable (age), making it difficult to capture complex interactions embedded within the features. Consequently, linear regression exhibited poorer performance as the number of features increased. Conversely, non-linear models such as random forest and XGBoost effectively handled feature selection due to their capability to capture complex, non-linear relationships between features and the target variable.
Regarding feature selection, we considered filter methods that select features prior to training machine learning models. Although embedded and wrapper methods, which integrate machine learning algorithms directly into the feature selection process, are available [30], they were not explored in this study. These alternative methods might be valuable in future studies aimed at further enhancing predictive performance.
Although numerous missing values existed in the PComm arterial segments, applying data imputation methods such as KNN imputation improved performance while preserving the total number of data samples. Given the high dimensionality of features (169 features) in this study, investigating more advanced imputation methods could potentially further enhance prediction performance. Additionally, certain subjects’ data were excluded due to insufficient field of view in the superior-inferior direction affecting ICA or BA. Interestingly, ICA-related features were not dominant in feature importance in this study. Thus, future studies should consider including these previously discarded subjects’ data with insufficient fields of view and evaluating their impact on the machine learning regression models’ performance.
Previous studies have also investigated age prediction using MRA image data [8,9] and reported better prediction performances compared to our approach. However, those studies employed end-to-end deep convolutional neural network models, characterized by black-box behavior, which limits model interpretability. In contrast, our approach explicitly performs feature extraction based on anatomical structures of segmental arteries, allowing machine learning models to directly identify and interpret important features.
Our study underscores the potential of segmental feature extraction to generate numerous morphological features available for analysis, and highlights the utility of feature subset selection for enhancing age prediction accuracy. Moreover, machine learning-derived feature importance scores directly pinpoint anatomical features most relevant to age prediction. For instance, the L-P1_TI consistently emerged as one of the most significant features across most models.
This study has several limitations. First, the sample size was relatively small. Expanding the dataset might require correcting inaccurate arterial centerlines and including data with smaller fields of view along with suitable imputation strategies. Second, clinical information such as history of stroke, hypertension, and dyslipidemia was unavailable. This clinical information was intentionally excluded from our model since our goal was to develop machine learning models based exclusively on morphological vascular features. Third, the complexity of image processing tasks— including segmentation, skeletonization, and measurements of tortuosity and diameter—poses practical challenges for large angiography datasets and requires manual correction of arterial centerlines.
In conclusion, this study demonstrated the feasibility of machine learning-based age prediction using high-dimensional features derived from arterial tortuosity and diameter measurements across 13 segments of the circle of Willis. Among tested models, random forest regression with 50% correlation-based feature subset selection achieved the best performance, characterized by the lowest RMSE and highest average R2. Although linear regression showed the weakest overall performance, it improved when using reduced feature subsets, highlighting the importance of effective feature selection. Evaluation of feature importance suggests future studies might benefit from including previously excluded data with missing or incomplete information due to limited fields of view in the inferior portions of the brain. Future research should also explore more sophisticated imputation methods and advanced feature selection techniques to potentially further enhance the performance of machine learning-based age prediction.

Notes

Conflict of Interest

No potential conflict of interest relevant to this article was reported.

Acknowledgments

This study was supported by “Regional Innovation Strategy (RIS)” through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (No. 2022RIS-005) and the Ministry of Science and ICT, Korea, under the National Program for Excellence in Software, supervised by the IITP (Institute of Information & communications Technology Planning & Evaluation) in 2024 (No. 2019-0-01219).

References

1. Xu X, Wang B, Ren C, Hu J, Greenberg DA, Chen T, et al. Age-related impairment of vascular structure and functions. Aging Dis. 2017; 8(5):590–610. https://doi.org/10.14336/AD.2017.0430.
crossref
2. Donato AJ, Machin DR, Lesniewski LA. Mechanisms of dysfunction in the aging vasculature and role in age-related disease. Circ Res. 2018; 123(7):825–48. https://doi.org/10.1161/CIRCRESAHA.118.312563.
crossref
3. Li Y, Choi WJ, Wei W, Song S, Zhang Q, Liu J, et al. Aging-associated changes in cerebral vasculature and blood flow as determined by quantitative optical coherence tomography angiography. Neurobiol Aging. 2018; 70:148–59. https://doi.org/10.1016/j.neurobiolaging.2018.06.017.
crossref
4. MacDonald ME, Pike GB. MRI of healthy brain aging: a review. NMR Biomed. 2021; 34(9):e4564. https://doi.org/10.1002/nbm.4564.
crossref
5. Ota Y, Shah G. Imaging of normal brain aging. Neuroimaging Clin N Am. 2022; 32(3):683–98. https://doi.org/10.1016/j.nic.2022.04.010.
crossref
6. Aanerud J, Borghammer P, Chakravarty MM, Vang K, Rodell AB, Jonsdottir KY, et al. Brain energy metabolism and blood flow differences in healthy aging. J Cereb Blood Flow Metab. 2012; 32(7):1177–87. https://doi.org/10.1038/jcbfm.2012.18.
crossref
7. Bullitt E, Zeng D, Mortamet B, Ghosh A, Aylward SR, Lin W, et al. The effects of healthy aging on intracerebral blood vessels visualized by magnetic resonance angiography. Neurobiol Aging. 2010; 31(2):290–300. https://doi.org/10.1016/j.neurobiolaging.2008.03.022.
crossref
8. Nam Y, Jang J, Lee HY, Choi Y, Shin NY, Ryu KH, et al. Estimating age-related changes in in vivo cerebral magnetic resonance angiography using convolutional neural network. Neurobiol Aging. 2020; 87:125–31. https://doi.org/10.1016/j.neurobiolaging.2019.12.008.
crossref
9. Mouches P, Wilms M, Rajashekar D, Langner S, Forkert ND. Multimodal biological brain age prediction using magnetic resonance imaging and angiography with the identification of predictive regions. Hum Brain Mapp. 2022; 43(8):2554–66. https://doi.org/10.1002/hbm.25805.
crossref
10. Ciurica S, Lopez-Sublet M, Loeys BL, Radhouani I, Natarajan N, Vikkula M, et al. Arterial tortuosity. Hypertension. 2019; 73(5):951–60. https://doi.org/10.1161/HYPERTENSIONAHA.118.11647.
crossref
11. Diedrich KT. Arterial tortuosity measurement system for examining correlations with vascular disease [dissertation]. Salt Lake City (UT): The University of Utah;2011.
12. Morris SA. Arterial tortuosity in genetic arteriopathies. Curr Opin Cardiol. 2015; 30(6):587–93. https://doi.org/10.1097/HCO.0000000000000218.
crossref
13. Yoon HS, Oh J, Kim YC. Assessing machine learning models for predicting age with intracranial vessel tortuosity and thickness information. Brain Sci. 2023; 13(11):1512. https://doi.org/10.3390/brainsci13111512.
crossref
14. Kim SO, Kim YC. Effects of path-finding algorithms on the labeling of the centerlines of circle of Willis arteries. Tomography. 2023; 9(4):1423–33. https://doi.org/10.3390/tomography9040113.
crossref
15. Javed A, Kim YC, Khoo MC, Ward SL, Nayak KS. Dynamic 3-D MR visualization and detection of upper airway obstruction during sleep using region-growing segmentation. IEEE Trans Biomed Eng. 2016; 63(2):431–7. https://doi.org/10.1109/TBME.2015.2462750.
crossref
16. van der Walt S, Schonberger JL, Nunez-Iglesias J, Boulogne F, Warner JD, Yager N, et al. scikit-image: image processing in Python. PeerJ. 2014; 2:e453. https://doi.org/10.7717/peerj.453.
crossref
17. Kim SO, Kim YC. An improved path-finding method for the tracking of centerlines of tortuous internal carotid arteries in MR angiography. J Imaging. 2024; 10(3):58. https://doi.org/10.3390/jimaging10030058.
crossref
18. Pedregosa F, Varoquaux G, Gramfort A, Michel V, Thirion B, Grisel O, et al. Scikit-learn: machine learning in Python. J Mach Learn Res. 2011; 12:2825–30.
19. Klis KM, Krzyzewski RM, Kwinta BM, Stachura K, Gasowski J. Tortuosity of the internal carotid artery and its clinical significance in the development of aneurysms. J Clin Med. 2019; 8:237. https://doi.org/10.3390/jcm8020237.
crossref
20. Virtanen P, Gommers R, Oliphant TE, Haberland M, Reddy T, Cournapeau D, et al. Author Correction: SciPy 1.0: fundamental algorithms for scientific computing in Python. Nat Methods. 2020; 17(3):352. https://doi.org/10.1038/s41592-020-0772-5.
crossref
21. Remeseiro B, Bolon-Canedo V. A review of feature selection methods in medical applications. Comput Biol Med. 2019; 112:103375. https://doi.org/10.1016/j.compbiomed.2019.103375.
crossref
22. Robnik-Sikonja M, Kononenko I. Theoretical and empirical analysis of ReliefF and RReliefF. Mach Learn. 2003; 53:23–69. https://doi.org/10.1023/A:1025667309714.
crossref
23. Kononenko I. Estimating attributes: analysis and extensions of RELIEF. Bergadano F, De Raedt L, editors. Machine learning: ECML-94. Heidelberg, Germany: Springer;1994. p. 171–82. https://doi.org/10.1007/3-540-57868-4_57.
crossref
24. Chen T, Guestrin C. XGBoost: a scalable tree boosting system. In : Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining; 2016 Aug 13–17; San Francisco, CA, USA. p. 785–94. https://doi.org/10.1145/2939672.2939785.
crossref
25. Ke G, Meng Q, Finley T, Wang T, Chen W, Ma W, et al. LightGBM: a highly efficient gradient boosting decision tree. Adv Neural Inf Process Syst. 2017; 30:3146–54.
26. Bergstra J, Bengio Y. Random search for hyper-parameter optimization. J Mach Learn Res. 2012; 13(1):281–305.
27. Lundberg SM, Lee SI. A unified approach to interpreting model predictions. Adv Neural Inf Process Syst. 2017; 30:4765–74.
28. Parsa AB, Movahedi A, Taghipour H, Derrible S, Mohammadian AK. Toward safer highways, application of XGBoost and SHAP for real-time accident detection and feature analysis. Accid Anal Prev. 2020; 136:105405. https://doi.org/10.1016/j.aap.2019.105405.
crossref
29. Mangalathu S, Hwang SH, Jeon JS. Failure mode and effects analysis of RC members based on machine-learning-based SHapley Additive exPlanations (SHAP) approach. Eng Struct. 2020; 219:110927. https://doi.org/10.1016/j.engstruct.2020.110927.
crossref
30. Tan PN, Steinbach M, Karpatne A, Kumar V. Introduction to data mining. 2nd ed. Harlow, UK: Pearson Education;2018.

Figure 1
Flowchart of data preprocessing to extract features. TOF: time-of-flight, MRA: magnetic resonance angiography.
hir-2025-31-3-284f1.gif
Figure 2
Flowchart of the proposed machine learning-based age prediction process. SOAM: sum of angles metric, PAD: product of angle distance, RL: relative length, TI: triangular index. See Table 1 for abbreviations of 13 segments.
hir-2025-31-3-284f2.gif
Figure 3
Representative example of the 13 arterial segments visualized in skeletonized arteries. Different colors were used to indicate each arterial segment. See Table 1 for abbreviations of 13 segments.
hir-2025-31-3-284f3.gif
Figure 4
Graphical illustration of measurements of arterial tortuosity features. (A) The Euclidean length (l) and the curve length (lc). (B) Triangular index. (C) Sum of angles metric.
hir-2025-31-3-284f4.gif
Figure 5
Histogram of age distribution.
hir-2025-31-3-284f5.gif
Figure 6
Box plots of the two most important features, (A) L-P1_TI and (B) L-P2_SOAM, obtained from training the random forest regression model with 10% selected features in Table 3. Three age groups were compared, where Groups A, B, and C correspond to 20–39 years (n = 56), 40–59 years (n = 57), and 60–89 years (n = 58), respectively.
hir-2025-31-3-284f6.gif
Figure 7
Shapley diagram. Each point on the plot corresponds to a Shapley additive explanation (SHAP) value for an individual data point within the validation sets, categorized by feature. The horizontal alignment of points reveals the influence of each feature on the model’s predictions (SHAP value). The color gradient from blue to red highlights the range of actual feature values, from low to high, illustrating how changes in feature values correspond to shifts in SHAP values. See Table 1 for abbreviations of arterial segments.
hir-2025-31-3-284f7.gif
Table 1
List of arterial segments
Abbreviation Meaning
R-ACA-A1 Right anterior cerebral artery A1
L-ACA-A1 Left anterior cerebral artery A1
R-MCA-M1 Right middle cerebral artery M1
L-MCA-M1 Left middle cerebral artery M1
R-ICA Right internal carotid artery
L-ICA Left internal carotid artery
R-PComm Right posterior communicating artery
L-PComm Left posterior communicating artery
R-PCA-P1 Right posterior cerebral artery P1
L-PCA-P1 Left posterior cerebral artery P1
R-PCA-P2 Right posterior cerebral artery P2
L-PCA-P2 Left posterior cerebral artery P2
BA Basilar artery
Table 2
Evaluation of model predictions using four-fold cross-validation
Model Feature selection Subset RMSE R2 MAPE PCC
Random forest Full 100% 14.146 ± 0.311 0.261±0.055 0.298 ± 0.025 0.554 ± 0.051
CFS 70% 14.151 ± 0.510 0.261±0.055 0.295 ± 0.028 0.546 ± 0.057
50% 14.001 ± 0.444 0.275±0.065 0.293 ± 0.026 0.560 ± 0.062
30% 14.188 ± 0.495 0.257±0.057 0.294 ± 0.027 0.530 ± 0.066
10% 14.242 ± 0.671 0.249±0.091 0.291 ± 0.029 0.518 ± 0.089
Relief-F 70% 14.512 ± 0.274 0.223±0.036 0.307 ± 0.023 0.512 ± 0.043
50% 14.473 ± 0.320 0.227±0.043 0.303 ± 0.022 0.518 ± 0.048
30% 14.828 ± 0.224 0.189±0.026 0.309 ± 0.018 0.457± 0.029
10% 15.366 ± 0.377 0.130±0.032 0.323 ± 0.025 0.382 ± 0.035

Linear regression Full 100% 35.375 ± 3.791 −3.627±0.791 0.602 ± 0.054 0.288 ± 0.148
CFS 70% 72.104 ± 35.014 −23.173±23.807 1.085 ± 0.406 0.125 ± 0.090
50% 46.440 ± 15.411 −8.013±5.289 0.711 ± 0.173 0.168 ± 0.215
30% 18.665 ± 1.692 −0.304±0.274 0.343 ± 0.035 0.397 ± 0.092
10% 15.756 ± 1.919 0.065±0.258 0.301 ± 0.025 0.454 ± 0.092
Relief-F 70% 69.747 ± 32.013 −21.087±20.258 1.113 ± 0.462 0.205 ± 0.115
50% 43.042 ± 25.759 −8.468±11.188 0.544 ± 0.152 0.194 ± 0.114
30% 20.790 ± 1.705 −0.616±0.308 0.347 ± 0.018 0.297 ± 0.158
10% 17.289 ± 1.068 −0.109±0.164 0.348 ± 0.012 0.230 ± 0.089

AdaBoost Full 100% 14.598 ± 1.120 0.206±0.147 0.290 ± 0.044 0.498 ± 0.112
CFS 70% 14.324 ± 1.055 0.237±0.131 0.282 ± 0.040 0.518 ± 0.106
50% 14.209 ± 1.303 0.247±0.158 0.280 ± 0.041 0.533 ± 0.102
30% 14.580± 1.255 0.208±0.159 0.287 ± 0.045 0.502 ± 0.118
10% 14.761± 1.818 0.182±0.227 0.292 ± 0.053 0.485 ± 0.152
Relief-F 70% 14.533± 0.606 0.218±0.091 0.290 ± 0.032 0.498 ± 0.066
50% 15.326± 0.904 0.127±0.131 0.303 ± 0.039 0.434 ± 0.092
30% 14.882± 1.027 0.176±0.140 0.297 ± 0.032 0.465 ± 0.113
10% 16.293± 0.778 0.021±0.067 0.348 ± 0.021 0.227 ± 0.108

XGBoost Full 100% 15.141± 0.818 0.149±0.118 0.290 ± 0.031 0.458 ± 0.079
CFS 70% 15.296± 0.935 0.240±0.123 0.283 ± 0.042 0.526 ± 0.092
50% 14.673± 0.820 0.202±0.106 0.290 ± 0.035 0.488 ± 0.086
30% 14.808± 0.801 0.188±0.104 0.285 ± 0.034 0.492 ± 0.081
10% 14.934± 1.174 0.169±0.154 0.292 ± 0.039 0.462 ± 0.120
Relief-F 70% 15.028± 0.558 0.165±0.075 0.300 ± 0.030 0.477 ± 0.030
50% 15.554± 0.550 0.104±0.098 0.303 ± 0.038 0.428 ± 0.070
30% 14.982± 1.245 0.162±0.162 0.294 ± 0.034 0.482 ± 0.102
10% 17.159± 1.662 −0.087±0.163 0.365 ± 0.039 0.175 ± 0.138

LightGBM Full 100% 14.560± 1.050 0.209±0.138 0.291 ± 0.027 0.506 ± 0.099
CFS 70% 14.437± 1.077 0.223±0.140 0.273 ± 0.038 0.527 ± 0.098
50% 14.149± 0.913 0.256±0.118 0.274 ± 0.042 0.556 ± 0.073
30% 14.221± 1.249 0.246±0.154 0.273 ± 0.038 0.530 ± 0.124
10% 14.699± 1.363 0.193±0.175 0.291 ± 0.046 0.489 ± 0.129
Relief-F 70% 14.876± 0.764 0.178±0.113 0.294 ± 0.023 0.474 ± 0.086
50% 15.150± 0.551 0.150±0.094 0.308 ± 0.024 0.450 ± 0.069
30% 15.360± 0.745 0.126±0.105 0.308 ± 0.024 0.427 ± 0.079
10% 15.396± 0.443 0.124±0.067 0.316 ± 0.025 0.386 ± 0.066

Values are presented as mean ± standard deviation.

Boldface indicates the best score of a given evaluation metric in each regression model.

CFS: correlation-based feature selection; RSME: root mean square error, MAPE: mean absolute percentage error, PCC: Pearson correlation coefficient.

Table 3
Results of feature importance
Model Feature subset Five most important features in order of importance

1 2 3 4 5
Random forest 100% L-P1_TI L-P2_TI L-P1_PAD R-ICA_d_mean L-P1_SOAM
70% L-P1_TI L-P2_PAD L-P2_TI L-P1_SOAM R-ICA_d_min
50% L-P1_TI R-ICA_d_min L-P2_RL L-P2_TI L-P1_PAD
30% L-P1_TI L-P2_PAD L-P2_TI L-P1_SOAM R-ICA_d_min
10% L-P1_TI L-P2_SOAM L-P1_SOAM R-ICA_d_min L-P2_RL

Linear regression 100% BA_TI L-A1_TI R-P2_TI R-P1_TI L-ICA_TI
70% R-P1_TI BA_TI L-PComm_TI R-M1_TI L-P1_TI
50% R-P1_TI BA_TI R-P2_TI R-M1_TI R-ICA_TI
30% R-PComm_d_min R-P1_TI L-M1_TI R-ICA_RL R-ICA_d_min
10% R-ICA_d_min L-P1_SOAM R-P1_d_min R-P1_d_50% R-P1_d_75%

AdaBoost 100% L-P1_TI L-P2_TI R-ICA_d_mean R-ICA_d_min L-P2_SOAM
70% L-P1_TI R-ICA_d_mean L-P2_TI L-P2_SOAM R-P1_d_mean
50% L-P1_TI L-P2_TI R-ICA_d_75% R-ICA_d_min L-P2_SOAM
30% L-P1_TI L-P2_TI R-ICA_d_75% R-ICA_d_min L-M1_RL
10% L-P1_TI R-ICA_d_min L-P2_RL R-P1_l L-P2_PAD

XGBoost 100% R-ICA_d_25% L-P2_TI L-P1_PAD R-ICA_d_75% L-P1_TI
70% R-ICA_d_75% L-P1_TI L-PComm_PAD L-P2_TI R-P2_d_50%
50% L-P2_TI R-P2_d_50% R-ICA_d_min L-P1_TI L-P1_PAD
30% L-P2_TI L-P1_TI L-M1_SOAM L-P1_PAD R-ICA_d_75%
10% L-P2_SOAM L-P1_PAD L-P2_PAD R-ICA_d_min L-P2_RL

LightGBM 100% R-PComm_l L-ICA_l L-ICA_RL L-P1_TI L-P2_TI
70% L-PComm_l R-A1_PAD R-PComm_l L-PComm_TI R-ICA_d_mean
50% R-PComm_l L-ICA_l L-P1_lc R-ICA_PAD L-A1_TI
30% R-ICA_d_mean L-ICA_RL R-ICA_d_75% L-M1_PAD L-P1_TI
10% R-ICA_d_mean L-M1_PAD L-P1_TI R-P1_l L-P2_SOAM

The correlation-based feature selection (CFS) method was used for feature subset selection.

TI: triangular index, PAD: product of angle distance, SOAM: sum of angles metric, RL: relative length, ICA: internal carotid artery, BA: basilar artery.

TOOLS
Similar articles