Journal List > Int J Thyroidol > v.16(1) > 1516082836

Cho, Lee, Lee, Lee, Yoon, Park, Rho, Yoon, and Kwak: Artificial Intelligence Computer-Assisted Diagnosis for Thyroid Nodules: Comparison of Diagnostic Performance Using Original and Mobile Ultrasonography Images

Abstract

Background and Objectives

This study investigated whether an artificial intelligence computer-assisted diagnosis (AI-CAD) software recently developed in our institution named the Severance Artificial intelligence program (SERA) could show similar diagnostic performance for thyroid cancers using ultrasonographic (US) images from a mobile phone (SERA_M) compared to using images directly downloaded from the pictures archive and communication system (PACS) (SERA_P).

Materials and Methods

From October 2019 to December 2019, 259 thyroid nodules from 259 patients were included. SERA was run on original and mobile images to evaluate SERA_P and SERA_M. Nodules were categorized according to the American College of Radiology Thyroid Imaging Reporting and Data System (ACR TI-RADS). To compare diagnostic performance, a logistic regression analysis was conducted using the Generalized Estimating Equation. The area under the curve (AUC) was calculated using the receiver operating characteristic (ROC) curve, and compared using the Delong Method.

Results

There were 40 cancers (15.4%) and 219 benign lesions (84.6%). The AUC and sensitivity of SERA_M (0.82 and 85%, respectively) were not statistically different from SERA_P (0.8 and 75%, respectively) (p=0.526 and p=0.091, respectively). The AUC of radiologists (0.856) was not significantly different compared to SERA_P and SERA_M (p=0.163 and p=0.414, respectively). The sensitivity of radiologists (77.5%) was not statistically different compared to SERA_P and SERA_M (p=0.739 and p=0.361, respectively).

Conclusion

AI-CAD software using pictures taken by a mobile phone showed comparable diagnostic performance with the same software using images directly from PACS.

Introduction

Recent advances in ultrasonography (US) technology and the higher number of thyroid and neck US performed for health examinations have increased the likelihood of thyroid nodules being detected.1) In real-time US, the accurate diagnosis of thyroid nodules requires the high concentration of physicians, and thus variations might occur in diagnosis due to several factors such as skill level and fatigue of specialists.2,3) With the rapid development of computer technology, artificial intelligence (AI)-based computer aided diagnosis (CAD) is impacting medical imaging diagnosis in diverse ways.4,5) In particular, convolutional neural network (CNN) technology enables multi-level self-education using medical images, and this can be applied to actual medical image analysis.6,7) CNN technology can yield accurate, objective diagnostic results, especially in cancer diagnosis, and assist doctors across various medical fields to increase efficiency.8,9) In the field of thyroid nodule diagnosis, CNN can diagnose and classify benign and malignant thyroid nodules on US images.10,11) Moreover, CNN can play an assistive role to improve the diagnostic performance of inexperienced physicians.12,13)
In our institution, we developed an AI-CAD software, Severance Artificial intelligence program (SERA), using 13,560 US images of thyroid nodules labeled as either benign or malignant.14) SERA can be run on the internet, making it possible to use it on both computers and mobile phones, thus raising user accessibility. When SERA is run on a computer, the user operates SERA with US images uploaded to the picture archiving and communication system (PACS). However, several issues can arise when downloading images from PACS elsewhere. Security issues forbid the direct linking of SERA with PACS images in in-hospital computers, and users may have to find ways to import US images to personal computers to operate SERA. Furthermore, local clinics without a connection to online PACS would have to use additional data transportation methods to access US images. Thus, alternative ways to access US images to operate SERA are needed.
Instead of the previous complex steps to access US images for SERA, it would be more convenient and accessible if users took a picture of the PACS monitor with their mobile phone and ran the image through SERA without saving it permanently on their device. However, pictures taken with a mobile phone may generate noise which can affect the resolution of US images and diagnostic ability of SERA. Therefore, in this study, we investigated whether SERA using pictures with a mobile phone would show similar diagnostic performance for thyroid cancers when compared SERA directly using images downloaded from PACS.

Materials and Methods

Ethics Statement

The Institutional Review Board approved this retrospective study and required neither patient approval nor informed consent for our review of patient images and records.

Study Population

From October 2019 to December 2019, 579 patients over the age of 19 who underwent US-guided fine-needle aspiration (US-FNA) on thyroid lesions in our institution were included. Among them, 211 patients who had thyroid nodules less than 1 cm in size were excluded. We also excluded 109 nodules in 109 patients who did not receive further management such as repeat FNA or surgery after nondiagnostic or indeterminate US-FNA results. Finally, 259 thyroid nodules in 259 patients were included (Fig. 1).

Acquisition of US Data

Five radiologists with 6 to 21 years of experience in thyroid imaging performed US and US-FNA, using EPIQ 7 (EPIQ 7, Phillips Medical Systems, Bothell, WA, USA) with a 7 to 17 MHz linear transducer. US images, which include composition, echogenicity, shape, margin, and calcification, were reviewed and analyzed by the radiology specialists before US-guided FNA.15) Composition was classified as solid, predominantly solid (solid component≥50%), or predominantly cystic (solid component<50%). Echogenecity was classified as hyperechoic (hyperechogenicity compared to the surrounding thyroid parenchyma) or isoechoic (isoechogenicity compared to the surrounding thyroid parenchyma), hypoechoic (hypoechogenicity compared to the surrounding thyroid parenchyma), and marked hypoechoic (hypoechogenicity compared to the strap muscles). Margin was classified as well-defined, microlobulated or irregular. Calcification was classified as negative, macrocalcifications, eggshell calcifications, or microcalcifications. Shape was classified as parallel or non-parallel (greater in the anteroposterior dimension than the transverse dimension, taller-than-wide shape). A representative US image of each thyroid nodule was selected by an experienced radiologist (J.Y.K.), and saved as a JPEG image in PACS. Afterwards, the same radiologist drew a rectangular region of interest including all thyroid nodules on the US image using the Windows 10 paint program. The US image of the thyroid nodules was displayed on a PACS monitor with a resolution of 1920*1080. Afterwards, the same radiologist used a mobile phone (Samsung Galaxy Note 8, 12 megapixel) to take a picture of the US image shown on the PACS monitor screen. Imaging of the PACS monitor was performed in a brightly lit room. Distance from the PACS monitor was set to about 20cm, and all other surrounding settings such as ambient light, background color, camera angle, exposure time and contrast value, were all manually controlled.16,17) In order to protect research ethics, all personal information and patient records that appears on the PACS monitor were excluded from the camera angle. All images were permanently discarded from the mobile phone after SERA diagnosis.

Experimental Design of CNN

SERA is a deep learning-based CAD that was trained with 13,560 US images of thyroid nodules that were either surgically confirmed or cytologically proven as benign or malignant by the Bethesda system and larger than 1 cm in size.14) The thyroid US images were collected at Severance Hospital from 2004 to 2019 and consisted of 7160 cases of malignant thyroid nodules and 6400 cases of benign thyroid nodules. To improve the training speed and accuracy of the transfer learning algorithm of the VGG16 network, each nodule was extracted from the thyroid image as a region of interest and irrelevant parts were removed. In addition, the extracted thyroid images were converted into a grayscale image and displayed as a value of 0 to 255. SERA analyzes the US image uploaded by the user and presents the malignancy probability of the image in continuous numbers between 0 and 100. Currently, SERA can be accessed through “http://seracse.yonsei.ac.kr” only by membership registration.

Data Analysis

Based on the review and analysis of US images, one radiologist (J.Y.K.) reclassified the US images into the American College of Radiology (ACR) Thyroid Imaging Reporting and Data System (TI-RADS, TR) 2 (not suspicious, 2 points), TR3 (mildly suspicious, 3 points), TR4 (moderately suspicious, 4-6 points), to TR5 (highly suspicious, 7 or more points).18) The same radiologist operated SERA using images directly downloaded from PACS and pictures taken with a mobile phone. Fig. 2 demonstrates each diagnostic process. The diagnostic results of SERA were defined as SERA_P and SERA_M depending on which images were analyzed. The malignancy rates from SERA were divided into categories of 2 to 5, according to the recommended malignancy rate ranges of ACR TI-RADS.

Statistical Analysis

Continuous variables were presented as means± standard deviations and categorical variables were presented as numbers (percentages). Youden’s index was used to determine the optimal cut-off point for SERA_P and SERA_M in the receiver operating characteristics (ROC) curve to maximize the sensitivity and specificity. To compare diagnostic performance such as sensitivity and specificity of SERA_P, SERA_M, and radiologists, a logistic regression analysis for clustered data (generalized estimated equation method) was performed. The area under the curve (AUC) was calculated using the ROC curve, and compared using the Delong Method.
For all statistical processing, SAS (version 9.4, SAS Inc., Cary, NC, USA) was used, and a two-sided test that resulted in a p-value of 0.05 or less was considered to have statistically significant results.

Results

Table 1 summarizes the baseline characteristics of the 259 patients that had 40 malignant and 219 benign ones. The average age of total patients, patients with malignant thyroid nodules, and patients with benign thyroid nodules was 50.59±13.26 years, 47±14.86 years, and 51.24±12.87 years, respectively. Neither gender nor age had a significant association with thyroid cancer (p=0.419 and p=0.063, respectively). The mean size of all, malignant, and benign thyroid nodules was 23.24±11.91 mm, 14.8±6.82 mm, and 24.78± 12.01 mm, respectively, indicating that benign thyroid nodules were significantly larger than malignant nodules (p<0.001). All US findings were significantly different between benign and malignant nodules (all p< 0.05) (Table 1). The mean value of SERA_P was 22.45±26.34% for benign nodules and 74.6±29.83% for malignant nodules (p<0.001). The mean value of SERA_M was 26.74±28.48% for benign nodules, and 76.41±27.62% for malignant nodules (p<0.001).
Malignancy risks calculated from SERA_P and SERA_M were categorized according to values recommended by ACR TI-RADS (Figs. 3, 4). According to the SERA_M results, the malignancy risk of TR2, TR3, TR4, and TR5 was 0% (0/39), 3.61% (3/83), 4.69% (3/64), and 42.5% (34/80), respectively. According to the SERA_P results, the malignancy risk of TR2, TR3, TR4, and TR5 was 0% (0/39), 3.26% (3/92), 10.77% (7/65), and 47.62% (30/63) respectively. According to the radiologists, the malignancy risk of TR2, TR3, TR4, TR5 was 0% (0/66), 0% (0/68), 13.64% (12/88) and 35.14 (13/37), respectively.
The optimal cutoff value was set at TR 4; nodules were considered malignant if the ACR TI-RADS category was 5 or greater. There was no difference between the AUC of SERA_P (0.8; 95% confidence interval [CI], 0.728 to 0.872) and SERA_M (0.82; 95% CI, 0.758 to 0.882) (p=0.526, Table 2, Fig. 5). The diagnostic sensitivity of SERA_P (75%; 95% CI, 61.58 to 88.42) and SERA_M (85%; 95% CI, 73.93 to 96.07) was not statistically different (p=0.091). SERA_P showed significantly higher diagnostic specificity (84.93%; 95% CI, 80.19 to 89.67) compared to SERA_M (79%; 95% CI, 73.6 to 84.39) (p=0.008). The AUC value of radiologists was 0.856 (95% CI; 0.788-0.923), which was not statistically different compared to SERA_P (0.8), and SERA_M (0.82) (p=0.163 and p=0.414, respectively). Radiologists showed a sensitivity of 77.5% (95% CI; 64.56-90.44) and specificity of 93.61% (95% CI; 90.37-96.85). Radiologists showed higher diagnostic specificity (93.61%) compared to SERA_P (84.93%) and SERA_M (79%) (p=0.001 and p<0.001, respectively). However, sensitivity was not statistically different between radiologists and SERA_P (77.5% vs. 75%; p=0.739), and between radiologists and SERA_M (77.5% vs. 85%; p=0.361) (Table 2).

Discussion

In this study, AUC and sensitivity of SERA_M were not significantly different to those of SERA_P. This suggests that pictures of original US images taken with a mobile phone can be used for AI-CAD in daily practice. Moreover, further technical support available for phone cameras, or external equipment may compensate for poor picture quality to increase the diagnostic ability of SERA_M.
Real-time US is the main diagnostic tool for thyroid nodules, but its diagnostic performance depends on the user’s skill.19,20) Meanwhile, AI-CAD is known to be more objective than human readers, and various studies have shown that its diagnostic performance is even comparable to those of experienced physicians.14,21) However, previous research was all conducted with US images stored on computers, and thus the images analyzed were of adequate quality and resolution for AI-CAD diagnosis.22) For better accessibility and user convenience in some facilities, pictures taken with a mobile phone can be a good alternative because the process does not require clinicians to undertake additional steps to access the PACS database to run SERA.
Rapid developments in the resolution and performance of mobile phone cameras have encouraged attempts to use pictures taken by mobile phones for medical use.23) Studies have shown that the diagnostic performance of physicians interpreting X-ray or US images transmitted by mobile phones can be similar to that of physicians interpreting original images.24,25) However, due to the imperfections and aberration of mobile phone lenses, there is a limitation to collecting pictures free of diffraction without additional devices and equipment.26) Using a mobile phone to take pictures may generate noise, depending on environmental factors, such as the distance from the monitor and lighting.17,27) Thus, the process of taking pictures of the PACS monitor would inevitably generate blurring caused by light diffraction, resulting in obstacles that would lower the diagnostic abilities of SERA compared to when SERA is run on original US images.28) Meanwhile, previous studies have shown that AI-CAD using pictures of X-rays taken by mobile phones can show excellent performance when detecting cardiac devices.29) Likewise, there is a possibility that even with lower resolution or more artifacts, US images taken by mobile phones may be of sufficient quality to achieve respectable diagnostic performance when used with AI-CAD. Therefore, our study aimed to evaluate whether SERA using pictures of lower quality taken by mobile phones could achieve comparable diagnostic value to SERA using original US images downloaded from PACS. Compared to SERA_P and radiologists, SERA_M showed satisfactory AUC and sensitivity. However, SERA_M showed the lowest specificity, followed by SERA_P and radiologists. This indicates that SERA still needs fine-tuning to improve its diagnostic performance. Based on the suggested risk ranges of ACR TI-RADS, categorization using SERA_P results showed satisfactory malignancy rates. The malignancy rates of SERA_M were generally within the suggested risk ranges of ACR TI-RADS, except for TR3 (4.69%). This suggests that SERA, even with pictures taken by mobile phones, has decent categorization abilities similar to ACR TI-RADS. Therefore, pictures taken by a mobile phone can be used to categorize nodules.
There are several limitations to our study. First, our study was a retrospective study conducted with data collected from a referral center. Thus, a selection bias was unavoidable with a high malignancy rate (15.4%) for the test set of thyroid nodules. Selection bias may have also arisen because we only included nodules that underwent surgery or FNA, and we also excluded patients with indeterminate results lost to follow-up. Second, our study population was not large with a total of 259 nodules after exclusion, with 40 malignant and 219 benign nodules. Our findings need to be confirmed with a larger study population. Third, while radiologists prospectively obtained the individual US features from thyroid US, only one radiologist assigned final assessments according to ACR TI-RADS through retrospective review. Also, we did not consider the differences in how the US descriptors were defined compared to ACR TI-RADS in our data analysis. If we had analyzed final assessments during real-time examinations, the results might have been different. Fourth, our study was conducted using a single mobile phone. Different models or camera settings may have affected the quality of images, resulting in deviation of diagnostic results. Fifth, this study was conducted in one tertiary hospital with high-quality US machines. When SERA is used in local hospitals with US images of lower quality, the performance of both SERA_P and SERA_M might be disturbed. Further assessment or validation of our findings should be conducted at an external institution. Finally, we only included thyroid nodules that underwent US-FNA but not surgery, and consequently, did not include surgical histopathology in our analysis. Thus, there is the possibility of false- negative or false-positive results, although those rates are expected to be very low as previous literature reports a false-negative rate of 3% and false-positive rates around 3-4%.30)
In conclusion, SERA performed with pictures taken by a mobile phone showed a comparable diagnostic performance to SERA performed with images directly downloaded from PACS. Using the results of this study, we hope that further research on operating AI-CAD programs on mobile devices can make AI-CAD more convenient and accessible to users.

Acknowledgments

This study was supported by the National Research Foundation of Korea (NRF) grant funded by the Korean government (MSIT) (2021R1A2C2007492). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

Notes

Conflicts of Interest

No potential conflict of interest relevant to this article was reported.

References

1. Bray F, Ferlay J, Soerjomataram I, Siegel RL, Torre LA, Jemal A. Global cancer statistics 2018: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J Clin. 2018; 68(6):394–424. DOI: 10.3322/caac.21492. PMID: 30207593.
crossref
2. Choi SH, Kim EK, Kwak JY, Kim MJ, Son EJ. 2010; Interobserver and intraobserver variations in ultrasound assessment of thyroid nodules. Thyroid. 20(2):167–72. DOI: 10.1089/thy.2008.0354. PMID: 19725777.
crossref
3. Koh J, Kim SY, Lee HS, Kim EK, Kwak JY, Moon HJ, et al. 2018; Diagnostic performances and interobserver agreement according to observer experience: a comparison study using three guidelines for management of thyroid nodules. Acta Radiol. 59(8):917–23. DOI: 10.1177/0284185117744001. PMID: 29181988.
4. Gurcan MN, Boucheron LE, Can A, Madabhushi A, Rajpoot NM, Yener B. 2009; Histopathological image analysis: a review. IEEE Rev Biomed Eng. 2:147–71. DOI: 10.1109/RBME.2009.2034865. PMID: 20671804. PMCID: PMC2910932.
crossref
5. Suzuki K. 2017; Overview of deep learning in medical imaging. Radiol Phys Technol. 10(3):257–73. DOI: 10.1007/s12194-017-0406-5. PMID: 28689314.
crossref
6. Litjens G, Sanchez CI, Timofeeva N, Hermsen M, Nagtegaal I, Kovacs I, et al. 2016; Deep learning as a tool for increased accuracy and efficiency of histopathological diagnosis. Sci Rep. 6:26286. DOI: 10.1038/srep26286. PMID: 27212078. PMCID: PMC4876324.
crossref
7. Liang X, Yu J, Liao J, Chen Z. 2020; Convolutional neural network for breast and thyroid nodules diagnosis in ultrasound imaging. Biomed Res Int. 2020:1763803. DOI: 10.1155/2020/1763803. PMID: 32420322. PMCID: PMC7199615.
crossref
8. Esteva A, Kuprel B, Novoa RA, Ko J, Swetter SM, Blau HM, et al. 2017; Dermatologist-level classification of skin cancer with deep neural networks. Nature. 542(7639):115–8. DOI: 10.1038/nature21056. PMID: 28117445. PMCID: PMC8382232.
crossref
9. Vandenberghe ME, Scott ML, Scorer PW, Soderberg M, Balcerzak D, Barker C. 2017; Relevance of deep learning to facilitate the diagnosis of HER2 status in breast cancer. Sci Rep. 7:45938. DOI: 10.1038/srep45938. PMID: 28378829. PMCID: PMC5380996.
crossref
10. Zhang X, Lee VC, Rong J, Lee JC, Liu F. 2022; Deep convolutional neural networks in thyroid disease detection: a multi-classification comparison by ultrasonography and computed tomography. Comput Methods Programs Biomed. 220:106823. DOI: 10.1016/j.cmpb.2022.106823. PMID: 35489145.
crossref
11. Chi J, Walia E, Babyn P, Wang J, Groot G, Eramian M. 2017; Thyroid nodule classification in ultrasound images by fine-tuning deep convolutional neural network. J Digit Imaging. 30(4):477–86. DOI: 10.1007/s10278-017-9997-y. PMID: 28695342. PMCID: PMC5537102.
crossref
12. Jin Z, Zhu Y, Zhang S, Xie F, Zhang M, Zhang Y, et al. 2020; Ultrasound computer-aided diagnosis (CAD) based on the Thyroid Imaging Reporting and Data System (TI-RADS) to distinguish benign from malignant thyroid nodules and the diagnostic performance of radiologists with different diagnostic experience. Med Sci Monit. 26:e918452. DOI: 10.12659/MSM.918452.
crossref
13. Youn I, Lee E, Yoon JH, Lee HS, Kwon MR, Moon J, et al. 2021; Diagnosing thyroid nodules with atypia of undetermined significance/follicular lesion of undetermined significance cytology with the deep convolutional neural network. Sci Rep. 11(1):20048. DOI: 10.1038/s41598-021-99622-0. PMID: 34625636. PMCID: PMC8501016.
crossref
14. Koh J, Lee E, Han K, Kim EK, Son EJ, Sohn YM, et al. 2020; Diagnosis of thyroid nodules on ultrasonography by a deep convolutional neural network. Sci Rep. 10(1):15245. DOI: 10.1038/s41598-020-72270-6. PMID: 32943696. PMCID: PMC7498581.
15. Kwak JY, Han KH, Yoon JH, Moon HJ, Son EJ, Park SH, et al. 2011; Thyroid Imaging Reporting and Data System for US features of nodules: a step in establishing better stratification of cancer risk. Radiology. 260(3):892–9. DOI: 10.1148/radiol.11110206. PMID: 21771959.
crossref
16. Quesada-González D, Merkoçi A. 2017; Mobile phone-based biosensing: an emerging "diagnostic and communication" technology. Biosens Bioelectron. 92:549–62. DOI: 10.1016/j.bios.2016.10.062.
crossref
17. Coleman B, Coarsey C, Asghar W. 2019; Cell phone based colorimetric analysis for point-of-care settings. Analyst. 144(6):1935–47. DOI: 10.1039/C8AN02521E. PMID: 30688961. PMCID: PMC6411445.
18. Tessler FN, Middleton WD, Grant EG, Hoang JK, Berland LL, Teefey SA, et al. 2017; ACR Thyroid Imaging, Reporting and Data System (TI-RADS): white paper of the ACR TI-RADS Committee. J Am Coll Radiol. 14(5):587–95. DOI: 10.1016/j.jacr.2017.01.046. PMID: 28372962.
crossref
19. Fujimoto Y, Oka A, Omoto R, Hirose M. 1967; Ultrasound scanning of the thyroid gland as a new diagnostic approach. Ultrasonics. 5:177–80. DOI: 10.1016/S0041-624X(67)80065-9. PMID: 6053923.
crossref
20. Gallo M, Pesenti M, Valcavi R. 2003; Ultrasound thyroid nodule measurements: the "gold standard" and its limitations in clinical decision making. Endocr Pract. 9(3):194–9. DOI: 10.4158/ep.9.3.194. PMID: 12917060.
crossref
21. Kim GR, Lee E, Kim HR, Yoon JH, Park VY, Kwak JY. 2021; Convolutional neural network to stratify the malignancy risk of thyroid nodules: diagnostic performance compared with the american college of radiology Thyroid Imaging Reporting and Data System implemented by experienced radiologists. AJNR Am J Neuroradiol. 42(8):1513–9. DOI: 10.3174/ajnr.A7149. PMID: 33985947. PMCID: PMC8367605.
22. Choi YJ, Baek JH, Park HS, Shim WH, Kim TY, Shong YK, et al. 2017; A computer-aided diagnosis system using artificial intelligence for the diagnosis and characterization of thyroid nodules on ultrasound: initial clinical assessment. Thyroid. 27(4):546–52. DOI: 10.1089/thy.2016.0372. PMID: 28071987.
crossref
23. Fawzy HED. 2015; The accuracy of mobile phone camera instead of high resolution camera in digital close range photogrammetry. Int J Civ Eng Technol. 6(1):76–85.
24. Blaivas M, Lyon M, Duggal S. 2005; Ultrasound image transmission via camera phones for overreading. Am J Emerg Med. 23(4):433–8. DOI: 10.1016/j.ajem.2004.09.037. PMID: 16032606.
25. Goost H, Witten J, Heck A, Hadizadeh DR, Weber O, Graff I, et al. 2012; Image and diagnosis quality of X-ray image transmission via cell phone camera: a project study evaluating quality and reliability. PLoS One. 7(10):e43402. DOI: 10.1371/journal.pone.0043402. PMID: 23082108. PMCID: PMC3474770.
crossref
26. Breslauer DN, Maamari RN, Switz NA, Lam WA, Fletcher DA. 2009; Mobile phone based clinical microscopy for global health applications. PLoS One. 4(7):e6320. DOI: 10.1371/journal.pone.0006320. PMID: 19623251. PMCID: PMC2709430.
crossref
27. Luo LB, Chong JW. 2012; Real‐time digital image stabilization for cell phone cameras in low‐light environments without frame memory. ETRI Journal. 34(1):138–41. DOI: 10.4218/etrij.12.0211.0338.
crossref
28. Gampala V, Kumar MS, Sushama C, Raj EFI. 2020. Deep learning based image processing approaches for image deblurring. Proceedings;Materials Today:
29. Weinreich M, Chudow JJ, Weinreich B, Krumerman T, Nag T, Rahgozar K, et al. 2019; Development of an artificially intelligent mobile phone application to identify cardiac devices on chest radiography. JACC Clin Electrophysiol. 5(9):1094–5. DOI: 10.1016/j.jacep.2019.05.013. PMID: 31537342.
crossref
30. Cibas ES, Ali SZ. 2017; The 2017 Bethesda system for reporting thyroid cytopathology. Thyroid. 27(11):1341–6. DOI: 10.1089/thy.2017.0500. PMID: 29091573.
crossref

Fig. 1
Flowchart of study population. AUS: atypia of undermined significance, FLUS: follicular lesion of undetermined significance, FNA: fine needle aspiration
ijt-16-1-111-f1.tif
Fig. 2
Diagram of overall process of image acquisition and diagnosis. ACR TI-RADS: American College of Radiology Thyroid Imaging Reporting and Data System
ijt-16-1-111-f2.tif
Fig. 3
Ultrasonography images and SERA results of a 71-year-old man with pathologically proven papillary thyroid cancer. The lesion was assessed as ACR TI-RADS category 5 by a radiologist. Compared with the original PACS image (A), the picture taken with a mobile phone (B) was blurry and had lower resolution, but the diagnostic results of SERA were both malignant (A, B, respectively). ACR TI-RADS: American College of Radiology Thyroid Imaging Reporting and Data System, PACS: picture archiving and communication system, SER: Severance Artificial intelligence program
ijt-16-1-111-f3.tif
Fig. 4
Ultrasonography images and SERA diagnostic results of a 58-year-old man with a FNA-proven benign nodule. The lesion was assessed as ACR TI-RADS category 3 nodule by a radiologist. Compared with the original PACS image (A), the picture taken with a mobile phone (B) was brighter and had lower resolution, but the diagnostic results of SERA were both benign (A, B, respectively). ACR TI-RADS: American College of Radiology Thyroid Imaging Reporting and Data System, PACS: picture archiving and communication system, SERA: Severance Artificial intelligence program
ijt-16-1-111-f4.tif
Fig. 5
Area under the ROC curve of SERA_P, SERA_M, and radiologists for differentiating thyroid nodules. ROC curve: receiver operating characteristic curve, SERA: Severance Artificial intelligence program, SERA_M: results from SERA using images taken with a mobile phone, SERA_P: results from SERA using images directly downloaded from PACS
ijt-16-1-111-f5.tif
Table 1
Baseline characteristics of the study cohort
Total
(n=259)
Malignant
(n=40)
Benign
(n=219)
p-value
Sex 0.419
Women 201 (77.61%) 33 (82.5%) 168 (76.71%)
Men 58 (22.39%) 7 (17.5%) 51 (23.29%)
Age (years) 50.59±13.26 47±14.86 51.24±12.87 0.063
Nodule size (mm) 23.24±11.92 14.8±6.82 24.78±12.01 <0.001
US features
Composition 0.001
Solid 167 (64.48%) 36 (90%) 131 (59.82%)
Predominantly solid 72 (27.8%) 4 (10%) 68 (31.05%)
Predominantly cystic 20 (7.72%) 0 (0%) 20 (9.13%)
Echogenicity <0.001
Hyperechoic or isoechoic 151 (58.3%) 3 (7.5%) 148 (67.58%)
Hypoechoic 103 (39.77%) 33 (82.5%) 70 (31.96%)
Marked hypoechoic 5 (1.93%) 4 (10%) 1 (0.46%)
Margin <0.001
Well-defined 215 (83.01%) 13 (32.5%) 202 (92.24%)
Microlobulated or irregular 44 (16.99%) 27 (67.5%) 17 (7.76%)
Calcification <0.001
Negative 190 (73.36%) 14 (35%) 176 (80.37%)
Macrocalcifications 31 (11.97%) 3 (7.5%) 28 (12.79%)
Eggshell calcifications 7 (2.7%) 1 (2.5%) 6 (2.74%)
Microcalcifications 31 (11.97%) 22 (55%) 9 (4.11%)
Shape <0.001
Parallel 238 (91.9%) 26 (65%) 212 (96.8%)
Non-parallel 21 (8.1%) 14 (35%) 7 (3.2%)
Table 2
Comparison of diagnostic performances for thyroid cancers among SERA_P, SERA_M, and radiologists
SERA_P SERA_M Radiologists p-value* p-value p-value
Sensitivity 75 (61.58-88.42) 85 (73.93-96.07) 77.5 (64.56-90.44) 0.091 0.739 0.361
Specificity 84.93 (80.19-89.67) 79 (73.6-84.39) 93.61 (90.37-96.85) 0.008 0.001 <0.001
AUC 0.8 (0.728-0.872) 0.82 (0.758-0.882) 0.856 (0.788-0.923) 0.526 0.163 0.414

The number in parentheses indicates the 95% confidence interval.

AUC: area under the receiver operating characteristic curve, SERA: Severance Artificial intelligence program, SERA_M: results from SERA using images taken from a mobile phone, SERA_P: results from SERA using images directly downloaded from PACS

*p-value for comparing SERA_P and SERA_M. p-value for comparing SERA_P and radiologists. p-value for comparing SERA_M and radiologists.

TOOLS
Similar articles