1.Diagnostic performance of a computer-aided system for tuberculosis screening in two Philippine cities.
Gabrielle P. FLORES ; Reiner Lorenzo J. TAMAYO ; Robert Neil F. LEONG ; Christian Sergio M. BIGLAEN ; Kathleen Nicole T. UY ; Renee Rose O. MAGLENTE ; Marlex Jorome M. NUGUID ; Jason V. ALACAP
Acta Medica Philippina 2025;59(2):33-40
BACKGROUND AND OBJECTIVES
The Philippines faces challenges in the screening of tuberculosis (TB), one of them being the shortage in the health workforce who are skilled and allowed to screen TB. Deep learning neural networks (DLNNs) have shown potential in the TB screening process utilizing chest radiographs (CXRs). However, local studies on AIbased TB screening are limited. This study evaluated qXR3.0 technology's diagnostic performance for TB screening in Filipino adults aged 15 and older. Specifically, we evaluated the specificity and sensitivity of qXR3.0 compared to radiologists' impressions and determined whether it meets the World Health Organization (WHO) standards.
METHODSA prospective cohort design was used to perform a study on comparing screening and diagnostic accuracies of qXR3.0 and two radiologist gradings in accordance with the Standards for Reporting Diagnostic Accuracy (STARD). Subjects from two clinics in Metro Manila which had qXR 3.0 seeking consultation at the time of study were invited to participate to have CXRs and sputum collected. Radiologists' and qXR3.0 readings and impressions were compared with respect to the reference standard Xpert MTB/RiF assay. Diagnostic accuracy measures were calculated.
RESULTSWith 82 participants, qXR3.0 demonstrated 100% sensitivity and 72.7% specificity with respect to the reference standard. There was a strong agreement between qXR3.0 and radiologists' readings as exhibited by the 0.7895 (between qXR 3.0 and CXRs read by at least one radiologist), 0.9362 (qXR 3.0 and CXRs read by both radiologists), and 0.9403 (qXR 3.0 and CXRs read as not suggestive of TB by at least one radiologist) concordance indices.
CONCLUSIONSqXR3.0 demonstrated high sensitivity to identify presence of TB among patients, and meets the WHO standard of at least 70% specificity for detecting true TB infection. This shows an immense potential for the tool to supplement the shortage of radiologists for TB screening in the country. Future research directions may consider larger sample sizes to confirm these findings and explore the economic value of mainstream adoption of qXR 3.0 for TB screening.
Human ; Tuberculosis ; Diagnostic Imaging ; Deep Learning
2.Deep learning-based radiomics allows for a more accurate assessment of sarcopenia as a prognostic factor in hepatocellular carcinoma.
Zhikun LIU ; Yichao WU ; Abid Ali KHAN ; L U LUN ; Jianguo WANG ; Jun CHEN ; Ningyang JIA ; Shusen ZHENG ; Xiao XU
Journal of Zhejiang University. Science. B 2024;25(1):83-90
Hepatocellular carcinoma (HCC) is one of the most common malignancies and is a major cause of cancer-related mortalities worldwide (Forner et al., 2018; He et al., 2023). Sarcopenia is a syndrome characterized by an accelerated loss of skeletal muscle (SM) mass that may be age-related or the result of malnutrition in cancer patients (Cruz-Jentoft and Sayer, 2019). Preoperative sarcopenia in HCC patients treated with hepatectomy or liver transplantation is an independent risk factor for poor survival (Voron et al., 2015; van Vugt et al., 2016). Previous studies have used various criteria to define sarcopenia, including muscle area and density. However, the lack of standardized diagnostic methods for sarcopenia limits their clinical use. In 2018, the European Working Group on Sarcopenia in Older People (EWGSOP) renewed a consensus on the definition of sarcopenia: low muscle strength, loss of muscle quantity, and poor physical performance (Cruz-Jentoft et al., 2019). Radiological imaging-based measurement of muscle quantity or mass is most commonly used to evaluate the degree of sarcopenia. The gold standard is to measure the SM and/or psoas muscle (PM) area using abdominal computed tomography (CT) at the third lumbar vertebra (L3), as it is linearly correlated to whole-body SM mass (van Vugt et al., 2016). According to a "North American Expert Opinion Statement on Sarcopenia," SM index (SMI) is the preferred measure of sarcopenia (Carey et al., 2019). The variability between morphometric muscle indexes revealed that they have different clinical relevance and are generally not applicable to broader populations (Esser et al., 2019).
Humans
;
Aged
;
Sarcopenia/diagnostic imaging*
;
Carcinoma, Hepatocellular/diagnostic imaging*
;
Muscle, Skeletal/diagnostic imaging*
;
Deep Learning
;
Prognosis
;
Radiomics
;
Liver Neoplasms/diagnostic imaging*
;
Retrospective Studies
3.Diagnostic performance of a computer-aided system for tuberculosis screening in two Philippine cities
Gabrielle P. Flores ; Reiner Lorenzo J. Tamayo ; Robert Neil F. Leong ; Christian Sergio M. Biglaen ; Kathleen Nicole T. Uy ; Renee Rose O. Maglente ; Marlex Jorome M. Nugui ; Jason V. Alacap
Acta Medica Philippina 2024;58(Early Access 2024):1-8
Background and Objectives:
The Philippines faces challenges in the screening of tuberculosis (TB), one of them being the shortage in the health workforce who are skilled and allowed to screen TB. Deep learning neural networks (DLNNs) have shown potential in the TB screening process utilizing chest radiographs (CXRs). However, local studies on AIbased TB screening are limited. This study evaluated qXR3.0 technology's diagnostic performance for TB screening in Filipino adults aged 15 and older. Specifically, we evaluated the specificity and sensitivity of qXR3.0 compared to radiologists' impressions and determined whether it meets the World Health Organization (WHO) standards.
Methods:
A prospective cohort design was used to perform a study on comparing screening and diagnostic accuracies of qXR3.0 and two radiologist gradings in accordance with the Standards for Reporting Diagnostic Accuracy (STARD). Subjects from two clinics in Metro Manila which had qXR 3.0 seeking consultation at the time of study were invited to participate to have CXRs and sputum collected. Radiologists' and qXR3.0 readings and impressions were compared with respect to the reference standard Xpert MTB/RiF assay. Diagnostic accuracy measures were calculated.
Results:
With 82 participants, qXR3.0 demonstrated 100% sensitivity and 72.7% specificity with respect to the
reference standard. There was a strong agreement between qXR3.0 and radiologists' readings as exhibited by
the 0.7895 (between qXR 3.0 and CXRs read by at least one radiologist), 0.9362 (qXR 3.0 and CXRs read by both
radiologists), and 0.9403 (qXR 3.0 and CXRs read as not suggestive of TB by at least one radiologist) concordance indices.
Conclusions
qXR3.0 demonstrated high sensitivity to identify presence of TB among patients, and meets the WHO standard of at least 70% specificity for detecting true TB infection. This shows an immense potential for the tool to supplement the shortage of radiologists for TB screening in the country. Future research directions may consider larger sample sizes to confirm these findings and explore the economic value of mainstream adoption of qXR 3.0 for TB screening.
Tuberculosis
;
Diagnostic Imaging
;
Deep Learning
4.The impact of anatomic racial variations on artificial intelligence analysis of Filipino retinal fundus photographs using an image-based deep learning model
Carlo A. Kasala ; Kaye Lani Rea B. Locaylocay ; Paolo S. Silva
Philippine Journal of Ophthalmology 2024;49(2):130-137
OBJECTIVES
This study evaluated the accuracy of an artificial intelligence (AI) model in identifying retinal lesions, validated its performance on a Filipino population dataset, and evaluated the impact of dataset diversity on AI analysis accuracy.
METHODSThis cross-sectional, analytical, institutional study analyzed standardized macula-centered fundus photos taken with the Zeiss Visucam®. The AI model’s output was compared with manual readings by trained retina specialists.
RESULTSA total of 215 eyes from 109 patients were included in the study. Human graders identified 109 eyes (50.7%) with retinal abnormalities. The AI model demonstrated an overall accuracy of 73.0% (95% CI 66.6% – 78.8%) in detecting abnormal retinas, with a sensitivity of 54.1% (95% CI 44.3% – 63.7%) and specificity of 92.5% (95% CI 85.7% – 96.7%).
CONCLUSIONThe availability and sources of AI training datasets can introduce biases into AI algorithms. In our dataset, racial differences in retinal morphology, such as differences in retinal pigmentation, affected the accuracy of AI image-based analysis. More diverse datasets and external validation on different populations are needed to mitigate these biases.
Human ; Artificial Intelligence ; Deep Learning
5.SPECT-MPI for Coronary Artery Disease: A deep learning approach
Vincent Peter C. Magboo ; Ma. Sheila A. Magboo
Acta Medica Philippina 2024;58(8):67-75
Background:
Worldwide, coronary artery disease (CAD) is a leading cause of mortality and morbidity and remains to be a top health priority in many countries. A non-invasive imaging modality for diagnosis of CAD such as single photon emission computed tomography-myocardial perfusion imaging (SPECT-MPI) is usually requested by cardiologists as it displays radiotracer distribution in the heart reflecting myocardial perfusion. The interpretation of SPECT-MPI is done visually by a nuclear medicine physician and is largely dependent on his clinical experience and showing significant inter-observer variability.
Objective:
The aim of the study is to apply a deep learning approach in the classification of SPECT-MPI for perfusion abnormalities using convolutional neural networks (CNN).
Methods:
A publicly available anonymized SPECT-MPI from a machine learning repository (https://www.kaggle.com/ selcankaplan/spect-mpi) was used in this study involving 192 patients who underwent stress-test-rest Tc99m MPI. An exploratory approach of CNN hyperparameter selection to search for optimum neural network model was utilized with particular focus on various dropouts (0.2, 0.5, 0.7), batch sizes (8, 16, 32, 64), and number of dense nodes (32, 64, 128, 256). The base CNN model was also compared with the commonly used pre-trained CNNs in medical images such as VGG16, InceptionV3, DenseNet121 and ResNet50. All simulations experiments were performed in Kaggle using TensorFlow 2.6.0., Keras 2.6.0, and Python language 3.7.10.
Results:
The best performing base CNN model with parameters consisting of 0.7 dropout, batch size 8, and 32 dense nodes generated the highest normalized Matthews Correlation Coefficient at 0.909 and obtained 93.75% accuracy, 96.00% sensitivity, 96.00% precision, and 96.00% F1-score. It also obtained higher classification performance as compared to the pre-trained architectures.
Conclusions
The results suggest that deep learning approaches through the use of CNN models can be deployed by nuclear medicine physicians in their clinical practice to further augment their decision skills in the interpretation of SPECT-MPI tests. These CNN models can also be used as a dependable and valid second opinion that can aid physicians as a decision-support tool as well as serve as teaching or learning materials for the less-experienced physicians particularly those still in their training career. These highlights the clinical utility of deep learning approaches through CNN models in the practice of nuclear cardiology.
Coronary Artery Disease
;
Deep Learning
6.Deep learning method for magnetic resonance imaging fluid-attenuated inversion recovery image synthesis.
Jianing ZHOU ; Hongyu GUO ; Hong CHEN
Journal of Biomedical Engineering 2023;40(5):903-911
Magnetic resonance imaging(MRI) can obtain multi-modal images with different contrast, which provides rich information for clinical diagnosis. However, some contrast images are not scanned or the quality of the acquired images cannot meet the diagnostic requirements due to the difficulty of patient's cooperation or the limitation of scanning conditions. Image synthesis techniques have become a method to compensate for such image deficiencies. In recent years, deep learning has been widely used in the field of MRI synthesis. In this paper, a synthesis network based on multi-modal fusion is proposed, which firstly uses a feature encoder to encode the features of multiple unimodal images separately, and then fuses the features of different modal images through a feature fusion module, and finally generates the target modal image. The similarity measure between the target image and the predicted image in the network is improved by introducing a dynamic weighted combined loss function based on the spatial domain and K-space domain. After experimental validation and quantitative comparison, the multi-modal fusion deep learning network proposed in this paper can effectively synthesize high-quality MRI fluid-attenuated inversion recovery (FLAIR) images. In summary, the method proposed in this paper can reduce MRI scanning time of the patient, as well as solve the clinical problem of missing FLAIR images or image quality that is difficult to meet diagnostic requirements.
Humans
;
Deep Learning
;
Magnetic Resonance Imaging/methods*
;
Image Processing, Computer-Assisted/methods*
7.Review on ultrasonographic diagnosis of thyroid diseases based on deep learning.
Fengyuan QI ; Min QIU ; Guohui WEI
Journal of Biomedical Engineering 2023;40(5):1027-1032
In recent years, the incidence of thyroid diseases has increased significantly and ultrasound examination is the first choice for the diagnosis of thyroid diseases. At the same time, the level of medical image analysis based on deep learning has been rapidly improved. Ultrasonic image analysis has made a series of milestone breakthroughs, and deep learning algorithms have shown strong performance in the field of medical image segmentation and classification. This article first elaborates on the application of deep learning algorithms in thyroid ultrasound image segmentation, feature extraction, and classification differentiation. Secondly, it summarizes the algorithms for deep learning processing multimodal ultrasound images. Finally, it points out the problems in thyroid ultrasound image diagnosis at the current stage and looks forward to future development directions. This study can promote the application of deep learning in clinical ultrasound image diagnosis of thyroid, and provide reference for doctors to diagnose thyroid disease.
Humans
;
Algorithms
;
Deep Learning
;
Image Processing, Computer-Assisted/methods*
;
Thyroid Diseases/diagnostic imaging*
;
Ultrasonography
8.Platelet RNA signature independently predicts ovarian cancer prognosis by deep learning neural network model.
Chun-Jie LIU ; Hua-Yi LI ; Yue GAO ; Gui-Yan XIE ; Jian-Hua CHI ; Gui-Ling LI ; Shao-Qing ZENG ; Xiao-Ming XIONG ; Jia-Hao LIU ; Lin-Li SHI ; Xiong LI ; Xiao-Dong CHENG ; Kun SONG ; Ding MA ; An-Yuan GUO ; Qing-Lei GAO
Protein & Cell 2023;14(8):618-622
9.Automated Classification of Inherited Retinal Diseases in Optical Coherence Tomography Images Using Few-shot Learning.
Qi ZHAO ; Si Wei MAI ; Qian LI ; Guan Chong HUANG ; Ming Chen GAO ; Wen Li YANG ; Ge WANG ; Ya MA ; Lei LI ; Xiao Yan PENG
Biomedical and Environmental Sciences 2023;36(5):431-440
OBJECTIVE:
To develop a few-shot learning (FSL) approach for classifying optical coherence tomography (OCT) images in patients with inherited retinal disorders (IRDs).
METHODS:
In this study, an FSL model based on a student-teacher learning framework was designed to classify images. 2,317 images from 189 participants were included. Of these, 1,126 images revealed IRDs, 533 were normal samples, and 658 were control samples.
RESULTS:
The FSL model achieved a total accuracy of 0.974-0.983, total sensitivity of 0.934-0.957, total specificity of 0.984-0.990, and total F1 score of 0.935-0.957, which were superior to the total accuracy of the baseline model of 0.943-0.954, total sensitivity of 0.866-0.886, total specificity of 0.962-0.971, and total F1 score of 0.859-0.885. The performance of most subclassifications also exhibited advantages. Moreover, the FSL model had a higher area under curves (AUC) of the receiver operating characteristic (ROC) curves in most subclassifications.
CONCLUSION
This study demonstrates the effective use of the FSL model for the classification of OCT images from patients with IRDs, normal, and control participants with a smaller volume of data. The general principle and similar network architectures can also be applied to other retinal diseases with a low prevalence.
Humans
;
Tomography, Optical Coherence
;
Deep Learning
;
Retinal Diseases/diagnostic imaging*
;
Retina/diagnostic imaging*
;
ROC Curve
10.Prediction of pulp exposure risk of carious pulpitis based on deep learning.
Li WANG ; Fei WU ; Mo XIAO ; Yu-Xin CHEN ; Ligeng WU
West China Journal of Stomatology 2023;41(2):218-224
OBJECTIVES:
This study aims to predict the risk of deep caries exposure in radiographic images based on the convolutional neural network model, compare the prediction results of the network model with those of senior dentists, evaluate the performance of the model for teaching and training stomatological students and young dentists, and assist dentists to clarify treatment plans and conduct good doctor-patient communication before surgery.
METHODS:
A total of 206 cases of pulpitis caused by deep caries were selected from the Department of Stomatological Hospital of Tianjin Medical University from 2019 to 2022. According to the inclusion and exclusion criteria, 104 cases of pulpitis were exposed during the decaying preparation period and 102 cases of pulpitis were not exposed. The 206 radiographic images collected were randomly divided into three groups according to the proportion: 126 radiographic images in the training set, 40 radiographic images in the validation set, and 40 radiographic images in the test set. Three convolutional neural networks, visual geometry group network (VGG), residual network (ResNet), and dense convolutional network (DenseNet) were selected to analyze the rules of the radiographic images in the training set. The radiographic images of the validation set were used to adjust the super parameters of the network. Finally, 40 radiographic images of the test set were used to evaluate the performance of the three network models. A senior dentist specializing in dental pulp was selected to predict whether the deep caries of 40 radiographic images in the test set were exposed. The gold standard is whether the pulp is exposed after decaying the prepared hole during the clinical operation. The prediction effect of the three network models (VGG, ResNet, and DenseNet) and the senior dentist on the pulp exposure of 40 radiographic images in the test set were compared using receiver operating characteristic (ROC) curve, area under the ROC curve (AUC), accuracy, sensitivity, specificity, positive predictive value, negative predictive value, and F1 score to select the best network model.
RESULTS:
The best network model was DenseNet model, with AUC of 0.97. The AUC values of the ResNet model, VGG model, and the senior dentist were 0.89, 0.78, and 0.87, respectively. Accuracy was not statistically different between the senior dentist (0.850) and the DenseNet model (0.850)(P>0.05). Kappa consistency test showed moderate reliability (Kappa=0.6>0.4, P<0.05).
CONCLUSIONS
Among the three convolutional neural network models, the DenseNet model has the best predictive effect on whether deep caries are exposed in imaging. The predictive effect of this model is equivalent to the level of senior dentists specializing in dental pulp.
Humans
;
Deep Learning
;
Neural Networks, Computer
;
Pulpitis/diagnostic imaging*
;
Reproducibility of Results
;
ROC Curve
;
Random Allocation


Result Analysis
Print
Save
E-mail