1.Adherence of Studies on Large Language Models for Medical Applications Published in Leading Medical Journals According to the MI-CLEAR-LLM Checklist
Ji Su KO ; Hwon HEO ; Chong Hyun SUH ; Jeho YI ; Woo Hyun SHIM
Korean Journal of Radiology 2025;26(4):304-312
Objective:
To evaluate the adherence of large language model (LLM)-based healthcare research to the Minimum Reporting Items for Clear Evaluation of Accuracy Reports of Large Language Models in Healthcare (MI-CLEAR-LLM) checklist, a framework designed to enhance the transparency and reproducibility of studies on the accuracy of LLMs for medical applications.
Materials and Methods:
A systematic PubMed search was conducted to identify articles on LLM performance published in high-ranking clinical medicine journals (the top 10% in each of the 59 specialties according to the 2023 Journal Impact Factor) from November 30, 2022, through June 25, 2024. Data on the six MI-CLEAR-LLM checklist items: 1) identification and specification of the LLM used, 2) stochasticity handling, 3) prompt wording and syntax, 4) prompt structuring, 5) prompt testing and optimization, and 6) independence of the test data—were independently extracted by two reviewers, and adherence was calculated for each item.
Results:
Of 159 studies, 100% (159/159) reported the name of the LLM, 96.9% (154/159) reported the version, and 91.8% (146/159) reported the manufacturer. However, only 54.1% (86/159) reported the training data cutoff date, 6.3% (10/159) documented access to web-based information, and 50.9% (81/159) provided the date of the query attempts. Clear documentation regarding stochasticity management was provided in 15.1% (24/159) of the studies. Regarding prompt details, 49.1% (78/159) provided exact prompt wording and syntax but only 34.0% (54/159) documented prompt-structuring practices. While 46.5% (74/159) of the studies detailed prompt testing, only 15.7% (25/159) explained the rationale for specific word choices. Test data independence was reported for only 13.2% (21/159) of the studies, and 56.6% (43/76) provided URLs for internet-sourced test data.
Conclusion
Although basic LLM identification details were relatively well reported, other key aspects, including stochasticity, prompts, and test data, were frequently underreported. Enhancing adherence to the MI-CLEAR-LLM checklist will allow LLM research to achieve greater transparency and will foster more credible and reliable future studies.
2.Proposal of age definition for early-onset gastric cancer based on the Korean Gastric Cancer Association nationwide survey data: a retrospective observational study
Seong-A JEONG ; Ji Sung LEE ; Ba Ool SEONG ; Seul-gi OH ; Chang Seok KO ; Sa-Hong MIN ; Chung Sik GONG ; Beom Su KIM ; Moon-Won YOO ; Jeong Hwan YOOK ; In-Seob LEE ;
Annals of Surgical Treatment and Research 2025;108(4):245-255
Purpose:
This study aimed to define an optimal age cutoff for early-onset gastric cancer (EOGC) and compare its characteristics with those of late-onset gastric cancer (LOGC) using nationwide survey data.
Methods:
Using data from a nationwide survey, this comprehensive population-based study analyzed data spanning 3 years (2009, 2014, and 2019). The joinpoint analysis and interrupted time series (ITS) methodology were employed to identify age cutoffs for EOGC based on the sex ratio and tumor histology. Clinicopathologic characteristics and surgical outcomes were compared between the EOGC and LOGC groups.
Results:
The age cutoff for defining EOGC was suggested to be 50 years, supported by joinpoint and ITS analyses. Early gastric cancer was predominantly present in the EOGC and LOGC groups. Patients with EOGC comprised 20.3% of the total study cohort and demonstrated a more advanced disease stage compared to patients with LOGC. However, patients with EOGC underwent more minimally invasive surgeries, experienced shorter hospital stays, and had lower postoperative morbidity and mortality rates.
Conclusion
This study proposes an age of ≤50 years as a criterion for defining EOGC and highlights its features compared to LOGC. Further research using this criterion should guide tailored treatment strategies and improve outcomes for young patients with gastric cancer.
3.CORRIGENDUM: Proposal of age definition for early-onset gastric cancer based on the Korean Gastric Cancer Association nationwide survey data: a retrospective observational study
Seong-A JEONG ; Ji Sung LEE ; Ba Ool SEONG ; Seul-gi OH ; Chang Seok KO ; Sa-Hong MIN ; Chung Sik GONG ; Beom Su KIM ; Moon-Won YOO ; Jeong Hwan YOOK ; In-Seob LEE ;
Annals of Surgical Treatment and Research 2025;108(5):331-331
4.Characteristics and Prevalence of Sequelae after COVID-19: A Longitudinal Cohort Study
Se Ju LEE ; Yae Jee BAEK ; Su Hwan LEE ; Jung Ho KIM ; Jin Young AHN ; Jooyun KIM ; Ji Hoon JEON ; Hyeri SEOK ; Won Suk CHOI ; Dae Won PARK ; Yunsang CHOI ; Kyoung-Ho SONG ; Eu Suk KIM ; Hong Bin KIM ; Jae-Hoon KO ; Kyong Ran PECK ; Jae-Phil CHOI ; Jun Hyoung KIM ; Hee-Sung KIM ; Hye Won JEONG ; Jun Yong CHOI
Infection and Chemotherapy 2025;57(1):72-80
Background:
The World Health Organization has declared the end of the coronavirus disease 2019 (COVID-19) public health emergency. However, this did not indicate the end of COVID-19. Several months after the infection, numerous patients complain of respiratory or nonspecific symptoms; this condition is called long COVID. Even patients with mild COVID-19 can experience long COVID, thus the burden of long COVID remains considerable. Therefore, we conducted this study to comprehensively analyze the effects of long COVID using multi-faceted assessments.
Materials and Methods:
We conducted a prospective cohort study involving patients diagnosed with COVID-19 between February 2020 and September 2021 in six tertiary hospitals in Korea. Patients were followed up at 1, 3, 6, 12, 18, and 24 months after discharge. Long COVID was defined as the persistence of three or more COVID-19-related symptoms. The primary outcome of this study was the prevalence of long COVID after the period of COVID-19.
Results:
During the study period, 290 patients were enrolled. Among them, 54.5 and 34.6% experienced long COVID within 6 months and after more than 18 months, respectively. Several patients showed abnormal results when tested for post-traumatic stress disorder (17.4%) and anxiety (31.9%) after 18 months. In patients who underwent follow-up chest computed tomography 18 months after COVID-19, abnormal findings remained at 51.9%. Males (odds ratio [OR], 0.17; 95% confidence interval [CI], 0.05–0.53; P=0.004) and elderly (OR, 1.04; 95% CI, 1.00–1.09; P=0.04) showed a significant association with long COVID after 12–18 months in a multivariable logistic regression analysis.
Conclusion
Many patients still showed long COVID after 18 months post SARS-CoV-2 infection. When managing these patients, the assessment of multiple aspects is necessary.
5.Proposal of age definition for early-onset gastric cancer based on the Korean Gastric Cancer Association nationwide survey data: a retrospective observational study
Seong-A JEONG ; Ji Sung LEE ; Ba Ool SEONG ; Seul-gi OH ; Chang Seok KO ; Sa-Hong MIN ; Chung Sik GONG ; Beom Su KIM ; Moon-Won YOO ; Jeong Hwan YOOK ; In-Seob LEE ;
Annals of Surgical Treatment and Research 2025;108(4):245-255
Purpose:
This study aimed to define an optimal age cutoff for early-onset gastric cancer (EOGC) and compare its characteristics with those of late-onset gastric cancer (LOGC) using nationwide survey data.
Methods:
Using data from a nationwide survey, this comprehensive population-based study analyzed data spanning 3 years (2009, 2014, and 2019). The joinpoint analysis and interrupted time series (ITS) methodology were employed to identify age cutoffs for EOGC based on the sex ratio and tumor histology. Clinicopathologic characteristics and surgical outcomes were compared between the EOGC and LOGC groups.
Results:
The age cutoff for defining EOGC was suggested to be 50 years, supported by joinpoint and ITS analyses. Early gastric cancer was predominantly present in the EOGC and LOGC groups. Patients with EOGC comprised 20.3% of the total study cohort and demonstrated a more advanced disease stage compared to patients with LOGC. However, patients with EOGC underwent more minimally invasive surgeries, experienced shorter hospital stays, and had lower postoperative morbidity and mortality rates.
Conclusion
This study proposes an age of ≤50 years as a criterion for defining EOGC and highlights its features compared to LOGC. Further research using this criterion should guide tailored treatment strategies and improve outcomes for young patients with gastric cancer.
6.CORRIGENDUM: Proposal of age definition for early-onset gastric cancer based on the Korean Gastric Cancer Association nationwide survey data: a retrospective observational study
Seong-A JEONG ; Ji Sung LEE ; Ba Ool SEONG ; Seul-gi OH ; Chang Seok KO ; Sa-Hong MIN ; Chung Sik GONG ; Beom Su KIM ; Moon-Won YOO ; Jeong Hwan YOOK ; In-Seob LEE ;
Annals of Surgical Treatment and Research 2025;108(5):331-331
7.Adherence of Studies on Large Language Models for Medical Applications Published in Leading Medical Journals According to the MI-CLEAR-LLM Checklist
Ji Su KO ; Hwon HEO ; Chong Hyun SUH ; Jeho YI ; Woo Hyun SHIM
Korean Journal of Radiology 2025;26(4):304-312
Objective:
To evaluate the adherence of large language model (LLM)-based healthcare research to the Minimum Reporting Items for Clear Evaluation of Accuracy Reports of Large Language Models in Healthcare (MI-CLEAR-LLM) checklist, a framework designed to enhance the transparency and reproducibility of studies on the accuracy of LLMs for medical applications.
Materials and Methods:
A systematic PubMed search was conducted to identify articles on LLM performance published in high-ranking clinical medicine journals (the top 10% in each of the 59 specialties according to the 2023 Journal Impact Factor) from November 30, 2022, through June 25, 2024. Data on the six MI-CLEAR-LLM checklist items: 1) identification and specification of the LLM used, 2) stochasticity handling, 3) prompt wording and syntax, 4) prompt structuring, 5) prompt testing and optimization, and 6) independence of the test data—were independently extracted by two reviewers, and adherence was calculated for each item.
Results:
Of 159 studies, 100% (159/159) reported the name of the LLM, 96.9% (154/159) reported the version, and 91.8% (146/159) reported the manufacturer. However, only 54.1% (86/159) reported the training data cutoff date, 6.3% (10/159) documented access to web-based information, and 50.9% (81/159) provided the date of the query attempts. Clear documentation regarding stochasticity management was provided in 15.1% (24/159) of the studies. Regarding prompt details, 49.1% (78/159) provided exact prompt wording and syntax but only 34.0% (54/159) documented prompt-structuring practices. While 46.5% (74/159) of the studies detailed prompt testing, only 15.7% (25/159) explained the rationale for specific word choices. Test data independence was reported for only 13.2% (21/159) of the studies, and 56.6% (43/76) provided URLs for internet-sourced test data.
Conclusion
Although basic LLM identification details were relatively well reported, other key aspects, including stochasticity, prompts, and test data, were frequently underreported. Enhancing adherence to the MI-CLEAR-LLM checklist will allow LLM research to achieve greater transparency and will foster more credible and reliable future studies.
8.Adherence of Studies on Large Language Models for Medical Applications Published in Leading Medical Journals According to the MI-CLEAR-LLM Checklist
Ji Su KO ; Hwon HEO ; Chong Hyun SUH ; Jeho YI ; Woo Hyun SHIM
Korean Journal of Radiology 2025;26(4):304-312
Objective:
To evaluate the adherence of large language model (LLM)-based healthcare research to the Minimum Reporting Items for Clear Evaluation of Accuracy Reports of Large Language Models in Healthcare (MI-CLEAR-LLM) checklist, a framework designed to enhance the transparency and reproducibility of studies on the accuracy of LLMs for medical applications.
Materials and Methods:
A systematic PubMed search was conducted to identify articles on LLM performance published in high-ranking clinical medicine journals (the top 10% in each of the 59 specialties according to the 2023 Journal Impact Factor) from November 30, 2022, through June 25, 2024. Data on the six MI-CLEAR-LLM checklist items: 1) identification and specification of the LLM used, 2) stochasticity handling, 3) prompt wording and syntax, 4) prompt structuring, 5) prompt testing and optimization, and 6) independence of the test data—were independently extracted by two reviewers, and adherence was calculated for each item.
Results:
Of 159 studies, 100% (159/159) reported the name of the LLM, 96.9% (154/159) reported the version, and 91.8% (146/159) reported the manufacturer. However, only 54.1% (86/159) reported the training data cutoff date, 6.3% (10/159) documented access to web-based information, and 50.9% (81/159) provided the date of the query attempts. Clear documentation regarding stochasticity management was provided in 15.1% (24/159) of the studies. Regarding prompt details, 49.1% (78/159) provided exact prompt wording and syntax but only 34.0% (54/159) documented prompt-structuring practices. While 46.5% (74/159) of the studies detailed prompt testing, only 15.7% (25/159) explained the rationale for specific word choices. Test data independence was reported for only 13.2% (21/159) of the studies, and 56.6% (43/76) provided URLs for internet-sourced test data.
Conclusion
Although basic LLM identification details were relatively well reported, other key aspects, including stochasticity, prompts, and test data, were frequently underreported. Enhancing adherence to the MI-CLEAR-LLM checklist will allow LLM research to achieve greater transparency and will foster more credible and reliable future studies.
9.Characteristics and Prevalence of Sequelae after COVID-19: A Longitudinal Cohort Study
Se Ju LEE ; Yae Jee BAEK ; Su Hwan LEE ; Jung Ho KIM ; Jin Young AHN ; Jooyun KIM ; Ji Hoon JEON ; Hyeri SEOK ; Won Suk CHOI ; Dae Won PARK ; Yunsang CHOI ; Kyoung-Ho SONG ; Eu Suk KIM ; Hong Bin KIM ; Jae-Hoon KO ; Kyong Ran PECK ; Jae-Phil CHOI ; Jun Hyoung KIM ; Hee-Sung KIM ; Hye Won JEONG ; Jun Yong CHOI
Infection and Chemotherapy 2025;57(1):72-80
Background:
The World Health Organization has declared the end of the coronavirus disease 2019 (COVID-19) public health emergency. However, this did not indicate the end of COVID-19. Several months after the infection, numerous patients complain of respiratory or nonspecific symptoms; this condition is called long COVID. Even patients with mild COVID-19 can experience long COVID, thus the burden of long COVID remains considerable. Therefore, we conducted this study to comprehensively analyze the effects of long COVID using multi-faceted assessments.
Materials and Methods:
We conducted a prospective cohort study involving patients diagnosed with COVID-19 between February 2020 and September 2021 in six tertiary hospitals in Korea. Patients were followed up at 1, 3, 6, 12, 18, and 24 months after discharge. Long COVID was defined as the persistence of three or more COVID-19-related symptoms. The primary outcome of this study was the prevalence of long COVID after the period of COVID-19.
Results:
During the study period, 290 patients were enrolled. Among them, 54.5 and 34.6% experienced long COVID within 6 months and after more than 18 months, respectively. Several patients showed abnormal results when tested for post-traumatic stress disorder (17.4%) and anxiety (31.9%) after 18 months. In patients who underwent follow-up chest computed tomography 18 months after COVID-19, abnormal findings remained at 51.9%. Males (odds ratio [OR], 0.17; 95% confidence interval [CI], 0.05–0.53; P=0.004) and elderly (OR, 1.04; 95% CI, 1.00–1.09; P=0.04) showed a significant association with long COVID after 12–18 months in a multivariable logistic regression analysis.
Conclusion
Many patients still showed long COVID after 18 months post SARS-CoV-2 infection. When managing these patients, the assessment of multiple aspects is necessary.
10.Adherence of Studies on Large Language Models for Medical Applications Published in Leading Medical Journals According to the MI-CLEAR-LLM Checklist
Ji Su KO ; Hwon HEO ; Chong Hyun SUH ; Jeho YI ; Woo Hyun SHIM
Korean Journal of Radiology 2025;26(4):304-312
Objective:
To evaluate the adherence of large language model (LLM)-based healthcare research to the Minimum Reporting Items for Clear Evaluation of Accuracy Reports of Large Language Models in Healthcare (MI-CLEAR-LLM) checklist, a framework designed to enhance the transparency and reproducibility of studies on the accuracy of LLMs for medical applications.
Materials and Methods:
A systematic PubMed search was conducted to identify articles on LLM performance published in high-ranking clinical medicine journals (the top 10% in each of the 59 specialties according to the 2023 Journal Impact Factor) from November 30, 2022, through June 25, 2024. Data on the six MI-CLEAR-LLM checklist items: 1) identification and specification of the LLM used, 2) stochasticity handling, 3) prompt wording and syntax, 4) prompt structuring, 5) prompt testing and optimization, and 6) independence of the test data—were independently extracted by two reviewers, and adherence was calculated for each item.
Results:
Of 159 studies, 100% (159/159) reported the name of the LLM, 96.9% (154/159) reported the version, and 91.8% (146/159) reported the manufacturer. However, only 54.1% (86/159) reported the training data cutoff date, 6.3% (10/159) documented access to web-based information, and 50.9% (81/159) provided the date of the query attempts. Clear documentation regarding stochasticity management was provided in 15.1% (24/159) of the studies. Regarding prompt details, 49.1% (78/159) provided exact prompt wording and syntax but only 34.0% (54/159) documented prompt-structuring practices. While 46.5% (74/159) of the studies detailed prompt testing, only 15.7% (25/159) explained the rationale for specific word choices. Test data independence was reported for only 13.2% (21/159) of the studies, and 56.6% (43/76) provided URLs for internet-sourced test data.
Conclusion
Although basic LLM identification details were relatively well reported, other key aspects, including stochasticity, prompts, and test data, were frequently underreported. Enhancing adherence to the MI-CLEAR-LLM checklist will allow LLM research to achieve greater transparency and will foster more credible and reliable future studies.

Result Analysis
Print
Save
E-mail