1.Adherence of Studies on Large Language Models for Medical Applications Published in Leading Medical Journals According to the MI-CLEAR-LLM Checklist
Ji Su KO ; Hwon HEO ; Chong Hyun SUH ; Jeho YI ; Woo Hyun SHIM
Korean Journal of Radiology 2025;26(4):304-312
Objective:
To evaluate the adherence of large language model (LLM)-based healthcare research to the Minimum Reporting Items for Clear Evaluation of Accuracy Reports of Large Language Models in Healthcare (MI-CLEAR-LLM) checklist, a framework designed to enhance the transparency and reproducibility of studies on the accuracy of LLMs for medical applications.
Materials and Methods:
A systematic PubMed search was conducted to identify articles on LLM performance published in high-ranking clinical medicine journals (the top 10% in each of the 59 specialties according to the 2023 Journal Impact Factor) from November 30, 2022, through June 25, 2024. Data on the six MI-CLEAR-LLM checklist items: 1) identification and specification of the LLM used, 2) stochasticity handling, 3) prompt wording and syntax, 4) prompt structuring, 5) prompt testing and optimization, and 6) independence of the test data—were independently extracted by two reviewers, and adherence was calculated for each item.
Results:
Of 159 studies, 100% (159/159) reported the name of the LLM, 96.9% (154/159) reported the version, and 91.8% (146/159) reported the manufacturer. However, only 54.1% (86/159) reported the training data cutoff date, 6.3% (10/159) documented access to web-based information, and 50.9% (81/159) provided the date of the query attempts. Clear documentation regarding stochasticity management was provided in 15.1% (24/159) of the studies. Regarding prompt details, 49.1% (78/159) provided exact prompt wording and syntax but only 34.0% (54/159) documented prompt-structuring practices. While 46.5% (74/159) of the studies detailed prompt testing, only 15.7% (25/159) explained the rationale for specific word choices. Test data independence was reported for only 13.2% (21/159) of the studies, and 56.6% (43/76) provided URLs for internet-sourced test data.
Conclusion
Although basic LLM identification details were relatively well reported, other key aspects, including stochasticity, prompts, and test data, were frequently underreported. Enhancing adherence to the MI-CLEAR-LLM checklist will allow LLM research to achieve greater transparency and will foster more credible and reliable future studies.
2.Adherence of Studies on Large Language Models for Medical Applications Published in Leading Medical Journals According to the MI-CLEAR-LLM Checklist
Ji Su KO ; Hwon HEO ; Chong Hyun SUH ; Jeho YI ; Woo Hyun SHIM
Korean Journal of Radiology 2025;26(4):304-312
Objective:
To evaluate the adherence of large language model (LLM)-based healthcare research to the Minimum Reporting Items for Clear Evaluation of Accuracy Reports of Large Language Models in Healthcare (MI-CLEAR-LLM) checklist, a framework designed to enhance the transparency and reproducibility of studies on the accuracy of LLMs for medical applications.
Materials and Methods:
A systematic PubMed search was conducted to identify articles on LLM performance published in high-ranking clinical medicine journals (the top 10% in each of the 59 specialties according to the 2023 Journal Impact Factor) from November 30, 2022, through June 25, 2024. Data on the six MI-CLEAR-LLM checklist items: 1) identification and specification of the LLM used, 2) stochasticity handling, 3) prompt wording and syntax, 4) prompt structuring, 5) prompt testing and optimization, and 6) independence of the test data—were independently extracted by two reviewers, and adherence was calculated for each item.
Results:
Of 159 studies, 100% (159/159) reported the name of the LLM, 96.9% (154/159) reported the version, and 91.8% (146/159) reported the manufacturer. However, only 54.1% (86/159) reported the training data cutoff date, 6.3% (10/159) documented access to web-based information, and 50.9% (81/159) provided the date of the query attempts. Clear documentation regarding stochasticity management was provided in 15.1% (24/159) of the studies. Regarding prompt details, 49.1% (78/159) provided exact prompt wording and syntax but only 34.0% (54/159) documented prompt-structuring practices. While 46.5% (74/159) of the studies detailed prompt testing, only 15.7% (25/159) explained the rationale for specific word choices. Test data independence was reported for only 13.2% (21/159) of the studies, and 56.6% (43/76) provided URLs for internet-sourced test data.
Conclusion
Although basic LLM identification details were relatively well reported, other key aspects, including stochasticity, prompts, and test data, were frequently underreported. Enhancing adherence to the MI-CLEAR-LLM checklist will allow LLM research to achieve greater transparency and will foster more credible and reliable future studies.
3.Adherence of Studies on Large Language Models for Medical Applications Published in Leading Medical Journals According to the MI-CLEAR-LLM Checklist
Ji Su KO ; Hwon HEO ; Chong Hyun SUH ; Jeho YI ; Woo Hyun SHIM
Korean Journal of Radiology 2025;26(4):304-312
Objective:
To evaluate the adherence of large language model (LLM)-based healthcare research to the Minimum Reporting Items for Clear Evaluation of Accuracy Reports of Large Language Models in Healthcare (MI-CLEAR-LLM) checklist, a framework designed to enhance the transparency and reproducibility of studies on the accuracy of LLMs for medical applications.
Materials and Methods:
A systematic PubMed search was conducted to identify articles on LLM performance published in high-ranking clinical medicine journals (the top 10% in each of the 59 specialties according to the 2023 Journal Impact Factor) from November 30, 2022, through June 25, 2024. Data on the six MI-CLEAR-LLM checklist items: 1) identification and specification of the LLM used, 2) stochasticity handling, 3) prompt wording and syntax, 4) prompt structuring, 5) prompt testing and optimization, and 6) independence of the test data—were independently extracted by two reviewers, and adherence was calculated for each item.
Results:
Of 159 studies, 100% (159/159) reported the name of the LLM, 96.9% (154/159) reported the version, and 91.8% (146/159) reported the manufacturer. However, only 54.1% (86/159) reported the training data cutoff date, 6.3% (10/159) documented access to web-based information, and 50.9% (81/159) provided the date of the query attempts. Clear documentation regarding stochasticity management was provided in 15.1% (24/159) of the studies. Regarding prompt details, 49.1% (78/159) provided exact prompt wording and syntax but only 34.0% (54/159) documented prompt-structuring practices. While 46.5% (74/159) of the studies detailed prompt testing, only 15.7% (25/159) explained the rationale for specific word choices. Test data independence was reported for only 13.2% (21/159) of the studies, and 56.6% (43/76) provided URLs for internet-sourced test data.
Conclusion
Although basic LLM identification details were relatively well reported, other key aspects, including stochasticity, prompts, and test data, were frequently underreported. Enhancing adherence to the MI-CLEAR-LLM checklist will allow LLM research to achieve greater transparency and will foster more credible and reliable future studies.
4.Adherence of Studies on Large Language Models for Medical Applications Published in Leading Medical Journals According to the MI-CLEAR-LLM Checklist
Ji Su KO ; Hwon HEO ; Chong Hyun SUH ; Jeho YI ; Woo Hyun SHIM
Korean Journal of Radiology 2025;26(4):304-312
Objective:
To evaluate the adherence of large language model (LLM)-based healthcare research to the Minimum Reporting Items for Clear Evaluation of Accuracy Reports of Large Language Models in Healthcare (MI-CLEAR-LLM) checklist, a framework designed to enhance the transparency and reproducibility of studies on the accuracy of LLMs for medical applications.
Materials and Methods:
A systematic PubMed search was conducted to identify articles on LLM performance published in high-ranking clinical medicine journals (the top 10% in each of the 59 specialties according to the 2023 Journal Impact Factor) from November 30, 2022, through June 25, 2024. Data on the six MI-CLEAR-LLM checklist items: 1) identification and specification of the LLM used, 2) stochasticity handling, 3) prompt wording and syntax, 4) prompt structuring, 5) prompt testing and optimization, and 6) independence of the test data—were independently extracted by two reviewers, and adherence was calculated for each item.
Results:
Of 159 studies, 100% (159/159) reported the name of the LLM, 96.9% (154/159) reported the version, and 91.8% (146/159) reported the manufacturer. However, only 54.1% (86/159) reported the training data cutoff date, 6.3% (10/159) documented access to web-based information, and 50.9% (81/159) provided the date of the query attempts. Clear documentation regarding stochasticity management was provided in 15.1% (24/159) of the studies. Regarding prompt details, 49.1% (78/159) provided exact prompt wording and syntax but only 34.0% (54/159) documented prompt-structuring practices. While 46.5% (74/159) of the studies detailed prompt testing, only 15.7% (25/159) explained the rationale for specific word choices. Test data independence was reported for only 13.2% (21/159) of the studies, and 56.6% (43/76) provided URLs for internet-sourced test data.
Conclusion
Although basic LLM identification details were relatively well reported, other key aspects, including stochasticity, prompts, and test data, were frequently underreported. Enhancing adherence to the MI-CLEAR-LLM checklist will allow LLM research to achieve greater transparency and will foster more credible and reliable future studies.
5.Adherence of Studies on Large Language Models for Medical Applications Published in Leading Medical Journals According to the MI-CLEAR-LLM Checklist
Ji Su KO ; Hwon HEO ; Chong Hyun SUH ; Jeho YI ; Woo Hyun SHIM
Korean Journal of Radiology 2025;26(4):304-312
Objective:
To evaluate the adherence of large language model (LLM)-based healthcare research to the Minimum Reporting Items for Clear Evaluation of Accuracy Reports of Large Language Models in Healthcare (MI-CLEAR-LLM) checklist, a framework designed to enhance the transparency and reproducibility of studies on the accuracy of LLMs for medical applications.
Materials and Methods:
A systematic PubMed search was conducted to identify articles on LLM performance published in high-ranking clinical medicine journals (the top 10% in each of the 59 specialties according to the 2023 Journal Impact Factor) from November 30, 2022, through June 25, 2024. Data on the six MI-CLEAR-LLM checklist items: 1) identification and specification of the LLM used, 2) stochasticity handling, 3) prompt wording and syntax, 4) prompt structuring, 5) prompt testing and optimization, and 6) independence of the test data—were independently extracted by two reviewers, and adherence was calculated for each item.
Results:
Of 159 studies, 100% (159/159) reported the name of the LLM, 96.9% (154/159) reported the version, and 91.8% (146/159) reported the manufacturer. However, only 54.1% (86/159) reported the training data cutoff date, 6.3% (10/159) documented access to web-based information, and 50.9% (81/159) provided the date of the query attempts. Clear documentation regarding stochasticity management was provided in 15.1% (24/159) of the studies. Regarding prompt details, 49.1% (78/159) provided exact prompt wording and syntax but only 34.0% (54/159) documented prompt-structuring practices. While 46.5% (74/159) of the studies detailed prompt testing, only 15.7% (25/159) explained the rationale for specific word choices. Test data independence was reported for only 13.2% (21/159) of the studies, and 56.6% (43/76) provided URLs for internet-sourced test data.
Conclusion
Although basic LLM identification details were relatively well reported, other key aspects, including stochasticity, prompts, and test data, were frequently underreported. Enhancing adherence to the MI-CLEAR-LLM checklist will allow LLM research to achieve greater transparency and will foster more credible and reliable future studies.
6.Korean Thyroid Association Guidelines on the Management of Differentiated Thyroid Cancers; Overview and Summary 2024
Young Joo PARK ; Eun Kyung LEE ; Young Shin SONG ; Bon Seok KOO ; Hyungju KWON ; Keunyoung KIM ; Mijin KIM ; Bo Hyun KIM ; Won Gu KIM ; Won Bae KIM ; Won Woong KIM ; Jung-Han KIM ; Hee Kyung KIM ; Hee Young NA ; Shin Je MOON ; Jung-Eun MOON ; Sohyun PARK ; Jun-Ook PARK ; Ji-In BANG ; Kyorim BACK ; Youngduk SEO ; Dong Yeob SHIN ; Su-Jin SHIN ; Hwa Young AHN ; So Won OH ; Seung Hoon WOO ; Ho-Ryun WON ; Chang Hwan RYU ; Jee Hee YOON ; Ka Hee YI ; Min Kyoung LEE ; Sang-Woo LEE ; Seung Eun LEE ; Sihoon LEE ; Young Ah LEE ; Joon-Hyop LEE ; Ji Ye LEE ; Jieun LEE ; Cho Rok LEE ; Dong-Jun LIM ; Jae-Yol LIM ; Yun Kyung JEON ; Kyong Yeun JUNG ; Ari CHONG ; Yun Jae CHUNG ; Chan Kwon JUNG ; Kwanhoon JO ; Yoon Young CHO ; A Ram HONG ; Chae Moon HONG ; Ho-Cheol KANG ; Sun Wook KIM ; Woong Youn CHUNG ; Do Joon PARK ; Dong Gyu NA ;
International Journal of Thyroidology 2024;17(1):1-20
Differentiated thyroid cancer demonstrates a wide range of clinical presentations, from very indolent cases to those with an aggressive prognosis. Therefore, diagnosing and treating each cancer appropriately based on its risk status is important. The Korean Thyroid Association (KTA) has provided and amended the clinical guidelines for thyroid cancer management since 2007. The main changes in this revised 2024 guideline include 1) individualization of surgical extent according to pathological tests and clinical findings, 2) application of active surveillance in low-risk papillary thyroid microcarcinoma, 3) indications for minimally invasive surgery, 4) adoption of World Health Organization pathological diagnostic criteria and definition of terminology in Korean, 5) update on literature evidence of recurrence risk for initial risk stratification, 6) addition of the role of molecular testing, 7) addition of definition of initial risk stratification and targeting thyroid stimulating hormone (TSH) concentrations according to ongoing risk stratification (ORS), 8) addition of treatment of perioperative hypoparathyroidism, 9) update on systemic chemotherapy, and 10) addition of treatment for pediatric patients with thyroid cancer.
7.The 2022 Annual Report on Toxicology Surveillance and Severe Poisoning Cases at Emergency Departments in Korea
Eun Sun LEE ; Su Jin KIM ; Gyu Chong CHO ; Mi Jin LEE ; Byung Hak SO ; Kyung Su KIM ; Juhyun SONG ; Sung Woo LEE
Journal of The Korean Society of Clinical Toxicology 2023;21(1):1-16
Purpose:
This study investigated the actual incidence of acute poisoning in Korea on a nationwide scale, with the aim of laying the groundwork for future initiatives in prevention, strategic antidote distribution, and the development of effective emergency treatment for acute poisoning.
Methods:
The study analyzed data from 3,038 patients who presented to emergency departments with poisoning-related conditions from June 1, 2022 to December 31, 2022 at 10 sites in nine cities across the country. We extracted data on general characteristics of the poisoning cases, including demographic characteristics (age and gender), place of exposure, reason for poisoning, route of exposure, and the substance involved in the poisoning incident. Age-related patterns in reasons for poisoning, medical outcomes, frequent and primary poisoning substances, and deaths were also analyzed.
Results:
The population analyzed in our study was predominantly female, with women constituting 54.74% of all cases. Among infants and children, non-intentional poisoning due to general accidents was the most common cause, accounting for 71.43% of cases. Conversely, suicidal poisoning was more prevalent among teenagers and adults over 20. Fifty-two patients died during the study period, with males comprising approximately two-thirds (67.31%) of these fatalities. Pesticides were the most common poisoning substance among those who died, accounting for 55.77% of such cases. Notably, a significant majority of the victims were elderly individuals aged 60 and above.
Conclusion
This study holds substantial significance, since it represents the first comprehensive investigation and analysis of the symptoms, treatment, and causes of death due to poisoning in Korea on a national scale. By substantially expanding the range and types of poisonous substances examined, we were able to more precisely identify the characteristics and clinical patterns of poisoning cases nationwide.
8.Korean guidelines for the management of gout
Jennifer Jooha LEE ; Ji Soo LEE ; Min Kyung CHUNG ; Joong Kyong AHN ; Hyo-Jin CHOI ; Seung-Jae HONG ; Chong-Hyeon YOON ; Su-Hyun KIM ; Kyung-Hwan JEONG ; Jong-Woo KIM ; Bo-Yeon KIM ; Jin-Ho SHIN ; Woo Gyu KIM ; Soo-Young KIM ; Hyun-Jung KIM ; Jeong-Soo SONG ; Jae-Bum JUN ; Hyun-Ah PARK ; Shung Chull CHAE ; Bum Soon CHOI ; Tae Nyun KIM ; Hyun Ah KIM
Journal of Rheumatic Diseases 2023;30(3):141-150
Gout is the most common form of arthritis, with the prevalence increasing worldwide. The present treatment guidelines provide recommendations for the appropriate treatment of acute gout, management during the inter-critical period, and prevention of chronic complications. The guidelines were developed based on evidence-based medicine and draft recommendations finalized after expert consensus. These guidelines are designed to provide clinicians with clinical evidence to enable efficient treatment of gout.
9.Korean guidelines for the management of gout
Jennifer Jooha LEE ; Ji Soo LEE ; Min Kyung CHUNG ; Joong Kyong AHN ; Hyo-Jin CHOI ; Seung-Jae HONG ; Chong-Hyeon YOON ; Su-Hyun KIM ; Kyung-Hwan JEONG ; Jong-Woo KIM ; Bo-Yeon KIM ; Jin-Ho SHIN ; Woo Gyu KIM ; Soo-Young KIM ; Hyun-Jung KIM ; Jeong-Soo SONG ; Jae-Bum JUN ; Hyun-Ah PARK ; Shung Chull CHAE ; Bum Soon CHOI ; Tae Nyun KIM ; Hyun Ah KIM
The Korean Journal of Internal Medicine 2023;38(5):641-650
Gout is the most common form of arthritis, with the prevalence increasing worldwide. The present treatment guidelines provide recommendations for the appropriate treatment of acute gout, management during the inter-critical period, and prevention of chronic complications. The guidelines were developed based on evidence-based medicine and draft recommendations finalized after expert consensus. These guidelines are designed to provide clinicians with clinical evidence to enable efficient treatment of gout.
10.A Position Statement of the Utilization and Support Status of Continuous Glucose Monitoring in Korea
Won Jun KIM ; Jae Hyun KIM ; Hye Jin YOO ; Jang Won SON ; Ah Reum KHANG ; Su Kyoung KWON ; Ji Hye KIM ; Tae Ho KIM ; Ohk Hyun RYU ; Kyeong Hye PARK ; Sun Ok SONG ; Kang-Woo LEE ; Woo Je LEE ; Jung Hwa JUNG ; Ho-Chan CHO ; Min Jeong GU ; Jeongrim LEE ; Dal Lae JU ; Yeon Hee LEE ; Eun Kyung KIM ; Young Sil EOM ; Sung Hoon YU ; Chong Hwa KIM ;
Journal of Korean Diabetes 2021;22(4):225-237
The accuracy and convenience of continuous glucose monitoring (CGM), which efficiently evaluates glycemic variability and hypoglycemia, are improving. There are two types of CGM: professional CGM and personal CGM. Personal CGM is subdivided into real-time CGM (rt-CGM) and intermittently scanned CGM (isCGM). CGM is being emphasized in both domestic and foreign diabetes management guidelines. Regardless of age or type of diabetes, CGM is useful for diabetic patients undergoing multiple insulin injection therapy or using an insulin pump. rt-CGM is recommended for all adults with type 1 diabetes (T1D), and can also be used in type 2 diabetes (T2D) treatments using multiple insulin injections. In some cases, short-term or intermittent use of CGM may be helpful for patients with T2D who use insulin therapy other than multiple insulin injections and/or oral hypoglycemic agents. CGM can help to achieve A1C targets in diabetes patients during pregnancy. CGM is a safe and cost-effective alternative to self-monitoring blood glucose in T1D and some T2D patients. CGM used in diabetes management works optimally with proper education, training, and follow up. To achieve the activation of CGM and its associated benefits, it is necessary to secure sufficient repetitive training and time for data analysis, management, and education. Various supports such as compensation, insurance coverage expansion, and reimbursement are required to increase the effectiveness of CGM while considering the scale of benefit recipients, policy priorities, and financial requirements.

Result Analysis
Print
Save
E-mail