98%
921
2 minutes
20
Background Generative artificial intelligence (AI) is anticipated to alter radiology workflows, requiring a clinical value assessment for frequent examinations like chest radiograph interpretation. Purpose To develop and evaluate the diagnostic accuracy and clinical value of a domain-specific multimodal generative AI model for providing preliminary interpretations of chest radiographs. Materials and Methods For training, consecutive radiograph-report pairs from frontal chest radiography were retrospectively collected from 42 hospitals (2005-2023). The trained domain-specific AI model generated radiology reports for the radiographs. The test set included public datasets (PadChest, Open-i, VinDr-CXR, and MIMIC-CXR-JPG) and radiographs excluded from training. The sensitivity and specificity of the model-generated reports for 13 radiographic findings, compared with radiologist annotations (reference standard), were calculated (with 95% CIs). Four radiologists evaluated the subjective quality of the reports in terms of acceptability, agreement score, quality score, and comparative ranking of reports from the domain-specific AI model, radiologists, and a general-purpose large language model (GPT-4Vision). Acceptability was defined as whether the radiologist would endorse the report as their own without changes. Agreement scores from 1 (clinically significant discrepancy) to 5 (complete agreement) were assigned using RADPEER; quality scores were on a 5-point Likert scale from 1 (very poor) to 5 (excellent). Results A total of 8 838 719 radiograph-report pairs (training) and 2145 radiographs (testing) were included (anonymized with respect to sex and gender). Reports generated by the domain-specific AI model demonstrated high sensitivity for detecting two critical radiographic findings: 95.3% (181 of 190) for pneumothorax and 92.6% (138 of 149) for subcutaneous emphysema. Acceptance rate, evaluated by four radiologists, was 70.5% (6047 of 8680), 73.3% (6288 of 8580), and 29.6% (2536 of 8580) for model-generated, radiologist, and GPT-4Vision reports, respectively. Agreement scores were highest for the model-generated reports (median = 4 [IQR, 3-5]) and lowest for GPT-4Vision reports (median = 1 [IQR, 1-3]; < .001). Quality scores were also highest for the model-generated reports (median = 4 [IQR, 3-5]) and lowest for the GPT-4Vision reports (median = 2 [IQR, 1-3]; < .001). From the ranking analysis, model-generated reports were most frequently ranked the highest (60.0%; 5146 of 8580), and GPT-4Vision reports were most frequently ranked the lowest (73.6%; 6312 of 8580). Conclusion A domain-specific multimodal generative AI model demonstrated potential for high diagnostic accuracy and clinical value in providing preliminary interpretations of chest radiographs for radiologists. © RSNA, 2025 See also the editorial by Little in this issue.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1148/radiol.241476 | DOI Listing |
Zhonghua Jie He He Hu Xi Za Zhi
September 2025
Pulmonary and Critical Care Medicine, The First Medical Center of Chinese PLA General Hospital, Beijing 100853, China.
To explore the feasibility and accuracy of predicting respiratory tract infections (RTIs) using physiological data obtained from consumer-grade smartwatches. The study used smartwatches and paired mobile applications to continuously collect physiological parameters while participants slept. A personalized baseline model was established using multi-day data, followed by the construction of RTIs risk prediction algorithm based on deviations from physiological parameter trends.
View Article and Find Full Text PDFWorld J Gastroenterol
August 2025
Department of Gastroenterology, Qilu Hospital of Shandong University, Jinan 250012, Shandong Province, China.
Background: Gastrointestinal diseases have complex etiologies and clinical presentations. An accurate diagnosis requires physicians to integrate diverse information, including medical history, laboratory test results, and imaging findings. Existing artificial intelligence-assisted diagnostic tools are limited to single-modality information, resulting in recommendations that are often incomplete and may be associated with clinical or legal risks.
View Article and Find Full Text PDFWorld J Methodol
December 2025
Department of Surgery and Cancer - Faculty of Medicine, Imperial College London, London W12 0BZ, United Kingdom.
Background: Relieving pain is central to the early management of knee osteoarthritis, with a plethora of pharmacological agents licensed for this purpose. Intra-articular corticosteroid injections are a widely used option, albeit with variable efficacy.
Aim: To develop a machine learning (ML) model that predicts which patients will benefit from corticosteroid injections.
Sci Rep
September 2025
Graduate School of Arts and Sciences, The University of Tokyo, Tokyo, 153-8902, Japan.
Recent studies have revealed that human emotions exhibit a high-dimensional, complex structure. A full capturing of this complexity requires new approaches, as conventional models that disregard high dimensionality risk overlooking key nuances of human emotions. Here, we examined the extent to which the latest generation of rapidly evolving Multimodal Large Language Models (MLLMs) capture these high-dimensional, intricate emotion structures, including capabilities and limitations.
View Article and Find Full Text PDFJMIR Cancer
August 2025
Department of Community Health Sciences, Max Rady College of Medicine, University of Manitoba, Winnipeg, MB, Canada.
Background: In Manitoba, Canada, the impact of the COVID-19 pandemic on cancer care delivery included, but was not limited to, disruption of many routine health care services, and the rapid introduction of both social distancing and virtual care. Little was known about how COVID-19-related changes to cancer care delivery would impact patient satisfaction with care and care coordination.
Objective: This report aims to present and interpret findings of an online survey of people with oncology-related conditions in Manitoba, Canada, during the COVID-19 pandemic, exploring patient satisfaction and care coordination.