98%
921
2 minutes
20
Artificial Intelligence (AI) has been increasingly integrated into healthcare settings, including the radiology department to aid radiographic image interpretation, including reporting by radiographers. Trust has been cited as a barrier to effective clinical implementation of AI. Appropriating trust will be important in the future with AI to ensure the ethical use of these systems for the benefit of the patient, clinician and health services. Means of explainable AI, such as heatmaps have been proposed to increase AI transparency and trust by elucidating which parts of image the AI 'focussed on' when making its decision. The aim of this novel study was to quantify the impact of different forms of AI feedback on the expert clinicians' trust. Whilst this study was conducted in the UK, it has potential international application and impact for AI interface design, either globally or in countries with similar cultural and/or economic status to the UK. A convolutional neural network was built for this study; trained, validated and tested on a publicly available dataset of MUsculoskeletal RAdiographs (MURA), with binary diagnoses and Gradient Class Activation Maps (GradCAM) as outputs. Reporting radiographers (n = 12) were recruited to this study from all four regions of the UK. Qualtrics was used to present each participant with a total of 18 complete examinations from the MURA test dataset (each examination contained more than one radiographic image). Participants were presented with the images first, images with heatmaps next and finally an AI binary diagnosis in a sequential order. Perception of trust in the AI systems was obtained following the presentation of each heatmap and binary feedback. The participants were asked to indicate whether they would change their mind (or decision switch) in response to the AI feedback. Participants disagreed with the AI heatmaps for the abnormal examinations 45.8% of the time and agreed with binary feedback on 86.7% of examinations (26/30 presentations).'Only two participants indicated that they would decision switch in response to all AI feedback (GradCAM and binary) (0.7%, n = 2) across all datasets. 22.2% (n = 32) of participants agreed with the localisation of pathology on the heatmap. The level of agreement with the GradCAM and binary diagnosis was found to be correlated with trust (GradCAM:-.515;-.584, significant large negative correlation at 0.01 level (p = < .01 and-.309;-.369, significant medium negative correlation at .01 level (p = < .01) for GradCAM and binary diagnosis respectively). This study shows that the extent of agreement with both AI binary diagnosis and heatmap is correlated with trust in AI for the participants in this study, where greater agreement with the form of AI feedback is associated with greater trust in AI, in particular in the heatmap form of AI feedback. Forms of explainable AI should be developed with cognisance of the need for precision and accuracy in localisation to promote appropriate trust in clinical end users.
Download full-text PDF |
Source |
---|---|
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC11305567 | PMC |
http://dx.doi.org/10.1371/journal.pdig.0000560 | DOI Listing |
ISA Trans
August 2025
School of Automation, Shenyang Aerospace University, Shenyang, Liaoning Province 110136, China. Electronic address:
When a failure occurs in bearings, vibration signals are characterized by strong non-stationarity and nonlinearity. Therefore, it is difficult to sufficiently dig fault features. 1D local binary pattern (1D-LBP) has the advantageous feature to effectively extract local information of signals.
View Article and Find Full Text PDFInt J Radiat Oncol Biol Phys
September 2025
Sir Peter MacCallum Department of Oncology, University of Melbourne, Melbourne, Australia; Department of Health Services Research, Peter MacCallum Cancer Centre, Melbourne, Australia; Department of Nursing, Faculty of Medicine, Dentistry, and Health Sciences, The University of Melbourne, Melbourne,
Purpose: This study examined head and neck cancer treatment outcome priorities in patients with human papillomavirus-associated oropharyngeal cancer (HPVOPC) before and 12 months (12m) after (chemo)radiotherapy ([C]RT).
Methods And Materials: Eligible patients were diagnosed with HPVOPC suitable for curative-intent primary [C]RT. Study data included responses to a modified version of the Chicago Priorities Scale (CPS-modified) and select items from the MDASI Head and Neck Cancer Module (MDASI-HN).
J Dent
September 2025
Dental Clinic Post-Graduate Program, University Center of State of Pará, Belém, Pará, Brazil. Electronic address:
Objective: This study evaluated the coherence, consistency, and diagnostic accuracy of eight AI-based chatbots in clinical scenarios related to dental implants.
Methods: A double-blind, clinical experimental study was carried out between February and March 2025, to evaluate eight AI-based chatbots using six fictional cases simulating peri-implant mucositis and peri-implantitis. Each chatbot answered five standardized clinical questions across three independent runs per case, generating 720 binary outputs.
STAR Protoc
September 2025
Department of Epidemiology, University Medical Center Groningen, University of Groningen, Groningen 9700 RB, the Netherlands. Electronic address:
Research on multimorbidity patterns promotes our understanding of the common pathological mechanisms that underlie co-occurring diseases. Here, we present a protocol to infer multimorbidity clusters in the form of disease topics from large-scale diagnosis data using treeLFA, a topic model based on the Bayesian binary non-negative matrix factorization. We describe steps for installing software, preparing input data, and training the model.
View Article and Find Full Text PDFSci Rep
September 2025
Institute of Diagnostic and Interventional Radiology and Neuroradiology, University Hospital Essen, Hufelandstraße 55, 45147, Essen, Germany.
In radiomics, feature selection methods are primarily used to eliminate redundant features and identify relevant ones. Feature projection methods, such as principal component analysis (PCA), are often avoided due to concerns that recombining features may compromise interpretability. However, since most radiomic features lack inherent semantic meaning, prioritizing interpretability over predictive performance may not be justified.
View Article and Find Full Text PDF