Category Ranking

98%

Total Visits

921

Avg Visit Duration

2 minutes

Citations

20

Article Abstract

Objective: To use artificial intelligence (AI) to automatically extract video clips of the fetal heart from a stream of ultrasound video, and to assess the performance of these when used for remote second review.

Methods: Using a dataset from a previous clinical trial of AI to assist in fetal ultrasound scanning, AI was used to automatically extract video clips of the fetal heart from ultrasound scans of 48 fetuses in which the diagnosis was known: 24 normal and 24 with congenital heart disease (CHD). These, and manually still saved images, were shown in a random order to expert clinicians, who were asked to detect cardiac abnormalities.

Results: The initial manual scan had a sensitivity of 0.792 and specificity of 0.917 for detecting CHD in this cohort. The addition of second review improved the sensitivity to 0.975 using video clips, which was significantly higher than using still images (0.892, p = 0.002). There was a significant drop in specificity to 0.767 and 0.833 (p < 0.001) for the video and still method, respectively, which were statistically similar to each other (p = 0.117). The median review time was 1.0 min (IQR 0.71) for the still images, and 3.75 min (IQR 3.12) for the AI-generated video clips.

Conclusion: AI can be used to automatically extract fetal cardiac video clips, and these can be used for remote second review to improve detection rates. Video clips are superior to still images, but both methods result in a significant drop in specificity.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC11987776PMC
http://dx.doi.org/10.1002/pd.6757DOI Listing

Publication Analysis

Top Keywords

video clips
12
fetal ultrasound
8
ultrasound scans
8
artificial intelligence
8
remote second
8
congenital heart
8
heart disease
8
automatically extract
8
extract video
8
clips fetal
8

Similar Publications

Posterior lumbar screw fixation is the most common surgical method for lumbar disc herniation, but patients often face multiple complications postoperatively. The occurrence of screw track loosening can lead to fusion failure and even life-threatening screw track extrusion. However, there is currently a lack of animal models specifically targeting changes in the screw track following lumbar screw fixation.

View Article and Find Full Text PDF

Humans navigate the social world by rapidly perceiving social features from other people and their interaction. Recently, large-language models (LLMs) have achieved high-level visual capabilities for detailed object and scene content recognition and description. This raises the question whether LLMs can infer complex social information from images and videos, and whether the high-dimensional structure of the feature annotations aligns with that of humans.

View Article and Find Full Text PDF

We aimed to report our experience with exoscopic keyhole clipping of unruptured middle cerebral artery aneurysms using multiple 4K 3-dimensional monitors.We performed sphenoid ridge keyhole clipping of unruptured middle cerebral artery aneurysms using the ORBEYE exoscope (Sony Olympus Medical Solutions, Inc., Tokyo, Japan) with multiple 4K 3-dimensional monitors in 19 patients in our institution from 2020 to 2023.

View Article and Find Full Text PDF

Purpose: To enhance the temporal feature learning capability of the laparoscopic cholecystectomy phase recognition model and address the class imbalance issue in the training data, this paper proposes an Xception-dual-channel LSTM fusion model based on a dynamic data balancing strategy.

Methods: The model dynamically adjusts the undersampling rate for each surgical phase, extracting short video clips from the original data as training samples to balance the data distribution and mitigate biased learning. The Xception model, utilizing depthwise separable convolutions, extracts fundamental visual features frame by frame, which are then passed to a dual-channel LSTM network.

View Article and Find Full Text PDF

Audio-visual event localization (AVEL) aims to recognize events in videos by associating audio-visual information. However, events involved in existing AVEL tasks are usually coarse-grained events. Actually, finer-grained events are sometimes necessary to be distinguished, especially in certain expert-level applications or rich-content-generation studies.

View Article and Find Full Text PDF