98%
921
2 minutes
20
Background: The current guidelines for managing screen-detected pulmonary nodules offer rule-based recommendations for immediate diagnostic work-up or follow-up at intervals of 3, 6, or 12 months. Customized visit plans are lacking.
Purpose: To develop individualized screening schedules using reinforcement learning (RL) and evaluate the effectiveness of RL-based policy models.
Methods: Using a nested case-control design, we retrospectively identified 308 patients with cancer who had positive screening results in at least two screening rounds in the National Lung Screening Trial. We established a control group that included cancer-free patients with nodules, matched (1:1) according to the year of cancer diagnosis. By generating 10,164 sequence decision episodes, we trained RL-based policy models, incorporating nodule diameter alone, combined with nodule appearance (attenuation and margin) and/or patient information (age, sex, smoking status, pack-years, and family history). We calculated rates of misdiagnosis, missed diagnosis, and delayed diagnosis, and compared the performance of RL-based policy models with rule-based follow-up protocols (National Comprehensive Cancer Network guideline; China Guideline for the Screening and Early Detection of Lung Cancer).
Results: We identified significant interactions between certain variables (e.g., nodule shape and patient smoking pack-years, beyond those considered in guideline protocols) and the selection of follow-up testing intervals, thereby impacting the quality of the decision sequence. In validation, one RL-based policy model achieved rates of 12.3% for misdiagnosis, 9.7% for missed diagnosis, and 11.7% for delayed diagnosis. Compared with the two rule-based protocols, the three best-performing RL-based policy models consistently demonstrated optimal performance for specific patient subgroups based on disease characteristics (benign or malignant), nodule phenotypes (size, shape, and attenuation), and individual attributes.
Conclusions: This study highlights the potential of using an RL-based approach that is both clinically interpretable and performance-robust to develop personalized lung cancer screening schedules. Our findings present opportunities for enhancing the current cancer screening system.
Download full-text PDF |
Source |
---|---|
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC11215689 | PMC |
http://dx.doi.org/10.1002/cam4.7436 | DOI Listing |
Sci Rep
September 2025
Department of Mechanical and Industrial Engineering, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal, 576104, Karnataka, India.
Solar photovoltaic (PV) systems, especially in dusty and high-temperature regions, suffer performance degradation due to dust accumulation, surface heating, and delayed maintenance. This study proposes an AI-integrated autonomous robotic system combining real-time monitoring, predictive analytics, and intelligent cleaning for enhanced solar panel performance. We developed a hybrid system that integrates CNN-LSTM-based fault detection, Reinforcement Learning (DQN)-driven robotic cleaning, and Edge AI analytics for low-latency decision-making.
View Article and Find Full Text PDFIEEE Trans Cybern
September 2025
This article aims to address the fixed-time optimal leader-following consensus issue for unknown multiagent systems (MASs) under Denial of Service (DoS) and false data injection (FDI) attacks. A novel fixed-time stability theorem under DoS attacks is presented to simplify the stability conditions and decrease the computational complexity of the settling time. Simultaneously, the deep neural networks (DNNs) structure with the projection operator are adopted in real-time to approximate the unknown system dynamics.
View Article and Find Full Text PDFArXiv
July 2025
Department of Electrical and Computer Engineering, Clemson University, Clemson, SC.
Deep brain stimulation (DBS) is an established intervention for Parkinson's disease (PD), but conventional open-loop systems lack adaptability, are energy-inefficient due to continuous stimulation, and provide limited personalization to individual neural dynamics. Adaptive DBS (aDBS) offers a closed-loop alternative, using biomarkers such as beta-band oscillations to dynamically modulate stimulation. While reinforcement learning (RL) holds promise for personalized aDBS control, existing methods suffer from high sample complexity, unstable exploration in binary action spaces, and limited deployability on resource-constrained hardware.
View Article and Find Full Text PDFSci Rep
July 2025
College of Computing and Information Technology, Shaqra University, Shaqra, Saudi Arabia.
This paper introduces a novel Reinforcement Learning-Based Hybrid Validation Protocol (RL-CC) that revolutionizes conflict resolution for time-sensitive IoT transactions through adaptive edge-cloud coordination. Efficient transaction management in sensor-based systems is crucial for maintaining data integrity and ensuring timely execution within the constraints of temporal validity. Our key innovation lies in dynamically learning optimal scheduling policies that minimize transaction aborts while maximizing throughput under varying workload conditions.
View Article and Find Full Text PDFComput Biol Med
June 2025
University Campus Bio-Medico of Rome, Via Alvaro del Portillo 21, Rome, 00128, RM, Italy. Electronic address:
Background: Managing blood glucose levels in Type 1 Diabetes Mellitus (T1DM) is essential to prevent complications. Traditional insulin delivery methods often require significant patient involvement, limiting automation. Reinforcement Learning (RL)-based controllers offer a promising approach for improving automated insulin administration.
View Article and Find Full Text PDF