Reinforcement learning for individualized lung cancer screening schedules: A nested case-control study.

Cancer Med

Department of Epidemiology and Biostatistics, Institute of Basic Medical Sciences, Chinese Academy of Medical Sciences & School of Basic Medicine, Peking Union Medical College, Beijing, China.

Published: July 2024


Category Ranking

98%

Total Visits

921

Avg Visit Duration

2 minutes

Citations

20

Article Abstract

Background: The current guidelines for managing screen-detected pulmonary nodules offer rule-based recommendations for immediate diagnostic work-up or follow-up at intervals of 3, 6, or 12 months. Customized visit plans are lacking.

Purpose: To develop individualized screening schedules using reinforcement learning (RL) and evaluate the effectiveness of RL-based policy models.

Methods: Using a nested case-control design, we retrospectively identified 308 patients with cancer who had positive screening results in at least two screening rounds in the National Lung Screening Trial. We established a control group that included cancer-free patients with nodules, matched (1:1) according to the year of cancer diagnosis. By generating 10,164 sequence decision episodes, we trained RL-based policy models, incorporating nodule diameter alone, combined with nodule appearance (attenuation and margin) and/or patient information (age, sex, smoking status, pack-years, and family history). We calculated rates of misdiagnosis, missed diagnosis, and delayed diagnosis, and compared the performance of RL-based policy models with rule-based follow-up protocols (National Comprehensive Cancer Network guideline; China Guideline for the Screening and Early Detection of Lung Cancer).

Results: We identified significant interactions between certain variables (e.g., nodule shape and patient smoking pack-years, beyond those considered in guideline protocols) and the selection of follow-up testing intervals, thereby impacting the quality of the decision sequence. In validation, one RL-based policy model achieved rates of 12.3% for misdiagnosis, 9.7% for missed diagnosis, and 11.7% for delayed diagnosis. Compared with the two rule-based protocols, the three best-performing RL-based policy models consistently demonstrated optimal performance for specific patient subgroups based on disease characteristics (benign or malignant), nodule phenotypes (size, shape, and attenuation), and individual attributes.

Conclusions: This study highlights the potential of using an RL-based approach that is both clinically interpretable and performance-robust to develop personalized lung cancer screening schedules. Our findings present opportunities for enhancing the current cancer screening system.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC11215689PMC
http://dx.doi.org/10.1002/cam4.7436DOI Listing

Publication Analysis

Top Keywords

rl-based policy
20
cancer screening
12
screening schedules
12
policy models
12
reinforcement learning
8
lung cancer
8
screening
8
nested case-control
8
missed diagnosis
8
delayed diagnosis
8

Similar Publications

AI-Integrated autonomous robotics for solar panel cleaning and predictive maintenance using drone and ground-based systems.

Sci Rep

September 2025

Department of Mechanical and Industrial Engineering, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal, 576104, Karnataka, India.

Solar photovoltaic (PV) systems, especially in dusty and high-temperature regions, suffer performance degradation due to dust accumulation, surface heating, and delayed maintenance. This study proposes an AI-integrated autonomous robotic system combining real-time monitoring, predictive analytics, and intelligent cleaning for enhanced solar panel performance. We developed a hybrid system that integrates CNN-LSTM-based fault detection, Reinforcement Learning (DQN)-driven robotic cleaning, and Edge AI analytics for low-latency decision-making.

View Article and Find Full Text PDF

This article aims to address the fixed-time optimal leader-following consensus issue for unknown multiagent systems (MASs) under Denial of Service (DoS) and false data injection (FDI) attacks. A novel fixed-time stability theorem under DoS attacks is presented to simplify the stability conditions and decrease the computational complexity of the settling time. Simultaneously, the deep neural networks (DNNs) structure with the projection operator are adopted in real-time to approximate the unknown system dynamics.

View Article and Find Full Text PDF

Deep brain stimulation (DBS) is an established intervention for Parkinson's disease (PD), but conventional open-loop systems lack adaptability, are energy-inefficient due to continuous stimulation, and provide limited personalization to individual neural dynamics. Adaptive DBS (aDBS) offers a closed-loop alternative, using biomarkers such as beta-band oscillations to dynamically modulate stimulation. While reinforcement learning (RL) holds promise for personalized aDBS control, existing methods suffer from high sample complexity, unstable exploration in binary action spaces, and limited deployability on resource-constrained hardware.

View Article and Find Full Text PDF

This paper introduces a novel Reinforcement Learning-Based Hybrid Validation Protocol (RL-CC) that revolutionizes conflict resolution for time-sensitive IoT transactions through adaptive edge-cloud coordination. Efficient transaction management in sensor-based systems is crucial for maintaining data integrity and ensuring timely execution within the constraints of temporal validity. Our key innovation lies in dynamically learning optimal scheduling policies that minimize transaction aborts while maximizing throughput under varying workload conditions.

View Article and Find Full Text PDF

Background: Managing blood glucose levels in Type 1 Diabetes Mellitus (T1DM) is essential to prevent complications. Traditional insulin delivery methods often require significant patient involvement, limiting automation. Reinforcement Learning (RL)-based controllers offer a promising approach for improving automated insulin administration.

View Article and Find Full Text PDF