Category Ranking

98%

Total Visits

921

Avg Visit Duration

2 minutes

Citations

20

Article Abstract

Deep reinforcement learning (DRL) has been widely applied to various applications, but improving the exploration and the accuracy of Q-value estimation remain key challenges. Recently, the double-actor architecture has emerged as a promising DRL framework that can enhance both exploration and Q-value estimation. Existing double-actor DRL methods sample from the replay buffer to update the two actors; however, the samples used to update each actor are generated by its previous versions and the other actor, resulting in a different data distribution compared with the current actor being updated, which can negatively impact the actor's update and lead to suboptimal policies. To this end, this work proposes a generic solution that can be seamlessly integrated into existing double-actor DRL methods to mitigate the adverse effects of data distribution differences on actor updates, thereby learning better policies. Specifically, we decompose the updates of double-actor DRL methods into two stages, each of which uses the same sampling approach to train a pair of actor-critic. This sampling approach classifies the samples in the replay buffer into distinct categories using a clustering technique, such as K-means, and subsequently employs the Jensen-Shannon (JS) divergence to evaluate the distributional differences between each sample category and the actor currently being updated. Samples are then prioritized from the categories with smaller distribution differences to the current actor to update it. In this way, we can effectively mitigate the distribution difference between the samples and the current actor being updated. Experiments demonstrate that our method enhances the performance of five state-of-the-art (SOTA) double-actor DRL methods and outperforms eight SOTA single-actor DRL methods across eight tasks.

Download full-text PDF

Source
http://dx.doi.org/10.1109/TNNLS.2025.3557930DOI Listing

Publication Analysis

Top Keywords

drl methods
20
double-actor drl
16
current actor
12
deep reinforcement
8
reinforcement learning
8
q-value estimation
8
existing double-actor
8
replay buffer
8
data distribution
8
actor updated
8

Similar Publications

Diagnostic reference levels (DRLs) are essential for optimizing radiologic practices and ensuring patient safety. This study aimed to establish typical DRLs for nuclear medicine (NM) procedures performed at a Brazilian public university hospital. A retrospective analysis of 2,609 patient records from 13 routine NM procedures was conducted.

View Article and Find Full Text PDF

The increasing dependence on cloud computing as a cornerstone of modern technological infrastructures has introduced significant challenges in resource management. Traditional load-balancing techniques often prove inadequate in addressing cloud environments' dynamic and complex nature, resulting in suboptimal resource utilization and heightened operational costs. This paper presents a novel smart load-balancing strategy incorporating advanced techniques to mitigate these limitations.

View Article and Find Full Text PDF

Turbulent convection governs heat transport in both natural and industrial settings, yet optimizing it under extreme conditions remains a significant challenge. Traditional control strategies, such as predefined temperature modulation, struggle to achieve substantial enhancement. Here, we introduce a deep reinforcement learning (DRL) framework that autonomously discovers optimal control policies to maximize heat transfer in turbulent Rayleigh-Bénard convection.

View Article and Find Full Text PDF

This review covers recent advances (2023-2024) in neuroimaging research into the pathophysiology, progression, and treatment of Alzheimer's disease (AD) and related dementias (ADRD). Despite the rapid emergence of blood-based biomarkers, neuroimaging continues to be a vital area of research in ADRD. Here, we discuss neuroimaging as a powerful tool to topographically visualize and quantify amyloid, tau, neurodegeneration, inflammation, and vascular disease in the brain.

View Article and Find Full Text PDF

Introduction: Diagnostic reference levels (DRLs) are essential for optimising ionising radiation use in medical imaging and minimising patient exposure. Radiographers play a key role in implementing DRLs to ensure dose optimisation and high-quality imaging. However, gaps in awareness and understanding can hinder effective application.

View Article and Find Full Text PDF