Severity: Warning
Message: file_get_contents(https://...@gmail.com&api_key=61f08fa0b96a73de8c900d749fcb997acc09&a=1): Failed to open stream: HTTP request failed! HTTP/1.1 429 Too Many Requests
Filename: helpers/my_audit_helper.php
Line Number: 197
Backtrace:
File: /var/www/html/application/helpers/my_audit_helper.php
Line: 197
Function: file_get_contents
File: /var/www/html/application/helpers/my_audit_helper.php
Line: 271
Function: simplexml_load_file_from_url
File: /var/www/html/application/helpers/my_audit_helper.php
Line: 3165
Function: getPubMedXML
File: /var/www/html/application/controllers/Detail.php
Line: 597
Function: pubMedSearch_Global
File: /var/www/html/application/controllers/Detail.php
Line: 511
Function: pubMedGetRelatedKeyword
File: /var/www/html/index.php
Line: 317
Function: require_once
98%
921
2 minutes
20
Operation optimization plays a crucial role in process control, directly influencing product quality and profitability. Reinforcement learning (RL), with its capabilities in autonomous learning and dynamic adaptability, has become a promising solution in this domain. However, its real-world application is constrained by the high costs and risks associated with its interactions with environments. Offline RL, which leverages fixed datasets without interactions, offers an alternative but faces significant challenges in the process industry due to imbalanced multioperating condition scenarios and heightened safety sensitivity. To address these challenges, this article introduces a novel offline actor-critic algorithm with expert knowledge guidance (EKG-AC). The method begins with a diffusion-transformer-based action generation framework that mitigates the out-of-distribution problem by capturing the evolution of decision sequences and the interdependencies between states and actions. An expert knowledge guidance mechanism is then integrated, steering the model to generate safe and adaptive candidate actions aligned with current operating conditions and expert knowledge. Subsequently, within the actor-critic framework, the optimal action is selected from the candidate pool based on the evaluated Q-value, thereby setting the operational variables for the optimization task. The proposed algorithm is validated through two real-world industrial processes, demonstrating superior optimization performance and behavior that is closely aligned with expert decision-making, underscoring its substantial practical value.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1109/TCYB.2025.3579361 | DOI Listing |