From natural language to RL formulation: a digital twin–centered optimization framework for manufacturing systems.
Published in CIRP Annals, 2026
Digital twins (DTs) enable simulation-driven optimization in manufacturing, yet reinforcement learning (RL) adoption remains limited due to the difficulty of formulating executable learning problems from high-level process objectives. This work presents a DT-centric framework that translates process-level optimization intent into RL formulations, including state, action space, and reward structure. An evolving playbook accumulates formulation knowledge across optimization cycles to improve robustness and efficiency. Experiments demonstrate a 28% improvement in final return, a 27% reduction in convergence time, and a stable success rate of 0.87.
Recommended citation: Fan, Haolin, et al. "From natural language to RL formulation: a digital twin–centered optimization framework for manufacturing systems." CIRP Annals (2026).
