← Back to Search

Impact of Markov Decision Process Design on Sim-to-Real Reinforcement Learning

β˜†β˜†β˜†β˜†β˜†Mar 10, 2026arxiv β†’

Tatjana Krau, Jorge Mandlmaier, Tobias Damm, Frieder Heieck

Abstract

Reinforcement Learning (RL) has demonstrated strong potential for industrial process control, yet policies trained in simulation often suffer from a significant sim-to-real gap when deployed on physical hardware. This work systematically analyzes how core Markov Decision Process (MDP) design choices -- state composition, target inclusion, reward formulation, termination criteria, and environment dynamics models -- affect this transfer. Using a color mixing task, we evaluate different MDP configurations and mixing dynamics across simulation and real-world experiments. We validate our findings on physical hardware, demonstrating that physics-based dynamics models achieve up to 50% real-world success under strict precision constraints where simplified models fail entirely. Our results provide practical MDP design guidelines for deploying RL in industrial process control.

Explain this paper

Ask this paper

Loading chat…

Rate this paper