Using Petri Nets as an Integrated Constraint Mechanism for Reinforcement Learning Tasks

The lack of trust in algorithms is usually an issue

when using Reinforcement Learning (RL) agents for control

in real-world domains such as production plants, autonomous

vehicles, or traffic-related infrastructure, partly due to the lack

of verifiability of the model itself. In such scenarios, Petri

nets (PNs) are often available for flowcharts or process steps,

as they are versatile and standardized. In order to facilitate

integration of RL models and as a step towards increasing AI

trustworthiness, we propose an approach that uses PNs with

three main advantages over typical RL approaches: Firstly, the

agent can now easily be modeled with a combined state including

both external environmental observations and agent-specific

state information from a given PN. Secondly, we can enforce

constraints for state-dependent actions through the inherent PN

model. And lastly, we can increase trustworthiness by verifying

PN properties through techniques such as model checking. We

test our approach on a typical four-way intersection traffic

light control setting and present our results, beating cycle-based

baselines.

Citation information

Sachweh, Timon; Haritz, Pierre; Liebig, Thomas: Using Petri Nets as an Integrated Constraint Mechanism for Reinforcement Learning Tasks, 2024 IEEE Intelligent Vehicles Symposium (IV), 2024, https://ieeexplore.ieee.org/document/10588380, Sachweh.etal.2024a,

Associated Lamarr Researchers

Portrait of Timo Sachweh.

Timon Sachweh

Autor to the profile
lamarr institute person Liebig Thomas - Lamarr Institute for Machine Learning (ML) and Artificial Intelligence (AI)

Prof. Dr. Thomas Liebig

Principal Investigator Trustworthy AI to the profile