Abstract
Deep reinforcement learning agents have achieved unprecedented results when learning to generalize from unstructured data. However, the “black-box” nature of the trained DRL agents makes it difficult to ensure that they adhere to various requirements posed by engineers. In this work, we put forth a novel technique for enhancing the reinforcement learning training loop, and specifically—its reward function, in a way that allows engineers to directly inject their expert knowledge into the training process. This allows us to make the trained agent adhere to multiple constraints of interest. Moreover, using scenario-based modeling techniques, our method allows users to formulate the defined constraints using advanced, well-established, behavioral modeling methods. This combination of such modeling methods together with ML learning tools produces agents that are both high performing and more likely to adhere to prescribed constraints. Furthermore, the resulting agents are more transparent and hence more maintainable. We demonstrate our technique by evaluating it on a case study from the domain of internet congestion control, and present promising results.
Original language | English |
---|---|
Article number | 156 |
Pages (from-to) | 1-13 |
Number of pages | 13 |
Journal | SN Computer Science |
Volume | 4 |
Issue number | 2 |
DOIs | |
State | Published - Mar 2023 |
Keywords
- Deep reinforcement learning
- Domain expertise
- Machine learning
- Rule-based specifications
- Scenario-based modeling
All Science Journal Classification (ASJC) codes
- General Computer Science
- Computer Science Applications
- Computer Networks and Communications
- Computer Graphics and Computer-Aided Design
- Computational Theory and Mathematics
- Artificial Intelligence