Abstract
The application of Deep Reinforcement Learning (DRL) for the management of engineering systems has shown very promising results in terms of optimality and scalability. The interpretability of these policies by decision-makers who are so far mostly familiar with traditional approaches is also needed for implementation. In this work, we address this topic by providing a comprehensive overview of POMDP- and DRL-based management policies, along with simulation-based implementation details, for facilitating their interpretation. By mapping a sufficient statistic, namely a belief state, to the current optimal action, POMDP-DRL strategies are able to automatically adapt in time considering long-term sought objectives and the prior history. Through simulated policy realizations, POMDP-DRL-based strategies identified for representative inspection and maintenance planning settings are thoroughly analyzed. The results reveal that if the decision-maker opts for an alternative, even suboptimal, action other than the one suggested by the DRL-based policy, the belief state will be accordingly updated and can still be used as input for the remainder of the planning horizon, without any requirements for model retraining.
Original language | English |
---|---|
Number of pages | 8 |
Publication status | Published - 2023 |
Event | 14th International Conference on Applications of Statistics and Probability in Civil Engineering 2023 - Trinity College Dublin, Dublin, Ireland Duration: 9 Jul 2023 → 13 Jul 2023 https://icasp14.com/ |
Conference
Conference | 14th International Conference on Applications of Statistics and Probability in Civil Engineering 2023 |
---|---|
Abbreviated title | ICASP14 |
Country/Territory | Ireland |
City | Dublin |
Period | 9/07/23 → 13/07/23 |
Internet address |