Correct-by-Construction Runtime Enforcement in AI – A Survey

Bettina Könighofer*, Roderick Bloem, Rüdiger Ehlers, Christian Pek

*Corresponding author for this work

Research output: Chapter in Book/Conference proceedings/Edited volumeChapterScientificpeer-review

2 Citations (Scopus)

Abstract

Runtime enforcement refers to the theories, techniques, and tools for enforcing correct behavior with respect to a formal specification of systems at runtime. In this paper, we are interested in techniques for constructing runtime enforcers for the concrete application domain of enforcing safety in AI. We discuss how safety is traditionally handled in the field of AI and how more formal guarantees on the safety of a self-learning agent can be given by integrating a runtime enforcer. We survey a selection of work on such enforcers, where we distinguish between approaches for discrete and continuous action spaces. The purpose of this paper is to foster a better understanding of advantages and limitations of different enforcement techniques, focusing on the specific challenges that arise due to their application in AI. Finally, we present some open challenges and avenues for future work.

Original languageEnglish
Title of host publicationLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
PublisherSpringer
Pages650-663
DOIs
Publication statusPublished - 2022
Externally publishedYes

Keywords

  • Formal methods
  • Reinforcement learning
  • Runtime enforcement
  • Safety in AI
  • Shielding

Fingerprint

Dive into the research topics of 'Correct-by-Construction Runtime Enforcement in AI – A Survey'. Together they form a unique fingerprint.

Cite this