Abstract
In recent years, safe reinforcement learning (RL) with the actor-critic structure has gained significant interest for continuous control tasks. However, achieving near-optimal control policies with safety and convergence guarantees remains challenging. Moreover, few works have focused on designing RL algorithms that handle time-varying safety constraints. This article proposes a safe RL algorithm for optimal control of nonlinear systems with time-varying state and control constraints. The algorithm's novelty lies in two key aspects. Firstly, the approach introduces a unique barrier force-based control policy structure to ensure control safety during learning. Secondly, a multistep policy evaluation mechanism is employed, enabling the prediction of policy safety risks under time-varying constraints and guiding safe updates. Theoretical results on learning convergence, stability, and robustness are proven. The proposed algorithm outperforms several state-of-the-art RL algorithms in the simulated Safety Gym environment. It is also applied to the real-world problem of integrated path following and collision avoidance for two intelligent vehicles - a differential-drive vehicle and an Ackermann-drive one. The experimental results demonstrate the impressive sim-to-real transfer capability of our approach, while showcasing satisfactory online control performance.
Original language | English |
---|---|
Pages (from-to) | 12744-12753 |
Number of pages | 10 |
Journal | IEEE Transactions on Industrial Electronics |
Volume | 71 |
Issue number | 10 |
DOIs | |
Publication status | Published - 2024 |
Keywords
- Barrier force
- Convergence
- Heuristic algorithms
- multistep policy evaluation
- Optimal control
- Reinforcement learning
- safe reinforcement learning (RL)
- Safety
- time-varying constraints
- Time-varying systems
- Vehicle dynamics