Conformal Reachability for Safe Control in Unknown Environments
AI 摘要
提出结合一致性预测和可达性分析的未知动力系统安全控制框架。
主要贡献
- 提出基于一致性预测的安全控制框架
- 开发优化名义奖励和最大化安全规划范围的控制策略训练算法
- 在多个安全控制场景验证了算法的有效性
方法论
结合一致性预测获取不确定性区间,利用可达性分析验证安全,训练控制策略优化奖励和安全。
原文摘要
Designing provably safe control is a core problem in trustworthy autonomy. However, most prior work in this regard assumes either that the system dynamics are known or deterministic, or that the state and action space are finite, significantly limiting application scope. We address this limitation by developing a probabilistic verification framework for unknown dynamical systems which combines conformal prediction with reachability analysis. In particular, we use conformal prediction to obtain valid uncertainty intervals for the unknown dynamics at each time step, with reachability then verifying whether safety is maintained within the conformal uncertainty bounds. Next, we develop an algorithmic approach for training control policies that optimize nominal reward while also maximizing the planning horizon with sound probabilistic safety guarantees. We evaluate the proposed approach in seven safe control settings spanning four domains -- cartpole, lane following, drone control, and safe navigation -- for both affine and nonlinear safety specifications. Our experiments show that the policies we learn achieve the strongest provable safety guarantees while still maintaining high average reward.