Statistically Model Checking PCTL Specifications on Markov Decision Processes via Reinforcement Learning

CDC(2020)

引用 4|浏览72
暂无评分
摘要
Probabilistic Computation Tree Logic (PCTL) is frequently used to formally specify control objectives such as probabilistic reachability and safety. In this work, we focus on model checking PCTL specifications statistically on Markov Decision Processes (MDPs) by sampling, e.g., checking whether there exists a feasible policy such that the probability of reaching certain goal states is greater than a threshold. We use reinforcement learning to search for such a feasible policy for PCTL specifications, and then develop a statistical model checking (SMC) method with provable guarantees on its error. Specifically, we first use upper-confidence-bound (UCB) based Q-learning to design an SMC algorithm for bounded-time PCTL specifications, and then extend this algorithm to unbounded-time specifications by identifying a proper truncation time by checking the PCTL specification and its negation at the same time. Finally, we evaluate the proposed method on case studies.
更多
查看译文
关键词
model checking PCTL specifications,Markov Decision Processes,reinforcement learning,Probabilistic Computation Tree Logic,probabilistic reachability,safety,PCTL specification,feasible policy,statistical model,upper-confidence-bound based Q-learning,bounded-time PCTL specifications,unbounded-time specifications
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要