Monte-Carlo exploration for deterministic planning

IJCAI(2009)

引用 132|浏览12
暂无评分
摘要
Search methods based on Monte-Carlo simulation have recently led to breakthrough performance improvements in difficult game-playing domains such as Go and General Game Playing. Monte-Carlo Random Walk (MRW) planning applies Monte-Carlo ideas to deterministic classical planning. In the forward chaining planner ARVAND, Monte-Carlo random walks are used to explore the local neighborhood of a search state for action selection. In contrast to the stochastic local search approach used in the recent planner Identidem, random walks yield a larger and unbiased sample of the search neighborhood, and require state evaluations only at the endpoints of each walk. On IPC-4 competition problems, the performance of ARVAND is competitive with state of the art systems.
更多
查看译文
关键词
monte-carlo exploration,search neighborhood,state evaluation,monte-carlo random walk,monte-carlo random,deterministic planning,breakthrough performance improvement,search state,search method,monte-carlo idea,stochastic local search approach,monte-carlo simulation,monte carlo,action selection,random walk,monte carlo simulation
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要