Learning Good State And Action Representations Via Tensor Decomposition

2021 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT)(2021)

引用 6|浏览8
暂无评分
摘要
The transition kernel of a continuous-state-action Markov decision process (MDP) admits a natural tensor structure. This paper proposes a tensor-inspired unsupervised learning method to identify meaningful low-dimensional state and action representations from empirical trajectories. The method exploits the MDP's tensor structure by kernelization, importance sampling and low-Tucker-rank approximation. This method can be further used to cluster states and actions respectively and find the best discrete MDP abstraction. We provide sharp statistical error bounds for tensor concentration and the preservation of diffusion distance after embedding.
更多
查看译文
关键词
good state,action representations,tensor decomposition,transition kernel,continuous-state-action Markov decision process,natural tensor structure,tensor-inspired unsupervised learning method,low-dimensional state,empirical trajectories,MDP's tensor structure,kernelization,low-Tucker-rank approximation,cluster states,discrete MDP abstraction,tensor concentration
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要