Lower Bound On the Computational Complexity of Discounted Markov Decision Problems

May 20, 2017 ยท The Ethereal ยท ๐Ÿ› arXiv.org

๐Ÿ”ฎ THE ETHEREAL: The Ethereal
Pure theory โ€” exists on a plane beyond code

"No code URL or promise found in abstract"

Evidence collected by the PWNC Scanner

Authors Yichen Chen, Mengdi Wang arXiv ID 1705.07312 Category cs.CC: Computational Complexity Cross-listed cs.LG Citations 19 Venue arXiv.org Last Checked 2 months ago
Abstract
We study the computational complexity of the infinite-horizon discounted-reward Markov Decision Problem (MDP) with a finite state space $|\mathcal{S}|$ and a finite action space $|\mathcal{A}|$. We show that any randomized algorithm needs a running time at least $ฮฉ(|\mathcal{S}|^2|\mathcal{A}|)$ to compute an $ฮต$-optimal policy with high probability. We consider two variants of the MDP where the input is given in specific data structures, including arrays of cumulative probabilities and binary trees of transition probabilities. For these cases, we show that the complexity lower bound reduces to $ฮฉ\left( \frac{|\mathcal{S}| |\mathcal{A}|}ฮต \right)$. These results reveal a surprising observation that the computational complexity of the MDP depends on the data structure of input.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computational Complexity