Simple Strategies in Multi-Objective MDPs (Technical Report)

October 24, 2019 ยท The Ethereal ยท ๐Ÿ› International Conference on Tools and Algorithms for Construction and Analysis of Systems

๐Ÿ”ฎ THE ETHEREAL: The Ethereal
Pure theory โ€” exists on a plane beyond code

"No code URL or promise found in abstract"

Evidence collected by the PWNC Scanner

Authors Florent Delgrange, Joost-Pieter Katoen, Tim Quatmann, Mickael Randour arXiv ID 1910.11024 Category cs.LO: Logic in CS Cross-listed cs.AI Citations 30 Venue International Conference on Tools and Algorithms for Construction and Analysis of Systems Last Checked 2 months ago
Abstract
We consider the verification of multiple expected reward objectives at once on Markov decision processes (MDPs). This enables a trade-off analysis among multiple objectives by obtaining the Pareto front. We focus on strategies that are easy to employ and implement. That is, strategies that are pure (no randomization) and have bounded memory. We show that checking whether a point is achievable by a pure stationary strategy is NP-complete, even for two objectives, and we provide an MILP encoding to solve the corresponding problem. The bounded memory case can be reduced to the stationary one by a product construction. Experimental results using \Storm and Gurobi show the feasibility of our algorithms.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Logic in CS