Debugging a Policy: Automatic Action-Policy Testing in AI Planning
Marcel Steinmetz, Daniel Fišer, Hasan Ferit Enişer, Patrick Ferber, Timo Gros, Philippe Heim, Daniel Höller, Xandra Schuler, Valentin Wüstholz, Maria Christakis and Joerg Hoffmann
Abstract: Testing is a promising way to gain trust in neural action policies Ï. Previous work on policy testing in sequential decision making targeted environment behavior leading to failure conditions. But if the failure is unavoidable given that behavior, then Ï is not actually to blame. For a situation to qualify as a "bug" in Ï, there must be an alternative policy Ï' that does better. We introduce a generic policy testing framework based on that intuition. This raises the bug confirmation problem, deciding whether or not a state is a bug. We analyze the use of optimistic and pessimistic bounds for the design of test oracles approximating that problem. We contribute an implementation of our framework in classical planning, experimenting with several test oracles and with random-walk methods generating test states biased to poor policy performance and/or state novelty. We evaluate these techniques on policies Ï learned with ASNets. We find that they are able to effectively identify bugs in these Ï, and that our random-walk biases improve over uninformed baselines.
*This password protected talk video will only be available after it was presented at the conference.