Contrastive explanations of plans through model restrictions

Krarup, Benjamin and Krivic, Senka and Magazzeni, Daniele and Long, Derek and Cashmore, Michael and Smith, David E (2021) Contrastive explanations of plans through model restrictions. Preprint / Working Paper. arXiv.org, Ithaca, N.Y..

[thumbnail of Krarup-etal-ArXiv-2021-Contrastive-explanations-of-plans-through-model]
Preview
Text. Filename: Krarup_etal_ArXiv_2021_Contrastive_explanations_of_plans_through_model.pdf
Final Published Version

Download (1MB)| Preview

Abstract

In automated planning, the need for explanations arises when there is a mismatch between a proposed plan and the user's expectation. We frame Explainable AI Planning in the context of the plan negotiation problem, in which a succession of hypothetical planning problems are generated and solved. The object of the negotiation is for the user to understand and ultimately arrive at a satisfactory plan. We present the results of a user study that demonstrates that when users ask questions about plans, those questions are contrastive, i.e. "why A rather than B?". We use the data from this study to construct a taxonomy of user questions that often arise during plan negotiation. We formally define our approach to plan negotiation through model restriction as an iterative process. This approach generates hypothetical problems and contrastive plans by restricting the model through constraints implied by user questions. We formally define model-based compilations in PDDL2.1 of each constraint derived from a user question in the taxonomy, and empirically evaluate the compilations in terms of computational complexity. The compilations were implemented as part of an explanation framework that employs iterative model restriction. We demonstrate its benefits in a second user study.

ORCID iDs

Krarup, Benjamin, Krivic, Senka, Magazzeni, Daniele, Long, Derek, Cashmore, Michael ORCID logoORCID: https://orcid.org/0000-0002-8334-4348 and Smith, David E;