In this paper we address the problem of explaining the recommendations returned by a Markov decision process (MDP) that is part of an intelligent assistant for operator training. When analyzing the explanations provided by human experts, we observed that they concentrated on the “most relevant variable”, i.e., the variable that in the current state of the system has the highest influence on the choice of the optimal action. We propose two heuristic rules for determining the most relevant variable based on a factored representation of an MDP. In the first one, we estimate the impact of each variable in the expected utility. The second rule evaluates the potential changes in the optimal action for each variable. We evaluated and compared each...
Markov decision problems (MDPs) provide the foundations for a number of problems of interest to AI r...
International audienceMarkov Decision Processes (MDPs) are a mathematical framework for modeling seq...
In this paper, we provide a unified presentation of the Configurable Markov Decision Process (Conf-M...
Explaining policies of Markov Decision Processes (MDPs) is complicated due to their probabilistic an...
Markov Decision Processes (MDPs) are a mathematical framework for modeling sequential decision probl...
A Markov decision process (MDP) relies on the notions of state, describing the current situation of ...
Decision-theoretic systems, such as Markov Decision Processes (MDPs), are used for sequential decisi...
General purpose intelligent learning agents cycle through (complex,non-MDP) sequences of observation...
An important goal in the design and development of Intelligent Tutoring Systems (ITSs) is to have a ...
<p>(A) Policy for the general MDP. In the fragment of MDP shown, the agent is in state <i>i</i> and ...
This paper introduces Advice-MDPs, an expansion of Markov Decision Processes for generating policies...
This paper concerns knowledge acquisition for supporting therapy decision making (TDM) within the fo...
It is over 30 years ago since D.J. White started his series of surveys on practical applications of ...
In control systems theory, the Markov decision process (MDP) is a widely used optimization model inv...
This paper deals with cognitive theories behind agent-based modeling of learning and information pro...
Markov decision problems (MDPs) provide the foundations for a number of problems of interest to AI r...
International audienceMarkov Decision Processes (MDPs) are a mathematical framework for modeling seq...
In this paper, we provide a unified presentation of the Configurable Markov Decision Process (Conf-M...
Explaining policies of Markov Decision Processes (MDPs) is complicated due to their probabilistic an...
Markov Decision Processes (MDPs) are a mathematical framework for modeling sequential decision probl...
A Markov decision process (MDP) relies on the notions of state, describing the current situation of ...
Decision-theoretic systems, such as Markov Decision Processes (MDPs), are used for sequential decisi...
General purpose intelligent learning agents cycle through (complex,non-MDP) sequences of observation...
An important goal in the design and development of Intelligent Tutoring Systems (ITSs) is to have a ...
<p>(A) Policy for the general MDP. In the fragment of MDP shown, the agent is in state <i>i</i> and ...
This paper introduces Advice-MDPs, an expansion of Markov Decision Processes for generating policies...
This paper concerns knowledge acquisition for supporting therapy decision making (TDM) within the fo...
It is over 30 years ago since D.J. White started his series of surveys on practical applications of ...
In control systems theory, the Markov decision process (MDP) is a widely used optimization model inv...
This paper deals with cognitive theories behind agent-based modeling of learning and information pro...
Markov decision problems (MDPs) provide the foundations for a number of problems of interest to AI r...
International audienceMarkov Decision Processes (MDPs) are a mathematical framework for modeling seq...
In this paper, we provide a unified presentation of the Configurable Markov Decision Process (Conf-M...