This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Inverse Decision Modeling:
Learning Interpretable Representations of Behavior

Daniel Jarrett    Alihan Hüyük    Mihaela van der Schaar
Abstract

Decision analysis deals with modeling and enhan- cing decision processes. A principal challenge in improving behavior is in obtaining a transparent description of existing behavior in the first place. In this paper, we develop an expressive, unifying perspective on inverse decision modeling: a framework for learning parameterized representations of sequential decision behavior. First, we formalize the forward problem (as a normative standard), subsuming common classes of control behavior. Second, we use this to formalize the inverse problem (as a descriptive model), generalizing existing work on imitation/reward learning—while opening up a much broader class of research problems in behavior representation. Finally, we instantiate this approach with an example (inverse bounded rational control), illustrating how this structure enables learning (interpretable) representations of (bounded) rationality—while naturally capturing intuitive notions of suboptimal actions, biased beliefs, and imperfect knowledge of environments.

Machine Learning, ICML
\contourlength

0.6pt


1 Introduction

Modeling and enhancing decision-making behavior is a fundamental concern in computational and behavioral science, with real-world applications to healthcare (li2015sequential, ), economics (clithero2018response, ), and cognition (drugowitsch2014relation, ). A principal challenge in improving decision processes is in obtaining a transparent understanding of existing behavior to begin with. In this pursuit, a key com- plication is that agents are often boundedly rational due to biological, psychological, and computational factors (wheeler2018bounded, ; griffiths2015rational, ; genewein2015bounded, ; augenblick2018belief, ; ortega2015information, ), the precise mechanics of which are seldom known. As such, how can we intelligibly characterize imperfect behavior?

Consider the “lifecycle” of decision analysis (keller1989role, ) in the real world. First, normative analysis deals with modeling rational decision-making. It asks the question: What constitutes ideal behavior? To this end, a prevailing approach is given by von Neumann-Morgenstern’s expected utility theory, and the study of optimal control is its incarnation in sequential decision-making (neumann1947theory, ). But judgment rendered by real-world agents is often imperfect, so prescriptive analysis deals with improving existing decision behavior. It asks the question: How can we move closer toward the ideal? To this end, the study of decision engineering seeks to design human-in-the-loop techniques that nudge or assist decision-makers, such as medical guidelines and best practices (mellers1998judgment, ). Importantly, however, this first requires a quantitative account of current practices and the imperfections that necessitate correcting.

To take this crucial first step, we must therefore start with descriptive analysis—that is, with understanding observed decision-making from demonstration. We ask the question: What does existing behavior look like—relative to the ideal? Most existing work on imitation learning (i.e. to replicate expert actions) (yue2018imitation, ) and apprenticeship learning (i.e. to match expert returns) (abbeel2004apprenticeship, ) offers limited help, as our objective is instead in understanding (i.e. to interpret imperfect behavior). In particular, beyond the utility-driven nature of rationality for agent behaviors, we wish to quantify intuitive notions of boundedness—such as the apparent flexibility of decisions, tolerance for surprise, or optimism in beliefs. At the same time, we wish that such representations be interpretable—that is, that they be projections of observed behaviors onto parameterized spaces that are meaningful and parsimonious.

Contributions  In this paper, our mission is to explicitly relax normative assumptions of optimality when modeling decision behavior from observations.333Our terminology is borrowed from economics: By “descriptive” models, we refer to those that capture observable decision-making behavior as-is (e.g. an imitator policy in behavioral cloning), and by “normative” models, we refer to those that specify optimal de- cision-making behavior (e.g. with respect to some utility function). First, we develop an expressive, unifying perspective on inverse decision modeling: a general framework for learning parameterized representations of sequential decision-making behavior. Specifically, we begin by formalizing the forward problem FF (as a normative standard), showing that this subsumes common classes of control behavior in literature. Second, we use this to formalize the inverse problem GG (as a descriptive model), showing that it generalizes existing work on imitation and reward learning. Importantly, this opens up a much broader variety of research problems in behavior representation learning—beyond simply learning optimal utility functions. Finally, we instantiate this approach with an example that we term inverse bounded rational control, illustrating how this structure enables learning (interpretable) representations of (bounded) rationality—capturing familiar notions of decision complexity, subjectivity, and uncertainty.

2 Related Work

Table 1: Inverse Decision Modeling. Comparison of primary class- es of imitation/reward learning (IL/IRL) versus our prototypical example (i.e. inverse bounded rational control) as instantiations of inverse decision modeling. Constraints on agent behavior include: environment dynamics (extrinsic), and bounded rationality (intrinsic). Legend: deterministic (Det.), stochastic (Stoc.), subjective dynamics (Subj.), behavioral cloning (BC), distribution matching (DM), risk-sensitive (RS), partially-observable (PO), maximum entropy (ME). All terms/notation are developed over Sections 34.

Inverse Decision Model Extrinsic Intrinsic Examples Partially Controllable Partially Observable Purposeful Behavior Subjective Dynamics Action Stochasticity Knowledge Uncertainty Decision Complexity Specification Complexity Recognition Complexity τenv\tau_{\text{env}} ωenv\omega_{\text{env}} υ\upsilon τ,ω\tau,\omega π\pi ρ,σ\rho,\sigma α\alpha β\beta η\eta BC-IL (pomerleau1991efficient, ; bain1999framework, ; syed2007imitation, ; ross2010efficient, ; syed2010reduction, ; ross2011reduction, ; piot2014boosted, ; jarrett2020strictly, ) Subj. BC-IL (huyuk2021explaining, ) Det. DM-IL (blonde2019sample, ; kostrikov2019discriminator, ) Stoc. DM-IL (ho2016generative, ; jeon2018bayesian, ; ghasemipour2019understanding, ; ghasemipour2019divergence, ; ke2019imitation, ; ke2020wafr, ; kim2018imitation, ; xiao2019wasserstein, ; dadashi2021primal, ; kostrikov2020imitation, ; arenz2020non, ; srinivasan2020interpretable, ; zhang2020f, ; baram2016model, ; baram2017model, ) Det. IRL (ng2000algorithms, ; syed2008game, ; syed2008apprenticeship, ; klein2011batch, ; mori2011model, ; lee2019truly, ; piot2017bridging, ) Stoc. IRL (klein2012inverse, ; klein2013cascaded, ; tossou2013probabilistic, ; jain2019model, ; neu2007apprenticeship, ; babes2011apprenticeship, ; ho2016model, ; finn2016guided, ; pirotta2016inverse, ; metelli2017compatible, ; tateo2017gradient, ; neu2009training, ; ramachandran2007bayesian, ; choi2011map, ; dimitrakakis2011bayesian, ; rothkopf2011preference, ; balakrishnan2020efficient, ; tanwani2013inverse, ; andrus2019inverse, ; belogolovsky2019learning, ) Subj. IRL (reddy2018you, ) RS-IRL (majumdar2017risk, ; singh2018risk, ) Det. PO-IRL (choi2009inverse, ; choi2011inverse, ; chinaei2012inverse, ; bica2021learning, ) Stoc. PO-IRL (makino2012apprenticeship, ; jarrett2020inverse, ; pattanayak2020inverse, ) Subj. PO-IRL (golub2013learning, ; wu2018inverse, ; daptardar2019inverse, ; kwon2020inverse, ) ME-IRL (ziebart2008maximum, ; boularias2011relative, ; kalakrishnan2013learning, ; wulfmeier2015maximum, ; finn2016connection, ; fu2018learning, ; qureshi2019adversarial, ; barde2020adversarial, ; ziebart2010modeling, ; zhou2017infinite, ; lee2018maximum, ; mai2019generalized, ) Subj. ME-IRL (herman2016inverse, ; herman2016thesis, ) Inverse Bounded Rational Control  ✓  ✓ Section 4

As specific forms of descriptive modeling, imitation learning and apprenticeship learning are popular paradigms for learning policies that mimic the behavior of a demonstrator. Imitation learning focuses on replicating an expert’s actions. Classically, “behavioral cloning” methods directly seek to learn a mapping from input states to output actions (pomerleau1991efficient, ; bain1999framework, ; syed2007imitation, ), using assistance from interactive experts or auxiliary regularization to improve generalization (ross2010efficient, ; syed2010reduction, ; ross2011reduction, ; piot2014boosted, ; jarrett2020strictly, ). More recently, “distribution-matching” methods have been proposed for learning an imitator policy whose induced state-action occupancy measure is close to that of the demonstrator (blonde2019sample, ; kostrikov2019discriminator, ; ho2016generative, ; jeon2018bayesian, ; ghasemipour2019understanding, ; ghasemipour2019divergence, ; ke2019imitation, ; ke2020wafr, ; kim2018imitation, ; xiao2019wasserstein, ; dadashi2021primal, ; kostrikov2020imitation, ; arenz2020non, ; srinivasan2020interpretable, ; zhang2020f, ; baram2016model, ; baram2017model, ). Apprenticeship learning focuses on matching the cumulative returns of the expert—on the basis of some ground-truth reward function not known to the imitator policy. This is most popularly approached by inverse reinforcement learning (IRL), which seeks to infer the reward function for which the demonstrated behavior appears most optimal, and using which an apprentice policy may itself be optimized via reinforcement learning. This includes maximum-margin methods based on feature expectations (ng2000algorithms, ; abbeel2004apprenticeship, ; syed2008game, ; syed2008apprenticeship, ; klein2011batch, ; mori2011model, ; lee2019truly, ), maximum likelihood soft policy matching (neu2007apprenticeship, ; babes2011apprenticeship, ), maximum entropy policies (ziebart2010modeling, ; zhou2017infinite, ; lee2018maximum, ; mai2019generalized, ; jain2019model, ), and Bayesian maximum a posteriori inference (ramachandran2007bayesian, ; choi2011map, ; dimitrakakis2011bayesian, ; rothkopf2011preference, ; balakrishnan2020efficient, ), as well as methods that leverage preference models and additional annotations for assistance (asri2016score, ; burchfiel2016distance, ; jacq2019learning, ; brown2019extrapolating, ; brown2020better, ). We defer to surveys of (yue2018imitation, ; osa2018imitation, ) for more detailed overviews of imitation learning and inverse reinforcement learning.

Inverse decision modeling subsumes most of the standard approaches to imitation and apprenticeship learning as specific instantiations, as we shall see (cf. Table 1). Yet—with very few exceptions (wu2018inverse, ; daptardar2019inverse, ; kwon2020inverse, )—the vast majority of these works are limited to cases where demonstrators are assumed to be ideal or close to ideal. Inference is therefore limited to that of a single utility function; after all, its primary purpose is less for introspection than simply as a mathematical intermediary for mimicking the demonstrator’s exhibited behavior. To the contrary, we seek to inspect and understand the demonstrator’s behavior, rather than simply producing a faithful copy of it. In this sense, the novelty of our work is two-fold. First, we shall formally define “inverse decision models” much more generally as projections in the space of behaviors. These projections depend on our conscious choices for forward and inverse planners, and the explicit structure we choose for their parameterizations allows asking new classes of targeted research questions based on normative factors (which we impose) and descriptive factors (which we learn). Second, we shall model an agent’s behavior as induced by both a recognition policy (committing observations to internal states) and a decision policy (emitting actions from internal states). Importantly, not only may an agent’s mapping from internal states into actions be suboptimal (viz. the latter), but that their mapping from observations into beliefs may also be subjective (viz. the former). This greatly generalizes the idea of “boundedness” in sequential decision-making—that is, instead of commonly-assumed forms of noisy optimality, we arrive at precise notions of subjective dynamics and biased belief-updates. Appendix A gives a more detailed treatment of related work.

3 Inverse Decision Modeling

First, we describe our formalism for planners (Section 3.1) and inverse planners (Section 3.2)—together constituting our framework for inverse decision modeling (Section 3.3). Next, we instantiate this with a prototypical example to spotlight the wider class of research questions that this unified perspective opens up (Section 4). Table 1 summarizes related work subsumed, and contextualizes our later example.

Table 2: Planners. Formulation of primary classes of planner algorithms in terms of our (forward) formalism, incl. the boundedly rational planner in our example (Section 4). Legend: controlled Markov process (CMP); Markov decision process (MDP); input-output hidden Markov model (IOHMM); partially-observable (PO); Dirac delta (δ\delta); any mapping into policies (ff); decision-rule parameterization (χ\chi).

Planner (FF) Setting (ψ\psi) Parameter (θ\theta) Optimization (π,ρ\pi^{*},\rho^{*}) Examples Decision-Rule CMP Policy 𝒮,𝒰,𝒯\mathcal{S},\mathcal{U},\mathcal{T} χ\chi argmaxπδ(πfdecision(χ))\text{argmax}_{\pi}\delta(\pi-f_{\text{decision}}(\chi)) (pomerleau1991efficient, ) Model-Free MDP Learner 𝒮,𝒰,𝒯\mathcal{S},\mathcal{U},\mathcal{T} υ,γ\upsilon,\gamma argmaxπEπ,τenv[tγtυ(st,ut)]\text{argmax}_{\pi}\mathbb{E}_{\pi,\tau_{\text{env}}}[\sum_{t}\gamma^{t}\upsilon(s_{t},u_{t})] (any RL agent) Max. Entropy MDP Learner 𝒮,𝒰,𝒯\mathcal{S},\mathcal{U},\mathcal{T} υ,γ,α\upsilon,\gamma,\alpha argmaxπEπ,τenv[tγtυ(st,ut)\text{argmax}_{\pi}\mathbb{E}_{\pi,\tau_{\text{env}}}[\sum_{t}\gamma^{t}\upsilon(s_{t},u_{t})++α(π(|st))]\alpha\mathcal{H}(\pi(\cdot|s_{t}))] (haarnoja2017reinforcement, ; haarnoja2018soft, ; eysenbach2019if, ; shi2019soft, ) Model-Based MDP Planner 𝒮,𝒰,𝒯\mathcal{S},\mathcal{U},\mathcal{T} υ,γ,τ\upsilon,\gamma,\tau argmaxπEπ,τ[tγtυ(st,ut)]\text{argmax}_{\pi}\mathbb{E}_{\pi,\tau}[\sum_{t}\gamma^{t}\upsilon(s_{t},u_{t})] (any MDP solver) Differentiable MDP Planner 𝒮,𝒰,𝒯\mathcal{S},\mathcal{U},\mathcal{T} υ,γ,τ\upsilon,\gamma,\tau argmaxπδ(πneural-network(ψ,υ,γ,τ))\text{argmax}_{\pi}\delta(\pi-\text{neural-network}(\psi,\upsilon,\gamma,\tau)) (shah2018inferring, ; shah2019feasibility, ) KL-Regularized MDP Planner 𝒮,𝒰,𝒯\mathcal{S},\mathcal{U},\mathcal{T} υ,γ,τ,α,π~\upsilon,\gamma,\tau,\alpha,\tilde{\pi} argmaxπEπ,τ[tγt(υ(st,ut)\text{argmax}_{\pi}\mathbb{E}_{\pi,\tau}[\sum_{t}\gamma^{t}(\upsilon(s_{t},u_{t})-αDKL(π(|st)π~))]\alpha D_{{}_{\text{KL}}}(\pi(\cdot|s_{t})\|\tilde{\pi}))] (rubin2012trading, ; galashov2019information, ; ho2020efficiency, ; tiomkin2017unified, ; leibfried2017information, ) Decision-Rule IOHMM Policy 𝒮,𝒳,𝒵,𝒰,𝒯,𝒪\mathcal{S},\mathcal{X},\mathcal{Z},\mathcal{U},\mathcal{T},\mathcal{O} χ,τ,ω\chi,\tau,\omega argmaxπδ(πfdecision(χ),ρfrecognition(τ,ω)\text{argmax}_{\pi}\delta(\pi-f_{\text{decision}}(\chi),\rho-f_{\text{recognition}}(\tau,\omega)) (huyuk2021explaining, ) Model-Free POMDP Learner 𝒮,𝒳,𝒵,𝒰,𝒯,𝒪\mathcal{S},\mathcal{X},\mathcal{Z},\mathcal{U},\mathcal{T},\mathcal{O} υ,γ\upsilon,\gamma argmaxπ,ρ{ρ is black-box}Eπ,τenv,ρ[tγtυ(st,ut)]\text{argmax}_{\pi,\rho\in\{\rho\text{~{}is black-box}\}}\mathbb{E}_{\pi,\tau_{\text{env}},\rho}[\sum_{t}\gamma^{t}\upsilon(s_{t},u_{t})] (hausknecht2015deep, ; zhu2017improving, ; igl2018deep, ; zhang2019learning, ; han2019variational, ; futoma2020popcorn, ) Model-Based POMDP Planner 𝒮,𝒳,𝒵,𝒰,𝒯,𝒪\mathcal{S},\mathcal{X},\mathcal{Z},\mathcal{U},\mathcal{T},\mathcal{O} υ,γ,τ,ω\upsilon,\gamma,\tau,\omega argmaxπ,ρ{ρ is unbiased}Eπ,τ,ρ[tγtυ(st,ut)]\text{argmax}_{\pi,\rho\in\{\rho\text{~{}is unbiased}\}}\mathbb{E}_{\pi,\tau,\rho}[\sum_{t}\gamma^{t}\upsilon(s_{t},u_{t})] (smallwood1973optimal, ; hauskrecht2000value, ; pineau2003point, ; kurniawati2008sarsop, ) Belief-Aware υ\upsilon-POMDP Planner 𝒮,𝒳,𝒵,𝒰,𝒯,𝒪\mathcal{S},\mathcal{X},\mathcal{Z},\mathcal{U},\mathcal{T},\mathcal{O} υ𝒵,γ,τ,ω\upsilon_{\scalebox{0.5}{$\mathcal{Z}$}},\gamma,\tau,\omega argmaxπ,ρ{ρ is unbiased}Eπ,τ,ρ[tγtυ𝒵(st,zt,ut)]\text{argmax}_{\pi,\rho\in\{\rho\text{~{}is unbiased}\}}\mathbb{E}_{\pi,\tau,\rho}[\sum_{t}\gamma^{t}\upsilon_{\scalebox{0.5}{$\mathcal{Z}$}}(s_{t},z_{t},u_{t})] (araya2010pomdp, ; fehr2018rho, ) Bounded Rational Control 𝒮,𝒳,𝒵,𝒰,𝒯,𝒪\mathcal{S},\mathcal{X},\mathcal{Z},\mathcal{U},\mathcal{T},\mathcal{O} υ,γ,α,β,\upsilon,\gamma,\alpha,\beta, η,π~,σ~,ϱ~\eta,\tilde{\pi},\tilde{\sigma},\tilde{\varrho} argmaxπ,ρ{ρ is possibly-biased}Eπ,ρ[tγtυ(st,ut)]\text{argmax}_{\pi,\rho\in\{\rho\text{~{}is possibly-biased}\}}\mathbb{E}_{\pi,\rho}[\sum_{t}\gamma^{t}\upsilon(s_{t},u_{t})] αIπ,ρ[π;π~]βIπ,ρ[σ;σ~]ηIπ,ρ[ϱ;ϱ~]-~{}\alpha\mathbb{I}_{\pi,\rho}[\pi;\tilde{\pi}]-\beta\mathbb{I}_{\pi,\rho}[\sigma;\tilde{\sigma}]-\eta\mathbb{I}_{\pi,\rho}[\varrho;\tilde{\varrho}] Theorems 45 General Formulation 𝒮,𝒳,𝒵,𝒰,𝒯,𝒪\mathcal{S},\mathcal{X},\mathcal{Z},\mathcal{U},\mathcal{T},\mathcal{O} (any) argmaxπ,ρψ(π,ρ;θ)\text{argmax}_{\pi,\rho}\mathcal{F}_{\psi}(\pi,\rho;\theta) Section 3.1

3.1 Forward Problem

Consider the standard setup for sequential decision-making, where an agent interacts with a (potentially partially-obser- vable) environment. First, let ψ\psi\doteq(𝒮,𝒳,𝒵,𝒰,𝒯,𝒪)(\mathcal{S},\mathcal{X},\mathcal{Z},\mathcal{U},\mathcal{T},\mathcal{O}) give the problem setting, where 𝒮\mathcal{S} denotes the space of (external) environment states, 𝒳\mathcal{X} of environment observables, 𝒵\mathcal{Z} of (internal) agent states, 𝒰\mathcal{U} of agent actions, 𝒯\mathcal{T}\doteq Δ(𝒮)𝒮×𝒰\Delta(\mathcal{S})^{\mathcal{S}\times\mathcal{U}} of environment transitions, and 𝒪\mathcal{O}\doteqΔ(𝒳)𝒰×𝒮\Delta(\mathcal{X})^{\mathcal{U}\times\mathcal{S}} of environment emissions. Second, denote with θ\theta the planning parameter: the parameterization of (subjective) factors that a planning algorithm uses to produce behavior, e.g. utility functions υ\upsilon\inR𝒮×𝒰\mathbb{R}^{\mathcal{S}\times\mathcal{U}}, discount factors γ\gamma\in[0,1)[0,1), or any other biases that an agent might be subject to, such as imperfect knowledge τ,ω\tau,\omega of true environment dynamics τenv,ωenv\tau_{\text{env}},\omega_{\text{env}}\in𝒯\mathcal{T}×\times𝒪\mathcal{O}. Note that access to the true dynamics is only (indirectly) possible via such knowledge, or by sampling online/from batch data. Now, a planner is a mapping producing observable behavior:

Definition 1 (name=Behavior)

Denote the space of (observation-action) trajectories with t=0(𝒳×𝒰)t×𝒳\mathcal{H}\doteq\cup_{t=0}^{\infty}(\mathcal{X}\times\mathcal{U})^{t}\times\mathcal{X}. Then a behavior ϕ\phi manifests as a distribution over trajectories (indu- ced by an agent’s policies interacting with the environment):

ΦΔ()\Phi\doteq\Delta(\mathcal{H}) (1)

Consider behaviors induced by an agent operating under a recognition policy ρ\rho\inΔ(𝒵)𝒵×𝒰×𝒳\Delta(\mathcal{Z})^{\mathcal{Z}\times\mathcal{U}\times\mathcal{X}} (i.e. committing obser- vation-action trajectories to internal states), together with a decision policy πΔ(𝒰)𝒵\pi\in\Delta(\mathcal{U})^{\mathcal{Z}} (i.e. emitting actions from internal states). We shall denote behaviors induced by π,ρ\pi,\rho:

ϕπ,ρ((x0,u0,))Puπ(|z)sτenv(|s,u)xωenv(|u,s)zρ(|z,u,x)(h=(x0,u0,))\phi_{\pi,\rho}\big{(}(x_{0},u_{0},...)\big{)}\doteq\mathbb{P}_{\begin{subarray}{c}u\sim\pi(\cdot|z)\\ s^{\prime}\sim\tau_{\text{env}}(\cdot|s,u)\\ x^{\prime}\sim\omega_{\text{env}}(\cdot|u,s^{\prime})\\ z^{\prime}\sim\rho(\cdot|z,u,x^{\prime})\end{subarray}}\big{(}h=(x_{0},u_{0},...)\big{)} (2)

(Note: Our notation may not be immediately familiar as we seek to unify terminology across multiple fields. For reference, a summary of notation is provided in Appendix E).

Definition 2 (name=Planner)

Given problem setting ψ\psi and planning parameter θ\theta, a planner is a mapping into behaviors:

F:Ψ×ΘΦF:\Psi\times\Theta\rightarrow\Phi (3)

where Ψ\Psi indicates the space of settings, and Θ\Theta the space of parameters. Often, behaviors of the form ϕπ,ρ\phi_{\pi,\rho} can be naturally expressed in terms of the solution to an optimization:

F(ψ,θ)ϕπ,ρ:π,ρargmaxπ,ρψ(π,ρ;θ)F(\psi,\theta)\doteq\phi_{\pi^{*},\rho^{*}}:\pi^{*},\rho^{*}\doteq\text{argmax}_{\pi,\rho}\mathcal{F}_{\psi}(\pi,\rho;\theta) (4)

of some real-valued function ψ\mathcal{F}_{\psi} (e.g. this includes all cases where a utility function υ\upsilon is an element of θ\theta). So, we shall write ϕϕπ,ρ\phi^{*}\doteq\phi_{\pi^{*},\rho^{*}} to indicate the behavior produced by FF.

This definition is very general: It encapsulates a wide range of standard algorithms in the literature (see Table 2), including decision-rule policies and neural-network planners. Importantly, however, observe that in most contexts, a global optimizer for ρ\rho is (trivially) either an identity function, or perfect Bayesian inference (with the practical caveat, of course, that in model-free contexts actually reaching such an optimum may be difficult, such as with a deep recurrent network). Therefore in addition to just π\pi, what Definition 2 makes explicit is the potential for ρ\rho to be biased—that is, to deviate from (perfect) Bayes updates; this will be one of the important developments made in our subsequent example.

Note that by equating a planner with such a mapping, we are implicitly assuming that the embedded optimization (Equation 4) is well-defined—that is, that there exists a single global optimum. In general if the optimization is non-trivial, this requires that the spaces of policies π,ρ𝒫\pi,\rho\in\mathcal{P}×\times\mathcal{R} be suitably restricted: This is satisfied by the usual (hard-/ soft-QQ) Boltzmann-rationality for decision policies, and by uniquely fixing the semantics of internal states as (subjective) beliefs, i.e. probability distributions over states, with recognition policies being (possibly-biased) Bayes updates.

A more practical question is whether this optimum is reachable. While this may seem more difficult (at least in the most general case), for our interpretative purposes it is rarely a problem, because (simple) human-understandable models are what we desire to be working with in the first instance. In healthcare, for example, diseases are often modeled in terms of discrete states, and subjective beliefs over those states are eminently transparent factors that medical practitioners can readily comprehend and reason about (sonnenberg1983markov, ; jackson2003multistate, ). This is prevalent in research and practice, e.g. two-to-four states in progressive dementia (obryant2008staging, ; jarrett2018match, ; jarrett2019dynamic, ), cancer screening (petousis2019using, ; cardoso2019early, ), cystic fibrosis (alaa2019attentive, ), as well as pulmonary disease (wang2014unsupervised, ). Of course, this is not to say our exposition is incompatible with model-free, online settings with complex spaces and black-box approximators. But our focus here is to establish an interpretative paradigm—for which simple state-based models are most amenable to human reasoning.

Table 3: Inverse Planners. Formulation of primary classes of identification strategies in terms of our (inverse) formalism. Legend: value functions for ϕ\phi under θ\theta (Vθϕ,QθϕV_{\theta}^{\phi},Q_{\theta}^{\phi}); regularizer (ζ\zeta); shaped-reward error (Δυ\Delta\upsilon); pp-norm (p\|\cdot\|_{p}); preference relation (\prec); ff-divergence (DfD_{f}). Note that while our notation is general, virtually all original works here have θdesc\theta_{\text{desc}}==υ\upsilon and assume full observability (whence 𝒮\mathcal{S}==𝒳\mathcal{X}==𝒵\mathcal{Z}).

Inverse Planner (GG) Demonstrator (ϕdemo\phi_{\text{demo}}) Helper Optimization (θdesc\theta_{\text{desc}}^{*}) Examples Minimum Perturbation Deterministic, Optimal Default θ~desc\tilde{\theta}_{\text{desc}} argminθdescθdesc\text{argmin}_{\theta_{\text{desc}}}\|\theta_{\text{desc}}-θ~descp\tilde{\theta}_{\text{desc}}\|_{p}::ϕdemo\phi_{\text{demo}}==F(ψ,θ)F(\psi,\theta) (heuberger2004inverse, ) Maximum Margin Deterministic, Optimal - argminθdescEzρ0[Vθϕimit(z)Vθϕdemo(z)]\text{argmin}_{\theta_{\text{desc}}}\kern 0.90005pt\mathbb{E}_{z\sim\rho_{0}}[\kern 0.90005ptV_{\theta}^{\phi_{\text{imit}}}(z)-V_{\theta}^{\phi_{\text{demo}}}(z)\kern 0.90005pt] (ng2000algorithms, ; syed2008game, ; syed2008apprenticeship, ; klein2011batch, ; mori2011model, ; lee2019truly, ; choi2009inverse, ; choi2011inverse, ; chinaei2012inverse, ; bica2021learning, ; ho2016model, ) Regularized Max. Margin Stochastic, Optimal - argminθdescEzρ0[Vsoft,θϕimit(z)Vθϕdemo(z)]+ζ(θ)\text{argmin}_{\theta_{\text{desc}}}\mathbb{E}_{z\sim\rho_{0}}[V_{\text{soft},\theta}^{\phi_{\text{imit}}}(z)-V_{\theta}^{\phi_{\text{demo}}}(z)]+\zeta(\theta) (ho2016generative, ) Multiple Experimentation Deterministic, Optimal Environments 𝒱\mathcal{V} argminθdesc\text{argmin}_{\theta_{\text{desc}}}\textstyle\intmax𝒱,u(Q𝒱,θϕdemo(z,u)\textstyle\text{max}_{\mathcal{V},u}(Q_{\mathcal{V},\theta}^{\phi_{\text{demo}}}(z,u)-V𝒱,θϕdemo(z))dxV_{\mathcal{V},\theta}^{\phi_{\text{demo}}}(z))dx (amin2016towards, ; amin2017repeated, ) Distance Minimization Individually-Scored Scores υ~(h)R\tilde{\upsilon}(h)\in\mathbb{R} argminθdescEhϕdemoυ~(h)s,uhυ(s,u)p\text{argmin}_{\theta_{\text{desc}}}\mathbb{E}_{h\sim\phi_{\text{demo}}}\|\tilde{\upsilon}(h)-\textstyle\sum_{s,u\in h}\upsilon(s,u)\|_{p} (asri2016score, ; burchfiel2016distance, ) Soft Policy Inversion Stoc., Batch-Ordered {ϕdemo(1),,ϕdemo(K)}\{\phi_{\text{demo}}^{\scalebox{0.6}{($1$)}},...,\phi_{\text{demo}}^{\scalebox{0.6}{($K$)}}\} argminθdesckEs,u,sϕdemo(k)\text{argmin}_{\theta_{\text{desc}}}\textstyle\sum_{k}\mathbb{E}_{s,u,s^{\prime}\sim\phi_{\text{demo}}^{\scalebox{0.6}{($k$)}}}Δυ(k)(s,u,s)p\|\Delta\upsilon^{\scalebox{0.6}{($k$)}}(s,u,s^{\prime})\|_{p} (jacq2019learning, ) Preference Extrapolation Stoc., Pairwise-Ranked {(i,j)|hihj}\{(i,j)|h_{i}\prec h_{j}\} argminθdescE(hihj)ϕdemologPυ(hihj)\text{argmin}_{\theta_{\text{desc}}}\mathbb{E}_{(h_{i}\prec h_{j})\sim\phi_{\text{demo}}}\log\mathbb{P}_{\upsilon}(h_{i}\prec h_{j}) (brown2019extrapolating, ; brown2020better, ) Soft Policy Matching Stochastic, Optimal - argminθdescDKL(Pϕdemo(u0:Tx0:T)Pϕimit(u0:Tx0:T))\text{argmin}_{\theta_{\text{desc}}}\raisebox{0.0pt}{$D_{\text{KL}}(\mathbb{P}_{\phi_{\text{demo}}}(u_{\scalebox{0.5}{0:$T$}}\|x_{\scalebox{0.5}{0:$T$}})\|\mathbb{P}_{\phi_{\text{imit}}}(u_{\scalebox{0.5}{0:$T$}}\|x_{\scalebox{0.5}{0:$T$}}))$} (ziebart2010modeling, ; zhou2017infinite, ; lee2018maximum, ; mai2019generalized, ; neu2007apprenticeship, ; babes2011apprenticeship, ; jain2019model, ; pattanayak2020inverse, ; klein2012inverse, ; klein2013cascaded, ; tossou2013probabilistic, ; herman2016inverse, ; herman2016thesis, ) Distribution Matching Stochastic, Optimal - argminθdescDf(ϕdemoϕimit)\text{argmin}_{\theta_{\text{desc}}}D_{f}(\phi_{\text{demo}}\|\phi_{\text{imit}}) (ho2016generative, ; jeon2018bayesian, ; ghasemipour2019understanding, ; ghasemipour2019divergence, ; ke2019imitation, ; ke2020wafr, ; kim2018imitation, ; xiao2019wasserstein, ; dadashi2021primal, ; kostrikov2020imitation, ; arenz2020non, ; srinivasan2020interpretable, ; zhang2020f, ; baram2016model, ; baram2017model, ; ziebart2008maximum, ; boularias2011relative, ; kalakrishnan2013learning, ; wulfmeier2015maximum, ; finn2016connection, ; fu2018learning, ; qureshi2019adversarial, ; barde2020adversarial, ; kostrikov2019discriminator, ; blonde2019sample, ; finn2016guided, ) General Formulation (any) (any) argminθdesc𝒢ψ(ϕdemo,ϕimit)\text{argmin}_{\theta_{\text{desc}}}\mathcal{G}_{\psi}(\phi_{\text{demo}},\phi_{\text{imit}}) Section 3.2

Figure 1: Forward, Inverse, and Projection Mappings. In the forward direction (i.e. generation): Given planning parameters θ\theta, a planner FF generates observable behavior ϕ\phi (Definition 2). In the opposite direction (i.e. inference): Given observed behavior ϕ\phi, an inverse planner GG infers the planning parameters θ\theta that produced it—subject to normative specifications (Definition 3). Finally, given observed behavior ϕ\phi, the composition of FF and GG gives its projection onto the space of behaviors that are parameterizable by θ\theta (Definition 4): This is the inverse decision model (Definition 5).
Refer to caption

3.2 Inverse Problem

Given any setting and appropriate planner, θ\theta gives a complete account of ϕ=F(ψ,θ)\phi^{*}=F(\psi,\theta): This deals with generation —that is, of behavior from its parameterization. In the opposite, given observed behavior ϕdemo\phi_{\text{demo}} produced by some planner, we can ask what its θ\theta appears to be: This now deals with inference—that is, of parameterizations from behavior.

First, note that absent any restrictions, this endeavor immediately falls prey to the celebrated “no free lunch” result: It is in general impossible to infer anything of use from ϕdemo\phi_{\text{demo}} alone, if we posit nothing about θ\theta (or FF) to begin with (armstrong2018occam, ; christiano2015medium, ). The only close attempt has recruited inductive biases requiring multiple environments, and is not interpretable due to the use of differentiable planners (shah2018inferring, ; shah2019feasibility, ).

On the other extreme, the vast literature on IRL has largely restricted attention to perfectly optimal agents—that is, with full visibility of states, certain knowledge of dynamics, and perfect ability to optimize υ\upsilon. While this indeed fends off the impossibility result, it is overly restrictive for understanding behavior: Summarizing ϕdemo\phi_{\text{demo}} using υ\upsilon alone is not informative as to specific types of biases we may be interested in. How aggressive does this clinician seem? How flexible do their actions appear? It is difficult to tease out such nuances from just υ\upsilon—let alone comparing between agents (michaud2020understanding, ; gleave2021quantifying, ).

We take a generalized approach to allow any middle ground of choice. While some normative specifications are required to fend off the impossibility result (armstrong2018occam, ; shah2019feasibility, ), they need not be so strong as to restrict us to perfect optimality. Formally:

Definition 3 (name=Inverse Planner)

Let ΘΘnorm×Θdesc\Theta\doteq\Theta_{\text{norm}}\times\Theta_{\text{desc}} decompose the parameter space into a normative component (i.e. whose values θnormΘnorm\theta_{\text{norm}}\in\Theta_{\text{norm}} we wish to clamp), and a descriptive component (i.e. whose values θdescΘdesc\theta_{\text{desc}}\in\Theta_{\text{desc}} we wish to infer). Then an inverse planner is given as follows:

G:Φ×ΘnormΘdescG:\Phi\times\Theta_{\text{norm}}\rightarrow\Theta_{\text{desc}} (5)

Often, the descriptive parameter can be naturally expressed as the solution to an optimization (of some real-valued 𝒢ψ\mathcal{G}_{\psi}):

G(ϕdemo,θnorm)argminθdesc𝒢ψ(ϕdemo,ϕimit)G(\phi_{\text{demo}},\theta_{\text{norm}})\doteq\text{argmin}_{\theta_{\text{desc}}}\mathcal{G}_{\psi}(\phi_{\text{demo}},\phi_{\text{imit}}) (6)

where we denote by ϕimit\phi_{\text{imit}}\doteqF(ψ,(θnorm,θdesc))F(\psi,(\theta_{\text{norm}},\theta_{\text{desc}})) the imitation behavior generated on the basis of θdesc\theta_{\text{desc}}. So, we shall write θdesc\theta_{\text{desc}}^{*}for the (minimizing) descriptive parameter output byGG.

As with the forward case, this definition is broad: It encapsulates a wide range of inverse optimization techniques in the literature (see Table 3). Although not all techniques entail learning imitating policies in the process, by far the most dominant paradigms do (i.e. maximum margin, soft policy matching, and distribution matching). Moreover, it is normatively flexible in the sense of the middle ground we wanted: θnorm\theta_{\text{norm}} can encode precisely the information we desire.444We can verify that θdesc\theta_{\text{desc}}==υ\upsilon alone recovers the usual IRL paradigm. This opens up new possibilities for interpretative research. For instance, contrary to IRL for imitation or apprenticeship, we may often not wish to recover υ\upsilon at all. Suppose—as an investigator—we believe that a certain υ\upsilon we defined is the “ought-to-be” ideal. By allowing υ\upsilon to be encoded in θnorm\theta_{\text{norm}} (instead of θdesc\theta_{\text{desc}}), we may now ask questions of the form: How “consistently” does ϕdemo\phi_{\text{demo}} appear to be in pursuing υ\upsilon? Does it seem “optimistic” or “pessimistic” relative to neutral beliefs about the world? All that is required is for appropriate measures of such notions (and any others) to be represented in θdesc\theta_{\text{desc}}. (Section 4 shall provide one such exemplar).

Note that parameter identifiability depends on the degrees of freedom in the target θdesc\theta_{\text{desc}} and the nature of the identifi- cation strategy GG. From our generalized standpoint, we sim- ply note that—beyond the usual restrictions (e.g. on scaling, shifting, reward shaping) in conjunction with GG—Bayesian inference remains a valid option to address ambiguities, as in (jeon2018bayesian, ) for distribution matching,(ramachandran2007bayesian, ; choi2011map, ; dimitrakakis2011bayesian, ; rothkopf2011preference, ; balakrishnan2020efficient, ; makino2012apprenticeship, ; jarrett2020inverse, ) for soft policy matching, and (brown2019deep, ; brown2020safe, ) for preference extrapolation.

3.3 Behavior Projection

Now we have the ingredients to formally define the business of inverse decision modeling. Compacting notation, denote Fθnorm()F_{\theta_{\text{norm}}}(\kern 1.00006pt\kern 1.00006pt\cdot\kern 1.00006pt\kern 1.00006pt)\doteqF(ψ,(θnorm,))F(\psi,(\theta_{\text{norm}},\kern 1.00006pt\cdot\kern 1.00006pt\kern 1.00006pt)), and Gθnorm()G_{\theta_{\text{norm}}}(\kern 1.00006pt\kern 1.00006pt\cdot\kern 1.00006pt\kern 1.00006pt)\doteqG(,θnorm)G(\kern 1.00006pt\kern 1.00006pt\kern 1.00006pt\cdot\kern 1.00006pt\kern 1.00006pt\kern 1.00006pt,\theta_{\text{norm}}). First, we require a projection operator that maps onto the sp- ace of behaviors that are parameterizable by θ\theta given FθnormF_{\theta_{\text{norm}}}:

Definition 4 (name=Behavior Projection)

Denote the image of Θdesc\Theta_{\text{desc}} under FθnormF_{\theta_{\text{norm}}} by the following: ΦθnormFθnorm[Θdesc]Φ\Phi_{\theta_{\text{norm}}}\doteq F_{\theta_{\text{norm}}}[\Theta_{\text{desc}}]\leq\Phi. Then the projection map onto this subspace is given by:

projΦθnormFθnormGθnorm\text{proj}_{\Phi_{\theta_{\text{norm}}}}\doteq F_{\theta_{\text{norm}}}\circ G_{\theta_{\text{norm}}} (7)
Definition 5 (name=Inverse Decision Model)

Given a specified method of parameterization Θ\Theta, normative standards θnorm\theta_{\text{norm}}, (and appropriate planner FF and identification strategy GG), the resulting inverse decision model of ϕdemo\phi_{\text{demo}} is given by:

ϕimitprojΦθnorm(ϕdemo)\phi_{\text{imit}}^{*}\doteq\text{proj}_{\Phi_{\theta_{\text{norm}}}}(\phi_{\text{demo}}) (8)

In other words, the model ϕimit\phi_{\text{imit}}^{*} serves as a complete (genera- tive) account of ϕdemo\phi_{\text{demo}} as its behavior projection onto Φθnorm\Phi_{\theta_{\text{norm}}}.

Interpretability  What dictates our choices? For pure imitation (i.e. replicating expert actions), a black-box decision-rule fitted by soft policy matching may do well. For apprenticeship (i.e. matching expert returns), a perfectly optimal planner inversed by distribution matching may do well. But for understanding, however, we wish to place appropriate structure on Θ\Theta depending on the question of interest: Precisely, the mission here is to choose some (interpretable) Fθnorm,GθnormF_{\theta_{\text{norm}}},G_{\theta_{\text{norm}}} such that ϕimit\phi_{\text{imit}}^{*} is amenable to human reasoning.

Note that these are not passive assumptions: We are not making the (factual) claim that θ\theta gives a scientific explanation of the complex neurobiological processes in a clinician’s head. Instead, these are active specifications: We are making the (effective) claim that the learned θ\theta is a parameterized “as-if” interpretation of the observed behavior. For instance, while there exist a multitude of commonly studied human biases in psychology, it is difficult to measure their magnitudes—much less compare them among agents. Section 4 shows an example of how inverse decision modeling can tackle this. (Figure 1 visualizes inverse decision modeling in a nutshell).

4 Bounded Rationality

We wish to understand observed behavior through the lens of bounded rationality. Specifically, let us account for the following facts: that (1) an agent’s knowledge of the environment is uncertain and possibly biased; that (2) the agent’s capacity for information processing is limited, both for decisions and recognition; and—as a result—that (3) the agent’s (subjective) beliefs and (suboptimal) actions deviate from those expected of a perfectly rational agent. We shall see, this naturally allows quantifying such notions as flexibility of decisions, tolerance for surprise, and optimism in beliefs.

First, Section 4.1 describes inference and control under environment uncertainty (cf. 1). Then, 4.2 develops the forward model (FF) for agents bounded by information constraints (cf. 2–3). Finally, 4.3 learns parameterizations of such bo- undedness from behavior by inverse decision modeling (GG).

4.1 Inference and Control

Consider that an agent has uncertain knowledge of the environment, captured by a prior over dynamics σ~\tilde{\sigma}\inΔ(𝒯\Delta(\mathcal{T}×\times𝒪)\mathcal{O}). As a normative baseline, let this be given by some (unbiased) posterior σ~p(τ,ω|)\tilde{\sigma}\doteq p(\tau,\omega|\mathcal{E}), where \mathcal{E} refers to any manner of experience (e.g. observed data about environment dynamics) with which we may come to form such a neutral belief.

Now, an agent may deviate from σ~\tilde{\sigma} depending on the situation, relying instead on τ,ω\tau,\omega\simσ(|z,u)\sigma(\cdot|z,u)—where z,uz,u allows the (biased) σ\sigma\inΔ(𝒯\Delta(\mathcal{T}×\times𝒪)𝒵×𝒰\mathcal{O})^{\mathcal{Z}\times\mathcal{U}} to be context-dependent. Consider recognition policies thereby parameterized by σ\sigma:

ρ(z|z,u,x)Eτ,ωσ(|z,u)ρτ,ω(z|z,u,x)\rho(z^{\prime}|z,u,x^{\prime})\doteq\mathbb{E}_{\tau,\omega\sim\sigma(\cdot|z,u)}\rho_{\tau,\omega}(z^{\prime}|z,u,x^{\prime}) (9)

where ρτ,ω\rho_{\tau,\omega} denotes the policy for adapting zz to xx^{\prime} given (a point value for) τ,ω\tau,\omega. For interpretability, we let ρτ,ω\rho_{\tau,\omega} be the usual Bayes belief-update. Importantly, however, ρ\rho can now effectively be biased (i.e. by σ\sigma) even while ρτ,ω\rho_{\tau,\omega} is Bayesian.

Forward Process  The forward (“inference”) process yields the occupancy measure. First, the stepwise conditional is:

p(z|z)=Euπ(|z)τ,ωσ(|z,u)sτ(|s,u)xω(|u,s)ρτ,ω(z|z,u,x)p(z^{\prime}|z)=\mathbb{E}_{\begin{subarray}{c}u\sim\pi(\cdot|z)\\ \tau,\omega\sim\sigma(\cdot|z,u)\\ s^{\prime}\sim\tau(\cdot|s,u)\\ x^{\prime}\sim\omega(\cdot|u,s^{\prime})\end{subarray}}\rho_{\tau,\omega}(z^{\prime}|z,u,x^{\prime}) (10)

Define Markov operator Mπ,ρ\mathbb{M}_{\pi,\rho}\inΔ(𝒵)Δ(𝒵)\Delta(\mathcal{Z})^{\Delta(\mathcal{Z})} such that for any distribution μ\mu\inΔ(𝒵)\Delta(\mathcal{Z}) : (Mπ,ρμ)(z)Ezμp(z|z)(\mathbb{M}_{\pi,\rho}\mu)(z^{\prime})\doteq\mathbb{E}_{z\sim\mu}p(z^{\prime}|z). Then

μπ,ρ(z)(1γ)t=0γtp(zt=z|z0ρ0)\mu_{\pi,\rho}(z)\doteq(1-\gamma)\textstyle\sum_{t=0}^{\infty}\gamma^{t}p(z_{t}=z|z_{0}\sim\rho_{0}) (11)

defines the occupancy measure μπ,ρ\mu_{\pi,\rho}\inΔ(𝒵)\Delta(\mathcal{Z}) for any initial (internal-state) distribution ρ0\rho_{0}, and discount rate γ[0,1)\gamma\in[0,1).

Lemma 1 (restate=forward,name=Forward Recursion)

Define the forward oper- ator Fπ,ρ:Δ(𝒵)Δ(𝒵)\mathbb{F}_{\pi,\rho}:\Delta(\mathcal{Z})\raisebox{-1.0pt}{${}^{\Delta(\mathcal{Z})}$} such that for any given μΔ(𝒵)\mu\in\Delta(\mathcal{Z}):

(Fπ,ρμ)(z)(1γ)ρ0(z)+γ(Mπ,ρμ)(z)(\mathbb{F}_{\pi,\rho}\mu)(z)\doteq(1-\gamma)\rho_{0}(z)+\gamma(\mathbb{M}_{\pi,\rho}\mu)(z) (12)

Then the occupancy μπ,ρ\mu_{\pi,\rho} is the (unique) fixed point of Fπ,ρ\mathbb{F}_{\pi,\rho}.

Backward Process  The backward (“control”) process yie- lds the value function. We want that μπ,ρ\mu_{\pi,\rho} maximize utility:

maximizeμπ,ρJπ,ρEzμπ,ρsp(|z)uπ(|z)υ(s,u)\text{maximize}_{\mu_{\pi,\rho}}J_{\pi,\rho}\doteq\mathbb{E}_{\begin{subarray}{c}z\sim\mu_{\pi,\rho}\\ s\sim p(\cdot|z)\\ u\sim\pi(\cdot|z)\end{subarray}}\upsilon(s,u) (13)

Using VV\inR𝒵\mathbb{R}^{\mathcal{Z}} to denote the multiplier, the Lagrangian is giv- en by π,ρ(μ,V)Jπ,ρV,μγMπ,ρμ(1γ)ρ0\mathcal{L}_{\pi,\rho}(\mu,V)\doteq J_{\pi,\rho}-\langle V,\mu-\gamma\mathbb{M}_{\pi,\rho}\mu-(1-\gamma)\rho_{0}\rangle.

Lemma 2 (restate=backward,name=Backward Recursion)

Define the backward o- perator Bπ,ρ:R𝒵R𝒵\mathbb{B}_{\pi,\rho}:\mathbb{R}^{\mathcal{Z}}\rightarrow\mathbb{R}^{\mathcal{Z}} such that for any given VR𝒵V\in\mathbb{R}^{\mathcal{Z}}:

(Bπ,ρV)(z)Esp(|z)uπ(|z)[υ(s,u)+Eτ,ωσ(|z,u)sτ(|s,u)xω(|u,s)zρτ,ω(|z,u,x)γV(z)]\displaystyle\begin{aligned} (\mathbb{B}_{\pi,\rho}V)(z)\doteq\mathbb{E}_{\begin{subarray}{c}s\sim p(\cdot|z)\\ u\sim\pi(\cdot|z)\end{subarray}}[\upsilon(s,u)+\mathbb{E}_{\begin{subarray}{c}\tau,\omega\sim\sigma(\cdot|z,u)\\ s^{\prime}\sim\tau(\cdot|s,u)\\ x^{\prime}\sim\omega(\cdot|u,s^{\prime})\\ z^{\prime}\sim\rho_{\tau,\omega}(\cdot|z,u,x^{\prime})\end{subarray}}\gamma V(z^{\prime})]\end{aligned} (14)

Then the (dual) optimal VV is the (unique) fixed point of Bπ,ρ\mathbb{B}_{\pi,\rho}; this is the value function considering knowledge uncertainty:

Vϕπ,ρ(z)t=0γtEstp(|zt)utπ(|zt)τ,ωσ(|zt,ut)st+1τ(|st,ut)xt+1ω(|ut,st+1)zt+1ρτ,ω(|zt,ut,xt+1)[υ(st,ut)|z0=z]V^{\phi_{\pi,\rho}}(z)\doteq\textstyle\sum_{t=0}^{\infty}\gamma^{t}\mathbb{E}_{\begin{subarray}{c}s_{t}\sim p(\cdot|z_{t})\\ u_{t}\sim\pi(\cdot|z_{t})\\ \tau,\omega\sim\sigma(\cdot|z_{t},u_{t})\\ s_{t+1}\sim\tau(\cdot|s_{t},u_{t})\\ x_{t+1}\sim\omega(\cdot|u_{t},s_{t+1})\\ z_{t+1}\sim\rho_{\tau,\omega}(\cdot|z_{t},u_{t},x_{t+1})\end{subarray}}[\upsilon(s_{t},u_{t})|z_{0}=z] (15)

so we can equivalently write targets Jπ,ρJ_{\pi,\rho}==Ezρ0Vϕπ,ρ(z)\mathbb{E}_{z\sim\rho_{0}}V^{\phi_{\pi,\rho}}(z). Likewise, we can also define the (state-action) value func- tion Qϕπ,ρQ^{\phi_{\pi,\rho}}\inR𝒵×𝒰\mathbb{R}^{\mathcal{Z}\times\mathcal{U}}—that is, Qϕπ,ρ(z,u)Q^{\phi_{\pi,\rho}}(z,u)\doteqEsp(|z)[υ(s,u)\mathbb{E}_{s\sim p(\cdot|z)}[\upsilon(s,u)++ Eτ,ωσ(|z,u),,zρτ,ω(|z,u,x)γVϕπ,ρ(z)]\mathbb{E}_{\tau,\omega\sim\sigma(\cdot|z,u),...,z^{\prime}\sim\rho_{\tau,\omega}(\cdot|z,u,x^{\prime})}\gamma V^{\phi_{\pi,\rho}}(z^{\prime})] given an action.

Table 4: Boundedly Rational Agents. Formulation of common decision agents as instantiations of our (boundedly rational) formalism. Note that either β1\beta^{-1}\rightarrow0 or σ~\tilde{\sigma}==δ\delta is sufficient to guarantee z,u:σ(|z,u)\forall z,u:\sigma(\cdot|z,u)==σ~\tilde{\sigma}.Softmax added on top of deterministic, optimal QQ-functions.

  Boundedly Rational Agent Flexibility Optimism Adaptivity (Action Prior) (Model Prior) (Belief Prior)   Observability   Examples α1\alpha^{-1} β1\beta^{-1} η1\eta^{-1} π~\tilde{\pi} σ~\tilde{\sigma} ϱ~\tilde{\varrho} Uniformly Random Agent 0\rightarrow~{}0~{} 0~{}\rightarrow~{}0~{} \rightarrow±\scalebox{0.7}{$\pm$}\infty~{} Uniform Dirac δ\delta - Full/Partial - Deterministic, Optimal Agent \rightarrow\infty 0~{}\rightarrow~{}0~{} \rightarrow±\scalebox{0.7}{$\pm$}\infty~{} - Dirac δ\delta - Full/Partial (any) Boltzmann-Exploratory Agent \rightarrow\infty 0~{}\rightarrow~{}0~{} \rightarrow±\scalebox{0.7}{$\pm$}\infty~{} - Dirac δ\delta - Full/Partial (heess2013actor, ; osband2016generalization, ; cesa2017boltzmann, ) Minimum-Information Agent =1=~{}\kern 0.90005pt1~{} 0~{}\rightarrow~{}0~{} =1=~{}~{}1~{}~{} (any) Dirac δ\delta (any) Full (globerson2009minimum, ; tishby2011information, ; ortega2013thermodynamics, ) Maximum Entropy Agent (0,)~{}\kern 0.90005pt(0,\infty) 0~{}\rightarrow~{}0~{} \rightarrow±\scalebox{0.7}{$\pm$}\infty~{} Uniform Dirac δ\delta - Full (haarnoja2017reinforcement, ; haarnoja2018soft, ; eysenbach2019if, ; shi2019soft, ) (Action) KL-Regularized Agent (0,)~{}\kern 0.90005pt(0,\infty) 0~{}\rightarrow~{}0~{} \rightarrow±\scalebox{0.7}{$\pm$}\infty~{} (any) Dirac δ\delta - Full (rubin2012trading, ; galashov2019information, ; ho2020efficiency, ; tiomkin2017unified, ; leibfried2017information, ) KL-Penalized Robust Agent \rightarrow\infty (,0)(-\infty,0)~{}\kern 0.90005pt\kern 0.90005pt \rightarrow±\scalebox{0.7}{$\pm$}\infty~{} - (any) - Full (petersen2000minimax, ; charalambous2004relations, ; osogami2012robustness, ; grau2016planning, ) General Formulation R{0}\mathbb{R}\setminus\{0\} R{0}\mathbb{R}\setminus\{0\} R{0}\mathbb{R}\setminus\{0\} (any) (any) (any) Full/Partial Section 4

4.2 Bounded Rational Control

For perfectly rational agents, the best decision policy given any zz simply maximizes Vϕπ,ρ(z)V^{\phi_{\pi,\rho}}(z), thus it selects actions according to argmaxuQϕπ,ρ(z,u)\text{argmax}_{u}Q^{\phi_{\pi,\rho}}(z,u). And the best recognition policy simply corresponds to their unbiased knowledge of the world, thus it sets σ(|z,u)=σ~,z,u\sigma(\cdot|z,u)=\tilde{\sigma},\forall z,u (in Equation 9).

Information Constraints  But control is resource-intensive. We formalize an agent’s boundedness in terms of capacities for processing information. First, decision complexity captures the informational effort in determining actions π(|z)\pi(\cdot|z), relative to some prior π~\tilde{\pi} (e.g. baseline clinical guidelines):

Iπ,ρ[π;π~]Ezμπ,ρDKL(π(|z)π~)\mathbb{I}_{\pi,\rho}[\pi;\tilde{\pi}]\doteq\mathbb{E}_{z\sim\mu_{\pi,\rho}}D_{\text{KL}}(\pi(\cdot|z)\|\tilde{\pi}) (16)

Second, specification complexity captures the average regret of their internal model σ(|z,u)\sigma(\cdot|z,u) deviating from their prior (i.e. unbiased knowledge σ~\tilde{\sigma}) about environment dynamics:

Iπ,ρ[σ;σ~]Ezμπ,ρuπ(|z)DKL(σ(|z,u)σ~)\mathbb{I}_{\pi,\rho}[\sigma;\tilde{\sigma}]\doteq\mathbb{E}_{\begin{subarray}{c}z\sim\mu_{\pi,\rho}\\ u\sim\pi(\cdot|z)\end{subarray}}D_{\text{KL}}(\sigma(\cdot|z,u)\|\tilde{\sigma}) (17)

Finally, recognition complexity captures the statistical surprise in adapting to successive beliefs about the partially-ob- servable states of the world (again, relative to some prior ϱ~\tilde{\varrho}):

Iπ,ρ[ϱ;ϱ~]Ezμπ,ρuπ(|z)τ,ωσ(|z,u)DKL(ϱτ,ω(|z,u)ϱ~)~{}~{}~{}~{}~{}\mathbb{I}_{\pi,\rho}[\varrho;\tilde{\varrho}]\doteq\mathbb{E}_{\begin{subarray}{c}z\sim\mu_{\pi,\rho}\\ u\sim\pi(\cdot|z)\\ \tau,\omega\sim\sigma(\cdot|z,u)\end{subarray}}D_{\text{KL}}(\varrho_{\tau,\omega}(\cdot|z,u)\|\tilde{\varrho}) (18)

where ϱτ,ω(|z,u)Esp(|z),sτ(|s,u),xω(|u,s)ρτ,ω(\varrho_{\tau,\omega}(\cdot|z,u)\doteq\mathbb{E}_{s\sim p(\cdot|z),s^{\prime}\sim\tau(\cdot|s,u),x^{\prime}\sim\omega(\cdot|u,s^{\prime})}\rho_{\tau,\omega}(\cdot z,u,x)z,u,x^{\prime}) gives the internal-state update. We shall see, these measures generalize information-theoretic ideas in control.

Backward Process  With capacity constraints, the maximi- zation in Equation 13 now becomes subject to Iπ,ρ[π;π~]\mathbb{I}_{\pi,\rho}[\pi;\tilde{\pi}]\leq AA, Iπ,ρ[σ;σ~]B\mathbb{I}_{\pi,\rho}[\sigma;\tilde{\sigma}]\leq B, and Iπ,ρ[ϱ;ϱ~]C\mathbb{I}_{\pi,\rho}[\varrho;\tilde{\varrho}]\leq C. So the Lagrangian (now with the additional multipliers α,β,η\alpha,\beta,\eta\inR\mathbb{R}) is given by π,ρ(μ,α,β,η,V)\mathcal{L}_{\pi,\rho}(\mu,\alpha,\beta,\eta,V)\doteqJπ,ρJ_{\pi,\rho}-V,μ\langle V,\mu-γMπ,ρμ\gamma\mathbb{M}_{\pi,\rho}\mu-(1(1-γ)ρ0\gamma)\rho_{0}\rangle- α\alpha\cdot(Iπ,ρ[π;π~](\mathbb{I}_{\pi,\rho}[\pi;\tilde{\pi}]-A)βA)-\beta\cdot(Iπ,ρ[σ;σ~](\mathbb{I}_{\pi,\rho}[\sigma;\tilde{\sigma}]-B)ηB)-\eta\cdot(Iπ,ρ[ϱ;ϱ~](\mathbb{I}_{\pi,\rho}[\varrho;\tilde{\varrho}]-C)C).

Proposition 3 (restate=backwardx,name=Backward Recursion)

Define the backwa- rd operator Bπ,ρ:R𝒵\mathbb{B}_{\pi,\rho}:\mathbb{R}^{\mathcal{Z}}\rightarrowR𝒵\mathbb{R}^{\mathcal{Z}} such that for any given function VV\inR𝒵\mathbb{R}^{\mathcal{Z}} and for any given coefficient values α,β,η\alpha,\beta,\eta\inR\mathbb{R}:

(Bπ,ρV)(z)Esp(|z)uπ(|z)[αlogπ(u|z)π~(u)+υ(s,u)+Eτ,ωσ(|z,u)[βlogσ(τ,ω|z,u)σ~(τ,ω)+Esτ(|s,u)xω(|u,s)zρτ,ω(|z,u,x)[ηlog

ϱτ,ω(|zz,u)~ϱ(z)

+γV(z)]]]
\displaystyle\begin{aligned} (\mathbb{B}_{\pi,\rho}V)(z)\doteq~{}\mathbb{E}_{\begin{subarray}{c}s\sim p(\cdot|z)\\ u\sim\pi(\cdot|z)\end{subarray}}&\big{[}-\alpha\log\scalebox{1.2}{$\frac{\pi(u|z)}{\tilde{\pi}(u)}$}+\upsilon(s,u)+\\[-8.61108pt] \mathbb{E}_{\tau,\omega\sim\sigma(\cdot|z,u)}&\big{[}-\beta\log\scalebox{1.2}{$\frac{\sigma(\tau,\omega|z,u)}{\tilde{\sigma}(\tau,\omega)}$}+\\[-3.44444pt] \mathbb{E}_{\begin{subarray}{c}s^{\prime}\sim\tau(\cdot|s,u)\\ x^{\prime}\sim\omega(\cdot|u,s^{\prime})\\ z^{\prime}\sim\rho_{\tau,\omega}(\cdot|z,u,x^{\prime})\end{subarray}}&\big{[}-\eta\log\raisebox{-2.0pt}{\scalebox{1.2}{$\frac{\varrho_{\tau,\omega}(z^{\prime}|z,u)}{\tilde{\varrho}(z^{\prime})}$}}\raisebox{-13.0pt}{$+\kern 1.00006pt\gamma V(z^{\prime})\big{]}\big{]}\big{]}$}\end{aligned}
(19)

Then the (dual) optimal VV is the (unique) fixed point of Bπ,ρ\mathbb{B}_{\pi,\rho}; as before, this is the value function Vϕπ,ρV^{\phi_{\pi,\rho}}—which now in- cludes the complexity terms. Likewise, we can also define the (state-action) Qϕπ,ρQ^{\phi_{\pi,\rho}}\inR𝒵×𝒰\mathbb{R}^{\mathcal{Z}\times\mathcal{U}} as the 1/3\nicefrac{{1}}{{3}}-step-ahead expectation, and the (state-action-model) Kϕπ,ρK^{\phi_{\pi,\rho}}\inR\mathbb{R}Z×U×T×O as the 2/3\nicefrac{{2}}{{3}}-steps-ahead expectation (which is new in this setup).

Policies and Values  The (dis-)/utility-seeking decision policy (min-)/maximizes Vϕπ,ρ(z)V^{\phi_{\pi,\rho}}(z), and a pessimistic/optimis- tic recognition policy min-/maximizes Qϕπ,ρ(z,u)Q^{\phi_{\pi,\rho}}(z,u) via σ\sigma.555In general, flipping the direction of optimization for π\pi or ρ\rho corre- sponds to the signs of α\alpha or β\beta, but does not change Theorems 45. These optimal policies depend on optimal value functions:

Theorem 4 (restate=values,name=Boundedly Rational Values)

Define the bac- kward operator B:R𝒵R𝒵\mathbb{B}^{*}:\mathbb{R}^{\mathcal{Z}}\rightarrow\mathbb{R}^{\mathcal{Z}} such that for any VR𝒵V\in\mathbb{R}^{\mathcal{Z}}:

(BV)(z)αlogEuπ~exp(1αQ(z,u))Q(z,u)βlogEτ,ωσ~exp(1βK(z,u,τ,ω))+Esp(|z)υ(s,u)K(z,u,τ,ω)Esp(|z)sτ(|s,u)xω(|u,s)zρτ,ω(|z,u,x)[ηlogϱτ,ω(z|z,u)ϱ~(z)+γV(z)]\displaystyle\begin{aligned} (\mathbb{B}^{*}&V)(z)\doteq\alpha\log\mathbb{E}_{u\sim\tilde{\pi}}\exp(\tfrac{1}{\alpha}Q(z,u))\\[-1.07639pt] &Q\hskip 0.65pt(z,u\hskip 0.5pt)\hskip 0.5pt\doteq\hskip 1.0pt\beta\log\mathbb{E}_{\tau,\omega\sim\tilde{\sigma}}\exp(\tfrac{1}{\beta}K(z,u,\tau,\omega))\raisebox{-15.0pt}{$+~{}\mathbb{E}_{s\sim p(\cdot|z)}\upsilon(s,u)$}\\[-13.99301pt] &K(z,u,\tau,\omega)\doteq\\[-2.15277pt] &\kern 1.00006pt~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}\mathbb{E}_{\begin{subarray}{c}s\sim p(\cdot|z)\\ s^{\prime}\sim\tau(\cdot|s,u)\\ x^{\prime}\sim\omega(\cdot|u,s^{\prime})\\ z^{\prime}\sim\rho_{\tau,\omega}(\cdot|z,u,x^{\prime})\end{subarray}}\big{[}-\eta\log\raisebox{-2.0pt}{\scalebox{1.2}{$\frac{\varrho_{\tau,\omega}(z^{\prime}|z,u)}{\tilde{\varrho}(z^{\prime})}$}}+\kern 1.00006pt\gamma V(z^{\prime})\big{]}\end{aligned} (20)

Then the boundedly rational value function VV^{*} for the (primal) optimal π,ρ\pi^{*},\rho^{*} is the (unique) fixed point of Bπ,ρ\mathbb{B}_{\pi,\rho}^{*}. (Note that both QQ^{*} and KK^{*} are immediately obtainable from this).

Theorem 5 (restate=policies,name=Boundedly Rational Policies)

The bounded- ly rational decision policy (i.e. primal optimal) is given by:

π(u|z)=π~(u)ZQ(z)exp(1αQ(z,u))\pi^{*}(u|z)=\scalebox{1.2}{$\frac{\tilde{\pi}(u)}{Z_{Q^{*}}(z)}$}\exp\big{(}\tfrac{1}{\alpha}Q^{*}(z,u)\big{)} (21)

and the boundedly rational recognition policy is given by:

ρ(z|z,u,x)=Eτ,ωσ(|z,u)ρτ,ω(z|z,u,x) , where σ(τ,ω|z,u)σ~(τ,ω)ZK(z,u)exp(1βK(z,u,τ,ω))\displaystyle\begin{aligned} \rho^{*}(z^{\prime}|z,u,x^{\prime})=&\hskip 2.0pt\mathbb{E}_{\tau,\omega\sim\sigma^{*}(\cdot|z,u)}\rho_{\tau,\omega}(z^{\prime}|z,u,x^{\prime})\text{ , where~{}~{}~{}}\\[-2.15277pt] \sigma^{*}(\tau,\omega|z,u)&\doteq\scalebox{1.2}{$\frac{\tilde{\sigma}(\tau,\omega)}{Z_{K^{*}}(z,u)}$}\exp\big{(}\tfrac{1}{\beta}K^{*}(z,u,\tau,\omega)\big{)}\end{aligned} (22)

where ZQ(z)=Euπ~exp(1αQ(z,u))Z_{Q^{*}}(z)=\mathbb{E}_{u\sim\tilde{\pi}}\exp(\tfrac{1}{\alpha}Q^{*}(z,u)) and ZK(z,u)=Z_{K^{*}}(z,u)= Eτ,ωσ~exp(1βK(z,u,τ,ω))\mathbb{E}_{\tau,\omega\sim\tilde{\sigma}}\exp(\tfrac{1}{\beta}K^{*}(z,u,\tau,\omega)) give the partition functions.

Interpretation of Parameters  This articulation of bound- ed rationality reflects the fact that imperfect behavior results from two sources of “boundedness”: Firstly, that (1) given a mental model ρ\rho for comprehending the world, an agent’s information-processing capacities distort their decision-ma- king π\pi (cf. suboptimal actions); and secondly, that (2) the agent’s mental model ρ\rho itself is an imperfect characteriza- tion of the world—because prior knowledge σ~\tilde{\sigma} is uncertain, and internal states can be biased by σ\sigma (cf. subjective beliefs).

Concretely, the parameters in Theorems 45 admit intuitive interpretations. First, α1\alpha^{-1} captures flexibility of decision-making, from a completely inflexible agent (α1\alpha^{-1}\rightarrow0) to an infinitely flexible, utility-seeking (α1\alpha^{-1}\rightarrow\infty) or disutility-seeking (α1\alpha^{-1}\rightarrow-\infty) one. Second, β1\beta^{-1} captures optimism in internal models, from a completely neutral agent (β1\beta^{-1}\rightarrow 0) to an infinitely optimistic (β1\beta^{-1}\rightarrow\infty) or pessimistic (β1\beta^{-1} \rightarrow-\infty) one. Lastly, η1\eta^{-1} captures adaptivity of beliefs, from a perfectly adaptive agent (η1\eta^{-1}±\rightarrow\pm\infty) to one with infinite intolerance (η1\eta^{-1}0+\rightarrow 0^{+}) or affinity (η1\eta^{-1}0\rightarrow 0^{-}) for surprise. Table 4 underscores the generality of this parameterization.

4.3 Inverse Bounded Rational Control

We hark back to our framework of Section 3: In bounded rational control (“BRC”), the planning parameter θBRC\theta^{\scalebox{0.6}{\kern 0.70004pt\text{BRC}}} represents {υ,γ,α,β,η,π~,σ~,ϱ~}\{\upsilon,\gamma,\alpha,\beta,\eta,\tilde{\pi},\tilde{\sigma},\tilde{\varrho}\}, and the space ΘBRC\Theta^{\scalebox{0.6}{\kern 0.70004pt\text{BRC}}} is again decomposable as ΘnormBRC\Theta_{\text{norm}}^{\scalebox{0.6}{\kern 0.70004pt\text{BRC}}}×\timesΘdescBRC\Theta_{\text{desc}}^{\scalebox{0.6}{\kern 0.70004pt\text{BRC}}}. The forward problem is encapsulated by Theorems 45 (which also yield a straight- forward algorithm, i.e. iterate 4 until convergence, then plug into 5). Therefore the forward planner is given as follows:

FθnormBRC(θdescBRC)ϕπ,ρ:π,ρTheorems 45F_{\theta_{\text{norm}}^{\scalebox{0.425}{\kern 0.50003pt\text{BRC}}}}(\theta_{\text{desc}}^{\scalebox{0.6}{\kern 0.70004pt\text{BRC}}})\doteq\phi_{\pi^{*},\rho^{*}}\hskip 1.0pt:\hskip 2.0pt\pi^{*},\rho^{*}\hskip 1.0pt\leftarrow\text{Theorems \ref{thm:values}--\ref{thm:policies}} (23)

In the opposite direction, the problem is of inverse bounded rational control. Consider a minimal setting where we are given access to logged data 𝒟{hnϕdemo}n=1N\mathcal{D}\doteq\{h_{n}\sim\phi_{\text{demo}}\}_{n=1}^{N} with no additional annotations. While several options from Table 3 are available, for simplicity we select soft policy matching for illustration. Thus the inverse planner is given as follows:

GθnormBRC(ϕ)argminθdescBRCEhϕlogPϕimit(u0:Tx0:T)G_{\theta_{\text{norm}}^{\scalebox{0.425}{\kern 0.50003pt\text{BRC}}}}(\phi\hskip 0.5pt)\doteq\text{argmin}_{\theta_{\text{desc}}^{\scalebox{0.425}{\kern 0.50003pt\text{BRC}}}}\mathbb{E}_{h\sim\phi}\log\mathbb{P}_{\phi_{\text{imit}}}(u_{0:T}\|x_{0:T}) (24)

where Pϕπ,ρ(u0:Tx0:T)\mathbb{P}_{\phi_{\pi,\rho}}(u_{0:T}\|x_{0:T}) is the causally-conditioned probability (ziebart2010thesis, ; kramer1998thesis, ; massey1990causality, ; marko1973bidirectional, ) t=0TPϕπ,ρ(ut|x1:t,u1:t1)\prod_{t=0}^{T}\mathbb{P}_{\phi_{\pi,\rho}}(u_{t}|x_{1:t},u_{1:t-1})—with the conditioning as induced by π,ρ\pi,\rho. In the most general case where ρτ,ω\rho_{\tau,\omega} may be stochastic, GθnormBRCG_{\theta_{\text{norm}}^{\scalebox{0.425}{\kern 0.50003pt\text{BRC}}}} would require an EM approach; however, since we selected ρτ,ω\rho_{\tau,\omega} to be the (deterministic) Bayes update for interpretability, the likelihood is:

logPϕπ,ρ(u0:Tx0:T)t=0Tlogπ(ut|zt)\begin{split}\log\mathbb{P}_{\phi_{\pi,\rho}}(u_{0:T}\|x_{0:T})&\propto\textstyle\sum_{t=0}^{T}\log\pi(u_{t}|z_{t})\end{split} (25)

where the ztz_{t} terms are computed recursively by ρ\rho (see Appendix C). Finally, here the inverse decision model of any ϕdemo\phi_{\text{demo}} is given by its projection ϕimit=FθnormBRCGθnormBRC(ϕdemo)\phi_{\text{imit}}^{*}=F_{\theta_{\text{norm}}^{\scalebox{0.425}{\kern 0.50003pt\text{BRC}}}}\circ G_{\theta_{\text{norm}}^{\scalebox{0.425}{\kern 0.50003pt\text{BRC}}}}(\phi_{\text{demo}}) onto the space ΦθnormBRC\Phi_{\theta_{\text{norm}}^{\scalebox{0.425}{\kern 0.50003pt\text{BRC}}}} of behaviors thereby interpretably para- meterized—i.e. by the structure we designed for ΘBRC\Theta^{\scalebox{0.6}{\kern 0.70004pt\text{BRC}}}, and by the normative standards θnormBRC\theta_{\text{norm}}^{\scalebox{0.6}{\kern 0.70004pt\text{BRC}}} we may choose to specify.

5 Illustrative Use Case

Refer to caption
(a)         Very Flexible Agent: α\alpha==10310^{-3}
Refer to caption
(b)      Optimistic Agent: β\beta==1.251.25
Refer to caption
(c)      Adaptive Agent: η\eta==10310^{-3}
Refer to caption
(d)        Inflexible Agent: α\alpha==1010
Refer to caption
(e)      Pessimistic Agent: β\beta==0.75-0.75
Refer to caption
(f)      Non-adaptive Agent: η\eta==7575

 

(g)

(a) Effect of Flexibility, for a neutral

   (β\beta==10310^{3}), adaptive (η\eta==10310^{-3}) agent

 

(h)

(b) Effect of Optimism, for a flexible

   (α\alpha==0.50.5), adaptive (η\eta==10310^{-3}) agent

 

(i)

(c) Effect of Adaptivity, for a flexible

   (α\alpha==0.50.5), neutral (β\beta==10310^{3}) agent

Figure 2: Bounded Rational Control. Decision agents in DIAG: In each panel, the boundedly rational decision policy π\pi is shown in terms of action probabilities (yy-axis) for different subjective beliefs (xx-axis). To visualize the boundedly rational recognition policy ρ\rho, each panel shows an example trajectory of beliefs (z0,z1,z2,z3z_{0},z_{1},z_{2},z_{3}) for the case where three consecutive positive outcomes are observed ( markers).

So far, we have argued for a systematic, unifying perspective on inverse decision modeling (“IDM”) for behavior representation learning, and presented inverse bounded rational control (“IBRC”) as a concrete example of the formalism. Three aspects of this approach deserve empirical illustration:

  • Interpretability\contourwhiteInterpretability: IBRC gives a transparent parameterization of behavior that can be successfully learned from data.

  • Expressivity\contourwhiteExpressivity: IBRC more finely differentiates between imperfect behaviors, while standard reward learning cannot.

  • Applicability\contourwhiteApplicability: IDM can be used in real-world settings, as an investigative device for understanding human decisions.

Normative-Descriptive Questions  Consider medical diagnosis, where there is often remarkable regional, institutional, and subgroup-level variability in practice (mckinlay2007sources, ; bock2016preoperative, ; osullivan2018variation, ), rendering detection and quantification of biases crucial (song2010regional, ; martin2017routine, ; allen2017unnecessary, ). Now in modeling an agent’s behavior, reward learning asks: (1) “What does this (perfectly rational) agent appear to be optimizing?” And the answer takes the form of a function υ\upsilon. However, while υ\upsilon alone is often sufficient as an intermediary for imitation/apprenticeship, it is seldom what we actually want as an end by itself—for introspective understanding. Importantly, we often can articulate some version of what our preferences υ\upsilon are. In medical diagnosis, for instance, from the view of an investigator, the average relative healthcare cost/benefit of in-/correct diagnoses is certainly spe- cifiable as a normative standard. So instead, we wish to ask: (2) “Given that this (boundedly rational) agent should optimize this υ\upsilon, how suboptimally do they appear to behave?” Clearly, such normative-descriptive questions are only poss- ible with the generalized perspective of IDM (and IBRC): Here, υ\upsilon is specified (in θnorm\theta_{\text{norm}}), whereas one or more behavio- ral parameters α,β,η\alpha,\beta,\eta are what we wish to recover (in θdesc\theta_{\text{desc}}).

Decision Environments  For our simulated setting (DIAG), we consider a POMDP where patients are diseased (s+s_{\raisebox{1.0pt}{\scalebox{0.6}{+}}}) or healthy (ss_{\raisebox{1.0pt}{\scalebox{0.6}{--}}}), and vital-signs measurements taken at each step are noisily indicative of being disease-positive (x+x_{\raisebox{1.0pt}{\scalebox{0.6}{+}}}) or negative (xx_{\raisebox{1.0pt}{\scalebox{0.6}{--}}}). Actions consist of the decision to continue monitoring the patient (u=u_{\raisebox{1.0pt}{\scalebox{0.6}{=}}})—which yields evidence, but is also costly; or stopping and declaring a final diagnosis—and if so, a diseased (u+u_{\raisebox{1.0pt}{\scalebox{0.6}{+}}}) or healthy (uu_{\raisebox{1.0pt}{\scalebox{0.6}{--}}}) call. Importantly, note that since we simulate τ,ω\tau,\omega\simσ(|z,u)\sigma(\cdot|z,u), DIAG is a strict generalization of the diagnostic environment from (huyuk2021explaining, ) with a point-valued, subjective τ,ω\tau,\omega\neqτenv,ωenv\tau_{\text{env}},\omega_{\text{env}}, and of the classic Tiger Problem in POMDP literature where τ,ω\tau,\omega==τenv,ωenv\tau_{\text{env}},\omega_{\text{env}} (kaelbling1998planning, ).

For our real-world setting, we consider 6-monthly clinical data for 1,737 patients in the Alzheimer’s Disease Neuroim- aging Initiative (marinescu2018tadpole, ) study (ADNI). The state space consists of normal function (snorms_{\text{norm}}), mild cognitive impairment (smilds_{\text{mild}}), and dementia (sdems_{\text{dem}}). For the action space, we consider ordering/not ordering an MRI—which yields evidence, but is costly. Results are classified per hippocampal volume: average (xavgMRIx_{\text{avg}}^{\scalebox{0.6}{\kern 0.70004pt\text{MRI}}}), high (xhighMRIx_{\text{high}}^{\scalebox{0.6}{\kern 0.70004pt\text{MRI}}}), low (xlowMRIx_{\text{low}}^{\scalebox{0.6}{\kern 0.70004pt\text{MRI}}}), not ordered (xnoneMRIx_{\text{none}}^{\scalebox{0.6}{\kern 0.70004pt\text{MRI}}}); separately, the cognitive dementia rating test result—which is always measured—is classified as normal (xnormCDRx_{\text{norm}}^{\scalebox{0.6}{\kern 0.70004pt\text{CDR}}}), questionable impairment (xquesCDRx_{\text{ques}}^{\scalebox{0.6}{\kern 0.70004pt\text{CDR}}}), and suspected dementia (xsuspCDRx_{\text{susp}}^{\scalebox{0.6}{\kern 0.70004pt\text{CDR}}}). So the observation space consists of such 12 combinations.

In DIAG, our normative specification (for υ\upsilon) is that diagnos- tic tests cost -11, correct diagnoses award 1010, incorrect -3636, and γ\gamma==0.950.95. Accuracies are 70%70\% in both directions (ωenv\omega_{\text{env}}), and patients arrive in equal proportions (τenv\tau_{\text{env}}). But this is un- known to the agent: We simulate σ~\tilde{\sigma} by discretizing the space of models such that probabilities vary in ±10%\pm 10\% increments from the (highest-likelihood) truth. In ADNI, the configuration is similar—except each MRI costs 1-1, while 2.52.5 is awarded once beliefs reach >>90%90\% certainty in any direction; also, σ~\tilde{\sigma} is centered at the IOHMM learned from the data. For simplicity, for π~,ϱ~\tilde{\pi},\tilde{\varrho} we use uniform priors in both settings.

Computationally, inference is performed via MCMC in log-parameter space (i.e. logα,logβ,logη\log\alpha,\log\beta,\log\eta) using standard methods, similar to e.g. Bayesian IRL (ramachandran2007bayesian, ; dimitrakakis2011bayesian, ; makino2012apprenticeship, ). In DIAG, we use 1,000 generated trajectories as basis for learning. Appendix B provides further details on experimental setup.

Refer to caption
(a)

(a) Learned α\alpha for Various Flexibility Levels

Refer to caption
(b)

(b) Learned β,η\beta,\eta for Non-adaptive Behavior

Refer to caption
(c)

(c) Learned β,η\beta,\eta for Optimistic Behavior

Figure 3: Inverse Bounded Rational Control. (a) Posteriors of α\alpha learned from extremely flexible (αtrue\alpha^{\text{true}}==10510^{-5}), flexible (αtrue\alpha^{\text{true}}==0.50.5), and inflexible (αtrue\alpha^{\text{true}}==1010) behaviors (with β,η\beta,\eta fixed as neutral and adaptive; similar plots can be obtained for those as well). (b) Joint posterior of β,η\beta,\eta for neutral but non-adaptive behavior (βtrue\beta^{\text{true}}==103,ηtrue10^{3},\eta^{\text{true}}==7575), and for (c) optimistic but adaptive behavior (βtrue\beta^{\text{true}}==1.25,ηtrue1.25,\eta^{\text{true}}==10310^{-3}).

5.1. Interpretability  Figure 2 verifies (for DIAG) that different BRC behaviors accord with our intuitions. First, cet- eris paribus, the flexibility (α\alpha) dimension manifests in how deterministically/stochastically optimal actions are selected (cf. willingness to deviate from action prior π~\tilde{\pi}): This is the notion of behavioral consistency (karni1991behavioral, ) in psychology. Sec- ond, the optimism (β\beta) dimension manifests in the illusion that diagnostic tests are more/less informative for subjective beliefs (cf. willingness to deviate from knowledge prior σ~\tilde{\sigma}): This is the phenomenon of over-/underreaction (daniel1998investor, ). Third, the adaptivity (η\eta) dimension manifests in how much/little evidence is required for declaring a final diagnosis: This corresponds to base-rate neglect/confirmation bias (tversky1981evidential, ). Hence by learning the parameters α,β,η\alpha,\beta,\eta from data, IBRC provides an eminently interpretable example of behavior re- presentation learning—one that exercises the IDM perspective (much more than just reward learning). Taking a Baye- sian approach to the likelihood (Equation 25), Figure 3(a) verifies that—as expected—IBRC is capable of recovering different parameter values from their generated behaviors.

5.2. Expressivity  Consider (i.) an agent who is biased tow- ards optimism, but otherwise flexible and adaptive (Figure 2(b), top), and (ii.) an agent who is non-adaptive, but otherwise flexible and neutral (2(c), bottom). Now, to an external observer, both types of boundedness lead to similar styles of behavior: They both tend to declare final diagnoses earlier than a neutral and adaptive agent would (2(c), top)—that is, π(u+|z)\pi(u_{\raisebox{1.0pt}{\scalebox{0.6}{+}}}|z)\approx11 after only 2 (not 3) positive tests. Of course, the former does so due to overreaction (evaluating the evidence incorrectly), whereas the latter does so due to a lower threshold for stopping (despite correctly evaluating the evidence). As shown by Figures 3(b)–(c), IBRC does differentiate between these two different types of biased behaviors: This is revealing, if not necessarily surprising. Crucially, however, this distinction is not possible with conventional IRL. All else equal, let us perform Bayesian IRL on the very same behaviors—that is, to learn an effectively skewed υ\upsilon (while implicitly setting α,β,η\alpha,\beta,\eta to their perfectly rational limits). As it turns out, the recovered υ\upsilon for (i.) gives a cost-benefit ratio (of incorrect/correct diagnoses) of 2.70-2.70±\pm0.310.31, and the recovered υ\upsilon for (ii.) gives a ratio of 2.60-2.60±\pm0.290.29. Both agents appear to penalize incorrect diagnoses much less than the normative specification of 3.60-3.60, which is consistent with them tending to commit to final diagnoses earlier than they should. However, this fails to differentiate between the two distinct underlying reasons for behaving in this manner.

5.3. Applicability  Lastly, we highlight the potential utility of IDM in real-world settings as an investigative device for auditing and understanding human decision-making. Consider diagnostic patterns for identifying dementia in ADNI, for patients from different risk groups. For instance, we discover that while β\beta==3.863.86 for all patients, clinicians appear to be significantly less optimistic when diagnosing patients with the ApoE4 genetic risk factor (β\beta==601.74601.74), for female patients (β\beta==920.70920.70), and even more so for patients aged >>7575 (β\beta==2,265.302,265.30). Note that such attitudes toward risk factors align with prevailing medical knowledge (allan2011influence, ; artero2008risk, ; hua2010sex, ). Moreover, in addition to obtaining such agent-level interpretations of biases (i.e. using the learned parameters), we can also obtain trajectory-level interpretations of decisions (i.e. using the evolution of beliefs). Appendix D gives examples of ADNI patients using diagrams of trajectories in the belief simplex, to contextualize actions the taken by clinical decision-makers and identify potentially belated diagnoses.

6 Conclusion

In this paper, we motivated the importance of descriptive models of behavior as the bridge between normative and prescriptive decision analysis, and formalized a unifying perspective on inverse decision modeling for behavior representation learning. For future work, an important question lies in exploring differently structured parameterizations Θ\Theta that are interpretable for different purposes. After all, IBRC is only one prototype that exercises the IDM formalism more fully. Another question is to what extent different forms of the inverse problem is identifiable to begin with. For instance, it is well-known that even with perfect knowledge of a demonstrator’s policy, in single environments we can only infer utility functions up to reward shaping. Thus balancing complexity, interpretability, and identifiability of decision models would be a challenging direction of work.

Acknowledgments

We would like to thank the reviewers for their generous feedback. This work was supported by Alzheimer’s Research UK, The Alan Turing Institute under the EPSRC grant EP/N510129/1, the US Office of Naval Research, as well as the National Science Foundation under grant numbers 1407712, 1462245, 1524417, 1533983, and 1722516.

References

  • [1] Aiping Li, Songchang Jin, Lumin Zhang, and Yan Jia. A sequential decision-theoretic model for medical diagnostic system. Technology and Healthcare, 2015.
  • [2] John A Clithero. Response times in economics: Looking through the lens of sequential sampling models. Journal of Economic Psychology, 2018.
  • [3] Jan Drugowitsch, Rubén Moreno-Bote, and Alexandre Pouget. Relation between belief and performance in perceptual decision making. PloS one, 2014.
  • [4] Gregory Wheeler. Bounded rationality. SEP: Stanford Center for the Study of Language and Information, 2018.
  • [5] Thomas L Griffiths, Falk Lieder, and Noah D Goodman. Rational use of cognitive resources: Levels of analysis between the computational and the algorithmic. Topics in cognitive science, 2015.
  • [6] Tim Genewein, Felix Leibfried, Jordi Grau-Moya, and Daniel Alexander Braun. Bounded rationality, abstraction, and hierarchical decision-making: An information-theoretic optimality principle. Frontiers in Robotics and AI, 2015.
  • [7] Ned Augenblick and Matthew Rabin. Belief movement, uncertainty reduction, and rational updating. UC Berkeley-Haas and Harvard University Mimeo, 2018.
  • [8] Pedro A Ortega, Daniel A Braun, Justin Dyer, Kee-Eung Kim, and Naftali Tishby. Information-theoretic bounded rationality. arXiv preprint, 2015.
  • [9] L Robin Keller. The role of generalized utility theories in descriptive, prescriptive, and normative decision analysis. Information and Decision Technologies, 1989.
  • [10] Ludwig Johann Neumann, Oskar Morgenstern, et al. Theory of games and economic behavior. Princeton university press Princeton, 1947.
  • [11] Barbara A Mellers, Alan Schwartz, and Alan DJ Cooke. Judgment and decision making. Annual review of psychology, 1998.
  • [12] Yisong Yue and Hoang M Le. Imitation learning (presentation). International Conference on Machine Learning (ICML), 2018.
  • [13] Pieter Abbeel and Andrew Y Ng. Apprenticeship learning via inverse reinforcement learning. International conference on Machine learning (ICML), 2004.
  • [14] Dean A Pomerleau. Efficient training of artificial neural networks for autonomous navigation. Neural computation (NC), 1991.
  • [15] Michael Bain and Claude Sammut. A framework for behavioural cloning. Machine Intelligence (MI), 1999.
  • [16] Umar Syed and Robert E Schapire. Imitation learning with a value-based prior. Conference on Uncertainty in Artificial Intelligence (UAI), 2007.
  • [17] Stéphane Ross and Drew Bagnell. Efficient reductions for imitation learning. International conference on artificial intelligence and statistics (AISTATS), 2010.
  • [18] Umar Syed and Robert E Schapire. A reduction from apprenticeship learning to classification. Advances in neural information processing systems (NeurIPS), 2010.
  • [19] Stéphane Ross, Geoffrey Gordon, and Drew Bagnell. A reduction of imitation learning and structured prediction to no-regret online learning. International conference on artificial intelligence and statistics (AISTATS), 2011.
  • [20] Bilal Piot, Matthieu Geist, and Olivier Pietquin. Boosted and reward-regularized classification for apprenticeship learning. International conference on Autonomous agents and multi-agent systems (AAMAS), 2014.
  • [21] Daniel Jarrett, Ioana Bica, and Mihaela van der Schaar. Strictly batch imitation learning by energy-based distribution matching. Advances in neural information processing systems (NeurIPS), 2020.
  • [22] Alihan Hüyük, Daniel Jarrett, Cem Tekin, and Mihaela van der Schaar. Explaining by imitating: Understanding decisions by interpretable policy learning. International Conference on Learning Representations (ICLR), 2021.
  • [23] Lionel Blondé and Alexandros Kalousis. Sample-efficient imitation learning via gans. International conference on artificial intelligence and statistics (AISTATS), 2019.
  • [24] Ilya Kostrikov, Kumar Krishna Agrawal, Debidatta Dwibedi, Sergey Levine, and Jonathan Tompson. Discriminator-actor-critic: Addressing sample inefficiency and reward bias in adversarial imitation. International Conference on Learning Representations (ICLR), 2019.
  • [25] Jonathan Ho and Stefano Ermon. Generative adversarial imitation learning. Advances in neural information processing systems (NeurIPS), 2016.
  • [26] Wonseok Jeon, Seokin Seo, and Kee-Eung Kim. A bayesian approach to generative adversarial imitation learning. Advances in Neural Information Processing Systems (NeurIPS), 2018.
  • [27] Seyed Kamyar Seyed Ghasemipour, Richard Zemel, and Shixiang Gu. Understanding the relation of bc and irl through divergence minimization. ICML Workshop on Deep Generative Models for Highly Structured Data, 2019.
  • [28] Seyed Kamyar Seyed Ghasemipour, Richard Zemel, and Shixiang Gu. A divergence minimization perspective on imitation learning methods. Conference on Robot Learning (CoRL), 2019.
  • [29] Liyiming Ke, Matt Barnes, Wen Sun, Gilwoo Lee, Sanjiban Choudhury, and Siddhartha Srinivasa. Imitation learning as ff-divergence minimization. arXiv preprint, 2019.
  • [30] Liyiming Ke, Matt Barnes, Wen Sun, Gilwoo Lee, Sanjiban Choudhury, and Siddhartha Srinivasa. Imitation learning as ff-divergence minimization. International Workshop on the Algorithmic Foundations of Robotics (WAFR), 2020.
  • [31] Kee-Eung Kim and Hyun Soo Park. Imitation learning via kernel mean embedding. AAAI Conference on Artificial Intelligence (AAAI), 2018.
  • [32] Huang Xiao, Michael Herman, Joerg Wagner, Sebastian Ziesche, Jalal Etesami, and Thai Hong Linh. Wasserstein adversarial imitation learning. arXiv preprint, 2019.
  • [33] Robert Dadashi, Leonard Hussenot, Matthieu Geist, and Olivier Pietquin. Primal wasserstein imitation learning. International Conference on Learning Representations (ICLR), 2021.
  • [34] Ilya Kostrikov, Ofir Nachum, and Jonathan Tompson. Imitation learning via off-policy distribution matching. International Conference on Learning Representations (ICLR), 2020.
  • [35] Oleg Arenz and Gerhard Neumann. Non-adversarial imitation learning and its connections to adversarial methods. arXiv preprint, 2020.
  • [36] Srivatsan Srinivasan and Finale Doshi-Velez. Interpretable batch irl to extract clinician goals in icu hypotension management. AMIA Summits on Translational Science Proceedings, 2020.
  • [37] Xin Zhang, Yanhua Li, Ziming Zhang, and Zhi-Li Zhang. ff-gail: Learning ff-divergence for generative adversarial imitation learning. Advances in Neural Information Processing Systems (NeurIPS), 2020.
  • [38] Nir Baram, Oron Anschel, and Shie Mannor. Model-based adversarial imitation learning. arXiv preprint, 2016.
  • [39] Nir Baram, Oron Anschel, and Shie Mannor. Model-based adversarial imitation learning. International Conference on Machine Learning (ICML), 2017.
  • [40] Andrew Y Ng, Stuart J Russell, et al. Algorithms for inverse reinforcement learning. International conference on Machine learning (ICML), 2000.
  • [41] Umar Syed and Robert E Schapire. A game-theoretic approach to apprenticeship learning. Advances in neural information processing systems (NeurIPS), 2008.
  • [42] Umar Syed, Michael Bowling, and Robert E Schapire. Apprenticeship learning using linear programming. International conference on Machine learning (ICML), 2008.
  • [43] Edouard Klein, Matthieu Geist, and Olivier Pietquin. Batch, off-policy and model-free apprenticeship learning. European Workshop on Reinforcement Learning (EWRL), 2011.
  • [44] Takeshi Mori, Matthew Howard, and Sethu Vijayakumar. Model-free apprenticeship learning for transfer of human impedance behaviour. IEEE-RAS International Conference on Humanoid Robots, 2011.
  • [45] Donghun Lee, Srivatsan Srinivasan, and Finale Doshi-Velez. Truly batch apprenticeship learning with deep successor features. International Joint Conference on Artificial Intelligence (IJCAI), 2019.
  • [46] Bilal Piot, Matthieu Geist, and Olivier Pietquin. Bridging the gap between imitation learning and irl. IEEE transactions on neural networks and learning systems, 2017.
  • [47] Edouard Klein, Matthieu Geist, Bilal Piot, and Olivier Pietquin. Irl through structured classification. Advances in neural information processing systems (NeurIPS), 2012.
  • [48] Edouard Klein, Bilal Piot, Matthieu Geist, and Olivier Pietquin. A cascaded supervised learning approach to inverse reinforcement learning. Joint European conference on machine learning and knowledge discovery in databases (ECML), 2013.
  • [49] Aristide CY Tossou and Christos Dimitrakakis. Probabilistic inverse reinforcement learning in unknown environments. Conference on Uncertainty in Artificial Intelligence (UAI), 2013.
  • [50] Vinamra Jain, Prashant Doshi, and Bikramjit Banerjee. Model-free irl using maximum likelihood estimation. AAAI Conference on Artificial Intelligence (AAAI), 2019.
  • [51] Gergely Neu and Csaba Szepesvári. Apprenticeship learning using irl and gradient methods. Conference on Uncertainty in Artificial Intelligence (UAI), 2007.
  • [52] Monica Babes, Vukosi Marivate, and Michael L Littman. Apprenticeship learning about multiple intentions. International conference on Machine learning (ICML), 2011.
  • [53] Jonathan Ho, Jayesh Gupta, and Stefano Ermon. Model-free imitation learning with policy optimization. International Conference on Machine Learning (ICML), 2016.
  • [54] Chelsea Finn, Sergey Levine, and Pieter Abbeel. Guided cost learning: Deep inverse optimal control via policy optimization. International conference on machine learning (ICML), 2016.
  • [55] Matteo Pirotta and Marcello Restelli. Inverse reinforcement learning through policy gradient minimization. AAAI Conference on Artificial Intelligence (AAAI), 2016.
  • [56] Alberto Maria Metelli, Matteo Pirotta, and Marcello Restelli. Compatible reward inverse reinforcement learning. Advances in Neural Information Processing Systems (NeurIPS), 2017.
  • [57] Davide Tateo, Matteo Pirotta, Marcello Restelli, and Andrea Bonarini. Gradient-based minimization for multi-expert inverse reinforcement learning. IEEE Symposium Series on Computational Intelligence (SSCI), 2017.
  • [58] Gergely Neu and Csaba Szepesvári. Training parsers by inverse reinforcement learning. Machine learning (ML), 2009.
  • [59] Deepak Ramachandran and Eyal Amir. Bayesian inverse reinforcement learning. International Joint Conference on Artificial Intelligence (IJCAI), 2007.
  • [60] Jaedeug Choi and Kee-Eung Kim. Map inference for bayesian irl. Advances in Neural Information Processing Systems (NeurIPS), 2011.
  • [61] Christos Dimitrakakis and Constantin A Rothkopf. Bayesian multitask irl. European workshop on reinforcement learning (EWRL), 2011.
  • [62] Constantin A Rothkopf and Christos Dimitrakakis. Preference elicitation and inverse reinforcement learning. Joint European conference on machine learning and knowledge discovery in databases (ECML), 2011.
  • [63] Sreejith Balakrishnan, Quoc Phong Nguyen, Bryan Kian Hsiang Low, and Harold Soh. Efficient exploration of reward functions in inverse reinforcement learning via bayesian optimization. Advances in Neural Information Processing Systems (NeurIPS), 2020.
  • [64] Ajay Kumar Tanwani and Aude Billard. Inverse reinforcement learning for compliant manipulation in letter handwriting. National Center of Competence in Robotics (NCCR), 2013.
  • [65] McKane Andrus. Inverse reinforcement learning for dynamics. Dissertation, University of California at Berkeley, 2019.
  • [66] Stav Belogolovsky, Philip Korsunsky, Shie Mannor, Chen Tessler, and Tom Zahavy. Learning personalized treatments via irl. arXiv preprint, 2019.
  • [67] Sid Reddy, Anca Dragan, and Sergey Levine. Where do you think you’re going?: Inferring beliefs about dynamics from behavior. Advances in Neural Information Processing Systems (NeurIPS), 2018.
  • [68] Anirudha Majumdar, Sumeet Singh, Ajay Mandlekar, and Marco Pavone. Risk-sensitive inverse reinforcement learning via coherent risk models. Robotics: Science and Systems, 2017.
  • [69] Sumeet Singh, Jonathan Lacotte, Anirudha Majumdar, and Marco Pavone. Risk-sensitive inverse reinforcement learning via semi-and non-parametric methods. International Journal of Robotics Research, 2018.
  • [70] Jaedeug Choi and Kee-Eung Kim. Inverse reinforcement learning in partially observable environments. International Joint Conference on Artificial Intelligence (IJCAI), 2009.
  • [71] Jaedeug Choi and Kee-Eung Kim. Inverse reinforcement learning in partially observable environments. Journal of Machine Learning Research (JMLR), 2011.
  • [72] Hamid R Chinaei and Brahim Chaib-Draa. An inverse reinforcement learning algorithm for partially observable domains with application on healthcare dialogue management. International Conference on Machine Learning and Applications, 2012.
  • [73] Ioana Bica, Daniel Jarrett, Alihan Hüyük, and Mihaela van der Schaar. Learning what-if explanations for sequential decision-making. International Conference on Learning Representations (ICLR), 2021.
  • [74] Takaki Makino and Johane Takeuchi. Apprenticeship learning for model parameters of partially observable environments. International Conference on Machine Learning (ICML), 2012.
  • [75] Daniel Jarrett and Mihaela van der Schaar. Inverse active sensing: Modeling and understanding timely decision-making. International Conference on Machine Learning, 2020.
  • [76] Kunal Pattanayak and Vikram Krishnamurthy. Inverse reinforcement learning for sequential hypothesis testing and search. International Conference on Information Fusion (FUSION), 2020.
  • [77] Matthew Golub, Steven Chase, and Byron Yu. Learning an internal dynamics model from control demonstration. International Conference on Machine Learning (ICML), 2013.
  • [78] Zhengwei Wu, Paul Schrater, and Xaq Pitkow. Inverse pomdp: Inferring what you think from what you do. arXiv preprint, 2018.
  • [79] Saurabh Daptardar, Paul Schrater, and Xaq Pitkow. Inverse rational control with partially observable continuous nonlinear dynamics. arXiv preprint, 2019.
  • [80] Minhae Kwon, Saurabh Daptardar, Paul Schrater, and Xaq Pitkow. Inverse rational control with partially observable continuous nonlinear dynamics. Advances in Neural Information Processing Systems (NeurIPS), 2020.
  • [81] Brian D Ziebart, Andrew L Maas, J Andrew Bagnell, and Anind K Dey. Maximum entropy inverse reinforcement learning. AAAI Conference on Artificial Intelligence (AAAI), 2008.
  • [82] Abdeslam Boularias, Jens Kober, and Jan Peters. Relative entropy inverse reinforcement learning. International conference on artificial intelligence and statistics (AISTATS), 2011.
  • [83] Mrinal Kalakrishnan, Peter Pastor, Ludovic Righetti, and Stefan Schaal. Learning objective functions for manipulation. International Conference on Robotics and Automation (ICRA), 2013.
  • [84] Markus Wulfmeier, Peter Ondruska, and Ingmar Posner. Maximum entropy deep inverse reinforcement learning. arXiv preprint, 2015.
  • [85] Chelsea Finn, Paul Christiano, Pieter Abbeel, and Sergey Levine. A connection between generative adversarial networks, inverse reinforcement learning, and energy-based models. NeurIPS Workshop on Adversarial Training, 2016.
  • [86] Justin Fu, Katie Luo, and Sergey Levine. Learning robust rewards with adversarial inverse reinforcement learning. International Conference on Learning Representations (ICLR), 2018.
  • [87] Ahmed H Qureshi, Byron Boots, and Michael C Yip. Adversarial imitation via variational inverse reinforcement learning. International Conference on Learning Representations (ICLR), 2019.
  • [88] Paul Barde, Julien Roy, Wonseok Jeon, Joelle Pineau, Christopher Pal, and Derek Nowrouzezahrai. Adversarial soft advantage fitting: Imitation learning without policy optimization. Advances in neural information processing systems (NeurIPS), 2020.
  • [89] Brian D Ziebart, J Andrew Bagnell, and Anind K Dey. Modeling interaction via the principle of maximum causal entropy. International conference on Machine learning (ICML), 2010.
  • [90] Zhengyuan Zhou, Michael Bloem, and Nicholas Bambos. Infinite time horizon maximum causal entropy inverse reinforcement learning. IEEE Transactions on Automatic Control (TACON), 2017.
  • [91] Kyungjae Lee, Sungjoon Choi, and Songhwai Oh. Maximum causal tsallis entropy imitation learning. Advances in Neural Information Processing Systems (NeurIPS), 2018.
  • [92] Tien Mai, Kennard Chan, and Patrick Jaillet. Generalized maximum causal entropy for inverse reinforcement learning. AAAI Conference on Artificial Intelligence (AAAI), 2019.
  • [93] Michael Herman, Tobias Gindele, Jörg Wagner, Felix Schmitt, and Wolfram Burgard. Inverse reinforcement learning with simultaneous estimation of rewards and dynamics. International conference on artificial intelligence and statistics (AISTATS), 2016.
  • [94] Michael Herman. Simultaneous estimation of rewards and dynamics in irl. Dissertation, Albert-Ludwigs-Universitat Freiburg, 2016.
  • [95] Layla El Asri, Bilal Piot, Matthieu Geist, Romain Laroche, and Olivier Pietquin. Score-based inverse reinforcement learning. International conference on Autonomous agents and multi-agent systems (AAMAS), 2016.
  • [96] Benjamin Burchfiel, Carlo Tomasi, and Ronald Parr. Distance minimization for reward learning from scored trajectories. AAAI Conference on Artificial Intelligence (AAAI), 2016.
  • [97] Alexis Jacq, Matthieu Geist, Ana Paiva, and Olivier Pietquin. Learning from a learner. International Conference on Machine Learning (ICML), 2019.
  • [98] Daniel S Brown, Wonjoon Goo, Prabhat Nagarajan, and Scott Niekum. Extrapolating beyond suboptimal demonstrations via inverse reinforcement learning from observations. International Conference on Machine Learning (ICML), 2019.
  • [99] Daniel S Brown, Wonjoon Goo, and Scott Niekum. Better-than-demonstrator imitation learning via automatically-ranked demonstrations. Conference on Robot Learning (CoRL), 2020.
  • [100] Takayuki Osa, Joni Pajarinen, Gerhard Neumann, J Andrew Bagnell, Pieter Abbeel, and Jan Peters. An algorithmic perspective on imitation learning. Foundations and Trends in Robotics, 2018.
  • [101] Tuomas Haarnoja, Haoran Tang, Pieter Abbeel, and Sergey Levine. Reinforcement learning with deep energy-based policies. International Conference on Machine Learning (ICML), 2017.
  • [102] Tuomas Haarnoja, Aurick Zhou, Pieter Abbeel, and Sergey Levine. Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. International Conference on Machine Learning (ICML), 2018.
  • [103] Benjamin Eysenbach and Sergey Levine. If maxent rl is the answer, what is the question? arXiv preprint, 2019.
  • [104] Wenjie Shi, Shiji Song, and Cheng Wu. Soft policy gradient method for maximum entropy deep reinforcement learning. International Joint Conference on Artificial Intelligence (IJCAI), 2019.
  • [105] Rohin Shah, Noah Gundotra, Pieter Abbeel, and Anca Dragan. Inferring reward functions from demonstrators with unknown biases. OpenReview, 2018.
  • [106] Rohin Shah, Noah Gundotra, Pieter Abbeel, and Anca D Dragan. On the feasibility of learning, rather than assuming, human biases for reward inference. International Conference on Machine Learning (ICML), 2019.
  • [107] Jonathan Rubin, Ohad Shamir, and Naftali Tishby. Trading value and information in mdps. Decision Making with Imperfect Decision Makers (Springer), 2012.
  • [108] Alexandre Galashov, Siddhant M Jayakumar, Leonard Hasenclever, Dhruva Tirumala, Jonathan Schwarz, Guillaume Desjardins, Wojciech M Czarnecki, Yee Whye Teh, Razvan Pascanu, and Nicolas Heess. Information asymmetry in kl-regularized rl. International Conference on Learning Representations (ICLR), 2019.
  • [109] Mark K Ho, David Abel, Jonathan D Cohen, Michael L Littman, and Thomas L Griffiths. The efficiency of human cognition reflects planned information processing. AAAI Conference on Artificial Intelligence (AAAI), 2020.
  • [110] Stas Tiomkin and Naftali Tishby. A unified bellman equation for causal information and value in markov decision processes. arXiv preprint arXiv:1703.01585, 2017.
  • [111] Felix Leibfried, Jordi Grau-Moya, and Haitham Bou-Ammar. An information-theoretic optimality principle for deep reinforcement learning. NeurIPS Workshop on Deep Reinforcement Learning, 2017.
  • [112] Matthew Hausknecht and Peter Stone. Deep recurrent q-learning for partially observable mdps. AAAI Conference on Artificial Intelligence (AAAI), 2015.
  • [113] Pengfei Zhu, Xin Li, Pascal Poupart, and Guanghui Miao. On improving deep reinforcement learning for pomdps. arXiv preprint, 2017.
  • [114] Maximilian Igl, Luisa Zintgraf, Tuan Anh Le, Frank Wood, and Shimon Whiteson. Deep variational reinforcement learning for pomdps. International Conference on Machine Learning (ICML), 2018.
  • [115] Amy Zhang, Zachary C Lipton, Luis Pineda, Kamyar Azizzadenesheli, Anima Anandkumar, Laurent Itti, Joelle Pineau, and Tommaso Furlanello. Learning causal state representations of partially observable environments. arXiv preprint, 2019.
  • [116] Dongqi Han, Kenji Doya, and Jun Tani. Variational recurrent models for solving partially observable control tasks. arXiv preprint arXiv:1912.10703, 2019.
  • [117] Joseph Futoma, Michael C Hughes, and Finale Doshi-Velez. Popcorn: Partially observed prediction constrained reinforcement learning. International conference on artificial intelligence and statistics (AISTATS), 2020.
  • [118] Richard D Smallwood and Edward J Sondik. The optimal control of partially observable markov processes over a finite horizon. Operations research, 1973.
  • [119] Milos Hauskrecht. Value-function approximations for partially observable markov decision processes. Journal of Artificial Intelligence Research (JAIR), 2000.
  • [120] Joelle Pineau, Geoff Gordon, Sebastian Thrun, et al. Point-based value iteration: An anytime algorithm for pomdps. International Joint Conference on Artificial Intelligence (IJCAI), 2003.
  • [121] Hanna Kurniawati, David Hsu, and Wee Sun Lee. Sarsop: Efficient point-based pomdp planning by approximating optimally reachable belief spaces. Robotics: Science and systems, 2008.
  • [122] Mauricio Araya, Olivier Buffet, Vincent Thomas, and Françcois Charpillet. A pomdp extension with belief-dependent rewards. Advances in Neural Information Processing Systems (NeurIPS), 2010.
  • [123] Mathieu Fehr, Olivier Buffet, Vincent Thomas, and Jilles Dibangoye. rho-pomdps have lipschitz-continuous epsilon-optimal value functions. Advances in Neural Information Processing Systems (NeurIPS), 2018.
  • [124] F A Sonnenberg and J R Beck. Markov models in medical decision making: a practical guide. Health Econ., 1983.
  • [125] C H Jackson, L D Sharples, S G Thompson, S W Duffy, and E Couto. Multistate Markov models for disease progression with classification error. Statistician, 2003.
  • [126] S E O’Bryant, S C Waring, C M Cullum, J Hall, L Lacritz, P J Massman, P J Lupo, J S Reisch, and R Doody. Staging dementia using Clinical Dementia Rating Scale Sum of Boxes scores: a Texas Alzheimer’s research consortium study. Arch. of Neurology, 2008.
  • [127] D Jarrett, J Yoon, and M van der Schaar. Match-net: Dynamic prediction in survival analysis using convolutional neural networks. NeurIPS Workshop on Machine Learning for Health, 2018.
  • [128] Daniel Jarrett, Jinsung Yoon, and Mihaela van der Schaar. Dynamic prediction in clinical survival analysis using temporal convolutional networks. IEEE Journal of Biomedical and Health Informatics, 2019.
  • [129] P Petousis, A Winter, W Speier, D R Aberle, W Hsu, and A A T Bui. Using sequential decision making to improve lung cancer screening performance. IEEE Access, 2019.
  • [130] F Cardoso, S Kyriakides, S Ohno, F Penault-Llorca, P Poortmans, I T Rubio, S Zackrisson, and E Senkus. Early breast cancer: ESMO Clinical Practice Guidelines for diagnosis, treatment and follow-up. Anna. Oncology, 2019.
  • [131] A M Alaa and M van der Schaar. Attentive state-space modeling of disease progression. Advances in neural information processing systems (NeurIPS), 2019.
  • [132] X Wang, D Sontag, and F Wang. Unsupervised learning of disease progression models. ACM SIGKDD Int. Conf. Knowl. Discovery Data Mining, 2014.
  • [133] Clemens Heuberger. Inverse combinatorial optimization. Journal of combinatorial optimization, 2004.
  • [134] Kareem Amin and Satinder Singh. Towards resolving unidentifiability in inverse reinforcement learning. arXiv preprint, 2016.
  • [135] Kareem Amin, Nan Jiang, and Satinder Singh. Repeated inverse reinforcement learning. Advances in neural information processing systems (NeurIPS), 2017.
  • [136] Stuart Armstrong and Sören Mindermann. Occam’s razor is insufficient to infer the preferences of irrational agents. Advances in neural information processing systems (NeurIPS), 2018.
  • [137] Paul Christiano. The easy goal inference problem is still hard. AI Alignment, 2015.
  • [138] Eric J Michaud, Adam Gleave, and Stuart Russell. Understanding learned reward functions. NeurIPS Workshop on Deep Reinforcement Learning, 2020.
  • [139] Adam Gleave, Michael Dennis, Shane Legg, Stuart Russell, and Jan Leike. Quantifying differences in reward functions. International Conference on Learning Representations (ICLR), 2021.
  • [140] Daniel S Brown and Scott Niekum. Deep bayesian reward learning from preferences. NeurIPS Workshop on Safety and Robustness in Decision-Making, 2019.
  • [141] Daniel S Brown, Russell Coleman, Ravi Srinivasan, and Scott Niekum. Safe imitation learning via fast bayesian reward inference from preferences. International Conference on Machine Learning (ICML), 2020.
  • [142] Nicolas Heess, David Silver, and Yee Whye Teh. Actor-critic reinforcement learning with energy-based policies. European Workshop on Reinforcement Learning (EWRL), 2013.
  • [143] Ian Osband, Benjamin Van Roy, and Zheng Wen. Generalization and exploration via randomized value functions. International Conference on Machine Learning (ICML), 2016.
  • [144] Nicolò Cesa-Bianchi, Claudio Gentile, Gábor Lugosi, and Gergely Neu. Boltzmann exploration done right. Advances in neural information processing systems (NeurIPS), 2017.
  • [145] Amir Globerson, Eran Stark, Eilon Vaadia, and Naftali Tishby. The minimum information principle and its application to neural code analysis. Proceedings of the National Academy of Sciences, 2009.
  • [146] Naftali Tishby and Daniel Polani. Information theory of decisions and actions. Perception-action cycle (Springer), 2011.
  • [147] Pedro A Ortega and Daniel A Braun. Thermodynamics as a theory of decision-making with information-processing costs. Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences, 2013.
  • [148] Ian R Petersen, Matthew R James, and Paul Dupuis. Minimax optimal control of stochastic uncertain systems with relative entropy constraints. IEEE Transactions on Automatic Control, 2000.
  • [149] Charalambos D Charalambous, Farzad Rezaei, and Andreas Kyprianou. Relations between information theory, robustness, and statistical mechanics of stochastic systems. IEEE Conference on Decision and Control (CDC), 2004.
  • [150] Takayuki Osogami. Robustness and risk-sensitivity in markov decision processes. Advances in Neural Information Processing Systems (NeurIPS), 2012.
  • [151] Jordi Grau-Moya, Felix Leibfried, Tim Genewein, and Daniel A Braun. Planning with information-processing constraints and model uncertainty in markov decision processes. Joint European conference on machine learning and knowledge discovery in databases (ECML), 2016.
  • [152] Brian D Ziebart. Modeling purposeful adaptive behavior with the principle of maximum causal entropy. Dissertation, Carnegie Mellon University, 2010.
  • [153] Gerhard Kramer. Directed information for channels with feedback. Dissertation, ETH Zurich, 1998.
  • [154] James Massey. Causality, feedback and directed information. International Symposium on Information Theory and Its Applications, 1990.
  • [155] Hans Marko. The bidirectional communication theory-a generalization of information theory. IEEE Transactions on Communications, 1973.
  • [156] John B McKinlay, Carol L Link, et al. Sources of variation in physician adherence with clinical guidelines. Journal of general internal medicine, 2007.
  • [157] Matthias Bock, Gerhard Fritsch, and David L Hepner. Preoperative laboratory testing. Anesthesiology clinics, 2016.
  • [158] Jack W O’Sullivan, Carl Heneghan, Rafael Perera, Jason Oke, Jeffrey K Aronson, Brian Shine, and Ben Goldacre. Variation in diagnostic test requests and outcomes: a preliminary metric for openpathology. net. Nature Scientific Reports, 2018.
  • [159] Yunjie Song, Jonathan Skinner, Julie Bynum, Jason Sutherland, John E Wennberg, and Elliott S Fisher. Regional variations in diagnostic practices. New England Journal of Medicine, (1), 2010.
  • [160] Shannon K Martin and Adam S Cifu. Routine preoperative laboratory tests for elective surgery. Journal of the American Medical Association (JAMA), 2017.
  • [161] M. Allen. Unnecessary tests and treatment explain why health care costs so much. Scientific American, 2017.
  • [162] Leslie Pack Kaelbling, Michael L Littman, and Anthony R Cassandra. Planning and acting in partially observable stochastic domains. Artificial intelligence, 1998.
  • [163] Razvan V Marinescu, Neil P Oxtoby, Alexandra L Young, Esther E Bron, Arthur W Toga, Michael W Weiner, Frederik Barkhof, Nick C Fox, Stefan Klein, Daniel C Alexander, et al. Tadpole challenge: Prediction of longitudinal evolution in alzheimer’s disease. arXiv preprint, 2018.
  • [164] Edi Karni and Zvi Safra. Behavioral consistency in sequential decisions. Progress in Decision, Utility and Risk Theory, 1991.
  • [165] Kent Daniel, David Hirshleifer, and Avanidhar Subrahmanyam. Investor psychology and security market under-and overreactions. The Journal of Finance, 1998.
  • [166] Amos Tversky and Daniel Kahneman. Evidential impact of base rates. Stanford University Department Of Psychology, 1981.
  • [167] Charlotte L Allan and Klaus P Ebmeier. The influence of apoe4 on clinical progression of dementia: a meta-analysis. International journal of geriatric psychiatry, 2011.
  • [168] Sylvaine Artero, Marie-Laure Ancelin, Florence Portet, A Dupuy, Claudine Berr, Jean-François Dartigues, Christophe Tzourio, Olivier Rouaud, Michel Poncet, Florence Pasquier, et al. Risk profiles for mild cognitive impairment and progression to dementia are gender specific. Journal of Neurology, Neurosurgery & Psychiatry, 2008.
  • [169] Xue Hua, Derrek P Hibar, Suh Lee, Arthur W Toga, Clifford R Jack Jr, Michael W Weiner, Paul M Thompson, Alzheimer’s Disease Neuroimaging Initiative, et al. Sex and age differences in atrophic rates: an adni study with n= 1368 mri scans. Neurobiology of aging, 2010.
  • [170] Brendan O’Donoghue, Remi Munos, Koray Kavukcuoglu, and Volodymyr Mnih. Combining policy gradient and q-learning. International Conference on Learning Representations (ICLR), 2017.
  • [171] Momchil Tomov. Structure learning and uncertainty-guided exploration in the human brain. Dissertation, Harvard University, 2020.
  • [172] Tianwei Ni, Harshit Sikchi, Yufei Wang, Tejus Gupta, Lisa Lee, and Benjamin Eysenbach. F-irl: Inverse reinforcement learning via state marginal matching. Conference on Robot Learning (CoRL), 2020.
  • [173] Hong Jun Jeon, Smitha Milli, and Anca D Dragan. Reward-rational (implicit) choice: A unifying formalism for reward learning. Advances in Neural Information Processing Systems (NeurIPS), 2020.
  • [174] Jeffrey Ely, Alexander Frankel, and Emir Kamenica. Suspense and surprise. Journal of Political Economy, 2015.
  • [175] Ahmed M Alaa and Mihaela van der Schaar. Balancing suspense and surprise: Timely decision making with endogenous information acquisition. Advances in neural information processing systems (NeurIPS), 2016.
  • [176] Owain Evans and Noah D Goodman. Learning the preferences of bounded agents. NeurIPS Workshop on Bounded Optimality, 2015.
  • [177] Tan Zhi-Xuan, Jordyn L Mann, Tom Silver, Joshua B Tenenbaum, and Vikash K Mansinghka. Online bayesian goal inference for boundedly-rational planning agents. Advances in neural information processing systems (NeurIPS), 2020.
  • [178] Osbert Bastani, Yewen Pu, and Armando Solar-Lezama. Verifiable reinforcement learning via policy extraction. Advances in neural information processing systems (NeurIPS), 2018.
  • [179] Herman Yau, Chris Russell, and Simon Hadfield. What did you think would happen? explaining agent behaviour through intended outcomes. Advances in neural information processing systems (NeurIPS), 2020.
  • [180] Tom Bewley, Jonathan Lawry, and Arthur Richards. Modelling agent policies with interpretable imitation learning. TAILOR Workshop at ECAI, 2020.
  • [181] Tathagata Chakraborti, Sarath Sreedharan, Yu Zhang, and Subbarao Kambhampati. Plan explanations as model reconciliation: Moving beyond explanation as soliloquy. International Joint Conference on Artificial Intelligence (IJCAI), 2017.
  • [182] Tathagata Chakraborti, Sarath Sreedharan, Sachin Grover, and Subbarao Kambhampati. Plan explanations as model reconciliation: an empirical study. International Conference on Human-Robot Interaction (HRI), 2019.
  • [183] Bradley Hayes and Julie A Shah. Improving robot controller transparency through autonomous policy explanation. International Conference on Human-Robot Interaction (HRI), 2017.
  • [184] Sarath Sreedharan, Utkash Soni, Mudit Verma, Siddharth Srivastava, and Subbarao Kambhampati. Bridging the gap: Providing post-hoc symbolic explanations for sequential decision-making problems with black box simulators. ICML Workshop on Human-in-the-Loop Learning, 2020.
  • [185] Roy Fox and Naftali Tishby. Minimum-information lqg control part i: Memoryless controllers. IEEE Conference on Decision and Control (CDC), 2016.
  • [186] Roy Fox and Naftali Tishby. Minimum-information lqg control part ii: Retentive controllers. IEEE Conference on Decision and Control (CDC), 2016.
  • [187] Robert Babuska. Model-based imitation learning. Springer Encyclopedia of the Sciences of Learning, 2012.
  • [188] Yoshua Bengio and Paolo Frasconi. An input output hmm architecture. Advances in neural information processing systems (NeurIPS), 1995.

Appendices

Appendix A gives a longer discussion of merits and caveats; Appendix B gives further experiment details; Appendix C gives derivations of propositions; Appendix D shows illustrative trajectories; Appendix E gives a summary of notation.

Appendix A Discussion

In this paper, we motivated the importance of descriptive models of behavior as the bridge between normative and prescriptive decision analysis [9, 10, 11] (Figure 4). On account of this, we formalized a unifying perspective on inverse decis- ion modeling for behavior representation learning. Precisely, the inverse decision model of any observed behavior ϕdemo\phi_{\text{demo}} is given by its projection ϕimit=FθnormGθnorm(ϕdemo)\phi_{\text{imit}}^{*}=F_{\theta_{\text{norm}}}\circ G_{\theta_{\text{norm}}}(\phi_{\text{demo}}) onto the space Φθnorm\Phi_{\theta_{\text{norm}}} of behaviors parameterizable by the structure designed for Θ\Theta and normative standards θnorm\theta_{\text{norm}} specified. This formulation is general. For instance, it is agnostic as to the nature of agent and environment state spaces (which— among other properties—are encoded in ψ\psi); it is also agnostic as to whether the underlying forward problem is model-free or model-based (which—among other properties—is encoded in θ\theta). Per the priorities of the investigator (cf. imitation, apprenticeship, understanding, and other objectives), different choices can and should be made to balance the expressivity, interpretability, and tractability of learned models.

Partial Observability  At first glance, our choice to accommodate partial observability may have appeared inconsequential. However, its significance becomes immediately apparent once we view an agent’s behavior as induced by both a decision policy π\pi as well as a recognition policy ρ\rho, and—importantly—that not only may an agent’s mapping from internal states into actions be suboptimal (viz. the former), but that their mapping from observations into beliefs may also be subjective (viz. the latter). Therefore in addition to the oft-studied, purely utility-centric nature of (perfectly rational) behavior, this generalized formalism immediately invites consideration of (boundedly rational) behaviors—that is, agents acting under knowledge uncertainty, biased by optimism/robustness, with policies distorted by the complexit- ies of information processing required for decision-making.

Bounded Rationality  While the IDM formalism subsumes most standard approaches to imitation learning, apprenticeship learning, and reward learning (cf. Table 1 and Table 3), we emphasize that—with very few exceptions [78, 79, 80]—the vast majority of original studies in these areas are limited to cases where θdesc\theta_{\text{desc}}==υ\upsilon alone, or assume fully-observable environments (whence 𝒮\mathcal{S}==𝒳\mathcal{X}==𝒵\mathcal{Z}, and ρ\rho simply being the identity function). Therefore our concrete example of inverse bounded rational control was presented as a prototypical instantiation of IDM that much more fully exercises the flexibility afforded by this generalized perspective. Importantly, while our notion of bounded rationality has (implicitly) been present to varying degrees in (forward) control and reinforcement learning (cf. Table 2 and Table 4), “boundedness” has largely been limited to mean “noisy actions”. To be precise, we may differentiate between three “levels” of boundedness:

  • Imperfect Response: This is the shallowest form of boundedness, and includes Boltzmann-exploratory [142, 143, 144] and (locally) entropy-regularized [170] behaviors: It considers first that agents are perfect in their ability to compute the optimal values/policies; however, their actions are ultimately executed with an artificial layer of stochasticity.

  • Capacity Constraints: Given an agent’s model (e.g. τ,Q\tau,Q- network, etc.), the information processing needed in computing actions on the go is costly. We may view soft-opt- imal [101, 102, 103, 104] and KL-regularized [107, 108, 109, 110, 111] planning and learning as examples. However, these do not model subjectivity of beliefs, adaptivity, or optimism/robustness.

  • Model Imperfection: The agent’s mental model itself is systematically flawed, due to uncertainty in knowledge, and to biases from optimism or pessimism. We may view certain robust MDPs (with penalties for deviating from priors) [148, 149, 151, 150] as examples. However, these still do not account for partial observability (and biased recognition).

Figure 4: Normative, Prescriptive, and Descriptive Modeling. Re- call the “lifecycle” of decision analysis (Section 1). As a paradigm of optimal behavior, normative standards serve as a theoretical benchmark. To guide imperfect agents toward this ideal, prescriptive advice serves to engineer behavior from humans in the loop. Importantly, however, this first requires an understanding of the imperfections—relative to the normative ideal—that require correcting. This is the goal of descriptive modeling—that is, to obtain an empirical account of existing behavior from observed data. Pre-cisely, inverse decision modeling (middle) leverages a normative standard (left) to obtain an interpretable account of demonstrated behavior, thereby enabling the introspection of existing practices, which  may  inform  construction  of  prescriptive  guidelines  (right).
Refer to caption

Now in the inverse direction, imitation/apprenticeship learning has typically viewed reward learning as but an intermediary, so classical methods have worked with perfectly rational planners [40, 13, 41, 42, 43, 44, 45, 70, 71, 72, 73]. Approaches that leverage probabilistic methods have usually simply used Boltzmann-exploratory policies on top of optimal action-value functions (viz. imperfect response) [59, 60, 61, 62, 63, 51, 52, 49, 50, 74, 75], or worked within maximum entropy planning/learning frameworks (viz. capacity constraints) [81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92]. Crucially, however, the corresponding parameters (i.e. inverse temperatures) have largely been treated as pre-specified parameters for learning υ\upsilon alone—not learnable parameters of interest by themselves. In contrast, what IDM allows (and what IBRC illustrates) is the “fullest” extent of boundedness—that is, where stochas- tic actions and subjective beliefs are endogenously the result of knowledge uncertainty and information processing constr- aints. Importantly, while recent work in imitation/appren- ticeship have studied aspects of subjective dynamics that can be jointly learnable [67, 93, 94, 68, 69], they are limited to environments that are fully-observable and/or agents that have point-valued knowledge of environments—substantial simplifications that ignore how humans can and do make imperfect inferences from recognizing environment signals.

A.1 Important Distinctions

Our goal of understanding in IDM departs from the standard objectives of imitation and apprenticeship learning. As a result, some caveats and distinctions warrant special attention as pertains assumptions, subjectivity, and model accuracy.

Decision-maker vs. Investigator  As noted in Section 3.3, the design of Θ\Theta (and specification of θnorm\theta_{\text{norm}}) are not assumptions: We are not making “factual” claims concerning the underlying psychological processes that govern human behavior; these are hugely complex, and are the preserve of neuroscience and biology [171]. Instead, such specifications are active design choices: We seek to make the “effective” claim that an agent is behaving as if their generative mechanism were parameterized by the (interpretable) structure we designed for Θ\Theta. Therefore when we speak of “assumptions”, it is important to distinguish between assumptions about the agent (of which we make none), versus assumptions about the investigator performing IDM (of which, by construction, we assume they have the ability to specify values for θnorm\theta_{\text{norm}}).

In IBRC, for example, in learning β\beta we are asking the question: “How much (optimistic/pessimistic) deviation from neutral knowledge does the agent appear to tolerate?” For this question to be meaningfully answered, we—as the inves- tigator—must be able to produce a meaningful value for σ~\tilde{\sigma} to specify as part of θnorm\theta_{\text{norm}}. In most cases, we are interested in deviations from some notion of “current medical knowledge”, or what knowledge an “ideal” clinician may be expected to possess; thus we may—for instance—supply a value for σ~\tilde{\sigma} via models learned a priori from data. Of course, coming up such values for θnorm\theta_{\text{norm}} is not trivial (not to mention entirely dependent on the problem and the investigator’s objectives regarding interpretability); however, we emphasize that this does not involve assumptions regarding the agent.

Figure 5: Graphical Model. In general, the environment’s states (top) are only accessible via its emissions in response to actions (middle), which the agent incorporates by way of internal states (bottom). However, note that—unlike classic POMDP/IOHMM settings, here the agent’s knowledge of the dynamics is subjective.
Refer to caption
Figure 6: Backup Diagram. In IBRC, the backup operation (Theorem 4) transfers value information across three recursive “layers”—that is, of successor values for agent states (VV), state-action pairs (QQ), and state-action-model tuples (KK). Indicated below are the utility and penalty terms collected along these backup operations.
Refer to caption

Subjective vs. Objective Dynamics  In imitation and app- renticeship learning, parameterizations of utilities and dyna- mics models are simply intermediaries for the downstream task (of replicating expert actions or matching expert returns). As a result, no distinction needs be made between the “external” environment (with objective dynamics τenv,ωenv\tau_{\text{env}},\omega_{\text{env}}) and the “internal” environment model that an agent works with (with subjective dynamics τ,ω\tau,\omega). Indeed, if the learned model were to be evaluated based on live deployment in the real environment (as is the case in IL/IRL), it only makes sense that we stipulate τ,ω=τenv,ωenv\tau,\omega=\tau_{\text{env}},\omega_{\text{env}} for the best results.

However, in IDM (and IBRC) we are precisely accounting for how an agent may appear to deviate from such perfect, point-valued knowledge of the environment. Disentangling subjective and objective dynamics is now critical: Both the forward recursion (Lemma 1) for occupancy measures and the backward recursion (Theorem 4) for value functions are computations internal to the agent’s mind—and need not correspond to any notion of true environment dynamics. The external dynamics only comes into play when considering the distribution of trajectories hϕπ,ρh\sim\phi_{\pi,\rho} induced by an agent’s policies, which—by definition—manifests through (actual or potential) interaction with the real environment.

Demonstrated vs. Projected Behavior  As advanced thro- ughout, a primary benefit of the generalized perspective we develop is that we may ask normative-descriptive questions taking the form: “Given that this (boundedly rational) agent should optimize this υ\upsilon, how suboptimally do they appear to behave?” Precisely, as pertains IBRC we noted that—as the investigator—we are free to specify (what we deem) “meaningful” values for υ\upsilon within θnorm\theta_{\text{norm}}, while recovering one or more behavioral parameters α,β,γ\alpha,\beta,\gamma from θdesc\theta_{\text{desc}}. Clearly, however, we are not at liberty to specify completely random values for υ\upsilon (or, more generally, that we are not at liberty to design Θ\Theta and θnorm\theta_{\text{norm}} in an entirely arbitrary fashion). For one, the resulting inverse decision model may simply be a poor reflection the original behavior (i.e. the projection ϕimit\phi_{\text{imit}}^{*} onto Φθnorm\Phi_{\theta_{\text{norm}}} may simply lose too much information from ϕdemo\phi_{\text{demo}}.666Abstractly, this is not dissimilar to any type of model fitting problem: If the mismatch between the (unknown) data generating pro- cess and the (imposed) structure of the model is too great, then the quality of the model—by any reasonable measure—would suffer.

Without doubt, the usefulness of the inverse decision model (i.e. in providing valid interpretations of observed behavior) depends entirely on the design and specification of Θ\Theta and θnorm\theta_{\text{norm}}, which requires care in practice. Most importantly, it should be verified that—under our designed parameteriza- tion—the projected behavior ϕimit\phi_{\text{imit}}^{*} is still a faithful model of the demonstrated behavior ϕdemo\phi_{\text{demo}}. In particular, compared with fitting a black-box model for imitating behavior—or any standard method for imitation/apprenticeship learning, for that matter—it should be verified that our (interpretably parameterized) model does not suffer inordinately in terms of accuracy measures (i.e. in predicting uu from hh); otherwise the model (and its interpretation) would not be meaningful. In Appendix B, we perform precisely such a sanity check for IBRC, using a variety of standard benchmarks (Table 5).

A.2 Further Related Work

While relevant works have been noted throughout the manus- cript, here we provide additional context for IDM and IBRC, and how notable techniques/frameworks relate to our work.

Inverse Decision Modeling  Pertinent methods subsumed by our forward and inverse formalisms have been noted in Tables 23. In particular, techniques that can be formalized as instantiations of IDM are enumerated in Table 1. Broadly, for imitation learning these include behavioral cloning-like methods [14, 15, 17, 18, 16, 19, 20, 21], as well as distribution-matching methods that directly match occupancy measures [23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39]; we defer to [12, 100] for more thorough surveys. For apprenticeship learning by inverse reinforcement learning, these include classic maximum-margin methods based on feature expec- tations [40, 13, 41, 42, 43, 44, 45], maximum likelihood soft policy matching using Boltzmann-rational policies [51, 52], maximum entropy policies [89, 90, 91, 92, 50], and Bayesian maximum a posteriori inference [59, 60, 61, 62, 63], as well as methods that leverage preference models and annotations for learning [95, 96, 97, 98, 99].

In this context, the novelty of the IDM formalism is two-fold. First, in defining a unifying framework that generalizes all prior techniques, IDM simultaneously opens up a new class of problems in behavior representation learning with consciously designed parameterizations. Specifically, in defining inverse decision models as projections in Φ\Phi-space induced by F,GF,G, and Θ\Theta, the structure and decomposition chosen for Θnorm×Θdesc\Theta_{\text{norm}}\times\Theta_{\text{desc}} allows asking normative-descriptive questions that seek to understand observed decision-making behavior. Second, in elevating recognition policies to first-class citizenship in partially-observable environments, IDM greatly generalizes the notion of “boundedness” in decision-making—that is, from the existing focus on noisy optimality in π\pi, to the ideas of subjective dynamics σ\sigma and biased belief-updates ρ\rho (viz. discussion in the beginning of this section).

Orthogonal Frameworks  Multiple studies have proposed frameworks that provide generalized treatments of different aspects of inverse reinforcement learning [58, 60, 28, 30, 172, 35, 173]. However, these are orthogonal to our purposes in the sense that they are primarily concerned with establishing connections between different aspects/subsets of the imitation/apprenticeship learning literature. These include loss-function perspectives [58] and Bayesian MAP perspectives [60] on inverse reinforcement learning, ff-divergence minimization perspectives [28, 30] on distribution matching, connections between adversarial and non-adversarial methods for distribution matching [35], as well as different problem settings for learning reward functions [173]. But relative to the IDM formalism, all such frameworks operate within the special case of θdesc\theta_{\text{desc}}==υ\upsilon (and full observability).

Case Study: GAIL  Beyond aforementioned distinctions, another implication is that IDM defines a single language for understanding key results in such prior works. For example, we revisit the well-known result in [25] that gives rise to generative adversarial imitation learning (“GAIL”): It is instructive to recast it in more general—but simpler—terms. First, consider a maximum entropy learner in the MDP setting (cf. Table 2), paired with a maximum margin identification strategy with a parameter regularizer ζ\zeta (cf. Table 3):

FθnormME(θdesc)ϕπwhereπargmaxπEzρ0Vsoft,θϕπ(z)\displaystyle\begin{aligned} F_{\theta_{\text{norm}}}^{\scalebox{0.6}{\kern 0.70004pt\text{ME}}}(\theta_{\text{desc}})\doteq\phi_{\pi^{*}}~{}\text{where}~{}\pi^{*}\doteq\text{argmax}_{\pi}\mathbb{E}_{z\sim\rho_{0}}V_{\text{soft},\theta}^{\phi_{\pi}}(z)~{}~{}~{}~{}~{}~{}~{}~{}\end{aligned} (26)
GθnormMM(ϕ)argminθdescEzρ0[Vsoft,θϕimit(z)Vθϕ(z)]+ζ(θdesc)\displaystyle\begin{aligned} G_{\theta_{\text{norm}}}^{\scalebox{0.6}{\kern 0.70004pt\text{MM}}}(\phi\hskip 0.25pt)\doteq\text{argmin}_{\theta_{\text{desc}}}\mathbb{E}_{z\sim\rho_{0}}[V_{\text{soft},\theta}^{\phi_{\text{imit}}}(z)-V_{\theta}^{\phi}(z)]+\zeta(\theta_{\text{desc}})~{}~{}~{}~{}~{}~{}~{}~{}\end{aligned} (27)

Second, consider a black-box decision-rule policy (cf. Table 2), where neural-network weights χ\chi directly parameterize a policy network fdecisionf_{\text{decision}} (and θdesc=χ\theta_{\text{desc}}=\chi); this is paired with a distribution matching identification strategy (cf. Table 3):

FθnormDR(θdesc)argmaxπδ(πfdecision(χ))F_{\theta_{\text{norm}}}^{\scalebox{0.6}{\kern 0.70004pt\text{DR}}}(\theta_{\text{desc}})\doteq\text{argmax}_{\pi}\kern 1.00006pt\delta(\pi-f_{\text{decision}}(\chi)) (28)
GθnormDM(ϕ)argminθdescζ(ϕdemoϕimit)imitG_{\theta_{\text{norm}}}^{\scalebox{0.6}{\kern 0.70004pt\text{DM}}}(\phi\hskip 0.25pt)\doteq\text{argmin}\raisebox{-4.0pt}{${}_{\theta_{\text{desc}}}$}\zeta^{*}(\phi_{\text{demo}}-\phi_{\text{imit}})-\mathcal{H}_{\text{imit}} (29)

where distance measures are given by the convex conjugate ζ\zeta^{*}, and imit\mathcal{H}_{\text{imit}} gives the causal entropy of the imitating policy. Now, the primary motivation behind generative adversarial imitation learning is the observation that ζ\zeta-regularized maximum-margin soft IRL implicitly seeks a policy whose occupancy is close to the demonstrator’s as measured by ζ\zeta^{*}. In IDM, this corresponds to a remarkably simple statement:

Proposition 6 (restate=gail,name=Ho and Ermon, Recast)

Define the beha- vior projections induced by the composition of each pairing:

projΦθnormME,MMFθnormMEGθnormMM\text{proj}_{\Phi_{\theta_{\text{norm}}}}^{\kern 0.70004pt\scalebox{0.6}{\text{ME},\kern 0.70004pt\text{MM}}}\doteq F_{\theta_{\text{norm}}}^{\scalebox{0.6}{\kern 0.70004pt\text{ME}}}\circ G_{\theta_{\text{norm}}}^{\scalebox{0.6}{\kern 0.70004pt\text{MM}}} (30)
projΦθnormDR,DMFθnormDRGθnormDM\text{proj}_{\Phi_{\theta_{\text{norm}}}}^{\kern 0.70004pt\scalebox{0.6}{\text{DR},\kern 0.70004pt\text{DM}}}\doteq F_{\theta_{\text{norm}}}^{\scalebox{0.6}{\kern 0.70004pt\text{DR}}}\circ G_{\theta_{\text{norm}}}^{\scalebox{0.6}{\kern 0.70004pt\text{DM}}} (31)

Then these projections are identical: projME,MMΦθnorm=projDR,DMΦθnorm\text{proj}\raisebox{1.0pt}{${}_{\Phi_{\theta_{\text{norm}}}}^{\kern 0.70004pt\scalebox{0.6}{\text{ME},\kern 0.70004pt\text{MM}}}$}=\text{proj}\raisebox{1.0pt}{${}_{\Phi_{\theta_{\text{norm}}}}^{\kern 0.70004pt\scalebox{0.6}{\text{DR},\kern 0.70004pt\text{DM}}}$} (and inverse decision models thereby obtained are identical).

In their original context, the significance of this lies in the fact that the first pairing explicitly requires parameterizations via reward functions (which—in classic apprenticeship methods—is restricted to be linear/convex), whereas the second pairing allows arbitrary parameterization by neural networks (which—while black-box—are more flexible). In our language, this simply means that the first projection req- uires θdesc\theta_{\text{desc}}==υ\upsilon, while the second projection allows θdesc\theta_{\text{desc}}==χ\chi.

Inverse Bounded Rational Control  Pertaining to IBRC, methods that are comparable and/or subsumed have been noted in Tables 1 and 4. In addition, the context of IBRC within existing notions of bounded rationality have been discussed in detail in the beginning of this section. Now, more broadly, we note that the study of imperfect behaviors [4] spans multiple disciplines: in cognitive science [5], biological systems [6], behavioral economics [7], and information theory [8]. Specifically, IBRC generalizes this latter class of information-theoretic approaches to bounded rationality. First, the notion of flexibility in terms of the informational effort in determining successive actions (cf. decision complexity) is present in maximum entropy [101, 102, 103, 104] and KL-regularized [107, 108, 109, 110, 111] agents. Second, the notion of toler- ance in terms of the statistical surprise in adapting to successive beliefs (cf. recognition complexity) is present in behavioral economics [174, 7] and decision theory [175, 146, 75]. Third, the notions of optimism and pessimism in terms of the average regret in deviating from prior knowledge (cf. specification complexity) are present in robust planners [148, 149, 151, 150].

On account of this, the novelty of the IBRC example is three-fold. First, it is the first to present generalized recursions incorporating all three notions of complexity—that is, in the mappings into internal states, models, and actions. Second, IBRC does so in the partially-observable setting, which—as noted in above discussions—crucially generalizes the idea of subjective dynamics into subjective beliefs, thereby accounting for boundedness in the recognition process itself. Third (perhaps most importantly), IBRC is the first to consider the inverse problem—that is, of turning the entire formalism on its head to learn the parameterizations of such boundedness, instead of simply assuming known parameters as required by the forward problem. Finally, it is important to note that IBRC is simply one example: There are of course many possibilities for formulating boundedness, including such aspects as myopia and temporal inconsistency [176, 176]; we leave such applications for future work.

Interpretable Behavior Representations  Lastly, a variety of works have approached the task of representing behaviors in an interpretable manner. In inverse reinforcement learning, multiple works have focused on the reward function itself, specifying interpretable structures that explicitly express a decision-maker’s preferences [62], behavior under time pressure [75], consideration of counterfactual outcomes [73], as well as intended goals [177]. Separately, another strand of research has focused on imposing interpretable structures onto policy functions themselves, such as representing policies in terms of decision trees [178] and intended outcomes [179] in the forward problem, or—in the inverse case—learning imitating policies based on decision trees [180] or decision boundaries [22]. In the context of IDM, both of these approaches can naturally be viewed as instantiations of our more general approach of learning representations of behavior through interpretably parameterized planners and inverse planners (as noted throughout Tables 13). Finally, for completeness also note that an orthogonal branch of research is dedicated to generating autonomous explanations of artificial behavior, as suggested updates to human models [181, 182], and also as responses to human queries in a shared [183] or user-specified vocabulary [184].

A.3 Future Work

A clear source of potential research lies in exploring differ- ently structured parameterizations Θ\Theta to allow interpretable representation learning of behaviors. After all, beyond the black-box and reward-centric approaches in Table 1 and the handful of works that have sought to account for subjective dynamics [22, 67, 93, 80], our example of IBRC is only one such prototype that exercises the IDM formalism more fully. In developing more complex and/or expressive forward mod- els, an important question to bear in mind is to what extent the inverse problem is identifiable. In most existing cases we have seen, the usual strategies—such as constraining scaling, shifting, reward shaping, as well as the use of Bayesian inference—is sufficient to recover meaningful values. How- ever, we have also seen that in the extreme case of an arbit- rary differentiable planner, any inverse problem immediately falls prey to the “no free lunch” result [136, 137, 105, 106]. Thus balancing aspects of complexity, interpretability, and identifiability of decision models would be an interesting direction of work. Finally, in this work we primarily focused on the idea of limited intentionality—that is, in the goal-seeking nature of an agent and how they may be constrained in this respect. But the flip side is also interesting: One can explore the idea of limited attentionality—that is, in how an agent may be constrained in their ability to focus on sequences of past events. This idea is explored in [185, 186] by analogy with information bottlenecks in sensors and memory capacities; however, there is much room for developing more human-interpretable parameterizations of how an agent may pay selective attention to observations over time.

Appendix B Experiment Details

Computation  In IBRC, we define the space of agent states (i.e. subjective beliefs) as 𝒵Rk\mathcal{Z}\doteq\mathbb{R}^{k}, where kk is the number of world states (kk==33 for ADNI, and kk==22 for DIAG). To implement the backward recursion (Theorem 4), each dimension of 𝒵\mathcal{Z} is discretized with a resolution of 100, and the values V(z)V(z) in the resulting lattice are updated iteratively exactly according to the backup operator B\mathbb{B}^{*}—until convergence (which is guaranteed by the fact that B\mathbb{B}^{*} is contractive, therefore the fixed point is unique; see Appendix C). For evaluation at any point zz, we (linearly) interpolate between the closest neighboring grid points. In terms of implement- ing the inverse problem in a Bayesian manner (i.e. to recover posterior distributions over Θdesc\Theta_{\text{desc}}), we perform MCMC in log-parameter space (i.e. logα,logβ,logη\log\alpha,\log\beta,\log\eta). Specifically, the proposal distribution is zero-mean Gaussian with standard deviation 0.1, with every 10th step collected as a sample. In each instance, the initial 1,000 burn-in samples are discarded, and a total of 10,000 steps are taken after burn-in.

Recognition  In the manuscript, we make multiple referenc- es to the Bayes update, in particular within the context of our (possibly-biased) belief-update (Equation 9). For completeness, we state this explicitly: Given point-valued knowledge of τ,ω\tau,\omega, update ρτ,ω(z|z,u,x)\rho_{\tau,\omega}(z^{\prime}|z,u,x^{\prime}) is the Dirac delta centered at

p(s|z,u,x,τ,ω)Esp(|z)[τ(s|s,u)ω(x|u,s)Esτ(|s,u)ω(x|u,s)]\displaystyle\begin{aligned} p(s^{\prime}|z,u,x^{\prime},\tau,\omega)\doteq\mathbb{E}_{s\sim p(\cdot|z)}\bigg{[}\frac{\tau(s^{\prime}|s,u)\omega(x^{\prime}|u,s^{\prime})}{\mathbb{E}_{s^{\prime}\sim\tau(\cdot|s,u)}\omega(x^{\prime}|u,s^{\prime})}\bigg{]}~{}~{}~{}~{}~{}~{}~{}\end{aligned} (32)

and the overall recognition policy is the expectation over such values of τ,ω\tau,\omega (Equation 9). As noted in Section 4.1, in general σ~\tilde{\sigma} represents any prior distribution the agent is specified to have, and in particular can be some Bayesian posterior p(τ,ω|)p(\tau,\omega|\mathcal{E}) given any form of experience \mathcal{E}. This can be modeled in any manner, and is not the focus of our work; what matters here is simply that the agent may deviate optimistically/pessimistically from such a prior. As noted in Section 5, for our purposes we simulate σ~\tilde{\sigma} by discretizing the space of models such that probabilities vary in ±10%\pm 10\% increments from the (highest-likelihood) truth. In ADNI, this means σ~\tilde{\sigma} is centered at the IOHMM learned from the data.

Model Accuracy  In Appendix A.1 we discussed the caveat: In order for an inverse decision model to provide valid interpretations of observed behavior, it should be verified that—under the designed parameterization—the projected behavior ϕimit\phi_{\text{imit}}^{*} is still an accurate model of the demonstrated behavior ϕdemo\phi_{\text{demo}}. Here we perform such a sanity check for our IBRC example using the ADNI environment. We consider the following standard benchmark algorithms. First, in terms of black-box models for imitation learning, we consider behavioral cloning [15] with a recurrent neural network for observation-action histories (RNN-Based BC-IL); an adaptation of model-based imitation learning [187] to partially-observable settings, using the learned IOHMM as model (IOHMM-Based BC-IL); and a recently-proposed model-based imitation learning that allows for subjective dynamics [22] by jointly learning the agent’s possibly-biased internal model and their probabilistic decision boundaries (Joint IOHMM-Based BC-IL). Second, in terms of classic reward-centric methods for apprenticeship learning, we consider Bayesian inverse reinforcement learning in partially-observable environments [75] equipped with the learned IOHMM as model (Bayesian PO-IRL); and—analogous to the black-box case—the equivalent of this method that trains the dynamics model jointly along with the agent’s apprenticeship policy [74] (Joint Bayesian PO-IRL). Algorithms requiring learned models are given IOHMMs estimated using conventional methods [188]—which is the same method by which the true model is estimated in IBRC (that is, as part of the space of candidate models in the support of σ~\tilde{\sigma}).

Table 5: Comparison of Model Accuracies. IBRC performs similarly to all benchmark algorithms in matching demonstrated actions. Results are computed using held-out samples based on 5-fold cross-validation. IBRC is slightly better-calibrated, and similar in precision-recall scores (differences are statistically insignificant).

Inverse Decision Model Calibration (Low is Better) PRC Score (High is Better) Black-Box Model: RNN-Based BC-IL 0.18 ±\pm 0.05 0.81 ±\pm 0.08 IOHMM-Based BC-IL 0.19 ±\pm 0.07 0.79 ±\pm 0.11 Joint IOHMM-Based BC-IL 0.17 ±\pm 0.05 0.81 ±\pm 0.09 Reward-Centric Model: Bayesian PO-IRL 0.23 ±\pm 0.01 0.78 ±\pm 0.09 Joint Bayesian PO-IRL 0.24 ±\pm 0.01 0.79 ±\pm 0.09 Boundedly Rational Model: IBRC (with learned α,β,η\alpha,\beta,\eta) 0.16 ±\pm 0.00 0.77 ±\pm 0.01

Results\contourwhiteResults. Table 5 shows results of this comparison on predicting actions, computed using held-out samples based on 5-fold cross-validation. Crucially, while IBRC has the advan- tage in terms of interpretability of parameterization, its per- formance—purely in terms of predicting actions—does not degrade: IBRC is slightly better in terms of calibration, and similar in precision-recall (differences are statistically insignificant), which—for our ADNI example—affirms the validity of IBRC as an (interpretable) representation of ϕdemo\phi_{\text{demo}}.

Data Selection  From the ADNI data, we first selected out anomalous cases without a cognitive dementia rating test result, which is almost always taken at every visit by every patient. Second, we also truncated patient trajectories at points where a visit is skipped (that is, if the next visit of a patient does not occur immediately after the 6-monthly period following the previous visit). This selection process leaves 1,626 patients out of the original 1,737, and the median number of consecutive visits for each patient is three. In measuring MRI outcomes, the “average” is defined to be within half a standard deviation of the population mean. Note that this is the same pre-processing method employed for ADNI in [22].

Implementation  Details of implementation for benchmark algorithms follow the setup in [22], and are reproduced here: RNN-Based BC-IL\contourwhiteRNN-Based BC-IL: We train an RNN whose inputs are the observed histories hh and whose outputs are the predicted probabilities π^(u|h)\hat{\pi}(u|h) of taking action uu given the observed history hh. The network consists of an LSTM unit of size 6464 and a fully-connected hidden layer of size 6464. The cross-entropy \mathcal{L}==n=1Nt=1Tu𝒰I{ut-\sum_{n=1}^{N}\sum_{t=1}^{T}\sum_{u\in\mathcal{U}}\mathbb{I}\{u_{t}==u}logπ^(u|h)u\}\log\hat{\pi}(u|h) is minimized using the Adam optimizer with a learning rate of 0.0010.001 until convergence (that is, when the loss does not improve for 100100 consecutive iterations). Bayesian PO-IRL\contourwhiteBayesian PO-IRL: The IOHMM parameters are initialized by sampling uniformly at random. Then, they are estimated and fixed using conventional IOHMM methods. The utility υ\upsilon is initialized as υ^0(s,u)\hat{\upsilon}^{0}(s,u)==εs,u\varepsilon_{s,u}, where εs,u\varepsilon_{s,u}\sim𝒩(0,0.0012)\mathcal{N}(0,0.001^{2}). Then, it is estimated via MCMC sampling, during which new candidate samples are generated by adding Gaussian noise with standard deviation 0.0010.001 to the previous sample. To form the final estimate, we average every 10th sample among the second set of 500500 samples, ignoring the first 500500 samples. To compute optimal QQ-values, we use an off-the-shelf POMDP solver https://www.pomdp.org/code/index.html . Joint Bayesian PO-IRL\contourwhiteJoint Bayesian PO-IRL: All parameters are initialized exactly the same way as in Bayesian PO-IRL. Then, both the IOHMM parameters and the utility are estimated jointly via MCMC sampling. In order to generate new candidate samples, with equal probabilities we either sample new IOHMM parameters from the posterior (but without changing υ\upsilon) or obtain a new υ\upsilon the same way we do in Bayesian PO-IRL (but without changing the IOHMM parameters). A final estimate is formed the same way as in Bayesian PO-IRL. IOHMM-Based BC-IL\contourwhiteIOHMM-Based BC-IL: The IOHMM parameters are initialized by sampling them uniformly at random. Then, they are estimated and fixed using conventional IOHMM methods. Given the IOHMM parameters, we parameterize policies using the method of [22], with the policy parameters {μu}u𝒰\{\mu_{u}\}_{u\in\mathcal{U}} (not to be confused with the occupancy measure “μ\mu” as defined in the present work) initialized as μ^u0(s)=(1/|S|+εu,s)/sS(1/|S|+εu,s)\hat{\mu}^{0}_{u}(s)=(1/|S|+\varepsilon_{u,s})/\sum_{s^{\prime}\in S}(1/|S|+\varepsilon_{u,s^{\prime}}), where εu,s\varepsilon_{u,s^{\prime}}\sim𝒩(0,0.0012)\mathcal{N}(0,0.001^{2}). Then, they are estimated according solely to the action likelihoods in using the EM algorithm. The expected log-posterior is maximized using the Adam optimizer with learning rate 0.0010.001 until convergence (that is, when the expected log-posterior does not improve for 100100 consecutive iterations). Joint IOHMM-Based BC-IL\contourwhiteJoint IOHMM-Based BC-IL: This corresponds exactly to the proposed method of [22] itself, which is similar to IOHMM-Based BC-IL except parameters are trained jointly. All parameters are initialized exactly the same way as before; then, the IOHMM parameters and the policy parameters are estimated jointly according to both the action likelihoods and the observation likelihoods simultaneously. The expected log-posterior is again maximized using the Adam optimizer with a learning rate of 0.0010.001 until convergence (non-improvement for 100100 consecutive iterations).

Appendix C Proofs of Propositions

\forward

*

Proof. Start from the definition of Mπ,ρ\mathbb{M}_{\pi,\rho}; episodes are res- tarted on completion ad infinitum, so we can write μπ,ρ\mu_{\pi,\rho} as:

μπ,ρ(z)(1γ)t=0γtp(zt=z|z0ρ0)=(1γ)t=0γt((Mπ,ρ)tρ0)(z)\displaystyle\begin{aligned} \mu_{\pi,\rho}(z)&\doteq(1-\gamma)\textstyle\sum_{t=0}^{\infty}\gamma^{t}p(z_{t}=z|z_{0}\sim\rho_{0})\\ &=(1-\gamma)\textstyle\sum_{t=0}^{\infty}\gamma^{t}((\mathbb{M}_{\pi,\rho})^{t}\rho_{0})(z)\end{aligned} (33)

Then we obtain the result by simple algebraic manipulation:

(1γ)ρ0(z)+γ(Mπ,ρμπ,ρ)(z)=(1γ)ρ0(z)+γ(1γ)t=0γt((Mπ,ρ)t+1ρ0)(z)=(1γ)(ρ0(z)+t=0γt+1((Mπ,ρ)t+1ρ0)(z))=(1γ)t=0γt((Mπ,ρ)tρ0)(z)=μπ,ρ(z)\displaystyle\begin{aligned} &~{}~{}~{}~{}~{}\hskip 0.75pt(1-\gamma)\rho_{0}(z)+\gamma(\mathbb{M}_{\pi,\rho}\mu_{\pi,\rho})(z)\\ &=(1-\gamma)\rho_{0}(z)+\gamma(1-\gamma)\textstyle\sum_{t=0}^{\infty}\gamma^{t}((\mathbb{M}_{\pi,\rho})^{t+1}\rho_{0})(z)\\ &=(1-\gamma)(\rho_{0}(z)+\textstyle\sum_{t=0}^{\infty}\gamma^{t+1}((\mathbb{M}_{\pi,\rho})^{t+1}\rho_{0})(z))\\ &=(1-\gamma)\textstyle\sum_{t=0}^{\infty}\gamma^{t}((\mathbb{M}_{\pi,\rho})^{t}\rho_{0})(z)\\ &=\mu_{\pi,\rho}(z)\end{aligned} (34)

For uniqueness, we use the usual conditions—that is, that the process induced by the environment and the agent’s policies is ergodic, with a single closed communicating class.

\backward

*

Proof. Start with the Lagrangian, with VV\inR𝒵\mathbb{R}^{\mathcal{Z}}: π,ρ(μ,V)\mathcal{L}_{\pi,\rho}(\mu,V)

Jπ,ρV,μγMπ,ρμ(1γ)ρ0=Ezμπ,ρsp(|z)uπ(|z)υ(s,u)V,μγMπ,ρμ(1γ)ρ0=Ezμπ,ρsp(|z)uπ(|z)υ(s,u)+Ezμπ,ρsp(|z)uπ(|z)τ,ωσ(|z,u)sτ(|s,u)xω(|u,s)zρ(|z,u,x)γV(z)Ezμπ,ρV(z)+V,(1γ)ρ0\displaystyle\begin{aligned} &\doteq J_{\pi,\rho}-\langle V,\mu-\gamma\mathbb{M}_{\pi,\rho}\mu-(1-\gamma)\rho_{0}\rangle\\ &=\mathbb{E}_{\begin{subarray}{c}z\sim\mu_{\pi,\rho}\\ s\sim p(\cdot|z)\\ u\sim\pi(\cdot|z)\end{subarray}}\upsilon(s,u)-\langle V,\mu-\gamma\mathbb{M}_{\pi,\rho}\mu-(1-\gamma)\rho_{0}\rangle\\ &=\mathbb{E}_{\begin{subarray}{c}z\sim\mu_{\pi,\rho}\\ s\sim p(\cdot|z)\\ u\sim\pi(\cdot|z)\end{subarray}}\upsilon(s,u)+\mathbb{E}_{\begin{subarray}{c}z\sim\mu_{\pi,\rho}\\ s\sim p(\cdot|z)\\ u\sim\pi(\cdot|z)\\ \tau,\omega\sim\sigma(\cdot|z,u)\\ s^{\prime}\sim\tau(\cdot|s,u)\\ x^{\prime}\sim\omega(\cdot|u,s^{\prime})\\ z^{\prime}\sim\rho(\cdot|z,u,x^{\prime})\end{subarray}}\gamma V(z^{\prime})\\ &~{}~{}~{}~{}-\mathbb{E}_{z\sim\mu_{\pi,\rho}}V(z)+\langle V,(1-\gamma)\rho_{0}\rangle\end{aligned} (35)
=Ezμπ,ρsp(|z)[Euπ(|z)[υ(s,u)+Eτ,ωσ(|z,u)sτ(|s,u)xω(|u,s)zρ(|z,u,x)γV(z)]V(z)]+V,(1γ)ρ0\displaystyle\begin{aligned} &=\mathbb{E}_{\begin{subarray}{c}z\sim\mu_{\pi,\rho}\\ s\sim p(\cdot|z)\end{subarray}}[\mathbb{E}_{u\sim\pi(\cdot|z)}[\upsilon(s,u)\\ &~{}~{}~{}~{}+\mathbb{E}_{\begin{subarray}{c}\tau,\omega\sim\sigma(\cdot|z,u)\\ s^{\prime}\sim\tau(\cdot|s,u)\\ x^{\prime}\sim\omega(\cdot|u,s^{\prime})\\ z^{\prime}\sim\rho(\cdot|z,u,x^{\prime})\end{subarray}}\gamma V(z^{\prime})]-V(z)]+\langle V,(1-\gamma)\rho_{0}\rangle\end{aligned} (36)

Then taking the gradient w.r.t. μ\mu and setting it to zero yields:

V(z)=Esp(|z)uπ(|z)[υ(s,u)+Eτ,ωσ(|z,u)sτ(|s,u)xω(|u,s)zρ(|z,u,x)γV(z)]\displaystyle\begin{aligned} V(z)=\mathbb{E}_{\begin{subarray}{c}s\sim p(\cdot|z)\\ u\sim\pi(\cdot|z)\end{subarray}}[\upsilon(s,u)+\mathbb{E}_{\begin{subarray}{c}\tau,\omega\sim\sigma(\cdot|z,u)\\ s^{\prime}\sim\tau(\cdot|s,u)\\ x^{\prime}\sim\omega(\cdot|u,s^{\prime})\\ z^{\prime}\sim\rho(\cdot|z,u,x^{\prime})\end{subarray}}\gamma V(z^{\prime})]~{}~{}~{}~{}~{}\end{aligned} (37)

For uniqueness, observe as usual that Bπ,ρ\mathbb{B}_{\pi,\rho} is γ\gamma-contracting:

Bπ,ρVBπ,ρV=maxz|Euπ(|z)τ,ωσ(|z,u)zϱτ,ω(|z,u)[γV(z)γV(z)]|maxzEuπ(|z)τ,ωσ(|z,u)zϱτ,ω(|z,u)[|γV(z)γV(z)|]maxz|γV(z)γV(z)|=γVV\displaystyle\begin{aligned} \|\mathbb{B}_{\pi,\rho}&V-\mathbb{B}_{\pi,\rho}V^{\prime}\|_{\infty}\\ &=\text{max}_{z}\big{|}\mathbb{E}_{\begin{subarray}{c}u\sim\pi(\cdot|z)\\ \tau,\omega\sim\sigma(\cdot|z,u)\\ z^{\prime}\sim\varrho_{\tau,\omega}(\cdot|z,u)\end{subarray}}\big{[}\gamma V(z^{\prime})-\gamma V^{\prime}(z^{\prime})\big{]}\big{|}\\ &\leq\text{max}_{z}\mathbb{E}_{\begin{subarray}{c}u\sim\pi(\cdot|z)\\ \tau,\omega\sim\sigma(\cdot|z,u)\\ z^{\prime}\sim\varrho_{\tau,\omega}(\cdot|z,u)\end{subarray}}\big{[}\big{|}\gamma V(z^{\prime})-\gamma V^{\prime}(z^{\prime})\big{|}\big{]}\\ &\leq\text{max}_{z^{\prime}}\big{|}\gamma V(z^{\prime})-\gamma V^{\prime}(z^{\prime})\big{|}\\ &=\gamma\|V-V^{\prime}\|_{\infty}\\ \end{aligned} (38)

which allows appealing to the contraction mapping theorem.

\backwardx

*

Proof. Start with the Lagrangian, now with the new multipliers α,β,ηR\alpha,\beta,\eta\in\mathbb{R} in addition to VR𝒵V\in\mathbb{R}^{\mathcal{Z}}: π,ρ(μ,α,β,η,V)\mathcal{L}_{\pi,\rho}(\mu,\alpha,\beta,\eta,V)

Jπ,ρV,μγMπ,ρμ(1γ)ρ0α(Iπ,ρ[π;π~]A)β(Iπ,ρ[σ;σ~]B)η(Iπ,ρ[ϱ;ϱ~]C)=Ezμπ,ρsp(|z)uπ(|z)υ(s,u)V,μγMπ,ρμ(1γ)ρ0α(Ezμπ,ρDKL(π(|z)π~)A)β(Ezμπ,ρuπ(|z)DKL(σ(|z,u)σ~)B)η(Ezμπ,ρuπ(|z)τ,ωσ(|z,u)DKL(ϱτ,ω(|z,u)ϱ~)C)\displaystyle\begin{aligned} &\doteq J_{\pi,\rho}-\langle V,\mu-\gamma\mathbb{M}_{\pi,\rho}\mu-(1-\gamma)\rho_{0}\rangle\\ &~{}~{}~{}~{}-\alpha\cdot(\mathbb{I}_{\pi,\rho}[\pi;\tilde{\pi}]-A)-\beta\cdot(\mathbb{I}_{\pi,\rho}[\sigma;\tilde{\sigma}]-B)\\ &~{}~{}~{}~{}-\eta\cdot(\mathbb{I}_{\pi,\rho}[\varrho;\tilde{\varrho}]-C)\\ &=\mathbb{E}_{\begin{subarray}{c}z\sim\mu_{\pi,\rho}\\ s\sim p(\cdot|z)\\ u\sim\pi(\cdot|z)\end{subarray}}\upsilon(s,u)-\langle V,\mu-\gamma\mathbb{M}_{\pi,\rho}\mu-(1-\gamma)\rho_{0}\rangle\\ &~{}~{}~{}~{}-\alpha\cdot(\mathbb{E}_{z\sim\mu_{\pi,\rho}}D_{\text{KL}}(\pi(\cdot|z)\|\tilde{\pi})-A)\\ &~{}~{}~{}~{}-\beta\cdot(\mathbb{E}_{\begin{subarray}{c}z\sim\mu_{\pi,\rho}\\ u\sim\pi(\cdot|z)\end{subarray}}D_{\text{KL}}(\sigma(\cdot|z,u)\|\tilde{\sigma})-B)\\ &~{}~{}~{}~{}-\eta\cdot(\mathbb{E}_{\begin{subarray}{c}z\sim\mu_{\pi,\rho}\\ u\sim\pi(\cdot|z)\\ \tau,\omega\sim\sigma(\cdot|z,u)\end{subarray}}D_{\text{KL}}(\varrho_{\tau,\omega}(\cdot|z,u)\|\tilde{\varrho})-C)\end{aligned} (39)
=Ezμπ,ρsp(|z)uπ(|z)υ(s,u)+Ezμπ,ρsp(|z)uπ(|z)τ,ωσ(|z,u)sτ(|s,u)xω(|u,s)zρ(|z,u,x)γV(z)Ezμπ,ρV(z)+V,(1γ)ρ0α(Ezμπ,ρuπ(|z)logπ(u|z)π~(u)A)β(Ezμπ,ρuπ(|z)τ,ωσ(|z,u)logσ(τ,ω|z,u)σ~(τ,ω)B)η(Ezμπ,ρuπ(|z)τ,ωσ(|z,u)sτ(|s,u)xω(|u,s)zρ(|z,u,x)logϱτ,ω(z|z,u)ϱ~(z)C)=Ezμπ,ρsp(|z)[Euπ(|z)[υ(s,u)α(logπ(u|z)π~(u)A)+Eτ,ωσ(|z,u)[β(logσ(τ,ω|z,u)σ~(τ,ω)B)+Esτ(|s,u)xω(|u,s)zρ(|z,u,x)[η(logϱτ,ω(z|z,u)ϱ~(z)C)+γV(z)]]]V(z)]+V,(1γ)ρ0\displaystyle\begin{aligned} &=\mathbb{E}_{\begin{subarray}{c}z\sim\mu_{\pi,\rho}\\ s\sim p(\cdot|z)\\ u\sim\pi(\cdot|z)\end{subarray}}\upsilon(s,u)+\mathbb{E}_{\begin{subarray}{c}z\sim\mu_{\pi,\rho}\\ s\sim p(\cdot|z)\\ u\sim\pi(\cdot|z)\\ \tau,\omega\sim\sigma(\cdot|z,u)\\ s^{\prime}\sim\tau(\cdot|s,u)\\ x^{\prime}\sim\omega(\cdot|u,s^{\prime})\\ z^{\prime}\sim\rho(\cdot|z,u,x^{\prime})\end{subarray}}\gamma V(z^{\prime})\\ &~{}~{}~{}~{}-\mathbb{E}_{z\sim\mu_{\pi,\rho}}V(z)+\langle V,(1-\gamma)\rho_{0}\rangle\\ &~{}~{}~{}~{}-\alpha\cdot(\mathbb{E}_{\begin{subarray}{c}z\sim\mu_{\pi,\rho}\\ u\sim\pi(\cdot|z)\end{subarray}}\log\scalebox{1.2}{$\frac{\pi(u|z)}{\tilde{\pi}(u)}$}-A)\\ &~{}~{}~{}~{}-\beta\cdot(\mathbb{E}_{\begin{subarray}{c}z\sim\mu_{\pi,\rho}\\ u\sim\pi(\cdot|z)\\ \tau,\omega\sim\sigma(\cdot|z,u)\end{subarray}}\log\scalebox{1.2}{$\frac{\sigma(\tau,\omega|z,u)}{\tilde{\sigma}(\tau,\omega)}$}-B)\\ &~{}~{}~{}~{}-\eta\cdot(\mathbb{E}_{\begin{subarray}{c}z\sim\mu_{\pi,\rho}\\ u\sim\pi(\cdot|z)\\ \tau,\omega\sim\sigma(\cdot|z,u)\\ s^{\prime}\sim\tau(\cdot|s,u)\\ x^{\prime}\sim\omega(\cdot|u,s^{\prime})\\ z^{\prime}\sim\rho(\cdot|z,u,x^{\prime})\end{subarray}}\log\scalebox{1.2}{$\frac{\varrho_{\tau,\omega}(z^{\prime}|z,u)}{\tilde{\varrho}(z^{\prime})}$}-C)\\ &=\mathbb{E}_{\begin{subarray}{c}z\sim\mu_{\pi,\rho}\\ s\sim p(\cdot|z)\end{subarray}}\big{[}\mathbb{E}_{u\sim\pi(\cdot|z)}\big{[}\upsilon(s,u)-\alpha\cdot(\log\scalebox{1.2}{$\frac{\pi(u|z)}{\tilde{\pi}(u)}$}-A)\\ &~{}~{}~{}~{}+\mathbb{E}_{\begin{subarray}{c}\tau,\omega\sim\sigma(\cdot|z,u)\end{subarray}}\big{[}-\beta\cdot(\log\scalebox{1.2}{$\frac{\sigma(\tau,\omega|z,u)}{\tilde{\sigma}(\tau,\omega)}$}-B)\\ &~{}~{}~{}~{}+\mathbb{E}_{\begin{subarray}{c}s^{\prime}\sim\tau(\cdot|s,u)\\ x^{\prime}\sim\omega(\cdot|u,s^{\prime})\\ z^{\prime}\sim\rho(\cdot|z,u,x^{\prime})\end{subarray}}\big{[}-\eta\cdot(\log\scalebox{1.2}{$\frac{\varrho_{\tau,\omega}(z^{\prime}|z,u)}{\tilde{\varrho}(z^{\prime})}$}-C)\\ &~{}~{}~{}~{}+\gamma V(z^{\prime})\big{]}\big{]}\big{]}-V(z)\big{]}+\langle V,(1-\gamma)\rho_{0}\rangle\end{aligned} (40)

Then taking the gradient w.r.t. μ\mu and setting it to zero yields:

V(z)=Esp(|z)uπ(|z)[αlogπ(u|z)π~(u)+υ(s,u)+Eτ,ωσ(|z,u)[βlogσ(τ,ω|z,u)σ~(τ,ω)+Esτ(|s,u)xω(|u,s)zρτ,ω(|z,u,x)[ηlogϱτ,ω(z|z,u)ϱ~(z)+γV(z)]]]\displaystyle\begin{aligned} V(z)=~{}\mathbb{E}_{\begin{subarray}{c}s\sim p(\cdot|z)\\ u\sim\pi(\cdot|z)\end{subarray}}&\big{[}-\alpha\log\scalebox{1.2}{$\frac{\pi(u|z)}{\tilde{\pi}(u)}$}+\upsilon(s,u)+\\[-8.61108pt] \mathbb{E}_{\tau,\omega\sim\sigma(\cdot|z,u)}&\big{[}-\beta\log\scalebox{1.2}{$\frac{\sigma(\tau,\omega|z,u)}{\tilde{\sigma}(\tau,\omega)}$}+\\[-3.44444pt] \mathbb{E}_{\begin{subarray}{c}s^{\prime}\sim\tau(\cdot|s,u)\\ x^{\prime}\sim\omega(\cdot|u,s^{\prime})\\ z^{\prime}\sim\rho_{\tau,\omega}(\cdot|z,u,x^{\prime})\end{subarray}}&\big{[}-\eta\log\raisebox{-2.0pt}{\scalebox{1.2}{$\frac{\varrho_{\tau,\omega}(z^{\prime}|z,u)}{\tilde{\varrho}(z^{\prime})}$}}+\kern 1.00006pt\gamma V(z^{\prime})\big{]}\big{]}\big{]}\end{aligned} (41)

For uniqueness, observe as before that Bπ,ρ\mathbb{B}_{\pi,\rho} is γ\gamma-contracting: Bπ,ρVBπ,ρVγVV\|\mathbb{B}_{\pi,\rho}V-\mathbb{B}_{\pi,\rho}V^{\prime}\|_{\infty}\leq\gamma\|V-V^{\prime}\|_{\infty}; then appeal to the contraction mapping theorem for uniqueness of fixed point. The only change from before is the additional log terms, which—like the utility term—cancel out of the differences.

For Theorems 4 and 5, we give a single derivation for both:

\values

*

\policies

*

Proof. From Proposition 3, the (state) value Vϕπ,ρR𝒵V^{\phi_{\pi,\rho}}\in\mathbb{R}^{\mathcal{Z}} is:

Vϕπ,ρ(z)=Esp(|z)uπ(|z)[αlogπ(u|z)π~(u)+υ(s,u)+Eτ,ωσ(|z,u)[βlogσ(τ,ω|z,u)σ~(τ,ω)+Esτ(|s,u)xω(|u,s)zρτ,ω(|z,u,x)[ηlog

ϱτ,ω(|zz,u)~ϱ(z)

+γVϕπ,ρ(z)]]]
\displaystyle\begin{aligned} V^{\phi_{\pi,\rho}}(z)=~{}\mathbb{E}_{\begin{subarray}{c}s\sim p(\cdot|z)\\ u\sim\pi(\cdot|z)\end{subarray}}&\big{[}-\alpha\log\scalebox{1.2}{$\frac{\pi(u|z)}{\tilde{\pi}(u)}$}+\upsilon(s,u)+\\[-8.61108pt] \mathbb{E}_{\tau,\omega\sim\sigma(\cdot|z,u)}&\big{[}-\beta\log\scalebox{1.2}{$\frac{\sigma(\tau,\omega|z,u)}{\tilde{\sigma}(\tau,\omega)}$}+\\[-3.44444pt] \mathbb{E}_{\begin{subarray}{c}s^{\prime}\sim\tau(\cdot|s,u)\\ x^{\prime}\sim\omega(\cdot|u,s^{\prime})\\ z^{\prime}\sim\rho_{\tau,\omega}(\cdot|z,u,x^{\prime})\end{subarray}}&\big{[}-\eta\log\raisebox{-2.0pt}{\scalebox{1.2}{$\frac{\varrho_{\tau,\omega}(z^{\prime}|z,u)}{\tilde{\varrho}(z^{\prime})}$}}\raisebox{-13.0pt}{$+\kern 1.00006pt\gamma V^{\phi_{\pi,\rho}}(z^{\prime})\big{]}\big{]}\big{]}$}\end{aligned}
(42)

Define (state-action) Qϕπ,ρQ^{\phi_{\pi,\rho}}\inR𝒵×𝒰\mathbb{R}^{\mathcal{Z}\times\mathcal{U}} to be ahead by 1/3\nicefrac{{1}}{{3}} steps:

Qϕπ,ρ(z,u)Esp(|z)[υ(s,u)+Eτ,ωσ(|z,u)[βlogσ(τ,ω|z,u)σ~(τ,ω)+Esτ(|s,u)xω(|u,s)zρτ,ω(|z,u,x)[ηlog

ϱτ,ω(|zz,u)~ϱ(z)

+γVϕπ,ρ(z)]]]
\displaystyle\begin{aligned} Q^{\phi_{\pi,\rho}}(z,u)\doteq\mathbb{E}_{\begin{subarray}{c}s\sim p(\cdot|z)\end{subarray}}&\big{[}\upsilon(s,u)+\\[-1.07639pt] \mathbb{E}_{\tau,\omega\sim\sigma(\cdot|z,u)}&\big{[}-\beta\log\scalebox{1.2}{$\frac{\sigma(\tau,\omega|z,u)}{\tilde{\sigma}(\tau,\omega)}$}+\\[-3.44444pt] \mathbb{E}_{\begin{subarray}{c}s^{\prime}\sim\tau(\cdot|s,u)\\ x^{\prime}\sim\omega(\cdot|u,s^{\prime})\\ z^{\prime}\sim\rho_{\tau,\omega}(\cdot|z,u,x^{\prime})\end{subarray}}&\big{[}-\eta\log\raisebox{-2.0pt}{\scalebox{1.2}{$\frac{\varrho_{\tau,\omega}(z^{\prime}|z,u)}{\tilde{\varrho}(z^{\prime})}$}}\raisebox{-17.0pt}{$+\kern 1.00006pt\gamma V^{\phi_{\pi,\rho}}(z^{\prime})\big{]}\big{]}\big{]}$}\end{aligned}
(43)

and (state-action-model) Kϕπ,ρK^{\phi_{\pi,\rho}}\inR𝒵×𝒰×𝒯×𝒪\mathbb{R}^{\mathcal{Z}\times\mathcal{U}\times\mathcal{T}\times\mathcal{O}} by 2/3\nicefrac{{2}}{{3}} steps:

Kϕπ,ρ(z,u,τ,ω)Esp(|z)sτ(|s,u)xω(|u,s)zρτ,ω(|z,u,x)[ηlog

ϱτ,ω(|zz,u)~ϱ(z)

+γVϕπ,ρ(z)]]]
\displaystyle\begin{aligned} K^{\phi_{\pi,\rho}}(z,u,\tau,\omega)\doteq~{}~{}~{}~{}~{}\kern 1.00006pt&\\ \mathbb{E}_{\begin{subarray}{c}s\sim p(\cdot|z)\\ s^{\prime}\sim\tau(\cdot|s,u)\\ x^{\prime}\sim\omega(\cdot|u,s^{\prime})\\ z^{\prime}\sim\rho_{\tau,\omega}(\cdot|z,u,x^{\prime})\end{subarray}}&\big{[}-\eta\log\raisebox{-2.0pt}{\scalebox{1.2}{$\frac{\varrho_{\tau,\omega}(z^{\prime}|z,u)}{\tilde{\varrho}(z^{\prime})}$}}\raisebox{-17.0pt}{$+\kern 1.00006pt\gamma V^{\phi_{\pi,\rho}}(z^{\prime})\big{]}\big{]}\big{]}$}\end{aligned}
(44)

The decision and recognition policies seek the optimizations:

extremizeπVϕπ,ρ(z)s.t.Euπ(|z)1=1\displaystyle\begin{aligned} &\text{extremize}_{\pi}V^{\phi_{\pi,\rho}}(z)\\ &\kern 1.00006pt~{}~{}~{}~{}~{}~{}\text{s.t.}~{}~{}~{}~{}~{}~{}~{}\mathbb{E}_{u\sim\pi(\cdot|z)}1=1\end{aligned} (45)
extremizeσQϕπ,ρ(z,u)s.t.Eτ,ωσ(|z,u)1=1\displaystyle\begin{aligned} &\text{extremize}_{\sigma}Q^{\phi_{\pi,\rho}}(z,u)\\ &\kern 1.00006pt~{}~{}~{}~{}~{}~{}\text{s.t.}~{}~{}~{}~{}~{}~{}~{}\mathbb{E}_{\tau,\omega\sim\sigma(\cdot|z,u)}1=1\end{aligned} (46)

Equations 4244 are true in particular for optimal values, so

V(z)=E[αlogπ(u|z)π~(u)+Q(z,u)]uπ(|z)\displaystyle\begin{aligned} V^{*}(z)=\mathbb{E}&{}_{u\sim\pi^{*}(\cdot|z)}\big{[}-\alpha\log\scalebox{1.2}{$\frac{\pi^{*}(u|z)}{\tilde{\pi}(u)}$}+Q^{*}(z,u)\big{]}\end{aligned} (47)
Q(z,u)=Esp(|z)[υ(s,u)]+Eτ,ωσ(|z,u)[βlogσ(τ,ω|z,u)σ~(τ,ω)+K(z,u,τ,ω)]\displaystyle\begin{aligned} Q^{*}(z,u)&=\mathbb{E}_{s\sim p(\cdot|z)}\big{[}\upsilon(s,u)\big{]}+\mathbb{E}_{\tau,\omega\sim\sigma^{*}(\cdot|z,u)}\big{[}~{}~{}\\ &-\beta\log\scalebox{1.2}{$\frac{\sigma^{*}(\tau,\omega|z,u)}{\tilde{\sigma}(\tau,\omega)}$}+K^{*}(z,u,\tau,\omega)\big{]}\end{aligned} (48)

Therefore for the extremizations we write the Lagrangians

(π,λ)V(z)+λ(Euπ(|z)11)\displaystyle\begin{aligned} \mathcal{L}(\pi^{*},\lambda)\doteq V^{*}(z)+\lambda\cdot(\mathbb{E}_{u\sim\pi^{*}(\cdot|z)}1-1)\end{aligned} (49)
(σ,ν)Q(z,u)+ν(Eτ,ωσ(|z,u)11)\displaystyle\begin{aligned} \mathcal{L}(\sigma^{*},\nu)\doteq Q^{*}(z,u)+\nu\cdot(\mathbb{E}_{\tau,\omega\sim\sigma^{*}(\cdot|z,u)}1-1)\end{aligned} (50)

Straightforward algebraic manipulation yields the policies:

π(u|z)=π~(ut)ZQ(z)exp(1αQ(z,u))\pi^{*}(u|z)=\scalebox{1.2}{$\frac{\tilde{\pi}(u_{t})}{Z_{Q^{*}}(z)}$}\exp\big{(}\tfrac{1}{\alpha}Q^{*}(z,u)\big{)} (51)
σ(τ,ω|z,u)=σ~(τ,ω)ZK(z,u)exp(1βK(z,u,τ,ω))\sigma^{*}(\tau,\omega|z,u)=\scalebox{1.2}{$\frac{\tilde{\sigma}(\tau,\omega)}{Z_{K^{*}}(z,u)}$}\exp\big{(}\tfrac{1}{\beta}K^{*}(z,u,\tau,\omega)\big{)} (52)

where partition functions ZQ(z)Z_{Q^{*}}(z) and ZK(z)Z_{K^{*}}(z) are given by:

ZQ(z)=Euπ~exp(1αQ(z,u))\displaystyle\begin{aligned} Z_{Q^{*}}(z)=\mathbb{E}_{u\sim\tilde{\pi}}\exp(\tfrac{1}{\alpha}Q^{*}(z,u))\end{aligned} (53)
ZK(z,u)=Eτ,ωσ~exp(1βK(z,u,τ,ω))\displaystyle\begin{aligned} Z_{K^{*}}(z,u)=\mathbb{E}_{\tau,\omega\sim\tilde{\sigma}}\exp(\tfrac{1}{\beta}K^{*}(z,u,\tau,\omega))\end{aligned} (54)

which proves Theorem 5. Then Theorem 4 is obtained by plugging back into the backward recursion (Proposition 3).

For uniqueness, we want BV\|\mathbb{B}V-BV\mathbb{B}V^{\prime}\|_{\infty}\leqγV\gamma\|V-VV^{\prime}\|_{\infty}. Let V\|V-VV^{\prime}\|_{\infty}==ε\varepsilon (maxz|V(z)\text{max}_{z^{\prime}}|V(z^{\prime})-V(z)|V^{\prime}(z^{\prime})|==ε\varepsilon). Now, (BV)(z)(\mathbb{B}^{*}V)(z)

αlogEuπ~[exp(1α(Esp(|z)υ(s,u)+βlogEτ,ωσ~[exp(1βEzϱτ,ω(|z,u)[ηlogϱτ,ω(z|z,u)ϱ~(z)+γV(z)])]))]αlogEuπ~[exp(1α(Esp(|z)υ(s,u)+βlogEτ,ωσ~[exp(1βEzϱτ,ω(|z,u)[ηlogϱτ,ω(z|z,u)ϱ~(z)+γ(V(z)+ε)])]))]=αlogEuπ~[exp(1α(Esp(|z)υ(s,u)+βlogEτ,ωσ~[exp(1βγε+1βEzϱτ,ω(|z,u)[ηlogϱτ,ω(z|z,u)ϱ~(z)+γV(z)])]))]=αlogEuπ~[exp(1α(Esp(|z)υ(s,u)+βlog(exp(1βγε)Eτ,ωσ~[exp(1βEzϱτ,ω(|z,u)[ηlogϱτ,ω(z|z,u)ϱ~(z)+γV(z)])])))]=αlogEuπ~[exp(1αγε+1α(Esp(|z)υ(s,u)+βlogEτ,ωσ~[exp(1βEzϱτ,ω(|z,u)[ηlogϱτ,ω(z|z,u)ϱ~(z)+γV(z)])]))]=αlog(exp(1αγε)Euπ~[exp(1α(Esp(|z)υ(s,u)+βlogEτ,ωσ~[exp(1βEzϱτ,ω(|z,u)[ηlogϱτ,ω(z|z,u)ϱ~(z)+γV(z)])]))])=γε+αlogEuπ~[exp(1α(Esp(|z)υ(s,u)+βlogEτ,ωσ~[exp(1βEzϱτ,ω(|z,u)[ηlogϱτ,ω(z|z,u)ϱ~(z)+γV(z)])]))]=γε+(BV)(z)\displaystyle\begin{aligned} &\doteq\alpha\log\mathbb{E}_{u\sim\tilde{\pi}}\big{[}\exp\big{(}\tfrac{1}{\alpha}\big{(}\mathbb{E}_{s\sim p(\cdot|z)}\upsilon(s,u)\\ &~{}~{}~{}~{}+\beta\log\mathbb{E}_{\tau,\omega\sim\tilde{\sigma}}\big{[}\exp\big{(}\tfrac{1}{\beta}\mathbb{E}_{z^{\prime}\sim\varrho_{\tau,\omega}(\cdot|z,u)}\big{[}\\ &~{}~{}~{}~{}-\eta\log\scalebox{1.2}{$\frac{\varrho_{\tau,\omega}(z^{\prime}|z,u)}{\tilde{\varrho}(z^{\prime})}$}+\kern 1.00006pt\gamma V(z^{\prime})\big{]}\big{)}\big{]}\big{)}\big{)}\big{]}\\ &\leq\alpha\log\mathbb{E}_{u\sim\tilde{\pi}}\big{[}\exp\big{(}\tfrac{1}{\alpha}\big{(}\mathbb{E}_{s\sim p(\cdot|z)}\upsilon(s,u)\\ &~{}~{}~{}~{}+\beta\log\mathbb{E}_{\tau,\omega\sim\tilde{\sigma}}\big{[}\exp\big{(}\tfrac{1}{\beta}\mathbb{E}_{z^{\prime}\sim\varrho_{\tau,\omega}(\cdot|z,u)}\big{[}\\ &~{}~{}~{}~{}-\eta\log\scalebox{1.2}{$\frac{\varrho_{\tau,\omega}(z^{\prime}|z,u)}{\tilde{\varrho}(z^{\prime})}$}+\kern 1.00006pt\gamma(V^{\prime}(z^{\prime})+\varepsilon)\big{]}\big{)}\big{]}\big{)}\big{)}\big{]}\\ &=\alpha\log\mathbb{E}_{u\sim\tilde{\pi}}\big{[}\exp\big{(}\tfrac{1}{\alpha}\big{(}\mathbb{E}_{s\sim p(\cdot|z)}\upsilon(s,u)\\ &~{}~{}~{}~{}+\beta\log\mathbb{E}_{\tau,\omega\sim\tilde{\sigma}}\big{[}\exp\big{(}\tfrac{1}{\beta}\gamma\varepsilon+\tfrac{1}{\beta}\mathbb{E}_{z^{\prime}\sim\varrho_{\tau,\omega}(\cdot|z,u)}\big{[}\\ &~{}~{}~{}~{}-\eta\log\scalebox{1.2}{$\frac{\varrho_{\tau,\omega}(z^{\prime}|z,u)}{\tilde{\varrho}(z^{\prime})}$}+\kern 1.00006pt\gamma V^{\prime}(z^{\prime})\big{]}\big{)}\big{]}\big{)}\big{)}\big{]}\\ &=\alpha\log\mathbb{E}_{u\sim\tilde{\pi}}\big{[}\exp\big{(}\tfrac{1}{\alpha}\big{(}\mathbb{E}_{s\sim p(\cdot|z)}\upsilon(s,u)\\ &~{}~{}~{}~{}+\beta\log\big{(}\exp(\tfrac{1}{\beta}\gamma\varepsilon)\mathbb{E}_{\tau,\omega\sim\tilde{\sigma}}\big{[}\exp\big{(}\tfrac{1}{\beta}\mathbb{E}_{z^{\prime}\sim\varrho_{\tau,\omega}(\cdot|z,u)}\big{[}\\ &~{}~{}~{}~{}-\eta\log\scalebox{1.2}{$\frac{\varrho_{\tau,\omega}(z^{\prime}|z,u)}{\tilde{\varrho}(z^{\prime})}$}+\kern 1.00006pt\gamma V^{\prime}(z^{\prime})\big{]}\big{)}\big{]}\big{)}\big{)}\big{)}\big{]}\\ &=\alpha\log\mathbb{E}_{u\sim\tilde{\pi}}\big{[}\exp\big{(}\tfrac{1}{\alpha}\gamma\varepsilon+\tfrac{1}{\alpha}\big{(}\mathbb{E}_{s\sim p(\cdot|z)}\upsilon(s,u)\\ &~{}~{}~{}~{}+\beta\log\mathbb{E}_{\tau,\omega\sim\tilde{\sigma}}\big{[}\exp\big{(}\tfrac{1}{\beta}\mathbb{E}_{z^{\prime}\sim\varrho_{\tau,\omega}(\cdot|z,u)}\big{[}\\ &~{}~{}~{}~{}-\eta\log\scalebox{1.2}{$\frac{\varrho_{\tau,\omega}(z^{\prime}|z,u)}{\tilde{\varrho}(z^{\prime})}$}+\kern 1.00006pt\gamma V^{\prime}(z^{\prime})\big{]}\big{)}\big{]}\big{)}\big{)}\big{]}\\ &=\alpha\log\big{(}\exp(\tfrac{1}{\alpha}\gamma\varepsilon)\mathbb{E}_{u\sim\tilde{\pi}}\big{[}\exp\big{(}\tfrac{1}{\alpha}\big{(}\mathbb{E}_{s\sim p(\cdot|z)}\upsilon(s,u)\\ &~{}~{}~{}~{}+\beta\log\mathbb{E}_{\tau,\omega\sim\tilde{\sigma}}\big{[}\exp\big{(}\tfrac{1}{\beta}\mathbb{E}_{z^{\prime}\sim\varrho_{\tau,\omega}(\cdot|z,u)}\big{[}\\ &~{}~{}~{}~{}-\eta\log\scalebox{1.2}{$\frac{\varrho_{\tau,\omega}(z^{\prime}|z,u)}{\tilde{\varrho}(z^{\prime})}$}+\kern 1.00006pt\gamma V^{\prime}(z^{\prime})\big{]}\big{)}\big{]}\big{)}\big{)}\big{]}\big{)}\\ &=\gamma\varepsilon+\alpha\log\mathbb{E}_{u\sim\tilde{\pi}}\big{[}\exp\big{(}\tfrac{1}{\alpha}\big{(}\mathbb{E}_{s\sim p(\cdot|z)}\upsilon(s,u)\\ &~{}~{}~{}~{}+\beta\log\mathbb{E}_{\tau,\omega\sim\tilde{\sigma}}\big{[}\exp\big{(}\tfrac{1}{\beta}\mathbb{E}_{z^{\prime}\sim\varrho_{\tau,\omega}(\cdot|z,u)}\big{[}\\ &~{}~{}~{}~{}-\eta\log\scalebox{1.2}{$\frac{\varrho_{\tau,\omega}(z^{\prime}|z,u)}{\tilde{\varrho}(z^{\prime})}$}+\kern 1.00006pt\gamma V^{\prime}(z^{\prime})\big{]}\big{)}\big{]}\big{)}\big{)}\big{]}\\ &=\gamma\varepsilon+(\mathbb{B}^{*}V^{\prime})(z)\end{aligned} (55)

Likewise, we can show that (BV)(z)(BV)(z)γε(\mathbb{B}^{*}V)(z)\geq(\mathbb{B}^{*}V^{\prime})(z)-\gamma\varepsilon. Hence maxz|(BV)(z)(BV)(z)|=BVBVγϵ\text{max}_{z}|(\mathbb{B}V)(z)-(\mathbb{B}V^{\prime})(z)|=\|\mathbb{B}V-\mathbb{B}V^{\prime}\|_{\infty}\leq\gamma\epsilon.

Note on Equation 24: Note that we originally formulated “soft policy matching” in Table 3 as a forward Kullback-Leibler divergence expression. However, analogously to maximum likelihood in supervised learning, the entropy terms drop out of the optimization, which yields Equation 24. To see this, note that the causally-conditioned probability is simply the product of conditional probabilities at each time step, and each conditional is “Markovianized” using beliefs ztz_{t} (i.e. Equation 25).

Appendix D Illustrative Trajectories

Here we direct attention to the potential utility of IBRC (and—more generally—instantiations of the IDM paradigm) as an “investigative device” for auditing and quantifying individual decisions. In Figure 7, we see that modeling the evolution of a decision-maker’s subjective beliefs provides a concrete basis for analyzing the corresponding sequence of actions chosen. Each vertex of the belief simplex corresponds to one of the three stable Alzheimer’s diagnoses, and each point within the simplex corresponds to a unique belief (i.e. probability distribution). The closer the point is to a vertex (i.e. disease state), the higher the probability assigned to that state. For instance, if the belief is located exactly in the middle of the simplex (i.e. equidistant from all vertices), then all states are believed to be equally likely. Note that this is visual presentation is done similarly to [22], where decision trajectories within belief simplices are first visualized in this manner—with the core difference here being that the decision policies (hence decision boundaries thereby induced) are computed using a different technique.

An MRI is less likely to be ordered.An MRI is more likely to be ordered.An MRI is not ordered.An MRI is ordered.Belief updatesFinal beliefsDecision boundaryBelief simplex
NLDementiaMCI
(a) Patient treated as if “rationally”
NLDementiaMCI
(b) Patient not treated “rationally”
NLDementiaMCI
(c) Patient is diagnosed belatedly
Figure 7: Decision Trajectories. Examples of apparent beliefs and actions of a clinical decision-maker regarding real patients, including cases where: (a) the clinician’s decisions coincide with those that would have been dictated by a “perfectly-rational” policy—despite their bounded rationality; (b) the clinician fails to make “perfectly-rational” decisions (in this context, the “boundedness” of the clinician could be due to any number of issues encountered during the diagnostic process); and (c) a patient who—apparently—could have been diagnosed much earlier than they actually were, but for the clinician not having followed the decisions prescribed by the “perfectly-rational” policy.

Appendix E Summary of Notation

Notation Meaning (first defined in) Notation Meaning (first defined in)
ψ\psi problem setting Section 3.1 ss environment state Section 3.1
xx environment emission Section 3.1 zz agent state, i.e. belief Section 3.1
uu agent emission, i.e. action Section 3.1 τenv\tau_{\text{env}} environment transition Section 3.1
τ\tau subjective transition Section 3.1 ωenv\omega_{\text{env}} environment emission Section 3.1
ω\omega subjective emission Section 3.1 υ\upsilon utility (i.e. reward) function Section 3.1
γ\gamma discount factor Section 3.1 ϕ\phi behavior Section 3.1
ϕdemo\phi_{\text{demo}} demonstrated behavior Section 3.2 ϕimit\phi_{\text{imit}} imitation behavior Section 3.2
θ\theta planning parameter Section 3.1 θnorm\theta_{\text{norm}} normative parameter Section 3.2
θdesc\theta_{\text{desc}} descriptive parameter Section 3.2 π\pi decision policy Section 3.1
ρ\rho recognition policy Section 3.1 σ\sigma specification policy Section 4.1
FF forward planner Section 3.1 GG inverse planner Section 3.2
α1\alpha^{-1} flexibility coefficient Section 4.2 β1\beta^{-1} optimism coefficient Section 4.2
η1\eta^{-1} adaptivity coefficient Section 4.2 π~\tilde{\pi} action prior Section 4.2
σ~\tilde{\sigma} model prior Section 4.2 ϱ~\tilde{\varrho} belief prior Section 4.2

References

  • [1] Aiping Li, Songchang Jin, Lumin Zhang, and Yan Jia. A sequential decision-theoretic model for medical diagnostic system. Technology and Healthcare, 2015.
  • [2] John A Clithero. Response times in economics: Looking through the lens of sequential sampling models. Journal of Economic Psychology, 2018.
  • [3] Jan Drugowitsch, Rubén Moreno-Bote, and Alexandre Pouget. Relation between belief and performance in perceptual decision making. PloS one, 2014.
  • [4] Gregory Wheeler. Bounded rationality. SEP: Stanford Center for the Study of Language and Information, 2018.
  • [5] Thomas L Griffiths, Falk Lieder, and Noah D Goodman. Rational use of cognitive resources: Levels of analysis between the computational and the algorithmic. Topics in cognitive science, 2015.
  • [6] Tim Genewein, Felix Leibfried, Jordi Grau-Moya, and Daniel Alexander Braun. Bounded rationality, abstraction, and hierarchical decision-making: An information-theoretic optimality principle. Frontiers in Robotics and AI, 2015.
  • [7] Ned Augenblick and Matthew Rabin. Belief movement, uncertainty reduction, and rational updating. UC Berkeley-Haas and Harvard University Mimeo, 2018.
  • [8] Pedro A Ortega, Daniel A Braun, Justin Dyer, Kee-Eung Kim, and Naftali Tishby. Information-theoretic bounded rationality. arXiv preprint, 2015.
  • [9] L Robin Keller. The role of generalized utility theories in descriptive, prescriptive, and normative decision analysis. Information and Decision Technologies, 1989.
  • [10] Ludwig Johann Neumann, Oskar Morgenstern, et al. Theory of games and economic behavior. Princeton university press Princeton, 1947.
  • [11] Barbara A Mellers, Alan Schwartz, and Alan DJ Cooke. Judgment and decision making. Annual review of psychology, 1998.
  • [12] Yisong Yue and Hoang M Le. Imitation learning (presentation). International Conference on Machine Learning (ICML), 2018.
  • [13] Pieter Abbeel and Andrew Y Ng. Apprenticeship learning via inverse reinforcement learning. International conference on Machine learning (ICML), 2004.
  • [14] Dean A Pomerleau. Efficient training of artificial neural networks for autonomous navigation. Neural computation (NC), 1991.
  • [15] Michael Bain and Claude Sammut. A framework for behavioural cloning. Machine Intelligence (MI), 1999.
  • [16] Umar Syed and Robert E Schapire. Imitation learning with a value-based prior. Conference on Uncertainty in Artificial Intelligence (UAI), 2007.
  • [17] Stéphane Ross and Drew Bagnell. Efficient reductions for imitation learning. International conference on artificial intelligence and statistics (AISTATS), 2010.
  • [18] Umar Syed and Robert E Schapire. A reduction from apprenticeship learning to classification. Advances in neural information processing systems (NeurIPS), 2010.
  • [19] Stéphane Ross, Geoffrey Gordon, and Drew Bagnell. A reduction of imitation learning and structured prediction to no-regret online learning. International conference on artificial intelligence and statistics (AISTATS), 2011.
  • [20] Bilal Piot, Matthieu Geist, and Olivier Pietquin. Boosted and reward-regularized classification for apprenticeship learning. International conference on Autonomous agents and multi-agent systems (AAMAS), 2014.
  • [21] Daniel Jarrett, Ioana Bica, and Mihaela van der Schaar. Strictly batch imitation learning by energy-based distribution matching. Advances in neural information processing systems (NeurIPS), 2020.
  • [22] Alihan Hüyük, Daniel Jarrett, Cem Tekin, and Mihaela van der Schaar. Explaining by imitating: Understanding decisions by interpretable policy learning. International Conference on Learning Representations (ICLR), 2021.
  • [23] Lionel Blondé and Alexandros Kalousis. Sample-efficient imitation learning via gans. International conference on artificial intelligence and statistics (AISTATS), 2019.
  • [24] Ilya Kostrikov, Kumar Krishna Agrawal, Debidatta Dwibedi, Sergey Levine, and Jonathan Tompson. Discriminator-actor-critic: Addressing sample inefficiency and reward bias in adversarial imitation. International Conference on Learning Representations (ICLR), 2019.
  • [25] Jonathan Ho and Stefano Ermon. Generative adversarial imitation learning. Advances in neural information processing systems (NeurIPS), 2016.
  • [26] Wonseok Jeon, Seokin Seo, and Kee-Eung Kim. A bayesian approach to generative adversarial imitation learning. Advances in Neural Information Processing Systems (NeurIPS), 2018.
  • [27] Seyed Kamyar Seyed Ghasemipour, Richard Zemel, and Shixiang Gu. Understanding the relation of bc and irl through divergence minimization. ICML Workshop on Deep Generative Models for Highly Structured Data, 2019.
  • [28] Seyed Kamyar Seyed Ghasemipour, Richard Zemel, and Shixiang Gu. A divergence minimization perspective on imitation learning methods. Conference on Robot Learning (CoRL), 2019.
  • [29] Liyiming Ke, Matt Barnes, Wen Sun, Gilwoo Lee, Sanjiban Choudhury, and Siddhartha Srinivasa. Imitation learning as ff-divergence minimization. arXiv preprint, 2019.
  • [30] Liyiming Ke, Matt Barnes, Wen Sun, Gilwoo Lee, Sanjiban Choudhury, and Siddhartha Srinivasa. Imitation learning as ff-divergence minimization. International Workshop on the Algorithmic Foundations of Robotics (WAFR), 2020.
  • [31] Kee-Eung Kim and Hyun Soo Park. Imitation learning via kernel mean embedding. AAAI Conference on Artificial Intelligence (AAAI), 2018.
  • [32] Huang Xiao, Michael Herman, Joerg Wagner, Sebastian Ziesche, Jalal Etesami, and Thai Hong Linh. Wasserstein adversarial imitation learning. arXiv preprint, 2019.
  • [33] Robert Dadashi, Leonard Hussenot, Matthieu Geist, and Olivier Pietquin. Primal wasserstein imitation learning. International Conference on Learning Representations (ICLR), 2021.
  • [34] Ilya Kostrikov, Ofir Nachum, and Jonathan Tompson. Imitation learning via off-policy distribution matching. International Conference on Learning Representations (ICLR), 2020.
  • [35] Oleg Arenz and Gerhard Neumann. Non-adversarial imitation learning and its connections to adversarial methods. arXiv preprint, 2020.
  • [36] Srivatsan Srinivasan and Finale Doshi-Velez. Interpretable batch irl to extract clinician goals in icu hypotension management. AMIA Summits on Translational Science Proceedings, 2020.
  • [37] Xin Zhang, Yanhua Li, Ziming Zhang, and Zhi-Li Zhang. ff-gail: Learning ff-divergence for generative adversarial imitation learning. Advances in Neural Information Processing Systems (NeurIPS), 2020.
  • [38] Nir Baram, Oron Anschel, and Shie Mannor. Model-based adversarial imitation learning. arXiv preprint, 2016.
  • [39] Nir Baram, Oron Anschel, and Shie Mannor. Model-based adversarial imitation learning. International Conference on Machine Learning (ICML), 2017.
  • [40] Andrew Y Ng, Stuart J Russell, et al. Algorithms for inverse reinforcement learning. International conference on Machine learning (ICML), 2000.
  • [41] Umar Syed and Robert E Schapire. A game-theoretic approach to apprenticeship learning. Advances in neural information processing systems (NeurIPS), 2008.
  • [42] Umar Syed, Michael Bowling, and Robert E Schapire. Apprenticeship learning using linear programming. International conference on Machine learning (ICML), 2008.
  • [43] Edouard Klein, Matthieu Geist, and Olivier Pietquin. Batch, off-policy and model-free apprenticeship learning. European Workshop on Reinforcement Learning (EWRL), 2011.
  • [44] Takeshi Mori, Matthew Howard, and Sethu Vijayakumar. Model-free apprenticeship learning for transfer of human impedance behaviour. IEEE-RAS International Conference on Humanoid Robots, 2011.
  • [45] Donghun Lee, Srivatsan Srinivasan, and Finale Doshi-Velez. Truly batch apprenticeship learning with deep successor features. International Joint Conference on Artificial Intelligence (IJCAI), 2019.
  • [46] Bilal Piot, Matthieu Geist, and Olivier Pietquin. Bridging the gap between imitation learning and irl. IEEE transactions on neural networks and learning systems, 2017.
  • [47] Edouard Klein, Matthieu Geist, Bilal Piot, and Olivier Pietquin. Irl through structured classification. Advances in neural information processing systems (NeurIPS), 2012.
  • [48] Edouard Klein, Bilal Piot, Matthieu Geist, and Olivier Pietquin. A cascaded supervised learning approach to inverse reinforcement learning. Joint European conference on machine learning and knowledge discovery in databases (ECML), 2013.
  • [49] Aristide CY Tossou and Christos Dimitrakakis. Probabilistic inverse reinforcement learning in unknown environments. Conference on Uncertainty in Artificial Intelligence (UAI), 2013.
  • [50] Vinamra Jain, Prashant Doshi, and Bikramjit Banerjee. Model-free irl using maximum likelihood estimation. AAAI Conference on Artificial Intelligence (AAAI), 2019.
  • [51] Gergely Neu and Csaba Szepesvári. Apprenticeship learning using irl and gradient methods. Conference on Uncertainty in Artificial Intelligence (UAI), 2007.
  • [52] Monica Babes, Vukosi Marivate, and Michael L Littman. Apprenticeship learning about multiple intentions. International conference on Machine learning (ICML), 2011.
  • [53] Jonathan Ho, Jayesh Gupta, and Stefano Ermon. Model-free imitation learning with policy optimization. International Conference on Machine Learning (ICML), 2016.
  • [54] Chelsea Finn, Sergey Levine, and Pieter Abbeel. Guided cost learning: Deep inverse optimal control via policy optimization. International conference on machine learning (ICML), 2016.
  • [55] Matteo Pirotta and Marcello Restelli. Inverse reinforcement learning through policy gradient minimization. AAAI Conference on Artificial Intelligence (AAAI), 2016.
  • [56] Alberto Maria Metelli, Matteo Pirotta, and Marcello Restelli. Compatible reward inverse reinforcement learning. Advances in Neural Information Processing Systems (NeurIPS), 2017.
  • [57] Davide Tateo, Matteo Pirotta, Marcello Restelli, and Andrea Bonarini. Gradient-based minimization for multi-expert inverse reinforcement learning. IEEE Symposium Series on Computational Intelligence (SSCI), 2017.
  • [58] Gergely Neu and Csaba Szepesvári. Training parsers by inverse reinforcement learning. Machine learning (ML), 2009.
  • [59] Deepak Ramachandran and Eyal Amir. Bayesian inverse reinforcement learning. International Joint Conference on Artificial Intelligence (IJCAI), 2007.
  • [60] Jaedeug Choi and Kee-Eung Kim. Map inference for bayesian irl. Advances in Neural Information Processing Systems (NeurIPS), 2011.
  • [61] Christos Dimitrakakis and Constantin A Rothkopf. Bayesian multitask irl. European workshop on reinforcement learning (EWRL), 2011.
  • [62] Constantin A Rothkopf and Christos Dimitrakakis. Preference elicitation and inverse reinforcement learning. Joint European conference on machine learning and knowledge discovery in databases (ECML), 2011.
  • [63] Sreejith Balakrishnan, Quoc Phong Nguyen, Bryan Kian Hsiang Low, and Harold Soh. Efficient exploration of reward functions in inverse reinforcement learning via bayesian optimization. Advances in Neural Information Processing Systems (NeurIPS), 2020.
  • [64] Ajay Kumar Tanwani and Aude Billard. Inverse reinforcement learning for compliant manipulation in letter handwriting. National Center of Competence in Robotics (NCCR), 2013.
  • [65] McKane Andrus. Inverse reinforcement learning for dynamics. Dissertation, University of California at Berkeley, 2019.
  • [66] Stav Belogolovsky, Philip Korsunsky, Shie Mannor, Chen Tessler, and Tom Zahavy. Learning personalized treatments via irl. arXiv preprint, 2019.
  • [67] Sid Reddy, Anca Dragan, and Sergey Levine. Where do you think you’re going?: Inferring beliefs about dynamics from behavior. Advances in Neural Information Processing Systems (NeurIPS), 2018.
  • [68] Anirudha Majumdar, Sumeet Singh, Ajay Mandlekar, and Marco Pavone. Risk-sensitive inverse reinforcement learning via coherent risk models. Robotics: Science and Systems, 2017.
  • [69] Sumeet Singh, Jonathan Lacotte, Anirudha Majumdar, and Marco Pavone. Risk-sensitive inverse reinforcement learning via semi-and non-parametric methods. International Journal of Robotics Research, 2018.
  • [70] Jaedeug Choi and Kee-Eung Kim. Inverse reinforcement learning in partially observable environments. International Joint Conference on Artificial Intelligence (IJCAI), 2009.
  • [71] Jaedeug Choi and Kee-Eung Kim. Inverse reinforcement learning in partially observable environments. Journal of Machine Learning Research (JMLR), 2011.
  • [72] Hamid R Chinaei and Brahim Chaib-Draa. An inverse reinforcement learning algorithm for partially observable domains with application on healthcare dialogue management. International Conference on Machine Learning and Applications, 2012.
  • [73] Ioana Bica, Daniel Jarrett, Alihan Hüyük, and Mihaela van der Schaar. Learning what-if explanations for sequential decision-making. International Conference on Learning Representations (ICLR), 2021.
  • [74] Takaki Makino and Johane Takeuchi. Apprenticeship learning for model parameters of partially observable environments. International Conference on Machine Learning (ICML), 2012.
  • [75] Daniel Jarrett and Mihaela van der Schaar. Inverse active sensing: Modeling and understanding timely decision-making. International Conference on Machine Learning, 2020.
  • [76] Kunal Pattanayak and Vikram Krishnamurthy. Inverse reinforcement learning for sequential hypothesis testing and search. International Conference on Information Fusion (FUSION), 2020.
  • [77] Matthew Golub, Steven Chase, and Byron Yu. Learning an internal dynamics model from control demonstration. International Conference on Machine Learning (ICML), 2013.
  • [78] Zhengwei Wu, Paul Schrater, and Xaq Pitkow. Inverse pomdp: Inferring what you think from what you do. arXiv preprint, 2018.
  • [79] Saurabh Daptardar, Paul Schrater, and Xaq Pitkow. Inverse rational control with partially observable continuous nonlinear dynamics. arXiv preprint, 2019.
  • [80] Minhae Kwon, Saurabh Daptardar, Paul Schrater, and Xaq Pitkow. Inverse rational control with partially observable continuous nonlinear dynamics. Advances in Neural Information Processing Systems (NeurIPS), 2020.
  • [81] Brian D Ziebart, Andrew L Maas, J Andrew Bagnell, and Anind K Dey. Maximum entropy inverse reinforcement learning. AAAI Conference on Artificial Intelligence (AAAI), 2008.
  • [82] Abdeslam Boularias, Jens Kober, and Jan Peters. Relative entropy inverse reinforcement learning. International conference on artificial intelligence and statistics (AISTATS), 2011.
  • [83] Mrinal Kalakrishnan, Peter Pastor, Ludovic Righetti, and Stefan Schaal. Learning objective functions for manipulation. International Conference on Robotics and Automation (ICRA), 2013.
  • [84] Markus Wulfmeier, Peter Ondruska, and Ingmar Posner. Maximum entropy deep inverse reinforcement learning. arXiv preprint, 2015.
  • [85] Chelsea Finn, Paul Christiano, Pieter Abbeel, and Sergey Levine. A connection between generative adversarial networks, inverse reinforcement learning, and energy-based models. NeurIPS Workshop on Adversarial Training, 2016.
  • [86] Justin Fu, Katie Luo, and Sergey Levine. Learning robust rewards with adversarial inverse reinforcement learning. International Conference on Learning Representations (ICLR), 2018.
  • [87] Ahmed H Qureshi, Byron Boots, and Michael C Yip. Adversarial imitation via variational inverse reinforcement learning. International Conference on Learning Representations (ICLR), 2019.
  • [88] Paul Barde, Julien Roy, Wonseok Jeon, Joelle Pineau, Christopher Pal, and Derek Nowrouzezahrai. Adversarial soft advantage fitting: Imitation learning without policy optimization. Advances in neural information processing systems (NeurIPS), 2020.
  • [89] Brian D Ziebart, J Andrew Bagnell, and Anind K Dey. Modeling interaction via the principle of maximum causal entropy. International conference on Machine learning (ICML), 2010.
  • [90] Zhengyuan Zhou, Michael Bloem, and Nicholas Bambos. Infinite time horizon maximum causal entropy inverse reinforcement learning. IEEE Transactions on Automatic Control (TACON), 2017.
  • [91] Kyungjae Lee, Sungjoon Choi, and Songhwai Oh. Maximum causal tsallis entropy imitation learning. Advances in Neural Information Processing Systems (NeurIPS), 2018.
  • [92] Tien Mai, Kennard Chan, and Patrick Jaillet. Generalized maximum causal entropy for inverse reinforcement learning. AAAI Conference on Artificial Intelligence (AAAI), 2019.
  • [93] Michael Herman, Tobias Gindele, Jörg Wagner, Felix Schmitt, and Wolfram Burgard. Inverse reinforcement learning with simultaneous estimation of rewards and dynamics. International conference on artificial intelligence and statistics (AISTATS), 2016.
  • [94] Michael Herman. Simultaneous estimation of rewards and dynamics in irl. Dissertation, Albert-Ludwigs-Universitat Freiburg, 2016.
  • [95] Layla El Asri, Bilal Piot, Matthieu Geist, Romain Laroche, and Olivier Pietquin. Score-based inverse reinforcement learning. International conference on Autonomous agents and multi-agent systems (AAMAS), 2016.
  • [96] Benjamin Burchfiel, Carlo Tomasi, and Ronald Parr. Distance minimization for reward learning from scored trajectories. AAAI Conference on Artificial Intelligence (AAAI), 2016.
  • [97] Alexis Jacq, Matthieu Geist, Ana Paiva, and Olivier Pietquin. Learning from a learner. International Conference on Machine Learning (ICML), 2019.
  • [98] Daniel S Brown, Wonjoon Goo, Prabhat Nagarajan, and Scott Niekum. Extrapolating beyond suboptimal demonstrations via inverse reinforcement learning from observations. International Conference on Machine Learning (ICML), 2019.
  • [99] Daniel S Brown, Wonjoon Goo, and Scott Niekum. Better-than-demonstrator imitation learning via automatically-ranked demonstrations. Conference on Robot Learning (CoRL), 2020.
  • [100] Takayuki Osa, Joni Pajarinen, Gerhard Neumann, J Andrew Bagnell, Pieter Abbeel, and Jan Peters. An algorithmic perspective on imitation learning. Foundations and Trends in Robotics, 2018.
  • [101] Tuomas Haarnoja, Haoran Tang, Pieter Abbeel, and Sergey Levine. Reinforcement learning with deep energy-based policies. International Conference on Machine Learning (ICML), 2017.
  • [102] Tuomas Haarnoja, Aurick Zhou, Pieter Abbeel, and Sergey Levine. Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. International Conference on Machine Learning (ICML), 2018.
  • [103] Benjamin Eysenbach and Sergey Levine. If maxent rl is the answer, what is the question? arXiv preprint, 2019.
  • [104] Wenjie Shi, Shiji Song, and Cheng Wu. Soft policy gradient method for maximum entropy deep reinforcement learning. International Joint Conference on Artificial Intelligence (IJCAI), 2019.
  • [105] Rohin Shah, Noah Gundotra, Pieter Abbeel, and Anca Dragan. Inferring reward functions from demonstrators with unknown biases. OpenReview, 2018.
  • [106] Rohin Shah, Noah Gundotra, Pieter Abbeel, and Anca D Dragan. On the feasibility of learning, rather than assuming, human biases for reward inference. International Conference on Machine Learning (ICML), 2019.
  • [107] Jonathan Rubin, Ohad Shamir, and Naftali Tishby. Trading value and information in mdps. Decision Making with Imperfect Decision Makers (Springer), 2012.
  • [108] Alexandre Galashov, Siddhant M Jayakumar, Leonard Hasenclever, Dhruva Tirumala, Jonathan Schwarz, Guillaume Desjardins, Wojciech M Czarnecki, Yee Whye Teh, Razvan Pascanu, and Nicolas Heess. Information asymmetry in kl-regularized rl. International Conference on Learning Representations (ICLR), 2019.
  • [109] Mark K Ho, David Abel, Jonathan D Cohen, Michael L Littman, and Thomas L Griffiths. The efficiency of human cognition reflects planned information processing. AAAI Conference on Artificial Intelligence (AAAI), 2020.
  • [110] Stas Tiomkin and Naftali Tishby. A unified bellman equation for causal information and value in markov decision processes. arXiv preprint arXiv:1703.01585, 2017.
  • [111] Felix Leibfried, Jordi Grau-Moya, and Haitham Bou-Ammar. An information-theoretic optimality principle for deep reinforcement learning. NeurIPS Workshop on Deep Reinforcement Learning, 2017.
  • [112] Matthew Hausknecht and Peter Stone. Deep recurrent q-learning for partially observable mdps. AAAI Conference on Artificial Intelligence (AAAI), 2015.
  • [113] Pengfei Zhu, Xin Li, Pascal Poupart, and Guanghui Miao. On improving deep reinforcement learning for pomdps. arXiv preprint, 2017.
  • [114] Maximilian Igl, Luisa Zintgraf, Tuan Anh Le, Frank Wood, and Shimon Whiteson. Deep variational reinforcement learning for pomdps. International Conference on Machine Learning (ICML), 2018.
  • [115] Amy Zhang, Zachary C Lipton, Luis Pineda, Kamyar Azizzadenesheli, Anima Anandkumar, Laurent Itti, Joelle Pineau, and Tommaso Furlanello. Learning causal state representations of partially observable environments. arXiv preprint, 2019.
  • [116] Dongqi Han, Kenji Doya, and Jun Tani. Variational recurrent models for solving partially observable control tasks. arXiv preprint arXiv:1912.10703, 2019.
  • [117] Joseph Futoma, Michael C Hughes, and Finale Doshi-Velez. Popcorn: Partially observed prediction constrained reinforcement learning. International conference on artificial intelligence and statistics (AISTATS), 2020.
  • [118] Richard D Smallwood and Edward J Sondik. The optimal control of partially observable markov processes over a finite horizon. Operations research, 1973.
  • [119] Milos Hauskrecht. Value-function approximations for partially observable markov decision processes. Journal of Artificial Intelligence Research (JAIR), 2000.
  • [120] Joelle Pineau, Geoff Gordon, Sebastian Thrun, et al. Point-based value iteration: An anytime algorithm for pomdps. International Joint Conference on Artificial Intelligence (IJCAI), 2003.
  • [121] Hanna Kurniawati, David Hsu, and Wee Sun Lee. Sarsop: Efficient point-based pomdp planning by approximating optimally reachable belief spaces. Robotics: Science and systems, 2008.
  • [122] Mauricio Araya, Olivier Buffet, Vincent Thomas, and Françcois Charpillet. A pomdp extension with belief-dependent rewards. Advances in Neural Information Processing Systems (NeurIPS), 2010.
  • [123] Mathieu Fehr, Olivier Buffet, Vincent Thomas, and Jilles Dibangoye. rho-pomdps have lipschitz-continuous epsilon-optimal value functions. Advances in Neural Information Processing Systems (NeurIPS), 2018.
  • [124] F A Sonnenberg and J R Beck. Markov models in medical decision making: a practical guide. Health Econ., 1983.
  • [125] C H Jackson, L D Sharples, S G Thompson, S W Duffy, and E Couto. Multistate Markov models for disease progression with classification error. Statistician, 2003.
  • [126] S E O’Bryant, S C Waring, C M Cullum, J Hall, L Lacritz, P J Massman, P J Lupo, J S Reisch, and R Doody. Staging dementia using Clinical Dementia Rating Scale Sum of Boxes scores: a Texas Alzheimer’s research consortium study. Arch. of Neurology, 2008.
  • [127] D Jarrett, J Yoon, and M van der Schaar. Match-net: Dynamic prediction in survival analysis using convolutional neural networks. NeurIPS Workshop on Machine Learning for Health, 2018.
  • [128] Daniel Jarrett, Jinsung Yoon, and Mihaela van der Schaar. Dynamic prediction in clinical survival analysis using temporal convolutional networks. IEEE Journal of Biomedical and Health Informatics, 2019.
  • [129] P Petousis, A Winter, W Speier, D R Aberle, W Hsu, and A A T Bui. Using sequential decision making to improve lung cancer screening performance. IEEE Access, 2019.
  • [130] F Cardoso, S Kyriakides, S Ohno, F Penault-Llorca, P Poortmans, I T Rubio, S Zackrisson, and E Senkus. Early breast cancer: ESMO Clinical Practice Guidelines for diagnosis, treatment and follow-up. Anna. Oncology, 2019.
  • [131] A M Alaa and M van der Schaar. Attentive state-space modeling of disease progression. Advances in neural information processing systems (NeurIPS), 2019.
  • [132] X Wang, D Sontag, and F Wang. Unsupervised learning of disease progression models. ACM SIGKDD Int. Conf. Knowl. Discovery Data Mining, 2014.
  • [133] Clemens Heuberger. Inverse combinatorial optimization. Journal of combinatorial optimization, 2004.
  • [134] Kareem Amin and Satinder Singh. Towards resolving unidentifiability in inverse reinforcement learning. arXiv preprint, 2016.
  • [135] Kareem Amin, Nan Jiang, and Satinder Singh. Repeated inverse reinforcement learning. Advances in neural information processing systems (NeurIPS), 2017.
  • [136] Stuart Armstrong and Sören Mindermann. Occam’s razor is insufficient to infer the preferences of irrational agents. Advances in neural information processing systems (NeurIPS), 2018.
  • [137] Paul Christiano. The easy goal inference problem is still hard. AI Alignment, 2015.
  • [138] Eric J Michaud, Adam Gleave, and Stuart Russell. Understanding learned reward functions. NeurIPS Workshop on Deep Reinforcement Learning, 2020.
  • [139] Adam Gleave, Michael Dennis, Shane Legg, Stuart Russell, and Jan Leike. Quantifying differences in reward functions. International Conference on Learning Representations (ICLR), 2021.
  • [140] Daniel S Brown and Scott Niekum. Deep bayesian reward learning from preferences. NeurIPS Workshop on Safety and Robustness in Decision-Making, 2019.
  • [141] Daniel S Brown, Russell Coleman, Ravi Srinivasan, and Scott Niekum. Safe imitation learning via fast bayesian reward inference from preferences. International Conference on Machine Learning (ICML), 2020.
  • [142] Nicolas Heess, David Silver, and Yee Whye Teh. Actor-critic reinforcement learning with energy-based policies. European Workshop on Reinforcement Learning (EWRL), 2013.
  • [143] Ian Osband, Benjamin Van Roy, and Zheng Wen. Generalization and exploration via randomized value functions. International Conference on Machine Learning (ICML), 2016.
  • [144] Nicolò Cesa-Bianchi, Claudio Gentile, Gábor Lugosi, and Gergely Neu. Boltzmann exploration done right. Advances in neural information processing systems (NeurIPS), 2017.
  • [145] Amir Globerson, Eran Stark, Eilon Vaadia, and Naftali Tishby. The minimum information principle and its application to neural code analysis. Proceedings of the National Academy of Sciences, 2009.
  • [146] Naftali Tishby and Daniel Polani. Information theory of decisions and actions. Perception-action cycle (Springer), 2011.
  • [147] Pedro A Ortega and Daniel A Braun. Thermodynamics as a theory of decision-making with information-processing costs. Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences, 2013.
  • [148] Ian R Petersen, Matthew R James, and Paul Dupuis. Minimax optimal control of stochastic uncertain systems with relative entropy constraints. IEEE Transactions on Automatic Control, 2000.
  • [149] Charalambos D Charalambous, Farzad Rezaei, and Andreas Kyprianou. Relations between information theory, robustness, and statistical mechanics of stochastic systems. IEEE Conference on Decision and Control (CDC), 2004.
  • [150] Takayuki Osogami. Robustness and risk-sensitivity in markov decision processes. Advances in Neural Information Processing Systems (NeurIPS), 2012.
  • [151] Jordi Grau-Moya, Felix Leibfried, Tim Genewein, and Daniel A Braun. Planning with information-processing constraints and model uncertainty in markov decision processes. Joint European conference on machine learning and knowledge discovery in databases (ECML), 2016.
  • [152] Brian D Ziebart. Modeling purposeful adaptive behavior with the principle of maximum causal entropy. Dissertation, Carnegie Mellon University, 2010.
  • [153] Gerhard Kramer. Directed information for channels with feedback. Dissertation, ETH Zurich, 1998.
  • [154] James Massey. Causality, feedback and directed information. International Symposium on Information Theory and Its Applications, 1990.
  • [155] Hans Marko. The bidirectional communication theory-a generalization of information theory. IEEE Transactions on Communications, 1973.
  • [156] John B McKinlay, Carol L Link, et al. Sources of variation in physician adherence with clinical guidelines. Journal of general internal medicine, 2007.
  • [157] Matthias Bock, Gerhard Fritsch, and David L Hepner. Preoperative laboratory testing. Anesthesiology clinics, 2016.
  • [158] Jack W O’Sullivan, Carl Heneghan, Rafael Perera, Jason Oke, Jeffrey K Aronson, Brian Shine, and Ben Goldacre. Variation in diagnostic test requests and outcomes: a preliminary metric for openpathology. net. Nature Scientific Reports, 2018.
  • [159] Yunjie Song, Jonathan Skinner, Julie Bynum, Jason Sutherland, John E Wennberg, and Elliott S Fisher. Regional variations in diagnostic practices. New England Journal of Medicine, (1), 2010.
  • [160] Shannon K Martin and Adam S Cifu. Routine preoperative laboratory tests for elective surgery. Journal of the American Medical Association (JAMA), 2017.
  • [161] M. Allen. Unnecessary tests and treatment explain why health care costs so much. Scientific American, 2017.
  • [162] Leslie Pack Kaelbling, Michael L Littman, and Anthony R Cassandra. Planning and acting in partially observable stochastic domains. Artificial intelligence, 1998.
  • [163] Razvan V Marinescu, Neil P Oxtoby, Alexandra L Young, Esther E Bron, Arthur W Toga, Michael W Weiner, Frederik Barkhof, Nick C Fox, Stefan Klein, Daniel C Alexander, et al. Tadpole challenge: Prediction of longitudinal evolution in alzheimer’s disease. arXiv preprint, 2018.
  • [164] Edi Karni and Zvi Safra. Behavioral consistency in sequential decisions. Progress in Decision, Utility and Risk Theory, 1991.
  • [165] Kent Daniel, David Hirshleifer, and Avanidhar Subrahmanyam. Investor psychology and security market under-and overreactions. The Journal of Finance, 1998.
  • [166] Amos Tversky and Daniel Kahneman. Evidential impact of base rates. Stanford University Department Of Psychology, 1981.
  • [167] Charlotte L Allan and Klaus P Ebmeier. The influence of apoe4 on clinical progression of dementia: a meta-analysis. International journal of geriatric psychiatry, 2011.
  • [168] Sylvaine Artero, Marie-Laure Ancelin, Florence Portet, A Dupuy, Claudine Berr, Jean-François Dartigues, Christophe Tzourio, Olivier Rouaud, Michel Poncet, Florence Pasquier, et al. Risk profiles for mild cognitive impairment and progression to dementia are gender specific. Journal of Neurology, Neurosurgery & Psychiatry, 2008.
  • [169] Xue Hua, Derrek P Hibar, Suh Lee, Arthur W Toga, Clifford R Jack Jr, Michael W Weiner, Paul M Thompson, Alzheimer’s Disease Neuroimaging Initiative, et al. Sex and age differences in atrophic rates: an adni study with n= 1368 mri scans. Neurobiology of aging, 2010.
  • [170] Brendan O’Donoghue, Remi Munos, Koray Kavukcuoglu, and Volodymyr Mnih. Combining policy gradient and q-learning. International Conference on Learning Representations (ICLR), 2017.
  • [171] Momchil Tomov. Structure learning and uncertainty-guided exploration in the human brain. Dissertation, Harvard University, 2020.
  • [172] Tianwei Ni, Harshit Sikchi, Yufei Wang, Tejus Gupta, Lisa Lee, and Benjamin Eysenbach. F-irl: Inverse reinforcement learning via state marginal matching. Conference on Robot Learning (CoRL), 2020.
  • [173] Hong Jun Jeon, Smitha Milli, and Anca D Dragan. Reward-rational (implicit) choice: A unifying formalism for reward learning. Advances in Neural Information Processing Systems (NeurIPS), 2020.
  • [174] Jeffrey Ely, Alexander Frankel, and Emir Kamenica. Suspense and surprise. Journal of Political Economy, 2015.
  • [175] Ahmed M Alaa and Mihaela van der Schaar. Balancing suspense and surprise: Timely decision making with endogenous information acquisition. Advances in neural information processing systems (NeurIPS), 2016.
  • [176] Owain Evans and Noah D Goodman. Learning the preferences of bounded agents. NeurIPS Workshop on Bounded Optimality, 2015.
  • [177] Tan Zhi-Xuan, Jordyn L Mann, Tom Silver, Joshua B Tenenbaum, and Vikash K Mansinghka. Online bayesian goal inference for boundedly-rational planning agents. Advances in neural information processing systems (NeurIPS), 2020.
  • [178] Osbert Bastani, Yewen Pu, and Armando Solar-Lezama. Verifiable reinforcement learning via policy extraction. Advances in neural information processing systems (NeurIPS), 2018.
  • [179] Herman Yau, Chris Russell, and Simon Hadfield. What did you think would happen? explaining agent behaviour through intended outcomes. Advances in neural information processing systems (NeurIPS), 2020.
  • [180] Tom Bewley, Jonathan Lawry, and Arthur Richards. Modelling agent policies with interpretable imitation learning. TAILOR Workshop at ECAI, 2020.
  • [181] Tathagata Chakraborti, Sarath Sreedharan, Yu Zhang, and Subbarao Kambhampati. Plan explanations as model reconciliation: Moving beyond explanation as soliloquy. International Joint Conference on Artificial Intelligence (IJCAI), 2017.
  • [182] Tathagata Chakraborti, Sarath Sreedharan, Sachin Grover, and Subbarao Kambhampati. Plan explanations as model reconciliation: an empirical study. International Conference on Human-Robot Interaction (HRI), 2019.
  • [183] Bradley Hayes and Julie A Shah. Improving robot controller transparency through autonomous policy explanation. International Conference on Human-Robot Interaction (HRI), 2017.
  • [184] Sarath Sreedharan, Utkash Soni, Mudit Verma, Siddharth Srivastava, and Subbarao Kambhampati. Bridging the gap: Providing post-hoc symbolic explanations for sequential decision-making problems with black box simulators. ICML Workshop on Human-in-the-Loop Learning, 2020.
  • [185] Roy Fox and Naftali Tishby. Minimum-information lqg control part i: Memoryless controllers. IEEE Conference on Decision and Control (CDC), 2016.
  • [186] Roy Fox and Naftali Tishby. Minimum-information lqg control part ii: Retentive controllers. IEEE Conference on Decision and Control (CDC), 2016.
  • [187] Robert Babuska. Model-based imitation learning. Springer Encyclopedia of the Sciences of Learning, 2012.
  • [188] Yoshua Bengio and Paolo Frasconi. An input output hmm architecture. Advances in neural information processing systems (NeurIPS), 1995.