This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Rationally Inattentive Utility Maximization for Interpretable Deep Image Classification

Kunal Pattanayak, , Vikram Krishnamurthy V. Krishnamurthy and K. Pattanayak are with the School of Electrical and Computer Engineering, Cornell University, Ithaca, NY, 14853 USA. e-mail: vikramk@cornell.edu, kp87@cornell.edu.
Abstract

Are deep convolutional neural networks (CNNs) for image classification explainable by utility maximization with information acquisition costs? We demonstrate that deep CNNs behave equivalently (in terms of necessary and sufficient conditions) to rationally inattentive utility maximizers, a generative model used extensively in economics for human decision making. Our claim is based by extensive experiments on 200 deep CNNs from 5 popular architectures. The parameters of our interpretable model are computed efficiently via convex feasibility algorithms. As an application, we show that our economics-based interpretable model can predict the classification performance of deep CNNs trained with arbitrary parameters with accuracy exceeding 94%94\%. This eliminates the need to re-train the deep CNNs for image classification. The theoretical foundation of our approach lies in Bayesian revealed preference studied in micro-economics. All our results are on GitHub and completely reproducible.

Index Terms:
Interpretable Machine Learning, Bayesian Revealed preference, Rational Inattention, Deep Neural Networks, Image Classification

1 Introduction

This paper considers interpretable deep image classification.111This paper builds substantially on our existing arXiv preprint https://arxiv.org/abs/2102.04594 uploaded in January, 2021. We show that image classification using deep Convolutional Neural Networks (CNNs) can be interpreted as a generative human decision-making model developed in microeconomics.

In micro- and behavioral economics222Micro-economics models the interaction of individual agents pursuing their private interests. Behavioral economics models human decision making in terms of subjective probabilities via prospect theory and framing. In the rest of this paper, we will use the term ‘agent’ to denote a human decision-maker., a fundamental question relating to human decision making is: How to model attention spans in humans (agents)? The area of rational inattention [1, 2], pioneered by Nobel laureate Christopher Sims models human attention in information-theoretic terms. The key hypothesis is that agents are “boundedly rational”- their perception of the environment is modeled as a Shannon capacity limited channel. In simple terms, rational inattention assigns a mutual information cost for human attention spans.

Building on the rational inattention model, the next key concept is that of a Bayesian agent with rational inattention that maximizes its expected utility. Such models are studied extensively in [3, 4, 5]. The intuition is this: more attentive decisions yield a higher expected utility at the expense of a larger attention cost. Hence, the Bayesian agent optimally trades off between minimizing its sensing cost and maximizing its expected utility. An important question is: How to test for rationally inattentive utility maximization given the decisions of a Bayesian agent? In the last decade, necessary and sufficient conditions have been developed in the area of Bayesian revealed preference [6, 7] to test if the decisions of a Bayesian agent are consistent with rationally inattentive utility maximization.

Summary of Results.

The question we address is: Can the decisions of deep CNNs in image classification be explained by a rationally inattentive Bayesian utility maximizer?

This paper uses a data-driven Bayesian revealed preference approach to interpretable deep CNN image classification. Bayesian revealed preference performs a post-hoc analysis of agent decisions. It constructs a generative333A generative model is image-independent, and hence provides a global explanation for deep image classification. In contrast, local approximation models for deep image classification are image-specific; they approximate model decisions via tractable functionals in a δ\delta-neighborhood of every input. explanatory model for the agent decisions, parameterized by a utility function and an information acquisition cost. Our approach draws important parallels between human decision making and deep neural networks; namely that deep neural networks satisfy economics based rationality.

By very nature, Bayesian revealed preference reconstructs a set of feasible utility functions and information acquisition costs. Every element in the feasible set explains the deep CNN decisions equally well. The computed utility function induces a preference ordering on the set of image classes. That is, how much a deep CNN prioritizes accurate classification over an inaccurate classification. The information acquisition cost abstracts the penalty incurred by the deep CNN to ‘learn’ an accurate latent feature representation. The key results in this paper are:

Deep ConvolutionalNeural Network CNN InputImage PredictedImage Class Rationally InattentiveUtility MaximizerEquivalentvia Theorem 1
Figure 1: Schematic illustration of rationally inattentive Bayesian utility maximization based interpretable image classification by deep CNNs. Theorem 1 establishes equivalence between the image classification behavior of a deep CNN and the decisions of a rationally inattentive maximizer. Hence, the deep CNN’s image classification behavior can be parsimoniously represented by a utility function and an information acquisition cost.
  1. 1.

    We show that the image classification decisions of deep CNNs satisfy the necessary and sufficient conditions for rationally inattentive utility maximization by a large margin. The margin by which the decisions satisfy these conditions are displayed in Table I. Hence, we establish that rationally inattentive utility maximization is a robust fit to deep image classification. This result is schematically shown in Fig. 1.

  2. 2.

    To aid visualization of our interpretable model, we provide a sparsity-enhanced decision test that computes the sparsest utility function and information acquisition cost which rationalizes deep CNN decisions. The sparsest solution yields a parsimonious representation of hundreds of thousands of layer weights of the deep CNNs in terms of a few hundred parameters. The utility function of the sparsest interpretable model also induces a useful preference ordering amongst the set of hypotheses (image labels) considered by the CNN; for example, how much additional priority is allocated to the classification of a cat as a cat compared to a cat as a dog. In classical deep learning, this preference ordering is not explicitly generated. The sparsity results for various deep CNN architectures are displayed in Table II and Fig. 2.

  3. 3.

    Our final result demonstrates the usefulness of our interpretable model. We show that the interpretable model computed from CNN decisions can predict the classification accuracy of a CNN trained with arbitrary parameters with accuracy exceeding 94%. This by-passes the need to re-train the deep CNN when its accuracy is observed for a finite number of training parameters. The prediction results are displayed in Table III.

The above results are backed by experiments performed on several deep CNN architectures using the benchmark image dataset, namely, CIFAR-10 [8]. The first two results use deep CNN decisions aggregated over varying training epochs. The third (prediction) result uses deep CNN decisions trained on noisy image datasets parameterized by the noise variance.

Related Works

Since we study interpretable deep learning using behavioral and micro- economics, we briefly discuss related works in these areas.

Bayesian revealed preference and Rational inattention. Estimating utility functions given a finite sequence of decisions and budget constraints is the central theme of revealed preference in micro-economics. The seminal work of [9, 10] (see also [11]) give necessary and sufficient conditions for the existence of a utility function that rationalizes a finite time series of consumption bundles of a decision-maker. Rationally inattentive models for Bayesian decision making have been studied extensively in [3, 4, 5]. In the last decade, the area of Bayesian revealed preference [6, 7] develops necessary and sufficient conditions to test for rationally inattentive Bayesian utility maximization.

Interpretable ML. Providing transparent models for de-obfuscating ‘black-box’ ML algorithms under the area of interpretable machine learning is a subject of extensive research [12, 13, 14]. Interpretable machine learning is defined in [15] as “the use of machine-learning models for the extraction of relevant knowledge about domain relationships contained in data”.

Since the literature is enormous, we only discuss a subset of works pertaining to interpretability of deep neural networks for image classification [16, 17, 18]. One prominent approach, namely, saliency maps, reconstructs the most preferred or typical image pertaining to each image class the deep neural network has learned [19, 20]. Related work includes creating hierarchical models for determining the importance of image features that determine its label [21]. In this paper, this feature importance is encoded into the utility function that parameterizes our interpretable model. Another approach seeks to provide local approximations to the trained model, local w.r.t the input image [22, 23]. In contrast, our generative interpretable model provides a global black-box approximation for deep image classification. A third approach approximates the decisions of the deep neural networks by a linear function of simplified individual image features [24, 25, 26, 23]. In contrast, our interpretable model fits a stochastic non-linear map that relates the true and predicted image labels. The parameters of the map are obtained by solving a convex feasibility problem parameterized by the deep CNN decisions. Finally, deep neural networks have also been modeled by Bayesian inference frameworks using probabilistic graphical methods [27].

To the best of our knowledge, an economics based approach for the post-hoc analysis of deep neural networks has not been explored in literature. However, we note that behavioral economics based interpretable models have been applied to domains outside interpretable machine learning, for example, in online finance platforms for efficient advertising [28, 29], training neural networks [30] and more recently in YouTube to rationalize user commenting behavior [31]. Finally, due to our recent equivalence result [32], our behavioral economics approach to interpretable deep image classification can be related to classical revealed preference methods [9, 10] in microeconomics.

2 Bayesian Revealed preference with Rational Inattention

This section describes the key ideas behind Bayesian revealed preference. Despite the abstract formulation below, the reader should keep in mind the deep learning context. In Sec. 3, we will use Bayesian revealed preference theory to construct an interpretable deep learning representation by showing that deep CNNs are equivalent to rationally inattentive Bayesian utility maximizers.

2.1 Utility Maximization with Rational Inattention (UMRI)

Bayesian revealed preference aims to determine if the decisions of a Bayesian agent are consistent with expected utility maximization subject to a rational inattention sensing cost. We start by describing the utility maximization model with rational inattention (henceforth called UMRI) for a collection of Bayesian decision makers/agents.

Abstractly, the UMRI model is parameterized by the tuple

Θ=(𝒦,𝒳,𝒴,𝒜,π0,C,{αk,uk,k𝒦}).\Theta=(\mathcal{K},\mathcal{X},\mathcal{Y},\mathcal{A},\pi_{0},C,\{\alpha_{k},u_{k},k\in\mathcal{K}\}). (1)

With respect to the abstract parametrization of the UMRI model for a collection of Bayesian agents, the following elements constitute the tuple Θ\Theta defined in (1).
Agents: 𝒦={1,2,,K}(K2)\mathcal{K}=\{1,2,\ldots,K\}~(K\geq 2) indexes the finite set of Bayesian agents.
State: 𝒳\mathcal{X} is the finite set of ground truths with prior probability distribution π0\pi_{0}. With respect to our image classification context, 𝒳={1,2,10}\mathcal{X}=\{1,2,\ldots 10\} is the set of image classes in the CIFAR-10 dataset and π0\pi_{0} is the empirical probability distribution of the image classes in the test dataset of CIFAR-10.
Observation and attention strategy: Agent k𝒦k\in\mathcal{K} chooses attention strategy αk:𝒳Δ(𝒴)\alpha_{k}:\mathcal{X}\rightarrow\Delta(\mathcal{Y}), a stochastic mapping from 𝒳\mathcal{X} to a finite set of observations 𝒴\mathcal{Y}. Given state xx and attention strategy αk\alpha_{k}, the agent samples observation yy with probability αk(y|x)\alpha_{k}(y|x). The agent then computes the posterior probability distribution p(x|y)p(x|y) via Bayes formula as

p(x|y)=π0(x)αk(y|x)x𝒳π0(x)αk(y|x).p(x|y)=\frac{\pi_{0}(x)\alpha_{k}(y|x)}{\sum_{x^{\prime}\in\mathcal{X}}\pi_{0}(x^{\prime})\alpha_{k}(y|x^{\prime})}. (2)

The observation and attention strategy are latent variables that abstractly represent the learned feature representations in the deep image classification context. Bayesian revealed preference theory tests their existence via the convex feasibility test in Theorem 1 below.
Action: Agent k𝒦k\in\mathcal{K} chooses action aa from a finite set of actions 𝒜\mathcal{A} after computing the posterior probability distribution p(x|y)p(x|y). In the image classification context, aa is the image class predicted by the neural network, hence 𝒜=𝒳\mathcal{A}=\mathcal{X}.
Utility function: Agent k𝒦k\in\mathcal{K} has a utility function uk(x,a)+u_{k}(x,a)\in\mathbb{R}^{+}, x𝒳,a𝒜x\in\mathcal{X},a\in\mathcal{A} and aims to maximize its expected value, with the expectation taken wrt the random state xx and random observation yy. A key feature in our approach is to show that the utility function rationalizes the decisions of the deep CNNs (made precise in Definition 1).
Information Acquisition Cost: The information acquisition cost C(α,π0)+C(\alpha,\pi_{0})\in\mathbb{R}^{+} depends on attention strategy α\alpha and prior pmf π0\pi_{0}. It is the sensing cost the agent incurs in order to estimate the underlying state (2). In the context of machine learning, C()C(\cdot) abstractly captures the ‘learning’ cost incurred during training of the deep neural networks. In rational inattention theory from behavioral economics, a higher information acquisition cost is incurred for more accurate attention strategies (equivalently, more accurate state estimates (2) given observation yy). We refer the reader to the influential work of [1, 2].

Each Bayesian agent k𝒦k\in\mathcal{K}, aims to maximize its expected utility while minimizing its cost of information acquisition. Hence, the action aa given observation yy, and attention strategy αk\alpha_{k} are chosen as follows:

Definition 1 (Rationally Inattentive Utility Maximization).

Consider a collection of Bayesian agents 𝒦\mathcal{K} parameterized by Θ\Theta in (1) under the UMRI model. Then,
(a)
Expected Utility Maximization: Given posterior probability distribution p(x|y)p(x|y), every agent k𝒦k\in\mathcal{K} chooses action aa that maximizes its expected utility. That is, with 𝔼\mathbb{E} denoting mathematical expectation, the action aa satisfies

aargmaxa𝒜𝔼x{uk(x,a)|y}=x𝒳p(x|y)uk(x,a)a\in\operatorname*{argmax}_{a^{\prime}\in\mathcal{A}}~\mathbb{E}_{x}\{u_{k}(x,a^{\prime})|y\}=\sum\limits_{x\in\mathcal{X}}p(x|y)u_{k}(x,a^{\prime}) (3)

(b) Attention Strategy Rationality: For agent kk, the attention strategy αk\alpha_{k} optimally trades off between maximizing the expected utility and minimizing the information acquisition cost.

αkargmaxα𝔼y{\displaystyle\alpha_{k}\in\operatorname*{argmax}_{\alpha^{\prime}}\mathbb{E}_{y}\{ maxa𝒜𝔼x{uk(x,a)|y}}C(α,π0)\displaystyle\operatorname*{max}_{a\in\mathcal{A}}\mathbb{E}_{x}\{u_{k}(x,a)|y\}\}-C(\alpha^{\prime},\pi_{0}) (4)

Eq. 3,4 in Definition 1 constitute a nested optimization problem. The lower-level optimization task is to choose the the ‘best’ action for any observation yy based on the computed posterior belief of the state. The upper-level optimization task is to sample the observations optimally by choosing the ‘best’ attention strategy.

Remark. The multiple Bayesian agents in Θ\Theta have the same state space 𝒳\mathcal{X}, observation space 𝒴\mathcal{Y}, action space 𝒜\mathcal{A}, prior π0\pi_{0} and cost of information acquisition CC, but only differ in their utility functions. Bayesian revealed preference theory relies on this crucial constraint on the optimization variables in (3), (4) for detecting optimal behavior in a finite number of agents.

2.2 Bayesian Revealed Preference (BRP) Test for Rationally Inattentive Utility Maximization

Having described the UMRI  model (collection of rationally inattentive utility maximizers), we are now ready to state our key result. Theorem 1 below says that the decisions of a collection of Bayesian agents is rationalized by a UMRI tuple Θ\Theta if and only if a set of convex inequalities have a feasible solution. These inequalities comprise our Bayesian Revealed Preference (henceforth called BRP) test for rationally inattentive utility maximization.

For notational convenience, the decisions of the Bayesian agents in the UMRI model are compacted into the dataset 𝔻\mathbb{D} defined as:

𝔻={π0,pk(a|x),x𝒳,a𝒜,k𝒦}.\mathbb{D}=\{\pi_{0},p_{k}(a|x),x\in\mathcal{X},a\in\mathcal{A},k\in\mathcal{K}\}. (5)

In (5), π0Δ|𝒳|1\pi_{0}\in\Delta^{|\mathcal{X}|-1} denotes the prior pmf over the set of states 𝒳\mathcal{X} in Θ\Theta (1). The variable pk(a|x)p_{k}(a|x) is the conditional probability that agent k𝒦={1,2,,K}k\in\mathcal{K}=\{1,2,\ldots,K\} takes action aa given state xx. 𝔻\mathbb{D} characterizes the input-output behavior of the collection of Bayesian agents and serves as the input for BRP feasibility test described below.

Theorem 1 (BRP Test for Rationally Inattentive Utility Maximization [7]).

Given the dataset 𝔻\mathbb{D} (5) obtained from a collection of Bayesian agents 𝒦\mathcal{K}. Then,
1. Existence: There exists a UMRI tuple Θ(𝔻)\Theta(\mathbb{D}) (1) that rationalizes dataset 𝔻\mathbb{D} if and only if there exists a feasible solution that satisfies the set of convex inequalities

BRP(𝔻,{uk,ck}k=1K)𝟎,uk+|𝒳|×|𝒜|,ck>0.\text{BRP}(\mathbb{D},\{u_{k},c_{k}\}_{k=1}^{K})\leq\mathbf{0},~u_{k}\in\mathbb{R}_{+}^{|\mathcal{X}|\times|\mathcal{A}|},~c_{k}>0. (6)

In (6), BRP()(\cdot) corresponds to a set of convex (in the variables {uk,ck}k=1K\{u_{k},c_{k}\}_{k=1}^{K}) inequalities, stated in Algorithm 1.
2. Reconstruction: Given a feasible solution {uk,ck}k=1K\{u_{k},c_{k}\}_{k=1}^{K} to BRP(𝔻,)BRP(\mathbb{D},\cdot), uku_{k} is the kthk^{\text{th}} Bayesian agent’s utility function in the feasible model tuple Θ(𝔻)\Theta(\mathbb{D}). The set of observations 𝒴=𝒜\mathcal{Y}=\mathcal{A}, the set of actions in 𝔻\mathbb{D}. The feasible cost of information acquisition CC in Θ(𝔻)\Theta(\mathbb{D}) is defined in terms of ckc_{k} as:

C(α)\displaystyle C(\alpha) =maxk𝒦ck+amaxb𝒜xp(x,a)uk(x,b)\displaystyle=\max_{k\in\mathcal{K}}c_{k}+\sum_{a}\max_{b\in\mathcal{A}}\sum_{x}p(x,a)u_{k}(x,b)
x,apk(x,a)uk(x,a),α={p(a|x)}.\displaystyle\quad\quad\quad~-\sum_{x,a}p_{k}(x,a)u_{k}(x,a),~\alpha=\{p(a|x)\}. (7)

The proof of Theorem 1 is in Appendix A.1. Before launching into a detailed discussion, we stress the “iff” in Theorem 1. Put simply: if the inequalities in (6) are not feasible, then the Bayesian agents that generate the dataset 𝔻\mathbb{D} are not rationally inattentive utility maximizers. If (6) has a feasible solution, then there exists a reconstructable family of viable utility functions and information acquisition costs that rationalize 𝔻\mathbb{D}444In terms of interpretable deep learning, of all parameters in the UMRI tuple, we are only interested in the utility functions of the agents and the cost of information acquisition, since the remaining parameters are immediately deduced from the decision dataset 𝔻\mathbb{D}.. A key feature of Theorem 1 is that the estimated utilities (and information costs) are set-valued; every utility and cost function in the feasible set explains 𝔻\mathbb{D} equally well. The estimated UMRI model parameters are set-valued due to the finite number of Bayesian agents whose decisions constitute the dataset 𝔻\mathbb{D}. The estimated parameter set converges to a point if and only if the inequality (6) holds as |𝒦||\mathcal{K}|\rightarrow\infty.

Computational Aspects of BRP Test. Suppose the dataset 𝔻\mathbb{D} is obtained from KK Bayesian agents. Then, BRP(𝔻)(\mathbb{D}) comprises a feasibility test with K(|𝒳||𝒜|+1)K~(|\mathcal{X}||\mathcal{A}|+1) free variables and K2+K(|𝒜|2|𝒜|1)K^{2}+K~(|\mathcal{A}|^{2}-|\mathcal{A}|-1) convex inequalities. Thus, the number of free variables and inequalities in the BRP feasibility test scale linearly and quadratically, respectively, with the number of observed Bayesian agents.

0: Dataset 𝔻={π0,pk(a|x),x,a𝒳,k𝒦}\mathbb{D}=\{\pi_{0},p_{k}(a|x),x,a\in\mathcal{X},k\in\mathcal{K}\} from a collection of Bayesian agents 𝒦\mathcal{K}. Find: Positive reals ck,uk(x,a)(0,1]c_{k},~u_{k}(x,a)\in(0,1] for all x𝒳,x\in\mathcal{X}, a𝒜,k𝒦a\in\mathcal{A},~k\in\mathcal{K} that satisfy the following inequalities:
NIAS¯:\displaystyle\hskip-42.67912pt\underline{\textbf{NIAS}}: xpk(x|a)(uk(x,b)uk(x,a))0,\displaystyle~\sum_{x}p_{k}(x|a)~(u_{k}(x,b)-u_{k}(x,a))\leq 0, (8)
a,b𝒜,k𝒦,\displaystyle~\forall a,b\in\mathcal{A},~k\in\mathcal{K},
NIAC¯:\displaystyle\hskip-42.67912pt\underline{\textbf{NIAC}}: a(maxbxpj(x,a)uk(x,b))cj\displaystyle~\sum_{a}\left(\max_{b}\sum_{x}p_{j}(x,a)u_{k}(x,b)\right)-c_{j} (9)
\displaystyle- x,apk(x,a)uk(x,a)+ck0,j,k𝒦,\displaystyle~\sum_{x,a}p_{k}(x,a)u_{k}(x,a)+c_{k}\leq 0,~\forall j,k\in\mathcal{K},
where pk(x,a)=π0(x)pk(a|x),pk(x|a)=pk(x,a)xpk(x,a)p_{k}(x,a)=\pi_{0}(x)p_{k}(a|x),~p_{k}(x|a)=\frac{p_{k}(x,a)}{\sum_{x^{\prime}}p_{k}(x^{\prime},a)}. Return: Set of feasible utility functions uku_{k} and information acquisition costs ckc_{k} incurred by agents k𝒦k\in\mathcal{K}.
Algorithm 1 BRP Convex Feasibility Test of Theorem 1

2.3 Relating UMRI and BRP test to Interpretable Deep Image Classification

We now discuss how the above BRP test relates to interpretable image classification using deep CNNs. The BRP convex feasibility test in Theorem 1 comprises two sets of inequalities, namely, the NIAS (No-Improving-Action-Switches) (8) and NIAC (No-Improving-Action-Cycles) (9) inequalities (Algorithm 1). NIAS ensures that the agent takes the best action given a posterior pmf. NIAC ensures that every agent chooses the best attention strategy. BRP test checks if there exist KK utility functions and KK positive reals that, together with 𝔻\mathbb{D}, satisfy the NIAS and NIAC inequalities.

Toy Example with 2 CNNs

The following discussion gives additional insight into our approach. Consider the simplest case involving two trained deep CNNs N1N_{1} and N2N_{2}; so 𝒦={1,2}\mathcal{K}=\{1,2\} in the above notation. Assume N1N_{1} and N2N_{2} have the same network architecture. Suppose an analyst observes that N1N_{1} makes accurate decisions on a rich input image dataset while N2N_{2} makes less accurate decisions on the same dataset.

Our UMRI model first abstracts the accuracy of the feature representations of the input image data learned by N1N_{1} and N2N_{2} via attention strategies α1\alpha_{1} and α2\alpha_{2} in (4). Second, the information acquisition cost function C()C(\cdot) abstracts the computational resources expended for learning the representations. The rationale is that learning an accurate latent feature representation is costly, and this is abstracted by the information acquisition cost.

Let the training cost incurred by N1N_{1} and N2N_{2} be C(α1)C(\alpha_{1}) and C(α2)C(\alpha_{2}) respectively. If the decisions of N1N_{1} and N2N_{2} can be explained by the UMRI model (and Theorem 1 below will give necessary and sufficient conditions for this), then there exist utility functions u1u_{1} and u2u_{2} for N1N_{1} and N2N_{2}, that satisfy:

𝔼αi{ui}C(αi)𝔼αj{ui}C(αj),i,j{1,2}\mathbb{E}_{\alpha_{i}}\{u_{i}\}-C(\alpha_{i})\geq\mathbb{E}_{\alpha_{j}}\{u_{i}\}-C(\alpha_{j}),~i,j\in\{1,2\}~ (10)

The above inequality says that CNNs N1N_{1} and N2N_{2} would be worse off (in an expected utility sense) if they make decisions based on swapping each other’s learned representations. That is, both N1N_{1} and N2N_{2} learn the ‘best’ feature representation of the input images given their training parameters.

Discussion

(i) Parsimonious Interpretable Representation of deep CNNs. In the deep image classification context, due to the UMRI model’s parsimonious parametrization in (1), the decisions of KK CNNs can be rationalized by just KK utility functions and an information acquisition cost function, thus bypassing the need of several million parameters to describe the deep CNNs.
(ii) Identifiability. The BRP feasibility test requires the dataset 𝔻\mathbb{D} to be generated from K>2K>2 Bayesian agents. If K=1K=1, then (6) holds trivially since any information acquisition cost satisfies the convex inequalities of BRP. Another intuitive way of motivating a collection of agents for the BRP is as follows. Reconstructing a feasible UMRI model tuple Θ\Theta that rationalizes the decisions of the deep CNNs is analogous to fitting a line to a finite number of points. One can fit infinitely many lines through a single point. The task becomes non-trivial if the number of points exceeds 22. In the Bayesian revealed preference context, the points correspond to the decisions from each Bayesian agent. The slope and intercept of the fitted line, in our case, corresponds to the utility function and cost of information acquisition that rationalize the agent decisions.
(iii) Relative Optimality implies Global Optimality. In the setting involving K>2K>2 deep CNNs (agents), the NIAS and NIAC inequalities of BRP test check for relative optimality - given utility function uku_{k}, does deep CNN kk performs at least as well as any other observed deep CNN in 𝒦\{k}\mathcal{K}\backslash\{k\}? Clearly, testing for relative optimality is weaker than testing for global optimality (4) which ideally requires access to decisions from an infinite number of deep CNNs. Setting the cost of information acquisition as a free variable bridges this gap. The proof of Theorem 1 shows that if the deep CNN decisions satisfy relative optimality, then there exists a cost of information acquisition such that the decisions are globally optimal. That is, Theorem 1 ensures relative optimality is sufficient for global optimality.
(iii) Generalization of [7]. Theorem 1 generalizes [7, Theorem 1] in two ways. (1) In [7], the utilities uku_{k} in UMRI model tuple Θ\Theta are assumed known, and only the information acquisition costs ckc_{k} are estimated, whereas Theorem 1 estimates both parameters. (2) The expression for the reconstructed model tuple Θ(𝔻)\Theta(\mathbb{D}) is novel; the discussion in [7] is only confined to the existence of such a tuple.
(vi) Single Utility UMRI (S-UMRI ). In Appendix A.2, we propose a sparse version of UMRI , namely, the S-UMRI model in (21). The key distinction of this model is that all agents have the same utility function uu and thus can be represented with substantially fewer parameters. In complete analogy to Theorem 1, we outline a decision test in Theorem 3 that states necessary and sufficient conditions for agent decisions to be consistent with the S-UMRI model of rationally inattentive utility maximization. We discuss this sparse parametrization in the appendix so as not to interrupt the flow of the main text.
(vii) Degenerate solution to BRP and S-BRP tests. The degenerate utility function of all zeros and cost of information acquisition C=0C=0 trivially satisfy the BRP and S-BRP tests and lie at the boundary of the feasible set of parameters.

Summary

This section formulated an economics-based decision-making model. Since this model may not be familiar to a machine learning reader, we summarize the main ideas. We introduced the rationally inattentive utility maximization model, namely, the UMRI model for a collection of Bayesian agents (decision makers). Our main result Theorem 1 outlines a decision test BRP for rationally inattentive utility maximization given decisions from a collection of agents. This BRP test comprises a set of convex inequalities that have a feasible solution if and only if the collection of agents are rationally inattentive utility maximizers. Theorem 1 also provides an explicit reconstruction of the feasible UMRI model parameters that rationalize input agent decisions. The set of feasible utility functions and information acquisition costs thus parsimoniously explain the decisions generated by the Bayesian agents. In Appendix A.2, we propose a single utility version of the UMRI model with fewer parameters. Due to fewer parameters, the decision test for this sparse model, given in Theorem 3, is computationally less expensive yet more restrictive than the BRP test for rationality in Theorem 1.

The rest of the paper focuses on computing interpretable UMRI models that rationalize deep CNN decisions. We will investigate through extensive experiments how well the UMRI fits the deep CNN decisions via robustness tests. We will also investigate how well the computed interpretable models, namely, UMRI and S-UMRI , predict the deep CNNs’ decisions when the training parameters are varied.

3 Bayesian Revealed Preference explains CIFAR-10 Image Classification by Deep CNNs

The experimental results in this section are divided into two parts: First, we show that the deep CNNs decisions pass the BRP and S-BRP tests formulated in Theorems 1 and 3 by a large margin. This implies that the rationally inattentive utility maximization model is a robust fit to the deep CNN decisions.

Our second result demonstrates an application of the reconstructed interpretable model. Training datasets are often noisy. We show that in such a noisy setting, the reconstructed interpretable model from Theorem 1 can accurately predict (with accuracy exceeding 94%) the image classification performance of the deep CNNs. This bypasses the need to train the deep CNN for various noise variances that corrupt the training dataset.

Experimental Setup: Deep CNN Architectures, Training Parameters and Construction of Dataset

Image Dataset. In our experiments, we trained and validated the deep CNNs using the CIFAR-10 benchmark image dataset [8]. This public dataset consists of 6000060000 3232x3232 colour images in 1010 distinct classes (for example, airplane, automobile, ship, cat, dog etc.), with 60006000 images per class. There are 5000050000 training images and 1000010000 test images. We will use the terms image classes and image labels interchangeably.555Our experiments are confined to the CIFAR-10 dataset for clarity of exposition. Our approach to interpretable deep learning can be easily extended to richer benchmark image datasets like ImageNet and CIFAR-100 (that comprise over a 100 image labels).

Network Architecture and Training Parameters. In this paper, we use 55 well-known deep CNN architectures for our experiments. 1. LeNet [33], 2. AlexNet [34] 3. VGG16 [35] 4. ResNet-50 [18] 5. Network-in-Network (NiN) [36] The deep CNNs are trained and validated on the CIFAR-10 image dataset, using 33 learning rate schedules, namely, L.R. 1, L.R. 2 and L.R. 3. All 33 schedules use the RMSprop optimizer [37] with the decay parameter and maximum training epochs (full passes of the training dataset) set to 10610^{-6} and 200200, respectively, and initial step size set to 0.010.01. The step size is halved every 20,30,4020,30,40 epochs, respectively, for L.R. 1, 2 and 3.

Relation to Bayesian revealed preference. We now relate the deep CNN setup to the Bayesian revealed preference framework in Sec. 2. For each CNN architecture, we use the decisions of K=20K=20 CNNs, i.e. , 2020 Bayesian agents in the terminology of Sec. 2, for our BRP and S-BRP decision tests. The CNN decisions from KK CNNs on the test image dataset of CIFAR-10 are aggregated into dataset 𝔻\mathbb{D} (5). The results of the decision tests are discussed below. In the deep image classification context, the parameter pk(a|x)p_{k}(a|x) in (5) is the probability that the kthk^{\text{th}} deep CNN classifies an image from category xx into category aa in the CIFAR-10 test image dataset. The prior π0\pi_{0} in 𝔻\mathbb{D} (5) is the empirical pmf over the set of image categories in the CIFAR-10 test dataset. Constructing 𝔻\mathbb{D} from raw CNN decisions is discussed in Appendix A.3.

3.1 BRP and S-BRP Tests for deep CNN datasets. Results and Insights

Network Architecture Learning Rate BRP\mathcal{R}_{\text{BRP}} (×104)(\times 10^{-4}) S-BRP\mathcal{R}_{\text{S-BRP}} (×104)(\times 10^{-4})
LeNet L. R.  1 30.3430.34 4.724.72
L. R.  2 35.1435.14 4.654.65
L. R.  3 37.9737.97 5.115.11
AlexNet L. R.  1 32.1032.10 3.213.21
L. R.  2 34.9834.98 3.913.91
L. R.  3 40.6040.60 4.624.62
VGG16 L. R.  1 96.3696.36 4.094.09
L. R.  2 107.4107.4 4.024.02
L. R.  3 119.8119.8 4.444.44
ResNet-50 L. R.  1 126.2126.2 2.822.82
L. R.  2 129.2129.2 3.453.45
L. R.  3 132.3132.3 3.833.83
Network-In-Network (NiN) L. R.  1 108.3108.3 3.593.59
L. R.  2 132.1132.1 3.363.36
L. R.  3 149.1149.1 5.575.57
TABLE I: How does increasing the number of degrees of freedom of the interpretable model improve robustness of fit to the CNN decisions? We see that BRP\mathcal{R}_{\text{BRP}} (11) is substantially higher (by an order of magnitude) than S-BRP\mathcal{R}_{\text{S-BRP}} (12) for all CNN architectures. We conclude that the UMRI model fits CNN decisions substantially better than the S-UMRI model, but with larger computing cost for evaluating the parameters of the interpretable model. Thus, if there are no computational constraints, we recommend using the UMRI model for interpreting CNN decisions.

A. Robustness Results on Deep CNN datasets

Our first key result is that image classifications of all 55 deep CNN architectures listed in Sec. 3 pass the BRP and S-BRP tests by a large margin. The results are tabulated in Table I. The robustness values BRP\mathcal{R}_{\text{BRP}} and S-BRP\mathcal{R}_{\text{S-BRP}} in Table I are defined in Definition 2 below which formalizes the notion of margin for the decision tests.

Network Architecture Learning Rate (L.R.) airplane auto bird cat deer dog frog horse ship truck
LeNet L.R. 1 17.6117.61 3.553.55 20.0620.06 1.881.88 17.1917.19 21.4221.42 42.0042.00 27.7927.79 1.911.91 9.559.55
L.R. 2 4.134.13 5.205.20 7.827.82 1.901.90 13.1813.18 18.6618.66 23.8423.84 8.168.16 2.482.48 2.472.47
L.R. 3 10.7910.79 8.278.27 18.6218.62 22.6722.67 19.9119.91 25.0125.01 47.7147.71 73.5273.52 2.652.65 1.011.01
AlexNet L.R. 1 210.78210.78 41.8441.84 49.7749.77 59.7159.71 51.2451.24 68.3168.31 83.9483.94 211.61211.61 60.4360.43 125.73125.73
L.R. 2 85.5185.51 47.8947.89 17.3817.38 1.001.00 25.3425.34 202.78202.78 21.3021.30 35.0135.01 533.62533.62 248.57248.57
L.R. 3 18.0018.00 49.5549.55 58.2558.25 28.3128.31 135.54135.54 29.2429.24 224.91224.91 214.51214.51 8.298.29 264.20264.20
VGG16 L.R. 1 164.48164.48 154.77154.77 15.4215.42 33.6733.67 6.286.28 123.89123.89 62.8362.83 26.2126.21 1.431.43 170.69170.69
L.R. 2 88.7388.73 154.10154.10 45.6345.63 297.61297.61 131.08131.08 136.52136.52 57.3457.34 229.80229.80 145.99145.99 11.9011.90
L.R. 3 24.3324.33 10.7810.78 93.9093.90 11.1111.11 91.9691.96 56.6456.64 77.3077.30 110.60110.60 20.2820.28 17.0917.09
ResNet-50 L.R. 1 50.8350.83 17.5517.55 16.0916.09 4.664.66 17.9217.92 3.673.67 4.924.92 3.953.95 15.4615.46 4.884.88
L.R. 2 7.517.51 8.408.40 72.7072.70 30.7230.72 32.4332.43 83.6583.65 221.27221.27 74.5974.59 99.0499.04 20.5120.51
L.R. 3 14.6114.61 367.59367.59 31.6131.61 9.209.20 16.3516.35 11.5811.58 41.4441.44 243.95243.95 222.67222.67 483.91483.91
Network-in-Network L.R. 1 5.025.02 30.9530.95 9.919.91 71.3871.38 63.6963.69 45.8845.88 31.3931.39 67.8667.86 17.0317.03 21.4121.41
L.R. 2 40.1740.17 60.3260.32 4.404.40 55.6755.67 95.0295.02 88.7288.72 91.1591.15 15.9815.98 176.75176.75 10.2710.27
L.R. 3 10.4710.47 75.3275.32 55.9755.97 24.1724.17 17.4117.41 8.948.94 23.0223.02 71.2771.27 29.9429.94 80.9180.91
TABLE II: The utility function of the sparsest interpretable model is a diagonal matrix. The diagonal elements yield a natural preference ordering amongst the set of image classes (classification hypotheses). For example, consider the VGG16 architecture trained using learning rate 11 (third row, first sub-row of table). The maximum utility is for trucks (170.69170.69, last column) and the minimum is for ships (1.431.43, second last column). This shows the sparsest interpretable model induces the following preference ordering for the VGG16 architecture: classifying trucks correctly is prioritized 100100 times more than classifying ships. Such a preference ordering is not explicitly generated by a CNN.
Definition 2 (Robustness (Goodness-of-fit) of BRP and S-BRP Tests.).

Given dataset 𝔻\mathbb{D} (5) aggregated from a collection of Bayesian agents, BRP(𝔻)\mathcal{R}_{\text{BRP}}(\mathbb{D}) and S-BRP(𝔻)\mathcal{R}_{\text{S-BRP}}(\mathbb{D}) measure the largest perturbation so that 𝔻\mathbb{D} passes the BRP and S-BRP decision tests:

BRP(𝔻)=maxε>0εKk=1Kuk22,BRP(𝔻,{uk,ck}k=1K)ε.\displaystyle\mathcal{R}_{\text{BRP}}(\mathbb{D})=\max_{\varepsilon>0}\frac{\varepsilon~K}{\sum_{k=1}^{K}\|u_{k}\|_{2}^{2}},~\text{BRP}(\mathbb{D},\{u_{k},c_{k}\}_{k=1}^{K})\leq-\varepsilon. (11)
S-BRP(𝔻)=maxε>0εu22,S-BRP(𝔻,u,{ck,λk}k=1K)ε.\displaystyle\mathcal{R}_{\text{S-BRP}}(\mathbb{D})=\max_{\varepsilon>0}\frac{\varepsilon}{\|u\|_{2}^{2}},\text{S-BRP}(\mathbb{D},u,\{c_{k},\lambda_{k}\}_{k=1}^{K})\leq-\varepsilon. (12)

In Definition 2, robustness values BRP\mathcal{R}_{\text{BRP}} and S-BRP\mathcal{R}_{\text{S-BRP}} measure, respectively, the smallest perturbation needed for 𝔻\mathbb{D} to fail the BRP and S-BRP decisions tests. Both BRP\mathcal{R}_{\text{BRP}} and S-BRP\mathcal{R}_{\text{S-BRP}} are normalized wrt the row-wise 2\mathcal{L}_{2} norm of the feasible utility functions. Higher robustness values imply a better fit of the UMRI , S-UMRI models to the decision dataset 666The robustness value for the non-informative dataset of uniformly distributed pmfs is 0. Hence, the robustness value measures the informativeness of the attention strategies in 𝔻\mathbb{D} relative to the uniform probability distribution..

Discussion and Insights. Robustness Results of Table I

(i) Deep CNN dataset: The deep CNN datasets used for the robustness tests (11), (12) comprise decisions of K=20K=20 deep CNNs for every network architecture, where CNN kk was trained for 10k10~k training epochs, k=1,2,,Kk=1,2,\ldots,K.
(ii) Comparison between BRP\mathcal{R}_{\text{BRP}} and S-BRP\mathcal{R}_{\text{S-BRP}} values for deep CNN datasets: The average value of S-BRP\mathcal{R}_{\text{S-BRP}} (12) over all 33 learning rate schedules and 55 network architectures was found to be 4.09×1044.09\times 10^{-4}. In contrast, the average value of BRP\mathcal{R}_{\text{BRP}} (11) was found to be 87.45×10487.45\times 10^{-4}, almost 2020 times the average value of S-BRP\mathcal{R}_{\text{S-BRP}}. This result shows that the UMRI model fits deep CNN decisions substantially better than the S-UMRI model. This result is expected since S-UMRI is parameterized using much fewer variables compared to the UMRI and hence, S-BRP test is more restrictive than BRP .
(iii) Sensitivity of BRP,S-BRP\mathcal{R}_{\text{BRP}},\mathcal{R}_{\text{S-BRP}} to Network Architecture: The average value of BRP\mathcal{R}_{\text{BRP}} is 122.29×104122.29\times 10^{-4} for the LeNet and AlexNet architectures, which is approximately 3.53.5 times the the average value of BRP\mathcal{R}_{\text{BRP}} for the VGG16, ResNet-50 and NiN architectures which is 35.18×10435.18\times 10^{-4}. The variation of S-BRP\mathcal{R}_{\text{S-BRP}} with network architecture is negligible compared to S-BRP\mathcal{R}_{\text{S-BRP}}. This shows the robustness test for UMRI model is more sensitive to network architecture compared to that for the S-UMRI model.
(iv) Computational aspects of BRP\mathcal{R}_{\text{BRP}} and S-BRP\mathcal{R}_{\text{S-BRP}}. The computation time for BRP\mathcal{R}_{\text{BRP}} is almost 3030 times that for S-BRP\mathcal{R}_{\text{S-BRP}}. This is expected since the UMRI model is parameterized by KK utility functions compared to a single utility function in S-UMRI .

B. Sparsity-enhanced Interpretable Model

Our next task is to determine the sparsest possible interpretable model that satisfies the decision tests BRP and S-BRP. The motivation is three fold:

  1. 1.

    The sparsest interpretable model explains the deep CNN decisions using the fewest number of parameters.

  2. 2.

    The sparsest interpretable model induces a useful preference ordering amongst the set of hypotheses (image labels) considered by the CNN; for example, how much additional priority is allocated to the classification of a cat as a cat compared to a cat as a dog. In classical deep learning, this preference ordering is not explicitly generated.

  3. 3.

    Third, the sparsest solution is a point valued estimate. Recall the BRP and S-BRP decision tests yield a set-valued estimate of feasible utility functions and cost of information acquisition that explain the deep CNN datasets. While every element in the set explains the dataset equally well, it is useful to have a single representative point.

Theorem 2 below computes the sparsest utility function out of all feasible utility functions.

Theorem 2 (Sparsity Enhanced BRP and S-BRP Tests for Deep CNN datasets).

Given dataset 𝔻\mathbb{D} (5) from a collection of KK Bayesian agents. The sparsest solutions to the BRP and S-BRP tests minimize the sum of row-wise 1\mathcal{L}_{1} norm of the feasible utility functions of the KK agents that generate 𝔻\mathbb{D}.

(u1:K)=argminu1:Kk=1Kuk1,BRP(𝔻,)𝟎,k=1Kuk22=K.\displaystyle(u_{1:K})^{\ast}=\operatorname*{argmin}_{u_{1:K}}\sum_{k=1}^{K}\lVert u_{k}\rVert_{1},\text{BRP}(\mathbb{D},\cdot)\leq\mathbf{0},~\sum_{k=1}^{K}\|u_{k}\|_{2}^{2}=K.
u=argminuu1,S-BRP(𝔻,)𝟎,u22=1.\displaystyle u^{\ast}=\operatorname*{argmin}_{u}\lVert u\rVert_{1},\text{S-BRP}(\mathbb{D},\cdot)\leq\mathbf{0},~\|u\|_{2}^{2}=1. (13)

where 1\lVert\cdot\rVert_{1} denotes the row-wise 1\mathcal{L}_{1} norm.

Results and Discussion. Sparsity Test for deep CNN datasets

The sparsest utility function from the S-BRP test are tabulated in Table II for all 55 deep CNN architectures. The corresponding information acquisition cost for all 55 architectures averaged over learning rates 1,2,31,2,3 are shown in Fig. 2. Together, the sparsest utility and information cost constitute the sparsest S-UMRI  interpretable model777For brevity, we have only included the sparsity results for the S-UMRI model. The sparsest utility functions of the UMRI model that explains deep CNN decisions are included in our public GitHub repository that contains all test results and codes. for the deep CNN decisions.
(i) The sparsest S-UMRI model is comprised of K(|𝒳||𝒜|+1)K(|\mathcal{X}||\mathcal{A}|+1) variables. (ii) Preference ordering induced from sparsest utility. The sparsest utility function for the S-UMRI model induces a useful preference ordering among the predicted image classes. That is, they measure how the deep CNN’s priority for accurate classification varies across image classes. For instance, consider the VGG16 architecture trained using learning rate schedule 11. Of all image categories, the maximum utility is observed for trucks (170.69170.69) and the minimum for ships (1.431.43). This shows the VGG16 architecture prioritizes classifying trucks correctly about 100100 times more than classifying ships.
(ii) Penalty for learning image features accurately. The computed information acquisition costs in Fig. 2 can be understood as the training cost the CNN incurs to learn latent image features accurately. The interpretable model cannot explain the variation in CNN classification accuracy versus variation in training parameters without an information acquisition cost. From Fig. 2, we can conclude that learning accurate image features is the most and least costly, respectively, for the AlexNet and ResNet architectures, respectively.

Refer to caption
Figure 2: The figure illustrates an important property of our approach to interpretable deep learning: in addition to the utility function (Table II), we also need a rational inattention term (cost of learning latent image features) to explain CNN decisions. Put differently, we cannot explain the variation in CNN classification accuracy versus variation in training parameters without an information acquisition cost. The figure displays the information acquisition cost CC (7) evaluated for the sparsest interpretable model. We also observe that learning accurate image features is most expensive for AlexNet, and least expensive for ResNet architectures.
Network Architecture airplane auto bird cat deer dog frog horse ship truck
LeNet 0.0420.042 0.0420.042 0.0410.041 0.0270.027 0.0460.046 0.0250.025 0.0490.049 0.0340.034 0.0400.040 0.0420.042
AlexNet 0.0250.025 0.0310.031 0.0340.034 0.0210.021 0.0460.046 0.0320.032 0.0490.049 0.0390.039 0.0450.045 0.0360.036
VGG16 0.0330.033 0.0350.035 0.0430.043 0.0410.041 0.0480.048 0.0480.048 0.0350.035 0.0460.046 0.0370.037 0.0480.048
ResNet-50 0.0300.030 0.0310.031 0.0270.027 0.0310.031 0.0200.020 0.0270.027 0.0400.040 0.0150.015 0.0230.023 0.0240.024
Network-in-Network 0.0510.051 0.0290.029 0.0250.025 0.0280.028 0.0560.056 0.0590.059 0.0300.030 0.0580.058 0.0450.045 0.0360.036
TABLE III: How well does our interpretable model predict CNN classification accuracy? The table displays the prediction error δη(x)\delta_{\eta}(x) defined in (14). Recall δη(x)\delta_{\eta}(x) is the error between the true CNN performance and the predicted performance using the interpretable model with Algorithm 2. The maximum error across all image classes and architectures was found to be 5.9%5.9\%. Hence, our interpretable model predicts CNN classification performance with accuracy exceeding 94%94\%.

3.2 Predicting deep CNN classification accuracy using our Interpretable Models

Training datasets are often noisy; for example, [38] considers noisy datasets for hand-written character recognition. We now exploit the proposed interpretable model to predict how the deep CNN will perform with a noisy training dataset without actually implementing the deep CNN.

Our predictive procedure is as follows. We first train the CNNs on noisy datasets that are generated by adding simulated Gaussian noise with noise variances chosen from a finite set. 888Injecting artificial noise in training datasets is also used in variational autoencoders for robust feature learning [39, 40]. Then given the CNN decisions, we compute our interpretable model over this finite set of noise variances. Finally, to predict how the CNN will perform for a noise variance not in the set, we interpolate the utility function of the interpretable model at this noise variance. Then given the interpolated utility function and information acquisition cost from our interpretable model, the predicted classification performance is computed by solving convex optimization problem (4). The above procedure is formalized in Algorithm 2. Hence, our interpetable model serves as a computationally efficient method for predicting the performance of a CNN without implementing the CNN. The interpretable model can be viewed as a low dimension projection of the high-dimension CNN with predictive accuracy exceeding 94%.

Remark. An alternative procedure is to directly interpolate the performance over the space of CNN weights (several hundreds of thousands). Due to the high dimensionality, this is an intractable interpolation. In comparison, interpolation over the utility functions in our interpretable model is over a few hundred variables.

Prediction Results of Algorithm 2 on Deep CNN Performance

Table III displays the prediction errors (difference between the true and predicted classification accuracy) for the deep CNNs for all 55 architectures and all image classes in CIFAR-10. For a fixed CNN architecture and noise variance η>0\eta>0, the prediction error δη(x)\delta_{\eta}(x) for image class xx is defined as:

δη(x)=|p^(x|x)pCNN(x|x)|.\delta_{\eta}(x)=|\hat{p}(x|x)-p_{\text{CNN}}(x|x)|. (14)

In (14), p^(|)\hat{p}(\cdot|\cdot) is the predicted CNN performance generated from Algorithm 2 and pCNN(|)p_{\text{CNN}}(\cdot|\cdot) is the true CNN performance obtained by implementing the CNN. Recall that p(x|x)p(x|x) is the probability that the CNN correctly classifies an image belonging to class xx.

0: Dataset 𝔻\mathbb{D} (26) from KK deep CNNs from a fixed network architecture. The kthk^{\text{th}} CNN is trained on a noisy dataset with added Gaussian noise with variance ηk=1+0.1×(k1)\eta_{k}=1+0.1\times(k-1).Step 1: Constructing Interpretable Model. The most robust utility functions {uk}k=1K\{u_{k}^{\ast}\}_{k=1}^{K} and information acquisition cost CC^{\ast} are computed by solving the following convex optimization problem.
{uk,ck}k=1K=argmaxu1:KεKk=1Kuk22,BRP(𝔻,)ε.\displaystyle\hskip-5.69046pt\{u_{k}^{\ast},c^{\ast}_{k}\}_{k=1}^{K}=\operatorname*{argmax}_{u_{1:K}}\frac{\varepsilon~K}{\sum_{k=1}^{K}\|u_{k}\|_{2}^{2}},~\text{BRP}(\mathbb{D},\cdot)\leq-\varepsilon.
C(p(a|x))=maxk=1ck+x,aπ0(x)(p(a|x)pk(a|x))u(x,a).\displaystyle\hskip-5.69046ptC^{\ast}(p(a|x))=\max_{k=1}c_{k}^{\ast}+\sum_{x,a}\pi_{0}(x)(p(a|x)-p_{k}(a|x))u^{\ast}(x,a).
Step 2: Predicting Classification Accuracy. For an arbitrary noise variance η1ηηK\eta_{1}\leq\eta\leq\eta_{K}, obtain index g,1g<Kg,~1\leq g<K such that η[ηg,ηg+1]\eta\in[\eta_{g},\eta_{g+1}]. Then, the predicted classification accuracy p^(a|x)\hat{p}(a|x) for noise variance η\eta is computed as follows:
p^(a|x)=argmaxp(a|x)amaxbxπ0(x)p(a|x)u^(x,a)C(p),\displaystyle\hskip-2.84544pt\hat{p}(a|x)=\operatorname*{argmax}_{p(a|x)}\sum_{a}\max_{b}\sum_{x}\pi_{0}(x)p(a|x)\hat{u}(x,a)-C^{\ast}(p),
u^=10×{(ηg+1η)ug+(ηηg)ug+1}.\displaystyle\hskip-2.84544pt\hat{u}=10\times\{(\eta_{g+1}-\eta)u^{\ast}_{g}+(\eta-\eta_{g})u^{\ast}_{g+1}\}. (15)
Return: Predicted CNN performance p^(a|x)\hat{p}(a|x) for noise variance η\eta.
Algorithm 2 Predicting Deep CNN Classification Accuracy via the S-UMRI model using Theorem 2.

Discussion and Insights

(i) Our interpretable model can predict CNN classification performance with accuracy exceeding 94% (see below).
(ii) The interpretable model (utility functions and information acquisition cost) for our predictive procedure (Algorithm 2) is evaluated on the set of noise variances G1={1+0.1×(k1),k=1,2,,11}G_{1}=\{1+0.1\times(k-1),k=1,2,\ldots,11\}. The predictive procedure of Algorithm 2 is applied on the the set of noise variances given by G2={1.05+0.1×(k1),k=1,2,,10}G_{2}=\{1.05+0.1\times(k-1),k=1,2,\ldots,10\}. Table III displays the prediction errors δη(x)\delta_{\eta}(x) averaged over all ηG2\eta\in G_{2}.
(iii) From Table III, the prediction error δη(x)\delta_{\eta}(x) averaged over all image classes xx for the 55 CNN architectures are:

  1. 1.

    LeNet - 0.0380.038

  2. 2.

    AlexNet - 0.0360.036

  3. 3.

    VGG16 - 0.0410.041

  4. 4.

    ResNet-50 - 0.0270.027

  5. 5.

    NiN- 0.0350.035

So the least accuracy is 95.9%, and highest accuracy is 97.3%.
(iv) The prediction error averaged over the network architectures was observed to be minimum for image class ‘cat’ (98.1%) and maximum for image class ‘deer’ (95.7%) over all image classes.
(iv) Statistical Similarity between Deep CNNs and Interpretable Model. We computed the Kullback-Leibler (KL) divergence between the true and predicted classification performances pimp(a|x)p_{\text{imp}}(a|x) and p^(a|x)\hat{p}(a|x). Recall p^(a|x)\hat{p}(a|x) is computed from the interpretable model via Algorithm 2 and pCNN(a|x)p_{\text{CNN}}(a|x) is obtained from the CNN. The KL divergence values for the 5 CNN architectures are:

  1. 1.

    LeNet - 0.0150.015

  2. 2.

    AlexNet - 0.0120.012

  3. 3.

    VGG16 - 0.0160.016

  4. 4.

    Resnet-50 - 0.0060.006

  5. 5.

    NiN - 0.0180.018.

Thus the decisions made by the deep CNNs are statistically similar to decisions generated by our interpretable model.

4 Conclusions and Extensions

This paper has developed an interpretable model for deep image classification using micro- and behavioral economics theory. By extensive analysis of the decisions of 200200 CNNs over 55 popular CNN architectures, we showed that deep CNNs can be explained by rationally inattentive Bayesian utility maximization.

Our main results were the following:
1. Using the theory of Bayesian revealed preference, Theorem 1 gave a necessary and sufficient condition for the actions of a collection of decision makers to be consistent with rationally inattentive Bayesian utility maximization. We showed that deep CNNs operating on the CIFAR-10 dataset satisfy these necessary and sufficient conditions.
2. Next we studied the robustness margin by which the deep CNNs satisfy Theorem 1; we found that the margins were sufficiently large implying robustness of the results. Our robustness results are summarized in Table I.
3. In Theorem 2, we constructed the sparsest interpretable model from the feasible set generated using Theorem 1. The sparsest interpretable model explains deep CNN decisions using the least number of parameters. The sparsest interpretable model introduces a useful preference ordering amongst the set of hypotheses (image labels) considered by the deep neural network; for example, how much additional priority is allocated to the classification of a cat as a cat compared to a cat as a dog. In classical deep learning, this preference ordering is not explicitly generated
4. Finally, we showed that our interpretable model can predict CNN performance with accuracy exceeding 94%, and the decisions generated by our interpretable model are statistically similar to that of a deep CNN. At a more conceptual level, our results suggest that deep CNNs for image classification are equivalent to an economics-based constrained Bayesian decision system (used in micro-economics to model human decision making).

Extensions. An immediate extension of this work is to construct appropriately designed image features to replace the image class label as the state. This would result in a richer descriptive model of the CNN due to more degrees of freedom in the utility function.

Our proposed interpretable model generates a concave utility function by design. This is an important feature of the revealed preference framework; even though the actual deep learner’s utility may not be convex. To quote Varian [11]: ”If data can be rationalized by any non-trivial utility function, then it can be rationalized by a nice utility function. Violations of concavity cannot be detected with only a finite number of observations.” A more speculative extension is to investigate the asymptotic behavior of the BRP and S-BRP decision tests for rationally inattentive utility maximization-do the tests pass when the number of deep CNNs tend to infinity? Recent results [41] show that an infinite dataset can at best be rationalized by a quasi-concave utility function.

Reproducibility: The computer programs and deep image classification datasets needed to reproduce all the results in this paper can be obtained from the public GitHub repository https://github.com/KunalP117/DL_RI.

Acknowledgement

This research was supported in part by the Army Research Office under grants W911NF-21-1-0093 and W911NF-19-1-0365, and the National Science Foundation under grant CCF-2112457.

References

  • [1] C. Sims. Implications of rational inattention. Journal of monetary Economics, 50(3):665–690, 2003.
  • [2] C. Sims. Rational inattention and monetary economics. Handbook of Monetary Economics, 3:155–181, 2010.
  • [3] M. Woodford. Inattentive valuation and reference-dependent choice. 2012.
  • [4] F. Matějka and A. McKay. Rational inattention to discrete choices: A new foundation for the multinomial logit model. American Economic Review, 105(1):272–98, 2015.
  • [5] H. De Oliveira, T. Denti, M. Mihm, and K. Ozbek. Rationally inattentive preferences and hidden information costs. Theoretical Economics, 12(2):621–654, 2017.
  • [6] A. Caplin and D. Martin. A testable theory of imperfect perception. The Economic Journal, 125(582):184–202, 2015.
  • [7] A. Caplin and M. Dean. Revealed preference, rational inattention, and costly information acquisition. The American Economic Review, 105(7):2183–2203, 2015.
  • [8] A. Krizhevsky, G. Hinton, et al. Learning multiple layers of features from tiny images. 2009.
  • [9] S. N. Afriat. The construction of utility functions from expenditure data. International economic review, 8(1):67–77, 1967.
  • [10] W. E. Diewert. Afriat and revealed preference theory. The Review of Economic Studies, 40(3):419–425, 1973.
  • [11] H. R. Varian. The nonparametric approach to demand analysis. Econometrica: Journal of the Econometric Society, pages 945–973, 1982.
  • [12] S. Chakraborty, R. Tomsett, R. Raghavendra, D. Harborne, M. Alzantot, F. Cerutti, M. Srivastava, A. Preece, S. Julier, R. M. Rao, et al. Interpretability of deep learning models: a survey of results. In 2017 IEEE smartworld, ubiquitous intelligence & computing, advanced & trusted computed, scalable computing & communications, cloud & big data computing, Internet of people and smart city innovation (smartworld/SCALCOM/UIC/ATC/CBDcom/IOP/SCI), pages 1–6. IEEE, 2017.
  • [13] F. Doshi-Velez and B. Kim. Towards a rigorous science of interpretable machine learning. arXiv preprint arXiv:1702.08608, 2017.
  • [14] R. Guidotti, A. Monreale, S. Ruggieri, F. Turini, F. Giannotti, and D. Pedreschi. A survey of methods for explaining black box models. ACM Computing Surveys (CSUR), 51(5):1–42, 2018.
  • [15] W. J. Murdoch, C. Singh, K. Kumbier, R. Abbasi-Asl, and B. Yu. Interpretable machine learning: definitions, methods, and applications. arXiv preprint arXiv:1901.04592, 2019.
  • [16] D. Ciregan, U. Meier, and J. Schmidhuber. Multi-column deep neural networks for image classification. In 2012 IEEE conference on computer vision and pattern recognition, pages 3642–3649. IEEE, 2012.
  • [17] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, et al. ImageNet large scale visual recognition challenge. International journal of computer vision, 115(3):211–252, 2015.
  • [18] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
  • [19] K. Simonyan, A. Vedaldi, and A. Zisserman. Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv preprint arXiv:1312.6034, 2013.
  • [20] A. Nguyen, A. Dosovitskiy, J. Yosinski, T. Brox, and J. Clune. Synthesizing the preferred inputs for neurons in neural networks via deep generator networks. arXiv preprint arXiv:1605.09304, 2016.
  • [21] P. Hase, C. Chen, O. Li, and C. Rudin. Interpretable image recognition with hierarchical prototypes. In Proceedings of the AAAI Conference on Human Computation and Crowdsourcing, volume 7-1, pages 32–40, 2019.
  • [22] T. Lei, R. Barzilay, and T. Jaakkola. Rationalizing neural predictions. arXiv preprint arXiv:1606.04155, 2016.
  • [23] S. Lundberg and S.-I. Lee. A unified approach to interpreting model predictions. arXiv preprint arXiv:1705.07874, 2017.
  • [24] S. Bach, A. Binder, G. Montavon, F. Klauschen, K.-R. Müller, and W. Samek. On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation. PloS one, 10(7):e0130140, 2015.
  • [25] M. T. Ribeiro, S. Singh, and C. Guestrin. Model-agnostic interpretability of machine learning. arXiv preprint arXiv:1606.05386, 2016.
  • [26] A. Shrikumar, P. Greenside, A. Shcherbina, and A. Kundaje. Not just a black box: Learning important features through propagating activation differences. arXiv preprint arXiv:1605.01713, 2016.
  • [27] H. Wang and D.-Y. Yeung. Towards Bayesian deep learning: A framework and some existing methods. IEEE Transactions on Knowledge and Data Engineering, 28(12):3395–3408, 2016.
  • [28] P. Milgrom. Good news and bad news: Representation theorems and applications. Bell Journal of Economics, 12(2):380–391, 1981.
  • [29] L. Huang and H. Liu. Rational inattention and portfolio selection. The Journal of Finance, 62(4):1999–2040, 2007.
  • [30] S. E. Mirsadeghi, A. Royat, and H. Rezatofighi. Unsupervised image segmentation by mutual information maximization and adversarial regularization. IEEE Robotics and Automation Letters, 2021.
  • [31] W. Hoiles, V. Krishnamurthy, and K. Pattanayak. Rationally Inattentive Inverse Reinforcement Learning Explains YouTube commenting behavior. The Journal of Machine Learning Research, 21(170):1–39, 2020.
  • [32] K. Pattanayak and V. Krishnamurthy. Unifying classical and bayesian revealed preference, 2021.
  • [33] Y. LeCun, B. Boser, J. Denker, D. Henderson, R. Howard, W. Hubbard, and L. Jackel. Handwritten digit recognition with a back-propagation network. Advances in neural information processing systems, 2, 1989.
  • [34] A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems, 25:1097–1105, 2012.
  • [35] K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556, 2014.
  • [36] M. Lin, Q. Chen, and S. Yan. Network in network. arXiv preprint arXiv:1312.4400, 2013.
  • [37] G. Hinton, N. Srivastava, and K. Swersky. Neural networks for machine learning lecture 6a overview of mini-batch gradient descent. Cited on, 14(8), 2012.
  • [38] R. Anand, T. Shanthi, R. Sabeenian, and S. Veni. Real time noisy dataset implementation of optical character identification using cnn. International Journal of Intelligent Enterprise, 7(1-3):67–80, 2020.
  • [39] Y. Bengio. Learning deep architectures for AI. Now Publishers Inc, 2009.
  • [40] P. Vincent, H. Larochelle, Y. Bengio, and P.-A. Manzagol. Extracting and composing robust features with denoising autoencoders. In Proceedings of the 25th international conference on Machine learning, pages 1096–1103, 2008.
  • [41] P. J. Reny. A characterization of rationalizable consumer behavior. Econometrica, 83(1):175–192, 2015.
  • [42] D. Blackwell. Equivalent comparisons of experiments. The annals of mathematical statistics, pages 265–272, 1953.

Appendix

A.1 Proof of Theorem 1

Proof of necessity of NIAS and NIAC:

  1. 1.

    NIAS (8): For agent k𝒦k\in\mathcal{K}, define the subset 𝒴a𝒴\mathcal{Y}_{a}\subseteq\mathcal{Y} so that for any observation y𝒴ay\in\mathcal{Y}_{a}, given posterior pmf pk(x|y)p_{k}(x|y), the optimal choice of action is aa (3). We define the revealed posterior pmf given action aa as pk(x|a)p_{k}(x|a). The revealed posterior pmf is a stochastically garbled version of the actual posterior pmf pk(x|y)p_{k}(x|y), that is,

    pk(x|a)=y𝒴pk(x,y,a)pk(a)=y𝒴pk(y|a)pk(x|y)p_{k}(x|a)=\sum_{y\in\mathcal{Y}}\frac{p_{k}(x,y,a)}{p_{k}(a)}=\sum_{y\in\mathcal{Y}}p_{k}(y|a)p_{k}(x|y) (16)

    Since the optimal action is aa for all y𝒴ay\in\mathcal{Y}_{a}, (3) implies:

    x𝒳pk(x|y)(uk(x,b)uk(x,a))0\displaystyle\hskip-11.38092pt\quad\quad~~\sum_{x\in\mathcal{X}}p_{k}(x|y)(u_{k}(x,b)-u_{k}(x,a))\leq 0
    y𝒴apk(y|a)x𝒳pk(x|y)(uk(x,b)uk(x,a))0\displaystyle\hskip-11.38092pt\implies\sum_{y\in\mathcal{Y}_{a}}p_{k}(y|a)\sum_{x\in\mathcal{X}}p_{k}(x|y)(u_{k}(x,b)-u_{k}(x,a))\leq 0
    y𝒴pk(y|a)x𝒳pk(x|y)(uk(x,b)uk(x,a))0\displaystyle\hskip-11.38092pt\implies\sum_{y\in\mathcal{Y}}p_{k}(y|a)\sum_{x\in\mathcal{X}}p_{k}(x|y)(u_{k}(x,b)-u_{k}(x,a))\leq 0
    (since pk(y|a)=0,y𝒴\𝒴a)\displaystyle\hskip-11.38092pt\quad\quad~~(\text{since }p_{k}(y|a)=0,~\forall y\in\mathcal{Y}\backslash\mathcal{Y}_{a})
    x𝒳y𝒴pk(y|a)pk(x|y)(uk(x,b)uk(x,a))0\displaystyle\hskip-11.38092pt\implies\sum_{x\in\mathcal{X}}\sum_{y\in\mathcal{Y}}p_{k}(y|a)p_{k}(x|y)(u_{k}(x,b)-u_{k}(x,a))\leq 0
    x𝒳pk(x|a)(uk(x,b)uk(x,a))0(from (16))\displaystyle\hskip-11.38092pt\implies\sum_{x\in\mathcal{X}}p_{k}(x|a)(u_{k}(x,b)-u_{k}(x,a))\leq 0~(\text{from }\eqref{eqn:revpos})

    This is precisely the NIAS inequality (8).

  2. 2.

    NIAC (9): Let ck=C(αk)>0c_{k}=C(\alpha_{k})>0, where C()C(\cdot) denotes the information acquisition cost of the collection of agents 𝒦\mathcal{K}. Also, let J(αk,uk)J(\alpha_{k},u_{k}) denote the expected utility of the kthk^{\text{th}} agent given attention strategy αk\alpha_{k} (first term in RHS of (4)). Here, the expectation is taken wrt both the state xx and observation yy. It can be verified that J(,uk)J(\cdot,u_{k}) is convex in the first argument. Finally, for the kthk^{\text{th}} agent, we define the revealed attention strategy αk\alpha_{k}^{\prime} over the set of actions 𝒜\mathcal{A} as

    αk(a|x)=pk(a|x),a𝒜,\alpha_{k}^{\prime}(a|x)=p_{k}(a|x),~\forall a\in\mathcal{A},

    where the variable pk(a|x)p_{k}(a|x) is obtained from the dataset 𝔻\mathbb{D}. Clearly, the revealed attention strategy is a stochastically garbled version of the true attention strategy since

    αk(a|x)=pk(a|x)=y𝒴pk(a|y)αk(y|x)\alpha_{k}^{\prime}(a|x)=p_{k}(a|x)=\sum_{y\in\mathcal{Y}}p_{k}(a|y)\alpha_{k}(y|x) (17)

    From Blackwell dominance [42] and the convexity of the expected utility functional J(,uk)J(\cdot,u_{k}), it follows that:

    J(αk,uj)J(αk,uj),J(\alpha_{k}^{\prime},u_{j})\leq J(\alpha_{k},u_{j}), (18)

    when αk\alpha_{k} Blackwell dominates αk\alpha_{k}^{\prime}. The above relationship holds with equality if k=jk=j (this is due to NIAS (8)). We now turn to condition (4) for optimality of attention strategy. The following inequalities hold for any pair of agents jkj\neq k:

    J(αk,uk)ck=(18)J(αk,uk)ck\displaystyle J(\alpha_{k}^{\prime},u_{k})-c_{k}\overset{\eqref{eqn:ineq_BLK}}{=}J(\alpha_{k},u_{k})-c_{k}
    (4)\displaystyle\overset{\eqref{eqn:attentionmaximization}}{\geq}~ J(αj,uk)cj(18)J(αj,uk)cj.\displaystyle J(\alpha_{j},u_{k})-c_{j}\overset{\eqref{eqn:ineq_BLK}}{\geq}~J(\alpha_{j}^{\prime},u_{k})-c_{j}. (19)

    This is precisely the NIAC inequality (9).

Proof for sufficiency of NIAS and NIAC: Let {uk,ck}k=1K\{u_{k},c_{k}\}_{k=1}^{K} denote a feasible solution to the NIAS and NIAC inequalities of Theorem 1. To prove sufficiency, we construct an UMRI tuple as a function of dataset 𝔻\mathbb{D} and the feasible solution that satisfies the optimality conditions (3),(4) of Definition 1.

Consider the following UMRI model tuple:

Θ=(𝒦,𝒳,𝒴=𝒜,𝒜,π0,C,{pk(a|x),uk,k𝒦}), where\displaystyle\Theta=(\mathcal{K},\mathcal{X},\mathcal{Y}=\mathcal{A},\mathcal{A},\pi_{0},C,\{p_{k}(a|x),u_{k},k\in\mathcal{K}\}),\text{ where}
C(p(a|x))=maxk𝒦ck+J(p(a|x),uk)J(pk(a|x),uk).\displaystyle C(p(a|x))=\max_{k\in\mathcal{K}}c_{k}+J(p(a|x),u_{k})-J(p_{k}(a|x),u_{k}). (20)

In (20), C()C(\cdot) is a convex cost since it is a point-wise maximum of monotone convex functions. Further, since NIAC is satisfied, (20) implies C(αk)=ckC(\alpha_{k})=c_{k}. It only remains to show that inequalities (3) and (4) in Definition 1 are satisfied for all agents in 𝒦\mathcal{K}.

  1. 1.

    NIAS implies (3) holds. This is straightforward to show since the observation and action sets are identical.

  2. 2.

    Information Acquisition Cost (20) implies (4) holds. Fix agent j𝒦j\in\mathcal{K}. Then, for any attention strategy p(a|x)p(a|x), the following inequalities hold.

    C(p(a|x))=maxk𝒦ck+J(p(a|x),uk)J(pk(a|x),uk)\displaystyle C(p(a|x))=\max_{k\in\mathcal{K}}~c_{k}+J(p(a|x),u_{k})-J(p_{k}(a|x),u_{k})
    J(pj(a|x))cjJ(p(a|x))C(p(a|x)),p(a|x)\displaystyle\hskip-11.38092pt\implies J(p_{j}(a|x))-c_{j}\geq J(p(a|x))-C(p(a|x)),~\forall~p(a|x)
    pk(a|x)argmaxp(a|x)J(p(a|x),uk)C(p(a|x))=(4).\displaystyle\hskip-11.38092pt\implies p_{k}(a|x)\in\operatorname*{argmax}_{p(a|x)}J(p(a|x),u_{k})-C(p(a|x))=\eqref{eqn:attentionmaximization}.

A.2 S-UMRI (Sparse UMRI ) Model for Rationally Inattentive Bayesian Utility Maximization

In Sec. 2.1, we outlined the UMRI model for rationally inattentive utility maximization of KK Bayesian agents parameterized by KK utility functions and a cost of information acquisition. This section proposes a sparse version of the UMRI model, namely, the S-UMRI model that is parameterized by a single utility function that rationalizes the decisions of KK Bayesian agents. Abstractly, the S-UMRI model is described by the tuple

Θ=(𝒦,𝒳,𝒴,𝒜,π0,C,u,{αk,λk,k𝒦}).\Theta=(\mathcal{K},\mathcal{X},\mathcal{Y},\mathcal{A},\pi_{0},C,u,\{\alpha_{k},\lambda_{k},k\in\mathcal{K}\}). (21)

All parameters in (21) are identical to that in (1) except for the additional parameter λk+\lambda_{k}\in\mathbb{R}_{+}. λk\lambda_{k} can be interpreted as the sensitivity to information acquisition of the kthk^{\text{th}} agent. We discuss the significance of λk\lambda_{k} in more detail below. In complete analogy to Definition 1, Definition 3 below specifies the optimal action and attention strategy policy of the Bayesian agents in 𝒦\mathcal{K}.

Definition 3 (Rationally Inattentive Utility Maximization for S-UMRI ).

Consider a collection of Bayesian agents 𝒦\mathcal{K} parameterized by Θ\Theta in (21) under the S-UMRI model. Then,
(a)
Expected Utility Maximization: Given posterior pmf p(x|y)p(x|y), agent k𝒦k\in\mathcal{K} chooses action aa that maximizes its expected utility.

aargmaxa𝒜𝔼x{uk(x,a)|y}=x𝒳p(x|y)u(x,a)a\in\operatorname*{argmax}_{a^{\prime}\in\mathcal{A}}~\mathbb{E}_{x}\{u_{k}(x,a^{\prime})|y\}=\sum\limits_{x\in\mathcal{X}}p(x|y)u(x,a^{\prime}) (22)

(b) Attention Strategy Rationality: Agent kk chooses attention strategy αk\alpha_{k} that optimally trades off between utility maximization and cost minimization.

αkargmaxα𝔼y{\displaystyle\alpha_{k}\in\operatorname*{argmax}_{\alpha^{\prime}}\mathbb{E}_{y}\{ maxa𝒜𝔼x{u(x,a)|y}}λkC(α,π0)\displaystyle\operatorname*{max}_{a\in\mathcal{A}}\mathbb{E}_{x}\{u(x,a)|y\}\}-\lambda_{k}C(\alpha^{\prime},\pi_{0}) (23)

Remarks. 1. Role of λk\lambda_{k}. In (23), λk\lambda_{k} is the differentiating parameter across agents. Even though all agents have the same utility function, different values of λk\lambda_{k} result in different optimal strategies αk\alpha_{k} (23).
2. Sparsity of S-UMRI . The UMRI model tuple for KK Bayesian agents is parameterized using K(|𝒳||𝒜|+1)K(|\mathcal{X}||\mathcal{A}|+1) variables. In comparison, the S-UMRI tuple is parameterized via |𝒳||𝒜|+K|\mathcal{X}||\mathcal{A}|+K variables. The difference in variables for parametrization is linear in KK.

Finally, in complete analogy to Theorem 1, we now state Theorem 3 that states necessary and sufficient conditions for the decisions of a collection of Bayesian agents to be rationalized by the S-UMRI model.

Theorem 3 (S-BRP Test for Rationally Inattentive Utility Maximization).

Given the dataset 𝔻\mathbb{D} (5) obtained from a collection of Bayesian agents 𝒦\mathcal{K}. Then,
1. Existence: There exists a S-UMRI tuple Θ(𝔻)\Theta(\mathbb{D}) (1) that rationalizes dataset 𝔻\mathbb{D} if and only if there exists a feasible solution that satisfies the set of inequalities

S-BRP(𝔻)𝟎.\text{S-BRP}(\mathbb{D})\leq\mathbf{0}. (24)

In (6), S-BRP()(\cdot) corresponds to a set of inequalities stated in Algorithm 3 below. The set-valued estimate of Θ\Theta that rationalizes 𝔻\mathbb{D} is the set of all feasible solutions to (6).
2. Reconstruction: Given a feasible solution {u,λk,ck}k=1K\{u,\lambda_{k},c_{k}\}_{k=1}^{K} to SBRP(𝔻,)S-BRP(\mathbb{D},\cdot), uu is the kthk^{\text{th}} Bayesian agent’s utility function, for all k=1,2,,Kk=1,2,\ldots,K. The set of observations 𝒴=𝒜\mathcal{Y}=\mathcal{A}, the set of actions in 𝔻\mathbb{D}. The feasible cost of information acquisition CC in Θ(𝔻)\Theta(\mathbb{D}) is defined in terms of the feasible variables ck,λkc_{k},\lambda_{k} as:

C(α)\displaystyle C(\alpha) =maxk𝒦ck+λkx,a(p(x,a)pk(x,a))u(x,a),\displaystyle=\max_{k\in\mathcal{K}}c_{k}+\lambda_{k}\sum_{x,a}(p(x,a)-p_{k}(x,a))u(x,a),
(α={p(a|x),a𝒜,x𝒳})\displaystyle~(\alpha=\{p(a|x),a\in\mathcal{A},x\in\mathcal{X}\}) (25)

The proof of Theorem 3 closely follows the proof of Theorem 1 and hence, omitted. In comparison to the BRP test of Theorem 1, the S-BRP test has the same number of inequalities but fewer decision variables. Hence, the set of feasible parameters generated from Algorithm 3 is smaller compared to Algorithm 1.

0: Dataset 𝔻={π0,pk(a|x),x,a𝒳,k𝒦}\mathbb{D}=\{\pi_{0},p_{k}(a|x),x,a\in\mathcal{X},k\in\mathcal{K}\} from a collection of Bayesian agents 𝒦\mathcal{K}. Find: Positive reals ck,λk,u(0,1]c_{k},\lambda_{k},u\in(0,1] for all x𝒳,x\in\mathcal{X}, a𝒜,k𝒦a\in\mathcal{A},~k\in\mathcal{K} that satisfy the following inequalities:
1.xpk(x|a)(u(x,b)u(x,a))0,a,b,k,\displaystyle\hskip-22.76228pt1.~\sum_{x}p_{k}(x|a)~(u(x,b)-u(x,a))\leq 0,\forall a,b,k,
2.x,a(pj(x,a)pk(x,a))u(x,a)+λk(ckcj)0,j,k,\displaystyle\hskip-22.76228pt2.~\sum_{x,a}(p_{j}(x,a)-p_{k}(x,a))u(x,a)+\lambda_{k}(c_{k}-c_{j})\leq 0,\forall j,k,
where pk(x,a)=π0(x)pk(a|x),pk(x|a)=pk(x,a)xpk(x,a)p_{k}(x,a)=\pi_{0}(x)p_{k}(a|x),~p_{k}(x|a)=\frac{p_{k}(x,a)}{\sum_{x^{\prime}}p_{k}(x^{\prime},a)}. Return: Set of feasible utility function uu, scalars λk\lambda_{k} and information acquisition costs ckc_{k} incurred by agents k𝒦k\in\mathcal{K}.
Algorithm 3 S-BRP Convex Feasibility Test of Theorem 3

A.3 Construction of Deep CNN Dataset

We now explain how the decisions of the deep CNNs are incorporated into our main theorems Theorems 1 and 3. Suppose KK deep CNNs indexed by k=1,,Kk=1,\ldots,K with different training parameters are trained on the CIFAR-10 dataset. For every trained deep CNN kk, given test image ii from CIFAR-10 test dataset with image class sis_{i}, let the vector fi,kΔ9f_{i,k}\in\Delta^{9} denote the corresponding softmax output of the deep CNN. The vector fi,kf_{i,k} is a 1010-dimensional probability vector where fi,k(j)f_{i,k}(j) is the probability that deep CNN kk classifies test image ii into image class jj.

The decisions of all KK deep CNNs on the CIFAR-10 test dataset are aggregated into dataset 𝔻\mathbb{D} for compatibility with Theorems 1 and 3 as follows:

𝔻={π0,pk(a|x),x,a𝒳,k{1,2,,K}}, where\displaystyle\mathbb{D}=\{\pi_{0},p_{k}(a|x),x,a\in\mathcal{X},k\in\{1,2,\ldots,K\}\},\text{ where}
π0(x)=i=1N𝟙{si=x}N,pk(a|x)=i=1N𝟙{si=x}fi,k(a)i=1N𝟙{si=x},\displaystyle\pi_{0}(x)=\sum_{i=1}^{N}\frac{\mathbbm{1}\{s_{i}=x\}}{N},~p_{k}(a|x)=\frac{\sum_{i=1}^{N}\mathbbm{1}\{s_{i}=x\}f_{i,k}(a)}{\sum_{i=1}^{N}\mathbbm{1}\{s_{i}=x\}},
N=10000,𝒳=𝒜={1,2,10}.\displaystyle N=10000,~\mathcal{X}=\mathcal{A}=\{1,2,\ldots 10\}. (26)

Here π0(x)\pi_{0}(x) is the empirical probability that the image class of a test image in the CIFAR-10 test dataset is xx. Since the output of the CNN is a probability vector, we compute pk(a|x)p_{k}(a|x) for the kthk^{\text{th}} CNN by averaging the atha^{\text{th}} component of the output over all test images in image class xx. Finally, NN is the number of test images in the CIFAR-10 test dataset, and the set of true and predicted image classes are the same, i.e., 𝒳=𝒜\mathcal{X}=\mathcal{A}. Although implicit in the above description, our Bayesian revealed preference approach to interpretable deep image classification assumes the deep CNN’s (agent’s) ground truth is the true image label, and its decision aa is the predicted image label.