This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Consistent Feature Selection for Analytic Deep Neural Networks

Vu Dinh
Department of Mathematical Sciences
University of Delaware
Delaware, USA
vucdinh@udel.edu
&Lam Si Tung Ho
Department of Mathematics and Statistics
Dalhousie University
Halifax, Nova Scotia, Canada
Lam.Ho@dal.ca
These authors contributed equally to this work.
Abstract

One of the most important steps toward interpretability and explainability of neural network models is feature selection, which aims to identify the subset of relevant features. Theoretical results in the field have mostly focused on the prediction aspect of the problem with virtually no work on feature selection consistency for deep neural networks due to the model’s severe nonlinearity and unidentifiability. This lack of theoretical foundation casts doubt on the applicability of deep learning to contexts where correct interpretations of the features play a central role.

In this work, we investigate the problem of feature selection for analytic deep networks. We prove that for a wide class of networks, including deep feed-forward neural networks, convolutional neural networks, and a major sub-class of residual neural networks, the Adaptive Group Lasso selection procedure with Group Lasso as the base estimator is selection-consistent. The work provides further evidence that Group Lasso might be inefficient for feature selection with neural networks and advocates the use of Adaptive Group Lasso over the popular Group Lasso.

1 Introduction

In recent years, neural networks have become one of the most popular models for learning systems for their strong approximation properties and superior predictive performance. Despite their successes, their “black box” nature provides little insight into how predictions are being made. This lack of interpretability and explainability stems from the fact that there has been little work that investigates statistical properties of neural networks due to its severe nonlinearity and unidentifiability.

One of the most important steps toward model interpretability is feature selection, which aims to identify the subset of relevant features with respect to an outcome. Numerous techniques have been developed in the statistics literature and adapted to neural networks over the years. For neural networks, standard Lasso is not ideal because a feature can only be dropped if all of its connections have been shrunk to zero together, an objective that Lasso does not actively pursue. Zhao et al. (2015), Scardapane et al. (2017) and Zhang et al. (2019) address this concern by utilizing Group Lasso and its variations for selecting features of deep neural networks (DNNs). Feng and Simon (2017) propose fitting a neural network with a Sparse Group Lasso penalty on the first-layer input weights while adding a sparsity penalty to later layers. Li et al. (2016) propose adding a sparse one-to-one linear layer between the input layer and the first hidden layer of a neural network and performing 1\ell_{1} regularization on the weights of this extra layer. Similar ideas are also applied to various types of networks and learning contexts (Tank et al., 2018; Ainsworth et al., 2018; Lemhadri et al., 2019).

Beyond regularization-based methods, Horel and Giesecke (2019) develop a test to assess the statistical significance of the feature variables of a neural network. Alternatively, a ranking of feature importance based on local perturbations, which includes fitting a model in the local region around the input or locally perturbing the input to see how predictions change, can also be used as a proxy for feature selection (Simonyan et al., 2013; Ibrahim et al., 2014; Ribeiro et al., 2016; Nezhad et al., 2016; Lundberg and Lee, 2017; Shrikumar et al., 2017; Ching et al., 2018; Taherkhani et al., 2018; Lu et al., 2018). “Though these methods can yield insightful interpretations, they focus on specific architectures of DNNs and can be difficult to generalize" (Lu et al., 2018).

Despite the success of these approaches, little is known about their theoretical properties. Results in the field have been either about shallow networks with one hidden layer (Dinh and Ho, 2020; Liang et al., 2018; Ye and Sun, 2018), or focus on posterior concentration, prediction consistency, parameter-estimation consistency and convergence of feature importance (Feng and Simon, 2017; Farrell et al., 2018; Polson and Ročková, 2018; Fallahgoul et al., 2019; Shen et al., 2019; Liu, 2019), with virtually no work on feature selection consistency for deep networks. This lack of a theoretical foundation for feature selection casts doubt on the applicability of deep learning to applications where correct interpretations play a central role such as medical and engineering sciences. This is problematic since works in other contexts indicated that Lasso and Group Lasso could be inconsistent/inefficient for feature selection (Zou, 2006; Zhao and Yu, 2006; Wang and Leng, 2008), especially when the model is highly-nonlinear (Zhang et al., 2018).

In this work, we investigate the problem of feature selection for deep networks. We prove that for a wide class of deep analytic neural networks, the GroupLasso + AdaptiveGroupLasso procedure (i.e., the Adaptive Group Lasso selection with Group Lasso as the base estimator) is feature-selection-consistent. The results also provide further evidence that Group Lasso might be inconsistent for feature selection and advocate the use of the Adaptive Group Lasso over the popular Group Lasso.

2 Feature selection with analytic deep neural networks

Throughout the paper, we consider a general analytic neural network model described as follows. Given an input xx that belongs to be a bounded open set 𝒳𝐑d0\mathcal{X}\subset\mathbf{R}^{d_{0}}, the output map fα(x)f_{\alpha}(x) of an LL-layer neural network with parameters α=(P,p,S,Q,q)\alpha=(P,p,S,Q,q) is defined by

  • input layer: h1(x)=Px+ph_{1}(x)=P\cdot x+p

  • hidden layers: hj(x)=ϕj1(S,hj1(x),hj2(x),,h1(x)),j=2,,L1.h_{j}(x)=\phi_{j-1}(S,h_{j-1}(x),h_{j-2}(x),\ldots,h_{1}(x)),~{}~{}j=2,\ldots,L-1.

  • output layer: fα(x)=hL(x)=QhL1(x)+qf_{\alpha}(x)=h_{L}(x)=Q\cdot h_{L-1}(x)+q

where did_{i} denotes the number of nodes in the ii-th hidden layer, P𝐑d1×d0P\in\mathbf{R}^{d_{1}\times d_{0}}, Q𝐑dL×dL1Q\in\mathbf{R}^{d_{L}\times d_{L-1}}, p𝐑d1p\in\mathbf{R}^{d_{1}}, q𝐑dLq\in\mathbf{R}^{d_{L}}, and ϕ1,ϕ2,,ϕL2\phi_{1},\phi_{2},\ldots,\phi_{L-2} are analytic functions parameterized by the hidden layers’ parameter SS. This framework allows interactions across layers of the network architecture and only requires that (i) the model interacts with inputs through a finite set of linear units, and (ii) the activation functions are analytic. This parameterization encompasses a wide class of models, including feed-forward networks, convolutional networks, and (a major subclass of) residual networks.

Hereafter, we assume that the set of all feasible vectors α\alpha of the model is a hypercube 𝒲=[A,A]nα\mathcal{W}=[-A,A]^{n_{\alpha}} and use the notation h[i]h^{[i]} to denote the ii-th component of a vector hh and u[i,k]u^{[i,k]} to denote the [i,k][i,k]-entry of a matrix uu. We study the feature selection problem for regression in the model-based setting:

Assumption 2.1.

Training data {(Xi,Yi)}i=1n\{(X_{i},Y_{i})\}_{i=1}^{n} are independent and identically distributed (i.i.d ) samples generated from PX,YP^{*}_{X,Y} such that the input density pXp_{X} is positive and continuous on its domain 𝒳\mathcal{X} and Yi=fα(Xi)+ϵiY_{i}=f_{\alpha^{*}}(X_{i})+\epsilon_{i} where ϵi𝒩(0,σ2)\epsilon_{i}\sim\mathcal{N}(0,\sigma^{2}) and α𝒲\alpha^{*}\in\mathcal{W}.

Remarks.

Although the analyses of the paper focus on Gaussian noise, the results also apply to all models for which ϵ2\epsilon^{2} is sub-Gaussian, including the cases of bounded noise. The assumption that the input density pXp_{X} is positive and continuous on its bounded domain 𝒳\mathcal{X} ensures that there is no perfect correlations among the inputs and plays an important role in our analysis. Throughout the paper, the network is assumed to be fixed, and we are interested in the asymptotic behaviors of estimators in the learning setting when the sample size nn increases.

We assume that the “true” model fα(x)f_{\alpha^{*}}(x) only depends on xx through a subset of significant features while being independent of the others. The goal of feature selection is to identify this set of significant features from the given data. For convenience, we separate the inputs into two groups snss\in\mathbb{R}^{n_{s}} and znzz\in\mathbb{R}^{n_{z}} (with ns+nz=d0n_{s}+n_{z}=d_{0}) that denote the significant and non-significant variables of fα(x)f_{\alpha^{*}}(x), respectively. Similarly, the parameters of the first layer (PP and pp) are grouped by their corresponding input, i.e.

x=(s,z),P=(u,v)andp=(b1,b2)x=(s,z),\hskip 17.34518ptP=(u,v)\hskip 17.34518pt\text{and}\hskip 17.34518ptp=(b_{1},b_{2})

where u𝐑d1×nsu\in\mathbf{R}^{d_{1}\times n_{s}}, v𝐑d1×nzv\in\mathbf{R}^{d_{1}\times n_{z}}, b1𝐑nsb_{1}\in\mathbf{R}^{n_{s}} and b2𝐑nzb_{2}\in\mathbf{R}^{n_{z}}. We note that this separation is simply for mathematical convenience and the training algorithm is not aware of such dichotomy.

We define significance and selection consistency as follows.

Definition 2.2 (Significance).

The ii-th input x[i]x^{[i]} of a network fα(x)f_{\alpha}(x) is referred to as non-significant iff the output of the network does not depend on the value of that input variable. That is, let gi(x,s)g_{i}(x,s) denote the vector obtained from xx by replacing the ii-th component of xx by ss\in\mathbb{R}, we have fα(x)=fα(gi(x,s))f_{\alpha}(x)=f_{\alpha}(g_{i}(x,s)) for all x𝒳,sx\in\mathcal{X},s\in\mathbb{R}.

Definition 2.3 (Feature selection consistency).

An estimator αn\alpha_{n} with first layer’s parameters (un,vn)(u_{n},v_{n}) is feature selection consistent if for any δ>0\delta>0, there exists NδN_{\delta} such that for n>Nδn>N_{\delta}, we have

un[:,k]0,k=1,,ns,andvn[:,l]=0,l=1,,nzu_{n}^{[:,k]}\neq 0,~{}\forall k=1,\ldots,n_{s},\hskip 17.34518pt\text{and}\hskip 17.34518ptv_{n}^{[:,l]}=0,~{}\forall l=1,\ldots,n_{z}

with probability at least 1δ1-\delta.

One popular method for feature selection with neural networks is the Group Lasso (GL). Moreover, since the model of our framework only interacts with the inputs through the first layer, it is reasonable that the penalty should be imposed only on these parameters. A simple GL estimator for neural networks is thus defined by

α^n:=argminα=(u,v,b1,b2,S,Q,q)1ni=1n(α,Xi,Yi)+λnL(α)whereL(α)=k=1nsu[:,k]+l=1nzv[:,l],\hat{\alpha}_{n}:=\operatornamewithlimits{argmin}_{\alpha=(u,v,b_{1},b_{2},S,Q,q)}{~{}~{}\frac{1}{n}\sum_{i=1}^{n}{\ell(\alpha,X_{i},Y_{i})}+\lambda_{n}L(\alpha)}~{}~{}\text{where}~{}~{}L(\alpha)=\sum_{k=1}^{n_{s}}{\|u^{[:,k]}\|}+\sum_{l=1}^{n_{z}}{\|v^{[:,l]}\|},

(α,x,y)=(yfα(x))2\ell(\alpha,x,y)=(y-f_{\alpha}(x))^{2} is the square-loss, λn>0\lambda_{n}>0, \|\cdot\| is the standard Euclidean norm and u[:,k]u^{[:,k]} is the vector of parameters associated with kk-th significant input.

While GL and its variation, the Sparse Group Lasso, have become the foundation for many feature selection algorithms in the field, there is no known result about selection consistency for this class of estimators. Furthermore, like the regular Lasso, GL penalizes groups of parameters with the same regularization strengths and it has been shown that excessive penalty applied to the significant variables can affect selection consistency (Zou, 2006; Wang and Leng, 2008). To address these issues, Dinh and Ho (2020) propose the “GroupLasso+AdaptiveGroupLasso” (GL+AGL) estimator for feature selection with neural networks, defined as

α~n:=argminα=(u,v,b1,b2,S,Q,q)1ni=1n(α,Xi,Yi)+ζnMn(α),\tilde{\alpha}_{n}:=\operatornamewithlimits{argmin}_{\alpha=(u,v,b_{1},b_{2},S,Q,q)}{~{}~{}\frac{1}{n}\sum_{i=1}^{n}{\ell(\alpha,X_{i},Y_{i})}+\zeta_{n}M_{n}(\alpha)},

where

Mn(α)=k=1ns1u^n[:,k]γu[:,k]+k=1nz1v^n[:,k]γv[:,k].M_{n}(\alpha)=\sum_{k=1}^{n_{s}}{\frac{1}{\|\hat{u}_{n}^{[:,k]}\|^{\gamma}}\|u^{[:,k]}\|}+\sum_{k=1}^{n_{z}}{\frac{1}{\|\hat{v}_{n}^{[:,k]}\|^{\gamma}}\|v^{[:,k]}\|}.

Here, we use the convention 0/0=10/0=1, γ>0\gamma>0, ζn\zeta_{n} is the regularizing constant and u^n\hat{u}_{n}, v^n\hat{v}_{n} denotes the uu and vv components of the GL estimate α^n\hat{\alpha}_{n}. As typical with adaptive lasso estimators, GL+AGL uses its base estimator to provide a rough data-dependent estimate to shrink groups of parameters with different regularization strengths. As nn grows, the weights for non-significant features get inflated (to infinity) while the weights for significant ones remain bounded (Zou, 2006). Dinh and Ho (2020) prove that GL+AGL is selection-consistent for irreducible shallow networks with hyperbolic tangent activation. In this work, we argue that the results can be extended to general analytic deep networks.

3 Consistent feature selection via GroupLasso+AdaptiveGroupLasso

The analysis of GL+AGL is decomposed into two steps. First, we establish that the Group Lasso provides a good proxy to estimate regularization strengths. Specifically, we show that

  • The uu-components of α^n\hat{\alpha}_{n} are bounded away from zero

  • The vv-components of α^n\hat{\alpha}_{n} converge to zero with a polynomial rate.

Second, we prove that a selection procedure based on the weights obtained from the first step can correctly select the set of significant variables given informative data with high probability.

One technical difficulty of the proof concerns with the geometry of the risk function around the set of risk minimizers

={α:R(α)=R(α)}\mathcal{H}^{*}=\{\alpha:R(\alpha)=R(\alpha^{*})\}

where R(α)R(\alpha) denotes the risk function R(α)=𝔼(X,Y)PX,Y[(fα(X)Y)2]R(\alpha)=\mathbb{E}_{(X,Y)\sim P_{X,Y}^{*}}[(f_{\alpha}(X)-Y)^{2}]. Since deep neural networks are highly unidentifiable, the set \mathcal{H}^{*} can be quite complex. For example:

  • (i)

    A simple rearrangement of the nodes in the same hidden layer leads to a new configuration that produces the same mapping as the generating network

  • (ii)

    If either all in-coming weights or all out-coming weights of a node is zero, changing the others also has no effect on the output

  • (iii)

    For hyperbolic tangent activation, multiplying all the in-coming and out-coming parameters associated with a node by -1 also lead to an equivalent configuration

These unidentifiability create a barrier in studying deep neural networks. Existing results in the field either are about equivalent graph transformation (Chen et al., 1993) or only hold in some generic sense (Fefferman and Markel, 1994) and thus are not sufficient to establish consistency. Traditional convergence analyses also often rely on local expansions of the risk function around isolated optima, which is no longer the case for neural networks since HH^{*} may contain subsets of high dimension (e.g., case (ii) above) and the Hessian matrix (and high-order derivatives) at an optimum might be singular.

In this section, we illustrate that these issues can be avoided with two adjustments. First, we show that while the behavior of a generic estimator may be erratic, the regularization effect of Group Lasso constraints it to converge to a subset 𝒦\mathcal{K}\subset\mathcal{H} of “well-behaved” optima. Second, instead of using local expansions, we employ Lojasewicz’s inequality for analytic functions to upper bound the distance from α^n\hat{\alpha}_{n} to \mathcal{H}^{*} by the excess risk. This removes the necessity of the regularity of the Hessian matrix and enlarges the class of networks for which selection consistency can be analyzed.

3.1 Characterizing the set of risk minimizers

Lemma 3.1.
  • (i)

    There exists c0>0c_{0}>0 such that uα[:,k]c0\|u_{\alpha}^{[:,k]}\|\geq c_{0} for all α\alpha\in\mathcal{H}^{*} and k=1,,nsk=1,\ldots,n_{s}.

  • (ii)

    For α\alpha\in\mathcal{H}^{*}, the vector ϕ(α)\phi(\alpha), obtained from α\alpha be setting its vv-components to zero, also belongs to \mathcal{H}^{*}.

Proof.

We first prove that α0\alpha_{0}\in\mathcal{H}^{*} if and only if fα0=fαf_{\alpha_{0}}=f_{\alpha^{*}}. Indeed, since fα(X)=EPX,Y[Y|X]f_{\alpha^{*}}(X)=E_{P^{*}_{X,Y}}[Y|X], we have

R(α)=ming𝔼YPX,Y[(g(X)Y)2]minα𝒲R(α)=R(α0)R(\alpha^{*})=\min_{g}{\mathbb{E}_{Y\sim P_{X,Y}^{*}}[(g(X)-Y)^{2}]}\leq\min_{\alpha\in\mathcal{W}}{R(\alpha)}=R(\alpha_{0})

with equality happens only if fα0=fαf_{\alpha_{0}}=f_{\alpha^{*}} a.s. on the support of pX(x)p_{X}(x). Since pX(x)p_{X}(x) is continuous and positive on its open domain 𝒳\mathcal{X} and the maps fαf_{\alpha} are analytic, we deduce that fα0=fαf_{\alpha_{0}}=f_{\alpha^{*}} everywhere.

(i) Assuming that no such c0c_{0} exists, since 𝒲\mathcal{W} is compact and fαf_{\alpha} is an analytic function in α\alpha and xx, we deduce that there exists α0\alpha_{0}\in\mathcal{H^{*}} and kk such that uα0[:,k]=0u_{\alpha_{0}}^{[:,k]}=0. This means fα0=fαf_{\alpha_{0}}=f_{\alpha^{*}} does not depend on significant input sks_{k}, which is a contradiction. (ii) Since α\alpha\in\mathcal{H^{*}}, we have fα(s,z)=fα(s,z)=fα(s,0)=fϕ(α)(s,z)f_{\alpha^{*}}(s,z)=f_{\alpha}(s,z)=f_{\alpha}(s,0)=f_{\phi(\alpha)}(s,z), which implies ϕ(α)\phi(\alpha)\in\mathcal{H^{*}}. ∎

Remarks.

Lemma 3.1 provides a way to study the behaviors of Group Lasso without a full characterization of the geometry of \mathcal{H}^{*} as in Dinh and Ho (2020). First, as long as d(α^n,)0d(\hat{\alpha}_{n},\mathcal{H}^{*})\to 0, it is straight forward that its uu-components are bounded away from zero (part (i)). Second, it shows that for all α\alpha\in\mathcal{H}^{*}, ϕ(α)\phi(\alpha) is a “better” hypothesis in terms of penalty while remaining an optimal hypothesis in terms of prediction. This enables us to prove that if we define

𝒦={α𝒲:fα=fαandvα=0},\mathcal{K}=\{\alpha\in\mathcal{W}:f_{\alpha}=f_{\alpha^{*}}~{}\text{and}~{}v_{\alpha}=0\},

and λn\lambda_{n} converges to zero slowly enough, the regularization term will force d(α^n,𝒦)0d(\hat{\alpha}_{n},\mathcal{K})\to 0. This helps establish that Group Lasso provides a good proxy to construct appropriate regularization strengths.

To provide a polynomial convergence rate of α^n\hat{\alpha}_{n}, we need the following Lemma.

Lemma 3.2.

There exist c2,ν>0c_{2},\nu>0 and such that R(β)R(α)c2d(β,)νR(\beta)-R(\alpha^{*})\geq c_{2}d(\beta,\mathcal{H}^{*})^{\nu} for all β𝒲\beta\in\mathcal{W}.

Proof.

We first note that since fα(x)f_{\alpha}(x) is analytic in both α\alpha and xx, the excess risk g(α)=R(α)R(α)g(\alpha)=R(\alpha)-R(\alpha^{*}) is also analytic in α\alpha. Thus ={α:R(α)=R(α)}\mathcal{H}^{*}=\{\alpha:R(\alpha)=R(\alpha^{*})\} is the zero level-set of the analytic function gg. By Lojasewicz’s inequality for algebraic varieties (Ji et al., 1992), there exist positive constants CC and ν\nu such that d(β,)νC|g(β)|β𝒲d(\beta,\mathcal{H}^{*})^{\nu}\leq C|g(\beta)|~{}~{}\forall\beta\in\mathcal{W}, which completes the proof. ∎

We note that for cases when \mathcal{H}^{*} is finite, Lemma 3.2 reduces to the standard Taylor’s inequality around a local optimum, with ν=2\nu=2 if the Hessian matrix at the optimum is non-singular (Dinh and Ho, 2020). When \mathcal{H}^{*} is a high-dimensional algebraic set, Lojasewicz’s inequality and Lemma 3.2 are more appropriate. For example, for g(x,y)=(xy)2g(x,y)=(x-y)^{2}, it is impossible to attain an inequality of the form

g(x,y)g(x0,x0)Cd((x,y),(x0,x0))νg(x^{\prime},y^{\prime})-g(x_{0},x_{0})\geq C\cdot d((x^{\prime},y^{\prime}),(x_{0},x_{0}))^{\nu}

in any neighborhood of any minimum (x0,x0)(x_{0},x_{0}), while it is straightforward that

g(x,y)g(x0,x0)d((x,y),Z)2whereZ={(x,x):x}.g(x^{\prime},y^{\prime})-g(x_{0},x_{0})\geq d((x^{\prime},y^{\prime}),Z)^{2}~{}~{}~{}\text{where}~{}~{}~{}Z=\{(x,x):x\in\mathbb{R}\}.

This approach provides a way to avoid dealing with model unidentfiability but requires some adaptation of the analysis to accommodate a new mode of convergence.

3.2 Convergence of Group Lasso

The two Lemmas in the previous section enables us to analyze the convergence of the Group Lasso estimate, in the sense that d(α^n,𝒦)0d(\hat{\alpha}_{n},\mathcal{K})\to 0. First, we define the empirical risk function

Rn(α)=1ni=1n(fα(Xi)Yi)2R_{n}(\alpha)=\frac{1}{n}\sum_{i=1}^{n}{(f_{\alpha}(X_{i})-Y_{i})^{2}}

and note that since the network of our framework is fixed, the learning problem is continuous and parametric, for which a standard generalization bound as follows can be obtained (proof in Appendix).

Lemma 3.3 (Generalization bound).

For any δ>0\delta>0, there exist c1(δ)>0c_{1}(\delta)>0 such that

|Rn(α)R(α)|c1lognn,α𝒲|R_{n}(\alpha)-R(\alpha)|\leq c_{1}\frac{\log n}{\sqrt{n}},\hskip 17.34518pt\forall\alpha\in\mathcal{W}

with probability at least 1δ1-\delta.

Combining Lemmas 3.2 and 3.3, we have

Theorem 3.4 (Convergence of Group Lasso).

For any δ>0\delta>0, there exist Cδ,C>0C_{\delta},C^{\prime}>0 and Nδ>0N_{\delta}>0 such that for all nNδn\geq N_{\delta},

d(α^n,)Cδ(λnν/(ν1)+lognn)1/νandv^n4c1lognλnn+Cd(α^n,)d(\hat{\alpha}_{n},\mathcal{H^{*}})\leq C_{\delta}\left(\lambda_{n}^{\nu/(\nu-1)}+\frac{\log n}{\sqrt{n}}\right)^{1/\nu}~{}~{}\text{and}~{}~{}\|\hat{v}_{n}\|\leq 4c_{1}\frac{\log n}{\lambda_{n}\sqrt{n}}+C^{\prime}~{}d(\hat{\alpha}_{n},\mathcal{H^{*}})

with probability at least 1δ1-\delta. Moreover, if λnn1/4\lambda_{n}\sim n^{-1/4}, then with probability at least 1δ1-\delta,

d(α^n,)C(lognn)14(ν1)andv^nC(lognn)14(ν1).d(\hat{\alpha}_{n},\mathcal{H^{*}})\leq C\left(\frac{\log n}{n}\right)^{\frac{1}{4(\nu-1)}}\hskip 17.34518pt\text{and}\hskip 17.34518pt\|\hat{v}_{n}\|\leq C\left(\frac{\log n}{n}\right)^{\frac{1}{4(\nu-1)}}.
Proof.

Let ϕ(α)\phi(\alpha) denote the weight vector obtained from α\alpha by setting the vv-components to zero. If we define βn=argminαα^nα\beta_{n}=\operatornamewithlimits{argmin}_{\alpha\in\mathcal{H}^{*}}{\|\hat{\alpha}_{n}-\alpha\|} then ϕ(βn)\phi(\beta_{n})\in\mathcal{H}^{*} and R(βn)=R(ϕ(βn))R(\beta_{n})=R(\phi(\beta_{n})).

Since L(α)L(\alpha) is a Lipschitz function, we have

c2d(α^n,)ν=c2βnα^nν\displaystyle c_{2}d(\hat{\alpha}_{n},\mathcal{H}^{*})^{\nu}=c_{2}\|\beta_{n}-\hat{\alpha}_{n}\|^{\nu} R(α^n)R(βn)\displaystyle\leq R(\hat{\alpha}_{n})-R(\beta_{n})
2c1lognn+λn(L(βn)L(α^n))2c1lognn+λnCβnα^n\displaystyle\leq 2c_{1}\frac{\log n}{\sqrt{n}}+\lambda_{n}\left(L(\beta_{n})-L(\hat{\alpha}_{n})\right)\leq 2c_{1}\frac{\log n}{\sqrt{n}}+\lambda_{n}C\|\beta_{n}-\hat{\alpha}_{n}\|

which implies (through Young’s inequality, details in Appendix) that

βnα^nνC1λnν/(ν1)+C2lognn.\|\beta_{n}-\hat{\alpha}_{n}\|^{\nu}\leq C_{1}\lambda_{n}^{\nu/(\nu-1)}+C_{2}\frac{\log n}{\sqrt{n}}.

Let KK denote the part of the regularization term without the vv-component. We note that KK is a Lipschitz function, K(ϕ(α))=K(α)K(\phi(\alpha))=K(\alpha) for all α\alpha, and R(ϕ(βn))=R(α^n)R(\phi(\beta_{n}))=R(\hat{\alpha}_{n}). Thus,

λnlv^n[:,l]\displaystyle\lambda_{n}\sum_{l}{\|\hat{v}^{[:,l]}_{n}\|} Rn(ϕ(βn))Rn(α^n)+λn[K(ϕ(βn))K(α^n)]\displaystyle\leq R_{n}(\phi(\beta_{n}))-R_{n}(\hat{\alpha}_{n})+\lambda_{n}[K(\phi(\beta_{n}))-K(\hat{\alpha}_{n})]
2c1lognn+R(ϕ(βn))R(α^n)+λn[K(βn)K(α^n)]\displaystyle\leq 2c_{1}\frac{\log n}{\sqrt{n}}+R(\phi(\beta_{n}))-R(\hat{\alpha}_{n})+\lambda_{n}[K(\beta_{n})-K(\hat{\alpha}_{n})]
2c1lognn+λnCβnα^n.\displaystyle\leq 2c_{1}\frac{\log n}{\sqrt{n}}+\lambda_{n}C\|\beta_{n}-\hat{\alpha}_{n}\|.

This completes the proof. ∎

Together, the two parts of Theorem 3.4 shows that the Group Lasso estimator converges to the set of “well-behaved” optimal hypotheses 𝒦={α𝒲:fα=fαandvα=0}\mathcal{K}=\{\alpha\in\mathcal{W}:f_{\alpha}=f_{\alpha^{*}}~{}\text{and}~{}v_{\alpha}=0\} (proof in Appendix)..

Corollary 3.5.

For any δ>0\delta>0, there exist Cδ>0C_{\delta}>0 and Nδ>0N_{\delta}>0 such that for all nNδn\geq N_{\delta},

[d(α^n,𝒦)4c1lognλnn+Cδ(λnν/(ν1)+lognn)1/ν]1δ.\mathbb{P}\left[d(\hat{\alpha}_{n},\mathcal{K})\leq 4c_{1}\frac{\log n}{\lambda_{n}\sqrt{n}}+C_{\delta}\left(\lambda_{n}^{\nu/(\nu-1)}+\frac{\log n}{\sqrt{n}}\right)^{1/\nu}\right]\geq 1-\delta.

3.3 Feature selection consistency of GroupLasso + AdaptiveGroupLasso

We are now ready to prove the main theorem of our paper.

Theorem 3.6 (Feature selection consistency of GL+AGL).

Let γ>0\gamma>0, ϵ>0\epsilon>0, λnn1/4\lambda_{n}\sim n^{-1/4}, and ζn=Ω(nγ/(4ν4)+ϵ)\zeta_{n}=\Omega(n^{-\gamma/(4\nu-4)+\epsilon}), then the GroupLasso+AdaptiveGroupLasso is feature selection consistent.

Proof.

Since d(α^n,)0d(\hat{\alpha}_{n},\mathcal{H^{*}})\to 0 (Theorem 3.4), we have minαu^n[:,k]u[:,k]0\min_{\alpha\in\mathcal{H}^{*}}\|\hat{u}_{n}^{[:,k]}-u^{[:,k]}\|\to 0 for all kk. By Lemma 3.1, we conclude that u^n[:,k]\hat{u}_{n}^{[:,k]} is bounded away from zero as nn\to\infty. Thus,

Mn(α)=i=1ns1u^nγu[:,k]<andM_{n}(\alpha^{*})=\sum_{i=1}^{n_{s}}\frac{1}{\|\hat{u}_{n}\|^{\gamma}}\|u^{*[:,k]}\|<\infty\hskip 17.34518pt\text{and}
c2d(α~n,)νR(α~n)R(α)2c1lognn+ζn(Mn(α)Mn(α^n))2c1lognn+ζnMn(α)c_{2}d(\tilde{\alpha}_{n},\mathcal{H}^{*})^{\nu}\leq R(\tilde{\alpha}_{n})-R(\alpha^{*})\leq 2c_{1}\frac{\log n}{\sqrt{n}}+\zeta_{n}\left(M_{n}(\alpha^{*})-M_{n}(\hat{\alpha}_{n})\right)\leq 2c_{1}\frac{\log n}{\sqrt{n}}+\zeta_{n}M_{n}(\alpha^{*})

which shows that d(α~n,)0d(\tilde{\alpha}_{n},\mathcal{H^{*}})\to 0. Thus u~n[:,k]\tilde{u}_{n}^{[:,k]} is also bounded away from zero for nn large enough.

We now assume that v~n[:,k]0\tilde{v}_{n}^{[:,k]}\neq 0 for some kk and define a new weight configuration gng_{n} obtained from α~n\tilde{\alpha}_{n} by setting the v[:,k]v^{[:,k]} component to 0. By definition of the estimator α~n\tilde{\alpha}_{n}, we have

Rn(α~n)+ζn1v^n[:,k]γv~n[:,k]Rn(gn).R_{n}(\tilde{\alpha}_{n})+\zeta_{n}\frac{1}{\|\hat{v}_{n}^{[:,k]}\|^{\gamma}}\|\tilde{v}_{n}^{[:,k]}\|\leq R_{n}(g_{n}).

By the (probabilistic) Lipschitzness of the empirical risk (proof in Appendix), there exists MδM_{\delta} s.t.

ζn1v^n[:,k]γv~n[:,k]Rn(gn)Rn(α~n)Mδgnα~n=Mδv~n[:,k]\zeta_{n}\frac{1}{\|\hat{v}_{n}^{[:,k]}\|^{\gamma}}\|\tilde{v}_{n}^{[:,k]}\|\leq R_{n}(g_{n})-R_{n}(\tilde{\alpha}_{n})\leq M_{\delta}\|g_{n}-\tilde{\alpha}_{n}\|=M_{\delta}\|\tilde{v}_{n}^{[:,k]}\|

with probability at least 1δ1-\delta. Since v~n[:,k]0\tilde{v}_{n}^{[:,k]}\neq 0, we deduce that ζn1v^n[:,k]γMδ\zeta_{n}\frac{1}{\|\hat{v}_{n}^{[:,k]}\|^{\gamma}}\leq M_{\delta}. This contradicts Theorem 3.4, which proves that for nn large enough

ζn1v^n[:,k]γCδγζn(nlogn)γ4(ν1)2Mδ\zeta_{n}\frac{1}{\|\hat{v}_{n}^{[:,k]}\|^{\gamma}}\geq C_{\delta}^{-\gamma}~{}\zeta_{n}\left(\frac{n}{\log n}\right)^{\frac{\gamma}{4(\nu-1)}}\geq 2M_{\delta}

with probability at least 1δ1-\delta. This completes the proof. ∎

Remark.

One notable aspect of Theorem 3.6 is the absence of regularity conditions on the correlation of the inputs often used in lasso-type analyses, such as Irrepresentable Conditions (Meinshausen and Bühlmann, 2006) and (Zhao and Yu, 2006), restricted isometry property (Candes and Tao, 2005), restricted eigenvalue conditions (Bickel et al., 2009; Meinshausen and Yu, 2009) or sparse Riesz condition (Zhang and Huang, 2008). We recall that by using different regularization strengths for individual parameters, adaptive lasso estimators often require less strict conditions for selection consistency than standard lasso. For linear models, the classical adaptive lasso only assumes that the limiting design matrix is positive definite, i.e., there is no perfect correlation among the inputs) (Zou, 2006). In our framework, this corresponds to the assumption that the density of XX is positive on its open domain (Assumption 2.1), which plays an essential role in the proof of Lemma 3.1.

3.4 Simulations

Refer to caption
Figure 1: The false positive rate and false negative rate of GL (left panel) and GL+AGL (right panel), respectively. Note that the true positive rate of GL and true negative rate of GL+AGL are 100% for all but one run and are not shown in the figure.

To further illustrate the theoretical findings of the paper, we use both synthetic and real data to investigate algorithmic behaviors of GL and GL+AGL. 111The code is available at https://github.com/vucdinh/alg-net. The simulations, implemented in Pytorch, focus on single-output deep feed-forward networks with three hidden layers of constant width. In these experiments, regularizing constants are chosen from a course grid {0.001,0.01,0.05,0.1,0.5,1,2}\{0.001,0.01,0.05,0.1,0.5,1,2\} with γ=2\gamma=2 using average test errors from random train-test splits of the corresponding dataset. The algorithms are trained over 20000 epochs using proximal gradient descent, which allows us to identify the exact support of estimators without having to use a cut-off value for selection.

In the first experiment, we consider a network with three hidden layers of 20 nodes. The input consists of 5050 features, 10 of which are significant while the others are rendered insignificant by setting the corresponding weights to zero. We generate 100100 datasets of size n=5000n=5000 from the generic model Y=fα(X)+ϵY=f_{\alpha^{*}}(X)+\epsilon where ϵ𝒩(0,1)\epsilon\sim\mathcal{N}(0,1) and non-zero weights of α\alpha^{*} are sampled independently from 𝒩(0,1)\mathcal{N}(0,1). We perform GL and GL+AGL on each simulated dataset with regularizing constants chosen using average test errors from three random three-fold train-test splits. We observe that overall, GL+AGL have a superior performance, selecting the correct support in 6363 out of 100100 runs, while GL cannot identify the support in any run. Except for one pathological case when both GL and GL+AGL choose a constant model, GL always selects the correct significant inputs but fail to de-select the insignificant ones (Figure 1, left panel) while GL+AGL always performs well with the insignificant inputs but sometimes over-shrinks the significant ones (Figure 1, right panel).

Next, we apply the methods to the Boston housing dataset 222http://lib.stat.cmu.edu/datasets/boston. This dataset consists of 506 observations of house prices and 13 predictors. To analyze the data, we consider a network with three hidden layers of 10 nodes. GL and GL+AGL are then performed on this dataset using average test errors from 20 random train-test splits (with the size of the test sets being 25%25\% of the original dataset). GL identifies all 13 predictors as important, while GL+AGL only selects 11 of them. To further investigate the robustness of the results, we follow the approach of Lemhadri et al. (2019) to add 13 random Gaussian noise predictors to the original dataset for analysis. 100100 such datasets are created to compare the performance of GL against GL+AGL using the same experimental setting as above. The results are presented in Figure 2, for which we observe that GL struggles to distinguish the random noises from the correct predictors. We note that Lemhadri et al. (2019) identifies 11 of the original predictors along with 2 random predictors as the optimal set of features for prediction, which is consistent with the performance of GL+AGL.

Refer to caption
Figure 2: Performance on the Boston Housing dataset: The frequency of being selected by GL (blue) and GL+AGL (red) of each of the 26 predictors. Note that the predictors 142614-26 are random Gaussian noise predictors while 1131-13 are from the original dataset.

4 Conclusions and Discussions

In this work, we prove that GL+AGL is feature-selection-consistent for all analytic deep networks that interact with inputs through a finite set of linear units. Both theoretical and simulation results of the work advocate the use of GL+AGL over the popular Group Lasso for feature selection.

4.1 Comparison to related works

To the best of our knowledge, this is the first work that establishes selection consistency for deep networks. This is in contrast to Dinh and Ho (2020), Liang et al. (2018) and Ye and Sun (2018), which only provide results for shallow networks with one hidden layer, or Polson and Ročková (2018), Feng and Simon (2017), Liu (2019), which focus on posterior concentration, prediction consistency, parameter-estimation consistency and convergence of feature importance. We note that for classical linear model with lasso estimate, it is known that (see Section 2 of Zou (2006) and the discussion therein) the lasso estimator is parameter-estimation consistent as long as the regularizing parameter λn0\lambda_{n}\to 0 (Zou (2006), Lemmas 2 and 3), but is not feature-selection consistent for λnn1/2\lambda_{n}\sim n^{-1/2} (Zou (2006), Proposition 1) or for all choices of λn\lambda_{n} if some necessary condition on the covariance matrix is not satisfied (Zou (2006), Theorem 1). For both linear model and neural network, parameter-estimation consistency directly implies prediction consistency and convergence of feature importance. In general, there’s no known trivial way to extending these approaches to obtain feature-selection-consistency.

Moreover, existing works usually assume that the network used for training has exactly the same size as the minimal network that generates the data. This is assumption is either made explicitly (as in Dinh and Ho (2020)) or implicitly implied by the regularity of the Hessian matrix at the optima (Assumption 5 in Ye and Sun (2018) and Condition 1 in Feng and Simon (2017)). We note that these latter two conditions cannot be satisfied if the size of the training network is not “correct” (for example, when data is generated by a one-hidden-layer network with 5 hidden nodes, but the training network is one with 10 nodes). The framework of our paper does not have this restriction. Finally, since our paper focus on model interpretability, the framework has been constructed in such a way that all assumptions are minimal/verifiable. This is in contrast to many previous results. For example, Ye and Sun (2018) takes Assumption 6 (which is difficult to check) as given while we can avoid this Assumption using Lemma 3.2.

4.2 Future works

There are many avenues for future directions. First, while simulations seem to hint that Group Lasso may not be optimal for feature selection with neural networks, a rigorous answer to this hypothesis requires a deeper understanding of the behavior of the estimator that is out of the scope of this paper. Second, the analyses in this work rely on the fact that the network model is analytic, which puts some restrictions on the type of activation function that can be used. While many tools to study non-analytic networks (as well as other learning settings, e.g., for classification, for regression with a different loss function, or networks that does not have a strict structure of layers) are already in existence, such extensions of the results require non-trivial efforts.

Throughout the paper, we assume that the network is fixed (pp) when the sample size (nn) increases. Although this assumption is reasonable in many application settings, in some contexts, for example for the task of identifying genes that increase the risk of a type of cancer, ones would be interested in the case of pnp\gg n. There have been existing results of this type for neural networks from the prediction aspect of the problem Feng and Simon (2017); Farrell et al. (2018) and it would be of general interest how analyses of selection consistency apply in those cases. Finally, since the main interest of this work is theoretical, many aspects of the performance of the GL+AGL across different experimental settings and types of networks are left as subjects of future work.

Acknowlegments

LSTH was supported by startup funds from Dalhousie University, the Canada Research Chairs program, and the Natural Sciences and Engineering Research Council of Canada (NSERC) Discovery Grant RGPIN-2018-05447. VD was supported by a startup fund from University of Delaware and National Science Foundation grant DMS-1951474.

Broader Impact

Deep learning has transformed modern science in an unprecedented manner and created a new force for technological developments. However, its black-box nature and the lacking of theoretical justifications have hindered its applications in fields where correct interpretations play an essential role. In many applications, a linear model with a justified confidence interval and a rigorous feature selection procedure is much more favored than a deep learning system that cannot be interpreted. Usage of deep learning in a process that requires transparency such as judicial and public decisions is still completely out of the question.

To the best of our knowledge, this is the first work that establishes feature selection consistency, an important cornerstone of interpretable statistical inference, for deep learning. The results of this work will greatly extend the set of problems to which statistical inference with deep learning can be applied. Medical sciences, public health decisions, and various fields of engineering, which depend upon well-founded estimates of uncertainty, fall naturally on the domain the work tries to explore. Researchers from these fields and the public alike may benefit from such a development and no one is put at disadvantage from this research.

By trying to select a parsimonious and transparent model out of an over-parametrized deep learning system, the approach of this work further provides a systematic way to detect and reduce bias in machine learning analysis. The analytical tools and the theoretical framework derived in this work may also be of independent interest in statistics, machine learning, and other fields of applied sciences.

References

  • Ainsworth et al. (2018) Samuel Ainsworth, Nicholas Foti, Adrian KC Lee, and Emily Fox. Interpretable VAEs for nonlinear group factor analysis. arXiv preprint arXiv:1802.06765, 2018.
  • Bickel et al. (2009) Peter J Bickel, Yaácov Ritov, and Alexandre B Tsybakov. Simultaneous analysis of Lasso and Dantzig selector. The Annals of Statistics, 37(4):1705–1732, 2009.
  • Candes and Tao (2005) Emmanuel J Candes and Terence Tao. Decoding by linear programming. IEEE Transactions on Information Theory, 51(12):4203–4215, 2005.
  • Chen et al. (1993) An Mei Chen, Haw-minn Lu, and Robert Hecht-Nielsen. On the geometry of feedforward neural network error surfaces. Neural computation, 5(6):910–927, 1993.
  • Ching et al. (2018) Travers Ching, Daniel S Himmelstein, Brett K Beaulieu-Jones, Alexandr A Kalinin, Brian T Do, Gregory P Way, Enrico Ferrero, Paul-Michael Agapow, Michael Zietz, and Michael M Hoffman. Opportunities and obstacles for deep learning in biology and medicine. Journal of The Royal Society Interface, 15(141):20170387, 2018.
  • Dinh and Ho (2020) Vu Dinh and Lam Ho. Consistent feature selection for neural networks via Adaptive Group Lasso. arXiv preprint arXiv:2006.00334, 2020.
  • Fallahgoul et al. (2019) Hasan Fallahgoul, Vincentius Franstianto, and Gregoire Loeper. Towards explaining the ReLU feed-forward network. Available at SSRN, 2019.
  • Farrell et al. (2018) Max H Farrell, Tengyuan Liang, and Sanjog Misra. Deep neural networks for estimation and inference. arXiv preprint arXiv:1809.09953, 2018.
  • Fefferman and Markel (1994) Charles Fefferman and Scott Markel. Recovering a feed-forward net from its output. In Advances in Neural Information Processing Systems, pages 335–342, 1994.
  • Feng and Simon (2017) Jean Feng and Noah Simon. Sparse-input neural networks for high-dimensional nonparametric regression and classification. arXiv preprint arXiv:1711.07592, 2017.
  • Horel and Giesecke (2019) Enguerrand Horel and Kay Giesecke. Towards Explainable AI: Significance tests for neural networks. arXiv preprint arXiv:1902.06021, 2019.
  • Ibrahim et al. (2014) Rania Ibrahim, Noha A Yousri, Mohamed A Ismail, and Nagwa M El-Makky. Multi-level gene/MiRNA feature selection using deep belief nets and active learning. In 2014 36th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, pages 3957–3960. IEEE, 2014.
  • Ji et al. (1992) Shanyu Ji, János Kollár, and Bernard Shiffman. A global Lojasiewicz inequality for algebraic varieties. Transactions of the American Mathematical Society, 329(2):813–818, 1992.
  • Lemhadri et al. (2019) Ismael Lemhadri, Feng Ruan, and Robert Tibshirani. A neural network with feature sparsity. arXiv preprint arXiv:1907.12207, 2019.
  • Li et al. (2016) Yifeng Li, Chih-Yu Chen, and Wyeth W Wasserman. Deep feature selection: theory and application to identify enhancers and promoters. Journal of Computational Biology, 23(5):322–336, 2016.
  • Liang et al. (2018) Faming Liang, Qizhai Li, and Lei Zhou. Bayesian neural networks for selection of drug sensitive genes. Journal of the American Statistical Association, 113(523):955–972, 2018.
  • Liu (2019) Jeremiah Zhe Liu. Variable selection with rigorous uncertainty quantification using deep bayesian neural networks: Posterior concentration and bernstein-von mises phenomenon. arXiv preprint arXiv:1912.01189, 2019.
  • Lu et al. (2018) Yang Lu, Yingying Fan, Jinchi Lv, and William Stafford Noble. DeepPINK: reproducible feature selection in deep neural networks. In Advances in Neural Information Processing Systems, pages 8676–8686, 2018.
  • Lundberg and Lee (2017) Scott M Lundberg and Su-In Lee. A unified approach to interpreting model predictions. In Advances in Neural Information Processing Systems, pages 4765–4774, 2017.
  • Meinshausen and Bühlmann (2006) Nicolai Meinshausen and Peter Bühlmann. High-dimensional graphs and variable selection with the Lasso. The Annals of Statistics, 34(3):1436–1462, 2006.
  • Meinshausen and Yu (2009) Nicolai Meinshausen and Bin Yu. Lasso-type recovery of sparse representations for high-dimensional data. The Annals of Statistics, 37(1):246–270, 2009.
  • Nezhad et al. (2016) Milad Zafar Nezhad, Dongxiao Zhu, Xiangrui Li, Kai Yang, and Phillip Levy. SAFS: A deep feature selection approach for precision medicine. In 2016 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), pages 501–506. IEEE, 2016.
  • Polson and Ročková (2018) Nicholas G Polson and Veronika Ročková. Posterior concentration for sparse deep learning. In Advances in Neural Information Processing Systems, pages 930–941, 2018.
  • Ribeiro et al. (2016) Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin. “Why should I trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data mining, pages 1135–1144, 2016.
  • Scardapane et al. (2017) Simone Scardapane, Danilo Comminiello, Amir Hussain, and Aurelio Uncini. Group sparse regularization for deep neural networks. Neurocomputing, 241:81–89, 2017.
  • Shen et al. (2019) Xiaoxi Shen, Chang Jiang, Lyudmila Sakhanenko, and Qing Lu. Asymptotic properties of neural network sieve estimators. arXiv preprint arXiv:1906.00875, 2019.
  • Shrikumar et al. (2017) Avanti Shrikumar, Peyton Greenside, and Anshul Kundaje. Learning important features through propagating activation differences. In Proceedings of the 34th International Conference on Machine Learning-Volume 70, pages 3145–3153, 2017.
  • Simonyan et al. (2013) Karen Simonyan, Andrea Vedaldi, and Andrew Zisserman. Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv preprint arXiv:1312.6034, 2013.
  • Taherkhani et al. (2018) Aboozar Taherkhani, Georgina Cosma, and T Martin McGinnity. Deep-FS: A feature selection algorithm for Deep Boltzmann Machines. Neurocomputing, 322:22–37, 2018.
  • Tank et al. (2018) Alex Tank, Ian Covert, Nicholas Foti, Ali Shojaie, and Emily Fox. Neural Granger causality for nonlinear time series. arXiv preprint arXiv:1802.05842, 2018.
  • Wang and Leng (2008) Hansheng Wang and Chenlei Leng. A note on Adaptive Group Lasso. Computational Statistics & Data Analysis, 52(12):5277–5286, 2008.
  • Ye and Sun (2018) Mao Ye and Yan Sun. Variable selection via penalized neural network: a drop-out-one loss approach. In International Conference on Machine Learning, pages 5620–5629, 2018.
  • Zhang et al. (2018) Cheng Zhang, Vu Dinh, and Frederick A Matsen IV. Non-bifurcating phylogenetic tree inference via the Adaptive Lasso. Journal of the American Statistical Association. arXiv preprint arXiv:1805.11073, 2018.
  • Zhang and Huang (2008) Cun-Hui Zhang and Jian Huang. The sparsity and bias of the Lasso selection in high-dimensional linear regression. The Annals of Statistics, 36(4):1567–1594, 2008.
  • Zhang et al. (2019) Huaqing Zhang, Jian Wang, Zhanquan Sun, Jacek M Zurada, and Nikhil R Pal. Feature selection for neural networks using Group Lasso regularization. IEEE Transactions on Knowledge and Data Engineering, 2019.
  • Zhao et al. (2015) Lei Zhao, Qinghua Hu, and Wenwu Wang. Heterogeneous feature selection with multi-modal deep neural networks and Sparse Group Lasso. IEEE Transactions on Multimedia, 17(11):1936–1948, 2015.
  • Zhao and Yu (2006) Peng Zhao and Bin Yu. On model selection consistency of Lasso. Journal of Machine learning Research, 7(Nov):2541–2563, 2006.
  • Zou (2006) Hui Zou. The Adaptive Lasso and its oracle properties. Journal of the American Statistical Association, 101(476):1418–1429, 2006.

5 Appendix

5.1 Details of the proof of Theorem 3.4

In the first part of the proof of Theorem 3.4, we established that

c2βnα^nν2c1lognn+λnCβnα^n.\displaystyle c_{2}\|\beta_{n}-\hat{\alpha}_{n}\|^{\nu}\leq 2c_{1}\frac{\log n}{\sqrt{n}}+\lambda_{n}C\|\beta_{n}-\hat{\alpha}_{n}\|.

Using Young’s inequality, we have

λnCβnα^n\displaystyle\lambda_{n}C\|\beta_{n}-\hat{\alpha}_{n}\| 1ν((c2ν)1/ν2βnα^n)ν+ν1ν(2C(c2ν)1/νλn)ν/(ν1)\displaystyle\leq\frac{1}{\nu}\left(\frac{(c_{2}\nu)^{1/\nu}}{2}\|\beta_{n}-\hat{\alpha}_{n}\|\right)^{\nu}+\frac{\nu-1}{\nu}\left(\frac{2C}{(c_{2}\nu)^{1/\nu}}\lambda_{n}\right)^{\nu/(\nu-1)}
=c22βnα^nν+2(ν1)Cν/(ν1)ν(c2ν)1/(ν1)λν/(ν1).\displaystyle=\frac{c_{2}}{2}\|\beta_{n}-\hat{\alpha}_{n}\|^{\nu}+\frac{2(\nu-1)C^{\nu/(\nu-1)}}{\nu(c_{2}\nu)^{1/(\nu-1)}}\lambda^{\nu/(\nu-1)}.

Combining the two estimates, we have

c22βnα^nν2c1lognn+2(ν1)Cν/(ν1)ν(c2ν)1/(ν1)λν/(ν1).\frac{c_{2}}{2}\|\beta_{n}-\hat{\alpha}_{n}\|^{\nu}\leq 2c_{1}\frac{\log n}{\sqrt{n}}+\frac{2(\nu-1)C^{\nu/(\nu-1)}}{\nu(c_{2}\nu)^{1/(\nu-1)}}\lambda^{\nu/(\nu-1)}.

5.2 Proof of Corollary 3.5

We note that ϕ(βn)𝒦\phi(\beta_{n})\in\mathcal{K}. Thus,

minα𝒦α^nαα^nϕ(βn)\displaystyle\min_{\alpha\in\mathcal{K}}\|\hat{\alpha}_{n}-\alpha\|\leq\|\hat{\alpha}_{n}-\phi(\beta_{n})\| α^nβn+βnϕ(βn)\displaystyle\leq\|\hat{\alpha}_{n}-\beta_{n}\|+\|\beta_{n}-\phi(\beta_{n})\|
α^nβn+vβn\displaystyle\leq\|\hat{\alpha}_{n}-\beta_{n}\|+\|v_{\beta_{n}}\|
α^nβn+v^n+Cα^nβn\displaystyle\leq\|\hat{\alpha}_{n}-\beta_{n}\|+\|\hat{v}_{n}\|+C\|\hat{\alpha}_{n}-\beta_{n}\|

and the bound can be obtained using the results of Theorem 3.4.

5.3 Probabilistic Lipschitzness of the empirical risk

Since both 𝒲\mathcal{W} and 𝒳\mathcal{X} are bounded and fαf_{\alpha} is analytic, there exist C1,C2>0C_{1},C_{2}>0 such that

|αfα(x)|C1and|fα(x)|C2α𝒲,x𝒳.|\nabla_{\alpha}f_{\alpha}(x)|\leq C_{1}\hskip 17.34518pt\text{and}\hskip 17.34518pt|f_{\alpha}(x)|\leq C_{2}\hskip 17.34518pt\forall\alpha\in\mathcal{W},x\in\mathcal{X}.

Therefore,

|R(α)R(β)|\displaystyle|R(\alpha)-R(\beta)| =|𝔼[(Yfα(X))2(Yfα(X))2]|\displaystyle=\left|\mathbb{E}\left[(Y-f_{\alpha}(X))^{2}-(Y-f_{\alpha}(X))^{2}\right]\right|
𝔼|(fα(X)fβ(X))(2Yfα(X)fβ(X))|\displaystyle\leq\mathbb{E}\left|(f_{\alpha}(X)-f_{\beta}(X))(2Y-f_{\alpha}(X)-f_{\beta}(X))\right|
C1αβ𝔼|2Yfα(X)fβ(X)|\displaystyle\leq C_{1}\|\alpha-\beta\|\cdot\mathbb{E}\left|2Y-f_{\alpha}(X)-f_{\beta}(X)\right|
C1αβ(2𝔼|Yfα(X)|+𝔼|fα(X)+fβ(X)2fα(X)|)\displaystyle\leq C_{1}\|\alpha-\beta\|\cdot\left(2\mathbb{E}\left|Y-f_{\alpha^{*}}(X)\right|+\mathbb{E}\left|f_{\alpha}(X)+f_{\beta}(X)-2f_{\alpha^{*}}(X)\right|\right)
C1αβ(2σ+4C2).\displaystyle\leq C_{1}\|\alpha-\beta\|\left(2\sigma+4C_{2}\right).

Similarly,

|Rn(α)Rn(β)|\displaystyle|R_{n}(\alpha)-R_{n}(\beta)| C1αβ(4C2+2ni=1n|Yifα(Xi)|)\displaystyle\leq C_{1}\|\alpha-\beta\|\left(4C_{2}+\frac{2}{n}\sum_{i=1}^{n}{|Y_{i}-f_{\alpha^{*}}(X_{i})|}\right)
=C1αβ(4C2+2ni=1n|ϵi|).\displaystyle=C_{1}\|\alpha-\beta\|\left(4C_{2}+\frac{2}{n}\sum_{i=1}^{n}{|\epsilon_{i}|}\right).

Thus, for all Mδ>4C1C2M_{\delta}>4C_{1}C_{2},

P[|Rn(α)Rn(β)|Mδαβα,β𝒲]\displaystyle\hskip 17.34518ptP\left[|R_{n}(\alpha)-R_{n}(\beta)|\leq M_{\delta}\|\alpha-\beta\|~{}\forall\alpha,\beta\in\mathcal{W}\right]
P(1ni=1n|ϵi|Mδ2C12C2)\displaystyle\leq P\left(\frac{1}{n}\sum_{i=1}^{n}{|\epsilon_{i}|}\leq\frac{M_{\delta}}{2C_{1}}-2C_{2}\right)
=1P(1ni=1n|ϵi|Mδ2C12C2)\displaystyle=1-P\left(\frac{1}{n}\sum_{i=1}^{n}{|\epsilon_{i}|}\geq\frac{M_{\delta}}{2C_{1}}-2C_{2}\right)
1E|ϵ1|Mδ2C12C2.\displaystyle\leq 1-\frac{E|\epsilon_{1}|}{\frac{M_{\delta}}{2C_{1}}-2C_{2}}.

5.4 Proof of Lemma 3.3

The proof of this Lemma is similar to that of Lemma 4.2 in Dinh and Ho [2020]. Since the network of our framework is fixed, a standard generalization bound (with constants depending on the dimension of the weight space 𝒲\mathcal{W}) can be obtained. For completeness, we include the proof of Lemma 4.2 in Dinh and Ho [2020] below.

Note that nRn(α)/σe2nR_{n}(\alpha)/\sigma^{2}_{e} follows a non-central chi-squared distribution with nn degrees of freedom and fα(X)f_{\alpha}(X) is bounded. By applying Theorem 7 in Zhang and Zhou (2018) 333Zhang, Anru and Yuchen Zhou. ”On the non-asymptotic and sharp lower tail bounds of random variables.” arXiv preprint arXiv:1810.09006 (2018)., we have

[|Rn(α)R(α)|>t/2]\displaystyle\mathbb{P}\left[|R_{n}(\alpha)-R(\alpha)|>t/2\right]
2exp(C1n2t2n+2i=1n[fα(X)fα(X)]2)\displaystyle\leq 2\exp\left(-\frac{C_{1}n^{2}t^{2}}{n+2\sum_{i=1}^{n}{[f_{\alpha}(X)-f_{\alpha^{*}}(X)]^{2}}}\right)
2exp(C2nt2),\displaystyle\leq 2\exp(-C_{2}nt^{2}),

for all

0<t<n+i=1n[fα(X)fα(X)]2n.0<t<\frac{n+\sum_{i=1}^{n}{[f_{\alpha}(X)-f_{\alpha^{*}}(X)]^{2}}}{n}.

We define the events

𝒜(α,t)={|Rn(α)R(α)|>t/2},\mathcal{A}(\alpha,t)=\{|R_{n}(\alpha)-R(\alpha)|>t/2\},
(α,t)={\displaystyle\mathcal{B}(\alpha,t)=\{ α𝒲such that\displaystyle\exists\alpha^{\prime}\in\mathcal{W}~{}\text{such that}~{}
ααt4Mδand|Rn(α)R(α)|>t},\displaystyle\|\alpha^{\prime}-\alpha\|\leq\frac{t}{4M_{\delta}}~{}\text{and}~{}|R_{n}(\alpha^{\prime})-R(\alpha^{\prime})|>t\},

and

𝒞={|Rn(α)Rn(α)|Mδαα,α,α𝒲}.\mathcal{C}=\{|R_{n}(\alpha)-R_{n}(\alpha^{\prime})|\leq M_{\delta}\|\alpha-\alpha^{\prime}\|,\forall\alpha,\alpha^{\prime}\in\mathcal{W}\}.

Here, MδM_{\delta} is defined in Lemma 3.5. By Lemma 3.5, (α,t)𝒞𝒜(α,t)\mathcal{B}(\alpha,t)\cap\mathcal{C}\subset\mathcal{A}(\alpha,t) and P(𝒞)1δP(\mathcal{C})\geq 1-\delta.

Let m=dim(𝒲)m=dim(\mathcal{W}), there exist C3(m)1C_{3}(m)\geq 1 and a finite set 𝒲\mathcal{H}\subset\mathcal{W} such that

𝒲α𝒱(α,ϵ)and||C3/ϵm\mathcal{W}\subset\bigcup_{\alpha\in\mathcal{H}}{\mathcal{V}(\alpha,\epsilon)}\hskip 17.34518pt\text{and}\hskip 17.34518pt|\mathcal{H}|\leq C_{3}/\epsilon^{m}

where ϵ=t/(4Mδ)\epsilon=t/(4M_{\delta}), 𝒱(α,ϵ)\mathcal{V}(\alpha,\epsilon) denotes the open ball centered at α\alpha with radius ϵ\epsilon, and |||\mathcal{H}| denotes the cardinality of \mathcal{H}. By a union bound, we have

[α:|Rn(α)R(α)|>t/2]2C3(4Mδ)mtmeC2nt2.\mathbb{P}\left[\exists\alpha\in\mathcal{H}:\left|R_{n}(\alpha)-R(\alpha)\right|>t/2\right]\leq 2\frac{C_{3}(4M_{\delta})^{m}}{t^{m}}e^{-C_{2}nt^{2}}.

Using the fact that (α,t)𝒞𝒜(α,t),α\mathcal{B}(\alpha,t)\cap\mathcal{C}\subset\mathcal{A}(\alpha,t),~{}\forall\alpha\in\mathcal{H}, we deduce

[{α𝒲:|Rn(α)R(α)|>t}𝒞]C4tmeC2nt2.\mathbb{P}\left[\{\exists\alpha\in\mathcal{W}:\left|R_{n}(\alpha)-R(\alpha)\right|>t\right\}\cap\mathcal{C}]\leq C_{4}t^{-m}e^{-C_{2}nt^{2}}.

Hence,

[{α𝒲:|Rn(α)R(α)|>t}]C4tmeC2nt2+δ.\mathbb{P}\left[\{\exists\alpha\in\mathcal{W}:\left|R_{n}(\alpha)-R(\alpha)\right|>t\right\}]\leq C_{4}t^{-m}e^{-C_{2}nt^{2}}+\delta.

To complete the proof, we chose tt in such a way that C4tmeC2nt2δC_{4}t^{-m}e^{-C_{2}nt^{2}}~{}\leq~{}\delta. This can be done by choosing t=𝒪(logn/n)t=\mathcal{O}(\log n/\sqrt{n}).