This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Group Spike and Slab Variational Bayes

Michael Komodromos , Kolyan Ray, Marina Evangelou and Sarah Filippi  
Department of Mathematics, Imperial College London
M.K. gratefully acknowledge EPSRC’s StatML CDT, Imperial’s CRUK center and Imperial’s Experimental Cancer Medicine center.
For the purpose of open access, the author has applied a Creative Commons Attribution (CC BY) licence to any Author Accepted Manuscript version arising.
Abstract

We introduce Group Spike-and-slab Variational Bayes (GSVB), a scalable method for group sparse regression. A fast co-ordinate ascent variational inference (CAVI) algorithm is developed for several common model families including Gaussian, Binomial and Poisson. Theoretical guarantees for our proposed approach are provided by deriving contraction rates for the variational posterior in grouped linear regression. Through extensive numerical studies, we demonstrate that GSVB provides state-of-the-art performance, offering a computationally inexpensive substitute to MCMC, whilst performing comparably or better than existing MAP methods. Additionally, we analyze three real world datasets wherein we highlight the practical utility of our method, demonstrating that GSVB provides parsimonious models with excellent predictive performance, variable selection and uncertainty quantification.


Keywords: High-dimensional regression, Group sparsity, Generalized Linear Models, Sparse priors, Bayesian Methods

1 Introduction

Group structures arise in various applications, such as genetics (Wang et al., 2007; Breheny and Huang, 2009), imaging (Lee and Cao, 2021), multi-factor analysis of variance (Meier et al., 2008), non-parametric regression (Huang et al., 2010), and multi-task learning, among others. In these settings, pp-dimensional feature vectors, xi=(xi1,,xip)px_{i}=(x_{i1},\dots,x_{ip})^{\top}\in\mathbb{R}^{p}, for i=1,,ni=1,\dots,n observations can be partitioned into groups of features. Formally this means that we can construct sub-vectors, xGk={xj:jGk}x_{G_{k}}=\{x_{j}:j\in G_{k}\} for k=1,,Mk=1,\dots,M, where the groups Gk={Gk,1,,Gk,mk}G_{k}=\{G_{k,1},\dots,G_{k,m_{k}}\} are disjoint sets of indices satisfying k=1MGk={1,,p}\bigcup_{k=1}^{M}G_{k}=\{1,\dots,p\}. Typically, these group structures are known beforehand, for example in genetics where biological pathways (gene sets) are known, or they are constructed artificially, for example, through a basis expansion in non-parametric additive models.

In the regression setting incorporating these group structures is crucial, as disregarding them can result in sub-optimal models (Huang and Zhang, 2010; Lounici et al., 2011). In this manuscript, we focus on the general linear regression model (GLM) where, for each observation i=1,,ni=1,\dots,n, the response, YiY_{i} can be modelled by

𝔼[Yi|xi,β]=f(k=1Mxi,GkβGk),i=1,,n\mathbb{E}[Y_{i}|x_{i},\beta]=f\left(\sum_{k=1}^{M}x_{i,G_{k}}^{\top}\beta_{G_{k}}\right),\quad i=1,\dots,n (1)

where f:f:\mathbb{R}\rightarrow\mathbb{R} represents the link function, β=(β1,,βp)p\beta=(\beta_{1},\dots,\beta_{p})^{\top}\in\mathbb{R}^{p} denotes the model coefficient vector with βGk={βj:jGk}\beta_{G_{k}}=\{\beta_{j}:j\in G_{k}\}. Beyond incorporating the group structure, it is often of practical importance to identify the groups of features that are associated with the response. This holds particularly true when there are a large number of them. To address this, various methods have been proposed over the years, with one of the most popular being the group LASSO (Yuan and Lin, 2006), which applies an 2,1\ell_{2,1} norm to groups of coefficients. Following Yuan and Lin (2006) there have been numerous extensions, including the group SCAD (Wang et al., 2007), the group LASSO for logistic regression (Meier et al., 2008), the group bridge (Huang et al., 2009), group LASSO with overlapping groups (Jacob et al., 2009), and the sparse group LASSO (Breheny and Huang, 2009; Simon et al., 2013), among others (see Huang et al. (2012) for a detailed review of frequentist methods).

In a similar vein, Bayesian group selection methods have arisen. The earliest of which being the Bayesian group LASSO (Raman et al., 2009; Kyung et al., 2010), which uses a multivariate double exponential distribution prior to impose shrinkage on groups of coefficients. Notably, the maximum a posteriori (MAP) estimate under this prior coincides with the estimate under the group LASSO. Other methods include Bayesian sparse group selection (Xu and Ghosh, 2015; Chen et al., 2016) and the group spike-and-slab LASSO (Bai et al., 2020) which approach the problem via stochastic search variable selection (Mitchell and Beauchamp, 1988; Chipman, 1996). Formally, these methods utilize a group spike-and-slab prior, a mixture distribution of a multivariate Dirac mass on zero and a continuous distribution over mk\mathbb{R}^{m_{k}}, where mkm_{k} is the size of the kkth group. Such priors have been shown to work excellently for variable selection as they are able to set coefficients exactly to zero, avoiding the use of shrinkage to enforce sparsity. For a comprehensive review see Lai and Chen (2021) and Jreich et al. (2022).

However, a serious drawback of these methods is that most use Markov Chain Monte Carlo (MCMC) to sample from the posterior (Raman et al., 2009; Xu and Ghosh, 2015; Chen et al., 2016). In general, MCMC approaches are known to be computationally expensive when there are a large number of variables, and within the group setting they can result in poor mixing when group sizes are large (Chen et al., 2016). To circumvent these issues, some authors proposed computing MAP estimates (Kyung et al., 2010; Bai et al., 2020), by relaxing the form of the prior, replacing the multivariate Dirac mass with a continuous distribution concentrated at zero (Ročková and George, 2018). Although these algorithms are fast to compute, they come at the sacrifice of interpretability, as posterior inclusion probabilities no longer guarantee the coefficient is zero but rather concentrated at zero. Beyond this, these algorithms only return a point estimate for β\beta and therefore do not provide uncertainty quantification – a task at the heart of Bayesian inference.

To bridge the gap between scalability and uncertainty quantification several authors have turned to variational inference (VI). An approach to inference wherein the posterior distribution is approximated by a tractable family of distributions known as the variational family (see Zhang et al. (2019) for a review). In the context of high-dimensional Bayesian inference, VI has proven particularly successful, and has been employed in linear regression (Carbonetto and Stephens, 2012; Ormerod et al., 2017; Ray and Szabó, 2022), logistic regression (Ray et al., 2020) and survival analysis (Komodromos et al., 2022) to name a few. Within the context of Bayesian group regression, to our knowledge only the Bayesian group LASSO has seen a variational counterpart (Babacan et al., 2014).

In this manuscript we examine the variational Bayes (VB) posterior arising from the group spike-and-slab prior with multivariate double exponential slab and Dirac spike, referring to our method as Group Spike-and-slab Variational Bayes (GSVB). We provide scalable variational approximations to three common classes of generalized linear models: the Gaussian with identity link function, Binomial with logistic link function, and Poisson with exponential link function. We outline a general scheme for computing the variational posterior via co-ordinate ascent variational inference. We further show that for specific cases, the variational family can be re-parameterized to allow for more efficient updates.

Through extensive numerical experiments we demonstrate that GSVB achieves state-of-the-art performance while significantly reducing computation time by several orders of magnitude compared to MCMC. Moreover, through our comparison with MCMC, we highlight that the variational posterior provides excellent uncertainty quantification through marginal credible sets with impressive coverage. Additionally, the proposed method is compared against the spike-and-slab group LASSO (Bai et al., 2020), a state-of-the-art Bayesian group selection algorithm that returns MAP estimates. Within this comparison our method demonstrates comparable or better performance in terms of group selection and effect estimation, whilst carrying the added benefit of providing uncertainty quantification, a feature not available by other scalable methods in the literature.

To highlight the practical utility of our method, we analyse three real datasets, demonstrating that our method provides excellent predictive accuracy, while also achieving parsimonious models. Additionally, we illustrate the usefulness of the VB posterior by showing its ability to provide posterior predictive intervals, a feature not available to methods that provide MAP estimates, and computationally prohibitive to compute via MCMC.

Theoretical guarantees for our method are provided in the form of posterior contraction rates, which quantify how far the posterior places most of its probability from the ‘ground truth’ generating the data for large sample sizes. This approach builds on the work of Castillo et al. (2015) and Ray and Szabó (2022), and extends previous Bayesian contraction rate results for true posteriors in the group sparse setting (Ning et al., 2020; Bai et al., 2020) to their variational approximation.

Concurrent to our work, Lin et al. (2023) proposed a similar spike-and-slab model for group variable selection in linear regression, which appeared on arXiv shortly after our preprint.

Notation. Let y=(y1,,yn)ny=(y_{1},\dots,y_{n})^{\top}\in\mathbb{R}^{n} denote the realization of the random vector Y=(Y1,,Yn)Y=(Y_{1},\dots,Y_{n}). Further, let X=(x1,,xn)n×pX=(x_{1},\dots,x_{n})^{\top}\in\mathbb{R}^{n\times p} denote the design matrix, where for a group GkG_{k} let XGk=(x1,Gk,,xn,Gk)n×mkX_{G_{k}}=(x_{1,G_{k}},\dots,x_{n,G_{k}})^{\top}\in\mathbb{R}^{n\times m_{k}}. Similarly, for Gkc={1,,p}GkG_{k}^{c}=\{1,\dots,p\}\setminus G_{k}, denote XGkc=(x1,Gkc,,xn,Gkc)n×(pmk)X_{G_{k}^{c}}=(x_{1,G_{k}^{c}},\dots,x_{n,G_{k}^{c}})^{\top}\in\mathbb{R}^{n\times(p-m_{k})} where xi,Gkc={xij:jGkc}x_{i,G_{k}^{c}}=\{x_{ij}:j\in G_{k}^{c}\}, and βGkc={βj:jGkc}\beta_{G_{k}^{c}}=\{\beta_{j}:j\in G_{k}^{c}\}. Wherein, without loss of generality we assume that the elements of the groups are ordered such that 1=G1,1<G1,2<<GM,mM=p1=G_{1,1}<G_{1,2}<\dots<G_{M,m_{M}}=p. Finally, the Kullback-Leibler divergence is defined as DKL=DKL(QP)=𝒳log(dQdP)𝑑QD_{\text{KL}}=D_{\text{KL}}(Q\|P)=\int_{\mathcal{X}}\log\left(\frac{dQ}{dP}\right)dQ, where QQ and PP are probability measures on 𝒳\mathcal{X}, such that QQ is absolutely continuous with respect to PP.

2 Prior and Variational family

To model the coefficients β\beta, we consider a group spike-and-slab prior. For each group GkG_{k}, the prior over βGk\beta_{G_{k}} consists of a mixture distribution of a multivariate Dirac mass on zero and a multivariate double exponential distribution, Ψ(βGk)\Psi(\beta_{G_{k}}), whose density is given by,

ψ(βGk;λ)=Ckλmkexp(λβGk)\psi(\beta_{G_{k}};\lambda)=C_{k}\lambda^{m_{k}}\exp\left(-\lambda\|\beta_{G_{k}}\|\right)

where Ck=[2mkπ(mk1)/2Γ((mk+1)/2)]1C_{k}=[2^{m_{k}}\pi^{(m_{k}-1)/2}\Gamma((m_{k}+1)/2)]^{-1} and \|\cdot\| is the 2\ell_{2}-norm. In the context of sparse Bayesian group regression the multivariate double exponential has been previously considered by Raman et al. (2009) and Kyung et al. (2010) as part of the Bayesian group LASSO and by Xu and Ghosh (2015) within a group spike-and-slab prior.

Formally the prior, which we consider throughout, is given by Π(β)=k=1MΠk(βGk)\Pi(\beta)=\bigotimes_{k=1}^{M}\Pi_{k}(\beta_{G_{k}}), where each Πk(βGk)\Pi_{k}(\beta_{G_{k}}) has the hierarchical representation,

βGk|zkind\displaystyle\beta_{G_{k}}|z_{k}\overset{\text{ind}}{\sim} zkΨ(βGk;λ)+(1zk)δ0(βGk)\displaystyle\ z_{k}\Psi(\beta_{G_{k}};\lambda)+(1-z_{k})\delta_{0}(\beta_{G_{k}}) (2)
zk|θkind\displaystyle z_{k}|\theta_{k}\overset{\text{ind}}{\sim} Bernoulli(θk)\displaystyle\ \text{Bernoulli}(\theta_{k})
θkiid\displaystyle\theta_{k}\overset{\text{iid}}{\sim} Beta(a0,b0)\displaystyle\ \text{Beta}(a_{0},b_{0})

for k=1,,Mk=1,\dots,M, where δ0(βGk)\delta_{0}(\beta_{G_{k}}) is the multivariate Dirac mass on zero with dimension mk=dim(βGk)m_{k}=\dim(\beta_{G_{k}}). In conjunction with the log-likelihood, (𝒟;β)\ell(\mathcal{D};\beta) for a dataset 𝒟={(yi,xi)}i=1n\mathcal{D}=\{(y_{i},x_{i})\}_{i=1}^{n}, we write the posterior density under the given prior as,

dΠ(β|𝒟)=ΠD1e(𝒟;β)dΠ(β)d\Pi(\beta|\mathcal{D})=\Pi_{D}^{-1}e^{\ell(\mathcal{D};\beta)}d\Pi(\beta) (3)

where Π𝒟=pe(𝒟;β)𝑑Π(β)\Pi_{\mathcal{D}}=\int_{\mathbb{R}^{p}}e^{\ell(\mathcal{D};\beta)}d\Pi(\beta) is a normalization constant known as the model evidence.

The posterior arising from the prior (2) and the dataset 𝒟\mathcal{D} assigns probability mass to all 2M2^{M} possible sub-models, i.e. each subset 𝒮{1,,M}\mathcal{S}\subseteq\{1,\dots,M\}, such that zk=1,k𝒮z_{k}=1,k\in\mathcal{S} and zk=0z_{k}=0 otherwise. As using MCMC procedures to sample from this complex posterior distribution is computationally prohibitive, even for a moderate number of groups, we resort to variational inference to approximate it. This approximation is known as the variational posterior which is an element of the variational family 𝒬\mathcal{Q} given by

Π~=argminQ𝒬DKL(QΠ(|𝒟)).\tilde{\Pi}=\underset{Q\in\mathcal{Q}}{{\arg\!\min}}\ D_{\text{KL}}\left(Q\|\Pi(\cdot|\mathcal{D})\right)\;. (4)

For our purposes the variational family we consider is a mean-field variational family,

𝒬={Q(μ,σ,γ)=k=1MQk(μGk,σGk,γk):=k=1M[γkNk(μGk,diag(σGk2))+(1γk)δ0]}\mathcal{Q}=\left\{Q(\mu,\sigma,\gamma)=\bigotimes_{k=1}^{M}Q_{k}(\mu_{G_{k}},\sigma_{G_{k}},\gamma_{k}):=\bigotimes_{k=1}^{M}\left[\gamma_{k}\ N_{k}\left(\mu_{G_{k}},\operatorname{diag}(\sigma_{G_{k}}^{2})\right)+(1-\gamma_{k})\delta_{0}\right]\right\} (5)

where μp\mu\in\mathbb{R}^{p} with μGk={μj:jGk}\mu_{G_{k}}=\{\mu_{j}:j\in G_{k}\}, σ2+p\sigma^{2}\in\mathbb{R}_{+}^{p} with σGk2={σj2:jGk}\sigma^{2}_{G_{k}}=\{\sigma^{2}_{j}:j\in G_{k}\}, γ=(γ1,,γM)[0,1]M\gamma=(\gamma_{1},\dots,\gamma_{M})^{\top}\in[0,1]^{M}. Nk(μ,Σ)N_{k}(\mu,\Sigma) denotes the multivariate Normal distribution with mean parameter μ\mu and covariance Σ\Sigma. Notably, under Q𝒬Q\in\mathcal{Q}, the vector of coefficients for each group GkG_{k} is a spike-and-slab distribution where the slab consists of the product of independent Normal distributions,

βGkindγk[jGkN(μj,σj2)]+(1γk)δ0,\beta_{G_{k}}\overset{\text{ind}}{\sim}\gamma_{k}\big{[}\bigotimes_{j\in G_{k}}N(\mu_{j},\sigma_{j}^{2})\big{]}+(1-\gamma_{k})\delta_{0},

meaning the structure (correlations) between elements within the same group are not captured. To mitigate this, a second variational family is introduced, where the covariance within groups is unrestricted. Formally,

𝒬={Q(μ,Σ,γ)=k=1MQk(μGk,ΣGk,γk):=k=1M[γkN(μGk,ΣGk)+(1γk)δ0]}\mathcal{Q}^{\prime}=\left\{Q^{\prime}(\mu,\Sigma,\gamma)=\bigotimes_{k=1}^{M}Q_{k}^{\prime}(\mu_{G_{k}},\Sigma_{G_{k}},\gamma_{k}):=\bigotimes_{k=1}^{M}\left[\gamma_{k}\ N\left(\mu_{G_{k}},\Sigma_{G_{k}}\right)+(1-\gamma_{k})\delta_{0}\right]\right\} (6)

where Σp×p\Sigma\in\mathbb{R}^{p\times p} is a covariance matrix for which Σij=0\Sigma_{ij}=0, for iGk,jGl,kli\in G_{k},j\in G_{l},k\neq l and ΣGk=(Σij)i,jGkmk×mk\Sigma_{G_{k}}=(\Sigma_{ij})_{i,j\in G_{k}}\in\mathbb{R}^{m_{k}\times m_{k}} denotes the covariance matrix of the kkth group. Notably, 𝒬\mathcal{Q}^{\prime} should provide greater flexibility when approximating the posterior, because unlike 𝒬𝒬\mathcal{Q}\subset\mathcal{Q}^{\prime}, it is able to capture the dependence between coefficients in the same group. The importance of which is highlighted empirically in Section 5 wherein the two families are compared.

Note that the posterior does not take the form QQ or QQ^{\prime}, as the use of these variational families replaces the 2M2^{M} model weights by MM VB group inclusion probabilities, γk\gamma_{k}, thereby introducing substantial additional independence. For example, information as to whether two groups of variables are selected together or not is lost. However, the form of the VB approximation retains many of the interpretable features of the original posterior such as the inclusion probabilities of particular groups.

3 Computing the Variational Posterior

Computing the variational posterior defined in (4) relies on optimizing a lower bound on the model evidence, Π𝒟\Pi_{\mathcal{D}}, known as the evidence lower bound (ELBO). The ELBO follows from the non-negativity of the Kullback-Leibler divergence and is given by 𝔼Q[(𝒟;β)logdQdΠ]\mathbb{E}_{Q}\left[\ell(\mathcal{D};\beta)-\log\frac{dQ}{d\Pi}\right]\ . Intuitively, the first term evaluates the model’s fit to the data, while the second term acts as a regularizer, ensuring the variational posterior is “close” to the prior distribution.

Various strategies exist to tackle the minimization of the negative ELBO,

Q(𝒟):=𝔼Q[logdQdΠ(𝒟;β)]\mathcal{L}_{Q}(\mathcal{D}):=\mathbb{E}_{Q}\left[\log\frac{dQ}{d\Pi}-\ell(\mathcal{D};\beta)\right] (7)

with respect to Q𝒬Q\in\mathcal{Q} (Zhang et al., 2019). One popular approach is co-ordinate ascent variational inference (CAVI), where sets of parameters of the variational family are optimized in turn while the remainder are kept fixed. Although this strategy does not (in general) guarantee the global optimum, it is easy to implement and often leads to scalable inference algorithms. In the following subsections we detail how this algorithm is constructed, noting that derivations are presented under the variational family 𝒬\mathcal{Q}^{\prime}, as the equations for 𝒬𝒬\mathcal{Q}\subset\mathcal{Q}^{\prime} follow by restricting ΣGk\Sigma_{G_{k}} to be a diagonal matrix.

3.1 Computing the Evidence lower bound

To compute the negative ELBO, note that the KL divergence between the prior and the variational distribution, DKL(QΠ)=𝔼Q[log(dQ/dΠ)]D_{\text{KL}}(Q^{\prime}\|\Pi)=\mathbb{E}_{Q^{\prime}}\left[\log(dQ^{\prime}/d\Pi)\right] is constant regardless of the form of the likelihood. To evaluate an expression for this quantity the group independence structure between the prior and variational distribution is exploited, allowing for the log Radon-Nikodym, log(dQ/dΠ)\log\left(dQ^{\prime}/d\Pi\right), to be expressed as,

logdQdΠ(β)=k=1MlogdQkdΠk(βGk)=k=1M𝕀zk=1logγkdNkw¯dΨk(βGk)+𝕀zk=0log(1γk)dδ0(1w¯)dδ0(βGk).\log\frac{dQ^{\prime}}{d\Pi}(\beta)=\sum_{k=1}^{M}\log\frac{dQ_{k}^{\prime}}{d\Pi_{k}}(\beta_{G_{k}})=\sum_{k=1}^{M}\mathbb{I}_{z_{k}=1}\log\frac{\gamma_{k}dN_{k}}{\bar{w}d\Psi_{k}}(\beta_{G_{k}})+\mathbb{I}_{z_{k}=0}\log\frac{(1-\gamma_{k})d\delta_{0}}{(1-\bar{w})d\delta_{0}}(\beta_{G_{k}}).

where w¯=α0/(α0+b0)\bar{w}={\alpha_{0}}/({\alpha_{0}+b_{0}}). Subsequently, it follows that,

DKL(QΠ)=\displaystyle D_{\text{KL}}(Q^{\prime}\|\Pi)= k=1M(γklogγkw¯γk2log(det(2πΣGk))γkmk2γklog(Ck)\displaystyle\ \sum_{k=1}^{M}\bigg{(}\gamma_{k}\log\frac{\gamma_{k}}{\bar{w}}-\frac{\gamma_{k}}{2}\log(\det(2\pi\Sigma_{G_{k}}))-\frac{\gamma_{k}m_{k}}{2}-\gamma_{k}\log(C_{k}) (8)
\displaystyle- γkmklog(λ)+𝔼Q[𝕀zk=1λβGk]+(1γk)log1γk1w¯)\displaystyle\ \gamma_{k}m_{k}\log(\lambda)+\mathbb{E}_{Q^{\prime}}\left[\mathbb{I}_{z_{k}=1}\lambda\|\beta_{G_{k}}\|\right]+(1-\gamma_{k})\log\frac{1-\gamma_{k}}{1-\bar{w}}\bigg{)}

where we have used the fact that 𝔼βGkN(μGk,ΣGk)[(βGkμGk)ΣGk1(βGkμGk)]=mk\mathbb{E}_{\beta_{G_{k}}\sim N(\mu_{G_{k}},\Sigma_{G_{k}})}[(\beta_{G_{k}}-\mu_{G_{k}})^{\top}\Sigma_{G_{k}}^{-1}(\beta_{G_{k}}-\mu_{G_{k}})]=m_{k}. Notably, there is no closed form for 𝔼Q[𝕀zk=1λβGk]\mathbb{E}_{Q^{\prime}}\left[\mathbb{I}_{z_{k}=1}\lambda\|\beta_{G_{k}}\|\right], meaning Q(𝒟)\mathcal{L}_{Q^{\prime}}(\mathcal{D}) is not tractable and optimization over this quantity would require costly Monte Carlo methods.

To circumvent this issue, a surrogate objective is constructed and used instead of the negative ELBO. This follows by applying Jensen’s inequality to 𝔼Q[𝕀zk=1λβGk]\mathbb{E}_{Q^{\prime}}\left[\mathbb{I}_{z_{k}=1}\lambda\|\beta_{G_{k}}\|\right], giving,

𝔼Q[𝕀zk=1λβGk]=γk𝔼Nk[λβGk]γkλ(iGkΣii+μi2)1/2.\mathbb{E}_{Q^{\prime}}\left[\mathbb{I}_{z_{k}=1}\lambda\|\beta_{G_{k}}\|\right]=\gamma_{k}\mathbb{E}_{N_{k}}\left[\lambda\|\beta_{G_{k}}\|\right]\leq\gamma_{k}\lambda\left(\sum_{i\in G_{k}}\Sigma_{ii}+\mu_{i}^{2}\right)^{1/2}. (9)

Thus, DKL(QΠ)D_{\text{KL}}(Q^{\prime}\|\Pi) can be upper-bounded by,

ϱ(\displaystyle\varrho( μ,Σ,γ):=k=1M(γklogγkw¯γk2log(det(2πΣGk))γkmk2\displaystyle\mu,\Sigma,\gamma)=\sum_{k=1}^{M}\bigg{(}\gamma_{k}\log\frac{\gamma_{k}}{\bar{w}}-\frac{\gamma_{k}}{2}\log(\det(2\pi\Sigma_{G_{k}}))-\frac{\gamma_{k}m_{k}}{2} (10)
\displaystyle- γklog(Ck)γkmklog(λ)+γkλ(iGkΣii+μi2)1/2+(1γk)log1γk1w¯)\displaystyle\ \gamma_{k}\log(C_{k})-\gamma_{k}m_{k}\log(\lambda)+\gamma_{k}\lambda\left(\sum_{i\in G_{k}}\Sigma_{ii}+\mu_{i}^{2}\right)^{1/2}+(1-\gamma_{k})\log\frac{1-\gamma_{k}}{1-\bar{w}}\bigg{)}

and in turn, Q(𝒟)ϱ(μ,Σ,γ)𝔼Q[(𝒟;β)]\mathcal{L}_{Q^{\prime}}(\mathcal{D})\leq\varrho(\mu,\Sigma,\gamma)-\mathbb{E}_{Q^{\prime}}\left[\ell(\mathcal{D};\beta)\right]. Via this upper bound, we are able to construct a tractable surrogate objective. Formally, this objective is denoted by, (μ,Σ,γ,θ)\mathcal{F}(\mu,\Sigma,\gamma,\theta), where we introduce ϑ\vartheta to parameterize additional hyperparameters, specifically those introduced to model the variance term under the Gaussian family, and those introduced to bound the Binomial likelihood. Under this parametrization we define,

(μ,Σ,γ,ϑ):=ϱ(μ,Σ,γ)+ϱ~(ϑ)+Λ(μ,Σ,γ,ϑ)Q(𝒟),\mathcal{F}(\mu,\Sigma,\gamma,\vartheta):=\varrho(\mu,\Sigma,\gamma)+\tilde{\varrho}(\vartheta)+\varLambda(\mu,\Sigma,\gamma,\vartheta)\geq\mathcal{L}_{Q^{\prime}}(\mathcal{D}), (11)

where ϱ~(ϑ):Θ\tilde{\varrho}(\vartheta):\varTheta\rightarrow\mathbb{R} and Λ(μ,Σ,γ,ϑ)𝔼Q[(𝒟;β)]\varLambda(\mu,\Sigma,\gamma,\vartheta)\geq\mathbb{E}_{Q^{\prime}}\left[-\ell(\mathcal{D};\beta)\right] (which is an inequality under non-tractable likelihoods). To this end, what remains to be computed is an expression for the expected negative log-likelihood and the term ϱ~(ϑ)\tilde{\varrho}(\vartheta) where appropriate. In the following three subsections we provide derivations of the objective function (μ,Σ,γ,θ)\mathcal{F}(\mu,\Sigma,\gamma,\theta) for the Gaussian, Binomial and Poisson likelihoods taking the canonical link functions for each.

3.1.1 Gaussian

Under the Gaussian family with identity link function, YiiidN(xiβ,τ2)Y_{i}\overset{\text{iid}}{\sim}N(x_{i}^{\top}\beta,\tau^{2}) for all i=1,,ni=1,\dots,n, the log\log-likelihood is given by (𝒟;β,τ2)=n2log(2πτ2)12τ2yXβ2\ell(\mathcal{D};\beta,\tau^{2})=-\frac{n}{2}\log(2\pi\tau^{2})-\frac{1}{2\tau^{2}}\|y-X\beta\|^{2}. To model τ2\tau^{2} an inverse Gamma prior is considered due to its popularity among practitioners, i.e. τ2indΓ1(a,b)\tau^{2}\overset{\text{ind}}{\sim}\Gamma^{-1}(a,b) which has density baΓ(a)xa1exp(bx)\frac{b^{a}}{\Gamma(a)}x^{-a-1}\exp(\frac{-b}{x}) where a,b>0a,b>0. The prior is therefore given by Π(β,τ2)=Π(β)Γ1(τ2)\Pi(\beta,\tau^{2})=\Pi(\beta)\Gamma^{-1}(\tau^{2}), and the posterior density by dΠ(β,τ2|𝒟)=ΠD1e(𝒟;β,τ2)dΠ(β,τ2)d\Pi(\beta,\tau^{2}|\mathcal{D})=\Pi_{D}^{-1}e^{\ell(\mathcal{D};\beta,\tau^{2})}d\Pi(\beta,\tau^{2}), where Π𝒟=p×+e(𝒟;β,τ2)𝑑Π(β,τ2)\Pi_{\mathcal{D}}=\int_{\mathbb{R}^{p}\times\mathbb{R}_{+}}e^{\ell(\mathcal{D};\beta,\tau^{2})}d\Pi(\beta,\tau^{2}) and 𝒟={(yi,xi)}i=1n\mathcal{D}=\{(y_{i},x_{i})\}_{i=1}^{n} is the observed data.

To include this term within the inference procedure, the variational families 𝒬\mathcal{Q} and 𝒬\mathcal{Q}^{\prime} are extended by 𝒬τ=𝒬×{Γ1(a,b):a,b>0}\mathcal{Q}_{\tau}=\mathcal{Q}\times\{\Gamma^{-1}(a^{\prime},b^{\prime}):a^{\prime},b^{\prime}>0\} and 𝒬τ=𝒬×{Γ1(a,b):a,b>0}\mathcal{Q}^{\prime}_{\tau}=\mathcal{Q}^{\prime}\times\{\Gamma^{-1}(a^{\prime},b^{\prime}):a^{\prime},b>0\} respectively. Under the extended variational family 𝒬τ\mathcal{Q}^{\prime}_{\tau}, the negative ELBO is given by

𝔼Qτ[logdQdΠ(β)+logdΓ1(a,b)dΓ1(a,b)(τ2)(𝒟;β,τ2)],\mathbb{E}_{Q^{\prime}_{\tau}}\bigg{[}\log\frac{dQ^{\prime}}{d\Pi}(\beta)+\log\frac{d\Gamma^{-1}(a^{\prime},b^{\prime})}{d\Gamma^{-1}(a,b)}(\tau^{2})-\ell(\mathcal{D};\beta,\tau^{2})\bigg{]},

where the second term follows from the independence of τ2\tau^{2} and β\beta in the prior and variational family and is given by,

ϱ~(ϑ):=𝔼Qτ[log(dΓ1(a,b)dΓ1(a,b))]=(aa)κ(a)+alogbb+logΓ(a)Γ(a)+(bb)ab\tilde{\varrho}(\vartheta):=\mathbb{E}_{Q^{\prime}_{\tau}}\left[\log\left(\frac{d\Gamma^{-1}(a^{\prime},b^{\prime})}{d\Gamma^{-1}(a,b)}\right)\right]=(a^{\prime}-a)\kappa(a^{\prime})+a\log\frac{b^{\prime}}{b}+\log\frac{\Gamma(a)}{\Gamma(a^{\prime})}+\frac{(b-b^{\prime})a^{\prime}}{b^{\prime}} (12)

where ϑ={(a,b)}\vartheta=\{(a^{\prime},b^{\prime})\}. The expectation of the negative log-likelihood is given by,

Λ(μ,Σ,\displaystyle\varLambda(\mu,\Sigma, γ,ϑ):=𝔼Qτ[(𝒟;β)]=n2(log(2π)+log(b)κ(a))\displaystyle\gamma,\vartheta)=\mathbb{E}_{Q^{\prime}_{\tau}}\left[-\ell(\mathcal{D};\beta)\right]=\frac{n}{2}(\log(2\pi)+\log(b^{\prime})-\kappa(a^{\prime})) (13)
+\displaystyle+ a2b(y2+(i=1pj=1p(XX)ij𝔼Qτ[βiβj])2k=1Mγky,XGkμGk)\displaystyle\ \frac{a^{\prime}}{2b^{\prime}}\Bigg{(}\|y\|^{2}+\left(\sum_{i=1}^{p}\sum_{j=1}^{p}(X^{\top}X)_{ij}\mathbb{E}_{Q^{\prime}_{\tau}}\left[\beta_{i}\beta_{j}\right]\right)-2\sum_{k=1}^{M}\gamma_{k}\langle y,X_{G_{k}}\mu_{G_{k}}\rangle\Bigg{)}

where,

𝔼Qτ[βiβj]={γk(Σij+μiμj)i,jGkγkγhμiμjiGk,jGh,hk\mathbb{E}_{Q^{\prime}_{\tau}}\left[\beta_{i}\beta_{j}\right]=\begin{cases}\gamma_{k}\left(\Sigma_{ij}+\mu_{i}\mu_{j}\right)&\quad i,j\in G_{k}\\ \gamma_{k}\gamma_{h}\mu_{i}\mu_{j}&\quad i\in G_{k},j\in G_{h},h\neq k\end{cases} (14)

Substituting (12) and (13) into (11) gives (μ,Σ,γ,θ)\mathcal{F}(\mu,\Sigma,\gamma,\theta) under the Gaussian family.

3.1.2 Binomial

Under the Binomial family with logistic link, YiiidBernoulli(pi)Y_{i}\overset{\text{iid}}{\sim}\text{Bernoulli}(p_{i}) for all i=1,,ni=1,\dots,n where pi=(Yi=1|xi)=exp(xiβ)/(1+exp(xiβ))p_{i}=\mathbb{P}(Y_{i}=1|x_{i})=\exp(x_{i}^{\top}\beta)/(1+\exp(x_{i}^{\top}\beta)). The log-likelihood is given by (𝒟,β)=i=1nyi(xiβ)log(1+exp(xiβ))\ell(\mathcal{D},\beta)=\sum_{i=1}^{n}y_{i}\left(x_{i}^{\top}\beta\right)-\log\left(1+\exp(x_{i}^{\top}\beta)\right) where 𝒟={(yi,xi)}i=1n\mathcal{D}=\{(y_{i},x_{i})\}_{i=1}^{n} with yi{0,1}y_{i}\in\{0,1\}.

Unlike the Gaussian family, variational inference in this setting is challenging because of the intractability of the expected log-likelihood under the variational family. To overcome this issue several authors have proposed bounds or approximations to maintain tractability (see Depraetere and Vandebroek (2017) for a review). Here we employ a bound introduced by Jaakkola and Jordan (1996), given as

s(x)s(t)exp{xt2a(t)2(x2t2)}s(x)\geq s(t)\exp\left\{\frac{x-t}{2}-\frac{a(t)}{2}(x^{2}-t^{2})\right\} (15)

where s(x)=(1+exp(x))1s(x)=(1+\exp(-x))^{-1} and a(t)=s(t)1/2ta(t)=\frac{s(t)-1/2}{t}, xx\in\mathbb{R} and tt\in\mathbb{R} is an additional parameter that must be optimized to ensure tightness of the bound.

Using (15) allows for the negative log-likelihood to be bounded by,

(𝒟;\displaystyle-\ell(\mathcal{D}; β)i=1nyixiβlogs(ti)+xiβ+ti2+a(ti)2((xiβ)2ti2)\displaystyle\ \beta)\leq\sum_{i=1}^{n}-y_{i}x_{i}^{\top}\beta-\log s(t_{i})+\frac{x_{i}^{\top}\beta+t_{i}}{2}+\frac{a(t_{i})}{2}((x_{i}^{\top}\beta)^{2}-t_{i}^{2}) (16)

where tit_{i}\in\mathbb{R} is a hyper-parameter for each observation. Taking the expectation of (16) with respect to the variational family gives

𝔼Q[(𝒟;β)]Λ(μ,Σ,γ,ϑ):=\displaystyle\mathbb{E}_{Q^{\prime}}\left[-\ell(\mathcal{D};\beta)\right]\leq\varLambda(\mu,\Sigma,\gamma,\vartheta)= i=1n(k=1Mγk(1/2yi)xi,GkμGk)+ti2logs(ti)\displaystyle\sum_{i=1}^{n}\left(\sum_{k=1}^{M}\gamma_{k}(1/2-y_{i})x_{i,G_{k}}^{\top}\mu_{G_{k}}\right)+\frac{t_{i}}{2}-\log s(t_{i}) (17)
+a(ti)2((j=1pl=1p(xijxil𝔼Q[βjβl])ti2)\displaystyle\qquad+\frac{a(t_{i})}{2}\left(\left(\sum_{j=1}^{p}\sum_{l=1}^{p}(x_{ij}x_{il}\mathbb{E}_{Q^{\prime}}\left[\beta_{j}\beta_{l}\right]\right)-t_{i}^{2}\right)

where ϑ={t1,,tn}\vartheta=\{t_{1},\dots,t_{n}\}. In turn substituting (17) and ϱ~(ϑ)=0\tilde{\varrho}(\vartheta)=0 into (11) gives the objective under the Binomial family.

3.1.3 Poisson

Finally, under the Poisson family with exponential link function, YiiidPoisson(λi)Y_{i}\overset{\text{iid}}{\sim}\text{Poisson}(\lambda_{i}) for all i=1,,ni=1,\dots,n with λi=exp(xiβ)>0\lambda_{i}=\exp(x_{i}^{\top}\beta)>0. The log-likelihood is given by, (𝒟;β)=i=1nyixiβexp(xiβ)log(y!)\ell(\mathcal{D};\beta)=\sum_{i=1}^{n}y_{i}x_{i}^{\top}\beta-\exp(x_{i}^{\top}\beta)-\log(y!), whose (negative) expectation is tractable and given by,

𝔼Q[(𝒟;β)]=Λ(μ,Σ,γ,ϑ):=i=1nlog(y!)MQ(xi)+(k=1Mγkyixi,GkμGk)\mathbb{E}_{Q^{\prime}}\left[-\ell(\mathcal{D};\beta)\right]=\varLambda(\mu,\Sigma,\gamma,\vartheta):=\sum_{i=1}^{n}\log(y!)-M_{Q^{\prime}}(x_{i})+\left(\sum_{k=1}^{M}\gamma_{k}y_{i}x_{i,G_{k}}^{\top}\mu_{G_{k}}\right) (18)

where MQ(xi)=k=1MMQk(xi,Gk)M_{Q^{\prime}}(x_{i})=\prod_{k=1}^{M}M_{Q_{k}}(x_{i,G_{k}}) is the moment generating function under the variational family, with MQk(xi,Gk):=γkMNk(xi,Gk)+(1γk)M_{Q_{k}}(x_{i,G_{k}}):=\gamma_{k}M_{N_{k}}(x_{i,G_{k}})+(1-\gamma_{k}) being the moment generating function for the kkth group and MNk(xi,Gk)=exp{xi,GkμGk+12xi,GkΣGkxi,Gk}M_{N_{k}}(x_{i,G_{k}})=\exp\left\{x_{i,G_{k}}^{\top}\mu_{G_{k}}+\frac{1}{2}x_{i,G_{k}}^{\top}\Sigma_{G_{k}}x_{i,G_{k}}\right\}. Unlike the previous two families the Poisson family, does not require any additional variational parameters, therefore ϑ={}\vartheta=\{\} and ϱ~(ϑ)=0\tilde{\varrho}(\vartheta)=0.

3.2 Coordinate ascent algorithm

Recall the aim is to approximate the posterior Π(|𝒟)\Pi(\cdot|\mathcal{D}) by a distribution from a given variational family. This approximation is obtained via the minimization of the objective \mathcal{F} derived in the previous section. To achieve this a CAVI algorithm is proposed (Murphy, 2007; Blei et al., 2017) as outlined in Algorithm 1.

In this context, the objective introduced in (11) is written as (μ,Σ,γ,ϑ)=(μGk,μGkc,ΣGk,ΣGkc,γk,γk,ϑ)\mathcal{F}(\mu,\Sigma,\gamma,\vartheta)=\mathcal{F}(\mu_{G_{k}},\mu_{G_{k}^{c}},\\ \Sigma_{G_{k}},\Sigma_{G_{k}^{c}},\gamma_{k},\gamma_{-k},\vartheta), highlighting the fact that optimization over the variational parameters occurs group-wise. Further, for each group kk, while the optimization of the objective function over the inclusion probability, γk\gamma_{k}, can be done analytically, we use the Limited memory Broyden–Fletcher–Goldfarb–Shanno optimization algorithm (L-BFGS) to update μGk\mu_{G_{k}} at each iteration of the CAVI procedure. Details for the optimization with respect to ΣGk\Sigma_{G_{k}} are presented in Section 3.2.1. The hyper-parameters, ϑ\vartheta, are updated using L-BFGS for the Gaussian family and analytically for those under the Binomial family.

Algorithm 1 Group sparse co-ordinate ascent variational inference
Initialize μ,Σ,γ,ϑ\mu,\Sigma,\gamma,\vartheta
while not converged 
  for k=1,,Mk=1,\dots,M 
   μGkargminμGkmk(μGk,μGkc,ΣGk,ΣGkc,γk=1,γk,ϑ)\mu_{G_{k}}\leftarrow{{\arg\!\min}}_{\mu_{G_{k}}\in\mathbb{R}^{m_{k}}}\ \mathcal{F}(\mu_{G_{k}},\mu_{G_{k}^{c}},\Sigma_{G_{k}},\Sigma_{G_{k}^{c}},\gamma_{k}=1,\gamma_{-k},\vartheta)
   ΣGkargminΣGkmk×mk(μGk,μGkc,ΣGk,ΣGkc,γk=1,γk,ϑ)\Sigma_{G_{k}}\leftarrow{{\arg\!\min}}_{\Sigma_{G_{k}}\in\mathbb{R}^{m_{k}\times m_{k}}}\ \mathcal{F}(\mu_{G_{k}},\mu_{G_{k}^{c}},\Sigma_{G_{k}},\Sigma_{G_{k}^{c}},\gamma_{k}=1,\gamma_{-k},\vartheta)
   γkargminγk[0,1](μGk,μGkc,ΣGk,ΣGkc,γk,γk,ϑ)\gamma_{k}\leftarrow{{\arg\!\min}}_{\gamma_{k}\in[0,1]}\ \mathcal{F}(\mu_{G_{k}},\mu_{G_{k}^{c}},\Sigma_{G_{k}},\Sigma_{G_{k}^{c}},\gamma_{k},\gamma_{-k},\vartheta)   
  ϑargminϑΘ(μGk,μGkc,ΣGk,ΣGkc,γk,γk,ϑ)\vartheta\ \leftarrow{{\arg\!\min}}_{\vartheta\in\varTheta}\ \mathcal{F}(\mu_{G_{k}},\mu_{G_{k}^{c}},\Sigma_{G_{k}},\Sigma_{G_{k}^{c}},\gamma_{k},\gamma_{-k},\vartheta)
return μ,Σ,γ,ϑ\mu,\Sigma,\gamma,\vartheta.

To assess convergence, the total absolute change in the parameters is tracked, terminating when this quantity falls below a specified threshold, set to 10310^{-3} in our implementation. Other methods involve monitoring the absolute change in the ELBO, however we found this prohibitively expensive to compute for this purpose.

3.2.1 Re-parameterization of ΣGk\Sigma_{G_{k}}

Our focus now turns to the optimization of (μGk,μGkc,ΣGk,ΣGkc,γk=1,γk,ϑ)\mathcal{F}(\mu_{G_{k}},\mu_{G_{k}^{c}},\Sigma_{G_{k}},\Sigma_{G_{k}^{c}},\gamma_{k}=1,\gamma_{-k},\vartheta) w.r.t. ΣGk\Sigma_{G_{k}}. By using similar ideas to those of Seeger (1999) and Opper and Archambeau (2009), it can be shown that only one free parameter is needed to describe the optimum of ΣGk\Sigma_{G_{k}} under the Gaussian (see below) and Binomial family (see Section A of the Supplementary material). The objective w.r.t ΣGk\Sigma_{G_{k}} under the Gaussian family, is given by,

a2btr(XGkXGkΣGk)12logdetΣGk+λ(iGkΣii+μi2)1/2+C\frac{a^{\prime}}{2b^{\prime}}\operatorname{tr}(X^{\top}_{G_{k}}X_{G_{k}}\Sigma_{G_{k}})-\frac{1}{2}\log\det\Sigma_{G_{k}}+\lambda(\sum_{i\in G_{k}}\Sigma_{ii}+\mu^{2}_{i})^{1/2}+C (19)

where CC is a constant that does not depend on ΣGk\Sigma_{G_{k}}. Differentiating (19) w.r.t. ΣGk\Sigma_{G_{k}}, setting to zero and re-arranging gives,

ΣGk=(abXGkXGk+2νkImk)1\Sigma_{G_{k}}=\left(\frac{a^{\prime}}{b^{\prime}}X_{G_{k}}^{\top}X_{G_{k}}+2\nu_{k}I_{m_{k}}\right)^{-1} (20)

where νk=12λ(iGkΣii+μi2)1/2\nu_{k}=\frac{1}{2}\lambda(\sum_{i\in G_{k}}\Sigma_{ii}+\mu_{i}^{2})^{-1/2}. Thus, substituting ΣGk=(abXGkXGk+wkImk)1\Sigma_{G_{k}}=\left(\frac{a^{\prime}}{b^{\prime}}X^{\top}_{G_{k}}X_{G_{k}}+w_{k}I_{m_{k}}\right)^{-1} where wkw_{k}\in\mathbb{R} into the objective and optimizing over wkw_{k} is equivalent to optimizing over ΣGk\Sigma_{G_{k}}, and carries the added benefit of requiring one free parameter to perform rather than mk(mk1)/2m_{k}(m_{k}-1)/2. Note that under this re-parametrization the inversion of an mk×mkm_{k}\times m_{k} matrix is required which can be a time consuming operation for large mkm_{k}.

For the Poisson family the same re-parameterization cannot be used, so ΣGk\Sigma_{G_{k}} is parameterized by UkUkU_{k}^{\top}U_{k} where Ukmk×mkU_{k}\in\mathbb{R}^{m_{k}\times m_{k}} is an upper triangular matrix. Optimization is then performed on the upper triangular elements of UkU_{k}.

3.2.2 Initialization

As with any gradient-descent based approach, our CAVI algorithm is sensitive to the choice of initial values. We suggest to initialize μ\mu using the group LASSO from the package gglasso using a small regularization parameter. This ensures many of the elements are non-zero. The covariance matrix Σ\Sigma can be initialised by using the re-parametrization outlined in Section 3.2.1 with an initial value of wk=1w_{k}=1 for k=1,,Mk=1,\dots,M for both the Gaussian and the Binomial families. For the Poisson family we propose the use of an initial covariance matrix Σ=diag(0.2,,0.2)\Sigma=\operatorname{diag}(0.2,\dots,0.2). Finally, the inclusion probabilities are initialised as γ=(0.5,,0.5)\gamma=(0.5,\dots,0.5)^{\top}. For the additional hyper-parameters ϑ\vartheta are used: a=b=103a^{\prime}=b^{\prime}=10^{-3} and ti=(k=1Mγk[μGk,xi,Gk2+xi,GkΣGkxi,Gk])1/2t_{i}=\left(\sum_{k=1}^{M}\gamma_{k}\left[\langle\mu_{G_{k}},x_{i,G_{k}}\rangle^{2}+x_{i,G_{k}}^{\top}\Sigma_{G_{k}}x_{i,G_{k}}\right]\right)^{1/2} for all i=1,,ni=1,\dots,n.

4 Theoretical results for grouped linear regression

4.1 Notation and assumptions

This section establishes frequentist theoretical guarantees for the proposed VB approach in sparse high-dimensional linear regression with group structure. The full proofs of the following results are provided in the Supplementary Material.

To simplify technicalities, the variance parameter τ2\tau^{2} is taken as known and equal to 1, giving model Y=Xβ+εY=X\beta+\varepsilon with YnY\in\mathbb{R}^{n}, Xn×pX\in\mathbb{R}^{n\times p} and εNn(0,In)\varepsilon\sim N_{n}(0,I_{n}). Under suitable conditions, contraction rates for the variational posterior are established, which quantify its spread around the ‘ground truth’ parameter β0p\beta_{0}\in\mathbb{R}^{p} generating the data as n,pn,p\to\infty.

Recall that the covariates are split into MM pre-specified disjoint groups G1,,GMG_{1},\dots,G_{M} of size |Gk|=mk|G_{k}|=m_{k} with k=1Mmk=p\sum_{k=1}^{M}m_{k}=p and maximal group size mmax=maxk=1,,Mmkm_{\text{max}}=\max_{k=1,\dots,M}m_{k}. The above model can then be written as

Y=k=1MXGkβGk+ε,Y=\sum_{k=1}^{M}X_{G_{k}}\beta_{G_{k}}+\varepsilon, (21)

with βGkmk\beta_{G_{k}}\in\mathbb{R}^{m_{k}} and XGkn×mkX_{G_{k}}\in\mathbb{R}^{n\times m_{k}}. Let PβP_{\beta} denote the law of YY under (21), Sβ{G1,,GM}S_{\beta}\subseteq\{G_{1},\dots,G_{M}\} be the set containing the indices of the non-zero groups in βp\beta\in\mathbb{R}^{p}. For a vector βp\beta\in\mathbb{R}^{p} and set SS, we also write βS=(βi)iGk:GkSGkSmk\beta_{S}=(\beta_{i})_{i\in G_{k}:G_{k}\in S}\in\mathbb{R}^{\sum_{G_{k}\in S}m_{k}} for its vector restriction to SS. We write β0\beta_{0} for the ground truth generating the data, S0=Sβ0S_{0}=S_{\beta_{0}} and s0=|S0|s_{0}=|S_{0}| for its group-sparsity. For a matrix Am×nA\in\mathbb{R}^{m\times n}, let AF2=i=1mj=1nAij2=Tr(ATA)\|A\|_{F}^{2}=\sum_{i=1}^{m}\sum_{j=1}^{n}A_{ij}^{2}=\text{Tr}(A^{T}A) be the Frobenius norm and define the group matrix norm of Xn×pX\in\mathbb{R}^{n\times p} by

X=maxk=1,,MXGkF.\|X\|=\max_{k=1,\dots,M}\|X_{G_{k}}\|_{F}.

If all the groups are singletons, X\|X\| reduces to the same norm as in Castillo et al. (2015). We further define the 2,1\ell_{2,1}-norm of a vector by β2,1=k=1MβGk2\|\beta\|_{2,1}=\sum_{k=1}^{M}\|\beta_{G_{k}}\|_{2}. We assume that the prior slab scale λ\lambda satisfies

λ¯λ2λ¯,λ¯=XM1/mmax,λ¯=3XlogM,\underline{\lambda}\leq\lambda\leq 2\bar{\lambda},\qquad\qquad\underline{\lambda}=\frac{\|X\|}{M^{1/m_{\text{max}}}},\qquad\qquad\bar{\lambda}=3\|X\|\sqrt{\log M}, (22)

mirroring the situation without grouping (Castillo et al., 2015; Ray and Szabó, 2022).

The parameter β\beta in (21) is not estimable without additional assumptions on the design matrix XX, for instance that XTXX^{T}X is invertible for sparse subspaces of p\mathbb{R}^{p}. These notions of invertibility can be precise via the following definitions, which are the natural adaptations of compatibility conditions to the group sparse setting (Bühlmann and van de Geer, 2011; Castillo et al., 2015).

Definition 1.

A model S{1,,M}S\subseteq\{1,\dots,M\} has compatibility number

ϕ(S)=inf{Xβ2|S|1/2XβS2,1:βSc2,17βS2,1,βS0}.\phi(S)=\inf\left\{\frac{\|X\beta\|_{2}|S|^{1/2}}{\|X\|\|\beta_{S}\|_{2,1}}:\|\beta_{S^{c}}\|_{2,1}\leq 7\|\beta_{S}\|_{2,1},\beta_{S}\neq 0\right\}.

Compatibility considers only vectors whose coordinates are small outside SS, and hence is a (weaker) notion of approximate rather than exact sparsity. For all β\beta in the above set, it holds that Xβ2|S|1/2ϕ(S)XβS2,1\|X\beta\|_{2}|S|^{1/2}\geq\phi(S)\|X\|\|\beta_{S}\|_{2,1}, which can be interpreted as a form of continuous invertibility of XX for approximately sparse vectors in the sense that changes in βS\beta_{S} lead to sufficiently large changes in XβX\beta that can be detected by the data. The number 7 is not important and is taken in Definition 2.1 of Castillo et al. (2015) to provide a specific numerical value; since we use several of their techniques, we employ the same convention. We next consider two further notions of invertibility for exact group sparsity.

Definition 2.

The compatibility number for vectors of dimension ss is

ϕ¯(s)=inf{Xβ2Sβ|1/2Xβ2,1:0|Sβ|s}.\bar{\phi}(s)=\inf\left\{\frac{\|X\beta\|_{2}\|S_{\beta}|^{1/2}}{\|X\|\|\beta\|_{2,1}}:0\neq|S_{\beta}|\leq s\right\}.
Definition 3.

The smallest scaled sparse singular value of dimension ss is

ϕ~(s)=inf{Xβ2Xβ2:0|Sβ|s}.\widetilde{\phi}(s)=\inf\left\{\frac{\|X\beta\|_{2}}{\|X\|\|\beta\|_{2}}:0\neq|S_{\beta}|\leq s\right\}.

While XTXX^{T}X is not generally invertible in the high-dimensional setting, these last two definitions weaken this requirement to sparse vectors. These are natural extensions of the definitions in Castillo et al. (2015) to the group setting, and similar interpretations and relations to the usual sparse setting apply also here, see Bühlmann and van de Geer (2011) or Section 2.2 in Castillo et al. (2015) for further discussion.

The interplay of the group structure and individual sparsity can lead to multiple regimes see for instance Lounici et al. (2011) and Bühlmann and van de Geer (2011). To make our results more interpretable, we restrict to the main case of practical interest where the group sizes are not too large, and hence the group sparsity drives the estimation rate.

Assumption (K).

There exists K>0K>0 such that mmaxlogmmaxKlogM.m_{\text{max}}\log m_{\text{max}}\leq K\log M.

While related works make similar assumptions (Bai et al., 2020), introducing an explicit constant K>0K>0 above allows us to clarify the uniformity in our results.

4.2 Asymptotic results

We now state our main result on variational posterior contraction for both prediction loss X(ββ0)2\|X(\beta-\beta_{0})\|_{2} and the usual 2\ell_{2}-loss. Our results are uniform over vectors in sets of the form

ρn,sn=ρn,sn(c0):={β0p:ϕ(Sβ0)c0,|Sβ0|sn,ϕ~(ρn|Sβ0|)c0},\displaystyle\mathcal{B}_{\rho_{n},s_{n}}=\mathcal{B}_{\rho_{n},s_{n}}(c_{0}):=\{\beta_{0}\in\mathbb{R}^{p}:\phi(S_{\beta_{0}})\geq c_{0},~|S_{\beta_{0}}|\leq s_{n},~\widetilde{\phi}(\rho_{n}|S_{\beta_{0}}|)\geq c_{0}\}, (23)

where sn1s_{n}\geq 1, c>0c>0 and ρn\rho_{n}\to\infty (arbitrarily slowly).

Theorem 1 (Contraction).

Suppose that Assumption (K) holds, the prior (2) satisfies (22) and sns_{n} satisfies mmaxlogsn=O(logM)m_{\text{max}}\log s_{n}=O(\log M). Then the variational posterior Π~\tilde{\Pi} based on either the variational family 𝒬\mathcal{Q} in (5) or 𝒬\mathcal{Q}^{\prime} in (6) satisfies, with s0=|Sβ0|s_{0}=|S_{\beta_{0}}|,

supβ0ρn,snEβ0Π~(β:X(ββ0)2H0ρn1/2s0logMϕ¯(ρns0))0\sup_{\beta_{0}\in\mathcal{B}_{\rho_{n},s_{n}}}E_{\beta_{0}}\tilde{\Pi}\left(\beta:\|X(\beta-\beta_{0})\|_{2}\geq\frac{H_{0}\rho_{n}^{1/2}\sqrt{s_{0}\log M}}{\bar{\phi}(\rho_{n}s_{0})}\right)\to 0
supβ0ρn,snEβ0Π~(β:ββ02H0ρn1/2s0logMXϕ~(ρns0)2)0,\sup_{\beta_{0}\in\mathcal{B}_{\rho_{n},s_{n}}}E_{\beta_{0}}\widetilde{\Pi}\left(\beta:\|\beta-\beta_{0}\|_{2}\geq\frac{H_{0}\rho_{n}^{1/2}\sqrt{s_{0}\log M}}{\|X\|\widetilde{\phi}(\rho_{n}s_{0})^{2}}\right)\to 0,

for any ρn\rho_{n}\to\infty (arbitrarily slowly), ρn,sn\mathcal{B}_{\rho_{n},s_{n}} defined in (23) and where H0H_{0} depends only on the prior.

Contraction rates for the full posterior based on a group spike and slab prior were established in Ning et al. (2020), as well as for the spike and slab group LASSO in Bai et al. (2020). Our proofs are instead based on those in Castillo et al. (2015) and Ray and Szabó (2022). We have extended their theoretical results from the coordinate sparse setting to the group sparse setting. This approach permits more explicit proofs and allows us to consider somewhat different assumptions from previous group sparse works (Ning et al., 2020; Bai et al., 2020), for instance removing the boundedness assumption for the parameter spaces.

Remark 1.

The optimization problem (4) is in general non-convex, and hence there is no guarantee that CAVI (or any other algorithm) will converge to the global minimizer Π~\tilde{\Pi}. However, an inspection of the proofs shows that the conclusions of Theorems 1 and 2 apply also to any element Q𝒬𝒬Q^{*}\in\mathcal{Q}\subset\mathcal{Q}^{\prime} in the variational families for which

0DKL(QΠ(|Y))DKL(Π~Π(|Y))=Q(𝒟)Π~(𝒟)=O(s0logM),0\leq D_{\text{KL}}(Q^{*}\|\Pi(\cdot|Y))-D_{\text{KL}}(\tilde{\Pi}\|\Pi(\cdot|Y))=\mathcal{L}_{Q^{*}}(\mathcal{D})-\mathcal{L}_{\tilde{\Pi}}(\mathcal{D})=O(s_{0}\log M),

where s0s_{0} is the true group sparsity and Q(𝒟)\mathcal{L}_{Q}(\mathcal{D}) is the negative ELBO. Thus, as long as the ELBO is within O(s0logM)O(s_{0}\log M) of its maximum, the resulting variational approximation will satisfy the above conclusions, even if it is not the global optimum.

The next result shows that the variational posterior puts most of its mass on models of size at most a multiple of the true number of groups, meaning that it concentrates on sparse sets.

Theorem 2 (Dimension).

Suppose that Assumption (K) holds, the prior (2) satisfies (22) and sns_{n} satisfies mmaxlogsn=O(logM)m_{\text{max}}\log s_{n}=O(\log M). Then the variational posterior Π~\tilde{\Pi} based on either the variational family 𝒬\mathcal{Q} in (5) or 𝒬\mathcal{Q}^{\prime} in (6) satisfies

supβ0ρn,snEβ0Π~(β:|Sβ|ρn|Sβ0|)0\sup_{\beta_{0}\in\mathcal{B}_{\rho_{n},s_{n}}}E_{\beta_{0}}\tilde{\Pi}\left(\beta:|S_{\beta}|\geq\rho_{n}|S_{\beta_{0}}|\right)\to 0

for any ρn\rho_{n}\to\infty (arbitrarily slowly) and ρn,sn\mathcal{B}_{\rho_{n},s_{n}} defined in (23).

5 Numerical experiments

In this section a numerical evaluation of our method, referred to as Group Spike-and-slab Variational Bayes (GSVB), is presented111Scripts to reproduce our results can be found at https://github.com/mkomod/p3. Where necessary we distinguish between the two families, 𝒬\mathcal{Q} and 𝒬\mathcal{Q}^{\prime}, by the suffix ‘–D’ and ‘–B’ respectively, i.e. GSVB–D denotes the method under the variational family 𝒬\mathcal{Q}. Notably, throughout all our numerical experiments the prior parameters are set to λ=1\lambda=1, α0=1\alpha_{0}=1, b0=Mb_{0}=M, and a=b=103a=b=10^{-3} for the inverse-Gamma prior on τ2\tau^{2} under the Gaussian family.

This section begins with a comparison of GSVB222 An R package is available at https://github.com/mkomod/gsvb. against MCMC to assess the quality of the variational posterior in terms of variable selection and uncertainty quantification. Following this, a large-scale comparison with the Spike-and-Slab Group LASSO (SSGL) proposed by Bai et al. (2020), a state-of-the-art MAP Bayesian group selection method described in detail in Section 5.3, is performed.

5.1 Simulation set-up

Data is simulated for i=1,,ni=1,\dots,n observations each having a response yiy_{i}\in\mathbb{R} and pp continuous predictors xipx_{i}\in\mathbb{R}^{p}. The response is sampled independently from the respective family with mean given by f(β0xi)f(\beta_{0}^{\top}x_{i}) where ff is the link function (and variance applicable to the Gaussian family of τ2=1\tau^{2}=1). The true coefficient vector β0=(β0,G1,,β0,GM)p\beta_{0}=(\beta_{0,G_{1}},\dots,\beta_{0,G_{M}})^{\top}\in\mathbb{R}^{p} consists of MM groups each of size mm i.e. M×m=pM\times m=p. Of these groups, ss are chosen at random to be non-zero and have each of their element values sampled independently and uniformly from [βmax,0.2][0.2,βmax][-\beta_{\max},0.2]\cup[0.2,\beta_{\max}] where βmax=1.5,1.0\beta_{\max}=1.5,1.0 and 0.450.45 for the Gaussian, Binomial and Poisson families respectively. Finally, the predictors are generated from one of four settings:

  • Setting 1: xiiidN(0p,Ip)x_{i}\overset{\text{iid}}{\sim}N(0_{p},I_{p})

  • Setting 2: xiiidN(0p,Σ)x_{i}\overset{\text{iid}}{\sim}N(0_{p},\Sigma) where Σij=0.6|ij|\Sigma_{ij}=0.6^{|i-j|} for i,j=1,,pi,j=1,\dots,p.

  • Setting 3: xiiidN(0,Σ)x_{i}\overset{\text{iid}}{\sim}N(0,\Sigma) where Σ\Sigma is a block diagonal matrix where each block AA is a 50×5050\times 50 square matrix such that Ajl=0.6,jlA_{jl}=0.6,j\neq l and Ajj=1A_{jj}=1 otherwise.

  • Setting 4: xiiidN(0,Σ)x_{i}\overset{\text{iid}}{\sim}N(0,\Sigma) where Σ=(1α)W1+αV1\Sigma=(1-\alpha)W^{-1}+\alpha V^{-1} with WWishart(p+ν,Ip)W\sim\text{Wishart}(p+\nu,I_{p}) and VV is a block diagonal matrix of MM blocks, where each block VkV_{k}, for k=1,,Mk=1,\dots,M, is an mk×mkm_{k}\times m_{k} matrix given by VkWishart(mk+ν,Imk)V_{k}\sim\text{Wishart}(m_{k}+\nu,I_{m_{k}}); we let (ν,α)=(3,0.9)(\nu,\alpha)=(3,0.9) so that predictors within groups are more correlated than variables between groups.

To evaluate the performance of the methods four different metrics are considered: (i) the 2\ell_{2}-error, β^β02\|\widehat{\beta}-\beta_{0}\|_{2}, between the true vector of coefficients and the estimated coefficient β^\widehat{\beta} defined as the posterior mean where applicable, or the maximum a posteriori (MAP) estimate if this is returned, (ii) the area under the curve (AUC) of the receiver operator characteristic curve, which compares true positive and false positive rate for different thresholds of the group posterior inclusion probability, (iii) the marginal coverage of the non-zero coefficients, which reports the proportion of times the true coefficient β0,j\beta_{0,j} is contained in the marginal credible set {j:β0,j0}\{j:\beta_{0,j}\neq 0\}, (iv) and the size of the marginal credible set for the non-zero coefficients, given by the Lebesgue measure of the set. The last two metrics can only be computed when a distribution for β\beta is available i.e. via MCMC or VB. The 95%95\% marginal credible sets for each variable jGkj\in G_{k} for k=1,,Mk=1,\dots,M are given by:

Sj={{0}if γk<α[μj±Σjj1/2Φ1(αγk2)]if γkα and 0[μj±Σjj1/2Φ1(αγk2)][μj±Σjj1/2Φ1(αγk2+1γk2)]{0}otherwiseS_{j}=\begin{cases}\{0\}&\text{if }\gamma_{k}<\alpha\\ \left[\mu_{j}\pm\Sigma_{jj}^{1/2}\Phi^{-1}(\frac{\alpha_{\gamma_{k}}}{2})\right]&\text{if }\gamma_{k}\geq\alpha\text{ and }0\notin[\mu_{j}\pm\Sigma_{jj}^{1/2}\Phi^{-1}(\frac{\alpha_{\gamma_{k}}}{2})]\\ \left[\mu_{j}\pm\Sigma_{jj}^{1/2}\Phi^{-1}(\frac{\alpha_{\gamma_{k}}}{2}+\frac{1-\gamma_{k}}{2})\right]\cup\{0\}&\text{otherwise}\\ \end{cases}

where αγk=1αγk\alpha_{\gamma_{k}}=1-\frac{\alpha}{\gamma_{k}} and Φ1\Phi^{-1} is the quantile function of the standard Normal distribution.

5.2 Comparison to MCMC

Our first numerical study compares GSVB to the posterior obtained via MCMC, often considered the gold standard in Bayesian inference. The details of the MCMC sampler used for this comparison are outlined in Section C of the Supplementary material333An implementation is available at https://github.com/mkomod/spsl.. The MCMC sampler is ran for 100,000100,\!000 iterations taking a burn-in period of 50,00050,\!000 iterations.

Within this comparison we set p=1,000p=1,\!000, m=5m=5, and vary the number of non-zero groups, ss. As highlighted in Figure 1 GSVB-B performs excellently in nearly all settings, demonstrating comparable results to MCMC in terms of 2\ell_{2}-error and AUC. This indicates that GSVB-B exhibits similar characteristics to MCMC, both in terms of the selected groups and the posterior mean. As anticipated, all the methods exhibit better performance in simpler settings and show a decline in performance as the problem complexity increases.

Refer to caption Refer to caption Refer to caption   Setting 1     Setting 2     Setting 3     Setting 4

  GSVB-D     GSVB-B     MCMC

Figure 1: Performance evaluation of GSVB and MCMC for Settings 1–4 with p=1,000p=1\!,000 across 100 runs. For each method the white diamond (\diamond) indicates the median of the metric, the thick black line ( ) the interquartile range, and the black line ( ) 1.5 times the interquartile range. Rows 1–2: Gaussian family with (n,m,s)=(200,5,5),(200,5,10)(n,m,s)=(200,5,5),(200,5,10). Row 3: Binomial family with (n,m,s)=(400,5,5)(n,m,s)=(400,5,5). Row 4: Poisson family with (n,m,s)=(400,5,2)(n,m,s)=(400,5,2). Note that for the Binomial family Setting 1 results in perfect separation of classes and is excluded from the study

Regarding coverage, whilst MCMC shows slightly better performance compared to GSVB, the proposed method still provides credible sets that capture a significant portion of the true non-zero coefficients (particularly GSVB-B). However, the credible sets of the variational posterior are sometimes not large enough to capture the true non-zero coefficients. This observation is further supported by the size of the marginal credible sets, with MCMC producing the largest sets, followed by GSVB-B and GSVB-D. These findings confirm the well-known fact that VB tends to underestimate the posterior variance (Carbonetto and Stephens, 2012; Blei et al., 2017; Zhang et al., 2019; Ray et al., 2020).

Interestingly, the set size is larger under 𝒬\mathcal{Q}^{\prime}, highlighting the fact that the mean field variational family (𝒬\mathcal{Q}) lacks the necessary flexibility to accurately capture the underlying structure in the data. Furthermore, this result indicates that the full marginal credible quantity improves through the consideration of the interactions within the group.

5.3 Large scale simulations

In this section, our proposed approach is compared against SSGL (Bai et al., 2020) a state-of-the-art Bayesian method. Notably, SSGL employs a similar prior to that introduced in (2), however the multivariate Dirac mass on zero is replaced with a multivariate double exponential distribution, giving a continuous mixture with one density acting as the spike and the other as the slab, parameterized by λ0\lambda_{0} and λ1\lambda_{1} respectively. Under this prior Bai et al. (2020) derive an EM algorithm, which allows for fast updates, however, only MAP estimates are returned, meaning a posterior distribution for β\beta is not available. In addition, the performance of the proposed method is evaluated on larger datasets. As both methods are scalable with pp, in this section it is increased to p=5,000p=5,000. Both the sample size and the number of active (non-zero) groups, ss are varied as illustrated in Figure 2.

In this comparison, we set λ1=1\lambda_{1}=1 for SSGL meaning the slab is identical between the two methods. For the spike for SSGL we took a value of λ0=100\lambda_{0}=100 for the Gaussian and Poisson family and λ0=20\lambda_{0}=20 under the Binomial family. These values were selected to ensure that sufficient mass is concentrated about zero without turning to cross validation to select the value. Finally, we let a0=1a_{0}=1 and b0=Mb_{0}=M for both methods.

Given that SSGL only returns a point estimate for the vector of coefficients β\beta, the methods are compared in terms of 2\ell_{2}-error between the estimated coefficient and the true one, as well as in terms of AUC using the group inclusion probabilities. Overall GSVB performs comparatively or better than SSGL in most settings, obtaining a lower 2\ell_{2}-error and higher AUC (Figure 2). As expected, across the different methods there is a decrease in performance as the difficulty of the setting increases, meaning all methods perform better when there is less correlation in the design matrix. We note that the runtime of SSGL is marginally faster than our method in Settings 1-3, and faster in Settings 4. This is explained by the fact that SSGL only provides point estimate for β\beta, rather than the full posterior distribution. A full breakdown of the runtimes is presented in Section D.2 of the Supplementary Material. Within this large scale simulation our method provides excellent uncertainty quantification. In particular, GSVB–B provides better coverage of the non-zero coefficients than GSVB–D, which can be justified by the set size. As in our comparison to MCMC we notice that there is an increase in the posterior set size as the difficulty of the setting increases, i.e. when there is an increase in the correlation of the design matrix.

Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption   Setting 1     Setting 2     Setting 3     Setting 4

  GSVB-D     GSVB-B     SSGL

Figure 2: Performance evaluation of GSVB and SSGL for Settings 1–4 with p=5,000p=5\!,000 across 100 runs. For each method the white diamond (\diamond) indicates the median of the metric, the thick black line ( ) the interquartile range, and the black line ( ) 1.5 times the interquartile range. Rows 1 – 2: Gaussian family with (n,m,s)=(500,10,10),(500,10,20)(n,m,s)=(500,10,10),(500,10,20). Row 3: Binomial family with (n,m,s)=(1000,5,5)(n,m,s)=(1000,5,5). Row 4: Poisson family with (n,m,s)=(1000,5,3)(n,m,s)=(1000,5,3). Note that for the Binomial family Setting 1 results in perfect separation of classes and is excluded from the study.

6 Real data analysis

This section presents the analysis of two real world datasets with a third dataset analysis included in the Supplementary Material444 R scripts used to produce these results are available at https://github.com/mkomod/p3 . The first problem is a linear regression problem wherein pnp\gg n for which GSVB and SSGL are applied. The second is a logistic regression problem where p<np<n, for which the logistic group LASSO is applied in addition to GSVB and SSGL. As before, a0=1a_{0}=1, b0=Mb_{0}=M and GSVB was ran with a value of λ=1\lambda=1. For SSGL, we let λ1=1\lambda_{1}=1 and λ0\lambda_{0} was chosen via five fold cross validation on the training set.

Overall, our results highlight that GSVB achieves state-of-the-art performance, producing parsimonious models with excellent predictive accuracy. Furthermore, in our analysis of the two datasets, variational posterior predictive distributions are constructed. These highlight the practical utility of GSVB, demonstrating how uncertainty in the predictions is quantified – an added benefit otherwise not available via MAP methods such as SSGL.

6.1 Bardet-Biedl Syndrome Gene Expression Study

The first dataset we examine is of microarray data, consisting of gene expression measurements for 120 laboratory rats555available at ftp://ftp.ncbi.nlm.nih.gov/geo/series/GSE5nnn/GSE5680/matrix/. Originally the dataset was studied by Scheetz et al. (2006) and has subsequently been used to demonstrate the performance of several group selection algorithms (Huang and Zhang, 2010; Breheny and Huang, 2015; Bai et al., 2020). Briefly, the dataset consists of normalized microarray data harvested from the eye tissue of 12 week old male rats. The outcome of interest is the expression of TRIM32 a gene known to cause Bardet-Biedl syndrome, a genetic disease of multiple organs including the retina.

To pre-process the original dataset, which consists of 31,09931,\!099 probe sets (the predictors), we follow Breheny and Huang (2015) and select the 5,0005,\!000 probe sets which exhibit the largest variation in expression on the log scale. Further, following Breheny and Huang (2015) and Bai et al. (2020) a non-parametric additive model is used, wherein, yi=μ+j=1pfj(xij)+ϵiy_{i}=\mu+\sum_{j=1}^{p}f_{j}(x_{ij})+\epsilon_{i} with ϵiN(0,τ2)\epsilon_{i}\sim N(0,\tau^{2}) and fj:f_{j}:\mathbb{R}\rightarrow\mathbb{R}. Here, yiy_{i}\in\mathbb{R} corresponds to the expression of TRIM32 for the iith observation, and xijx_{ij} the expression of the jjth probe set for the iith observation. To approximate fjf_{j} a three term natural cubic spline basis expansion is used. The resulting processing gave a group regression problem with n=120n=120 and M=1,000M=1,\!000 groups of size m=3m=3. Denoting by ϕj,k\phi_{j,k} the kkth basis for fjf_{j}, we have

yi=μ+j=1Mk=1mϕj,k(xij)+ϵi.y_{i}=\mu+\sum_{j=1}^{M}\sum_{k=1}^{m}\phi_{j,k}(x_{ij})+\epsilon_{i}\;. (24)

The performance of the methods is evaluated a ten-fold cross validation, where the methods are fit to the validation set and the test set is used for evaluation. More details about the computation of the metrics used to assess method performance can be found in the Supplementary Material.Overall, GSVB performed excellently, in particular GSVB–D obtained a smaller 10-fold cross validated MSE and model size than SSGL, meaning the models produced are parsimonious and comparably predictive to those of SSGL (Table 1). Furthermore, the PP coverage is excellent, particularly for GSVB–D which has a larger PP coverage than GSVB–B, whilst also having a smaller PP interval length.

Method MSE Num. selected groups PP. coverage PP. length GSVB-D 0.017 (0.019) 1.10 (0.316) 0.983 (0.035) 0.5460 (0.100) GSVB-B 0.018 (0.013) 1.50 (0.527) 0.950 (0.070) 0.5673 (0.091) SSGL 0.019 (0.001) 2.30 (1.060)

Table 1: Bardet-Biedl Syndrome Gene Expression Study, evaluation metrics computed on the held out test data. Average (sd.) over the 10 folds for the metrics are provided: mean squared error, number of selected groups, posterior predictive coverage, and mean posterior predictive interval length. Notably, GSVB failed to converge on the third fold.

To identify genes associated with TRIM32 expression, GSVB was ran on the full dataset. Both methods identify one gene each, with GSVB–D identifying Clec3a and GSVB–B identifying Slc25a34. Interestingly, the MSE was 0.0120.012 and 0.0080.008 for each method, suggesting Slc25a34 is more predictive of TRIM32 expression. Further, the PP coverage was 0.9600.960 for both methods and the PP interval length was 0.4600.460 and 0.3670.367 for GSVB–D and GSVB–B, reflecting there is less uncertainty in the prediction under GSVB–B.

6.2 MEMset splice site detection

The following application is based on the prediction of short read DNA motifs, a task which plays an important role in many areas of computational biology. For example, in gene finding, wherein algorithms such as GENIE (Burge and Karlin, 1997) rely on the prediction of splice sites (regions between coding and non-coding DNA segments).

The MEMset donor data has been used to build predictive models for splice sites and consists of a large training and test set666available at http://hollywood.mit.edu/burgelab/maxent/ssdata/. The original training set contains 8,415 true and 179,438 false donor sites and the test set 4,208 true and 89,717 false donor sites. The predictors are given by 7 factors with 4 levels each (A, T, C, G), a more detailed description is given in Yeo and Burge (2004). Initially analyzed by Yeo and Burge (2004), the data has subsequently been used by Meier et al. (2008) to evaluate the logistic group LASSO.

To create a predictive model we follow Meier et al. (2008) and consider all 33rd order interactions of the 7 factors, which gives a total of M=64M=64 groups and p=1,156p=1,156 predictors. To balance the sets we randomly sub-sampled the training set without replacement, creating a training set of 1,500 true (Y=1Y=1) and 1,500 false donor sites. Regarding the test set a balanced set of 4,208 true and 4,208 false donor sites was created.

In addition to GSVB, we fit SSGL and group LASSO (used in the original analysis), for which we performed 5-fold cross validation to tune the hyperparameters. To assess the different methods, we use the test data, dividing it into 10 folds, and reporting the: (i) precision (TPTP + FP)\left(\frac{\text{TP}}{\text{TP + FP}}\right), (ii) recall (TPTP + FN)\left(\frac{\text{TP}}{\text{TP + FN}}\right), (iii) F-score (TPTP + 0.5FP + 0.5FN)\left(\frac{\text{TP}}{\text{TP + 0.5FP + 0.5FN}}\right), (iv) AUC, and (v) the maximum correlation coefficient between true class membership and the predicted membership over a range of thresholds, ρmax\rho_{\max}, as used in Yeo and Burge (2004) and Meier et al. (2008). In addition we report (vi) the number selected groups, which for the group LASSO is given by the number which have a non-zero 2,1\ell_{2,1} norm.

Method Precision Recall F-score AUC ρmax\rho_{\max} Num. selected groups GSVB–D 0.915 (0.013) 0.951 (0.011) 0.933 (0.009) 0.975 (0.004) 0.870 (0.016) 10 GSVB–B 0.916 (0.011) 0.958 (0.013) 0.936 (0.008) 0.975 (0.004) 0.875 (0.016) 12 SSGL 0.921 (0.012) 0.950 (0.011) 0.935 (0.009) 0.977 (0.004) 0.879 (0.015) 48 Group LASSO 0.924 (0.013) 0.952 (0.010) 0.938 (0.010) 0.977 (0.004) 0.882 (0.016) 60

Table 2: Memset splice site detection, evaluation metrics computed on the held out test data. The test set was split into 10 folds and the mean (sd.) of the metrics are computed.

Overall the methods performed comparably. With GSVB–D obtaining the largest recall and the group LASSO the largest precision, F-score, AUC, and ρmax\rho_{\max} by a small margin (Table 2). However, GSVB returned models that were of smaller size and therefore more parsimonious than those of SSGL and the group LASSO. This is further highlighted in Figure 5 (found in the Supplementary Material), which showcases the fact that the models obtained by GSVB are far simpler, selecting groups 1–7 as well as the interactions between groups 2:4 and 6:7 for both methods, with GSVB–D selecting 5:7 and GSVB–B selecting 2:6, 4:6 and 2:4:6 in addition to these. Further, Figure 5 showcases the fact that uncertainty is available about the norms of the groups.

7 Discussion

In this manuscript we have introduced GSVB, a scalable method for group sparse general linear regression. We have shown how a fast co-ordinate ascent variational inference algorithms can be constructed and used to compute the variational posterior. We have provided theoretical guarantees for the proposed VB approach in the setting of grouped spare linear regression by extending the theoretical work of Castillo et al. (2015) and Ray and Szabó (2022) for group sparse settings. Through extensive numerical studies we have demonstrated that GSVB provides state-of-the-art performance, offering a computationally inexpensive substitute to MCMC, whilst also performing comparably or better than MAP methods. Additionally, through our analysis of real world datasets we have highlighted the practical utility of our method. Demonstrating, that GSVB provides parsimonious models with excellent predictive performance, and as demonstrated in Section E.1, selects variables with established biological significance. Finally, different to MAP and frequentist methods, GSVB provides scalable uncertainty quantification, which serves as a powerful tool in several application areas.

References

  • Babacan et al. (2014) Babacan, S. D., S. Nakajima, and M. N. Do (2014). Bayesian group-sparse modeling and variational inference. IEEE Transactions on Signal Processing 62(11), 2906–2921.
  • Bai et al. (2020) Bai, R. et al. (2020). Spike-and-Slab Group Lassos for Grouped Regression and Sparse Generalized Additive Models. Journal of the American Statistical Association, 1–41.
  • Blake et al. (2021) Blake, J. A. et al. (2021). Mouse Genome Database (MGD): Knowledgebase for mouse-human comparative biology. Nucleic Acids Research 49(D1), D981–D987.
  • Blei et al. (2017) Blei, D. M., A. Kucukelbir, and J. D. McAuliffe (2017). Variational Inference: A Review for Statisticians. Journal of the American Statistical Association 112(518), 859–877.
  • Breheny and Huang (2009) Breheny, P. and J. Huang (2009). Penalized methods for bi-level variable selection. Statistics and Its Interface 2(3), 369–380.
  • Breheny and Huang (2015) Breheny, P. and J. Huang (2015). Group descent algorithms for nonconvex penalized linear and logistic regression models with grouped predictors. Statistics and Computing 25(2), 173–187.
  • Bühlmann and van de Geer (2011) Bühlmann, P. and S. van de Geer (2011). Statistics for high-dimensional data. Springer Series in Statistics. Springer, Heidelberg. Methods, theory and applications.
  • Burge and Karlin (1997) Burge, C. and S. Karlin (1997). Prediction of complete gene structures in human genomic DNA11Edited by F. E. Cohen. Journal of Molecular Biology 268(1), 78–94.
  • Carbonetto and Stephens (2012) Carbonetto, P. and M. Stephens (2012). Scalable variational inference for bayesian variable selection in regression, and its accuracy in genetic association studies. Bayesian Analysis 7(1), 73–108.
  • Castillo et al. (2015) Castillo, I., J. Schmidt-Hieber, and A. van der Vaart (2015). Bayesian linear regression with sparse priors. Ann. Statist. 43(5), 1986–2018.
  • Chen et al. (2016) Chen, R. B., C. H. Chu, S. Yuan, and Y. N. Wu (2016). Bayesian Sparse Group Selection. Journal of Computational and Graphical Statistics 25(3), 665–683.
  • Chipman (1996) Chipman, H. (1996). Bayesian variable selection with related predictors. Canadian Journal of Statistics 24(1), 17–36.
  • Depraetere and Vandebroek (2017) Depraetere, N. and M. Vandebroek (2017). A comparison of variational approximations for fast inference in mixed logit models. Computational Statistics 32(1), 93–125.
  • Huang et al. (2012) Huang, J., P. Breheny, and S. Ma (2012). A selective review of group selection in high-dimensional models. Statistical Science 27(4), 481–499.
  • Huang et al. (2010) Huang, J., J. L. Horowitz, and F. Wei (2010). Variable selection in nonparametric additive models. Annals of Statistics 38(4), 2282–2313.
  • Huang et al. (2009) Huang, J., S. Ma, H. Xie, and C. H. Zhang (2009). A group bridge approach for variable selection. Biometrika 96(2), 339–355.
  • Huang and Zhang (2010) Huang, J. and T. Zhang (2010). The benefit of group sparsity. Annals of Statistics 38(4), 1978–2004.
  • Jaakkola and Jordan (1996) Jaakkola, T. S. and M. I. Jordan (1996). A variational approach to Bayesian logistic regression models and their extensions.
  • Jacob et al. (2009) Jacob, L., G. Obozinski, and J. P. Vert (2009). Group lasso with overlap and graph lasso. ACM International Conference Proceeding Series 382.
  • Jreich et al. (2022) Jreich, R., C. Hatte, and E. Parent (2022). Review of Bayesian selection methods for categorical predictors using JAGS. Journal of Applied Statistics 49(9), 2370–2388.
  • Komodromos et al. (2022) Komodromos, M., E. O. Aboagye, M. Evangelou, S. Filippi, and K. Ray (2022, aug). Variational Bayes for high-dimensional proportional hazards models with applications within gene expression. Bioinformatics 38(16), 3918–3926.
  • Kyung et al. (2010) Kyung, M., J. Gilly, M. Ghoshz, and G. Casellax (2010). Penalized regression, standard errors, and Bayesian lassos. Bayesian Analysis 5(2), 369–412.
  • Lai and Chen (2021) Lai, W. T. and R. B. Chen (2021). A review of Bayesian group selection approaches for linear regression models. Wiley Interdisciplinary Reviews: Computational Statistics 13(4), 1–22.
  • Lee and Cao (2021) Lee, K. and X. Cao (2021). Bayesian group selection in logistic regression with application to MRI data analysis. Biometrics 77(2), 391–400.
  • Lin et al. (2023) Lin, B., C. Ge, and J. S. Liu (2023). A variational spike-and-slab approach for group variable selection. arXiv:2309.16855.
  • Lounici et al. (2011) Lounici, K., M. Pontil, S. Van De Geer, and A. B. Tsybakov (2011). Oracle inequalities and optimal inference under group sparsity. Annals of Statistics 39(4), 2164–2204.
  • Matsui et al. (2010) Matsui, M. et al. (2010, dec). Activation of LDL Receptor Expression by Small RNAs Complementary to a Noncoding Transcript that Overlaps the LDLR Promoter. Chemistry & Biology 17(12), 1344–1355.
  • Meier et al. (2008) Meier, L., S. Van De Geer, and P. Bühlmann (2008). The group lasso for logistic regression. Journal of the Royal Statistical Society. Series B: Statistical Methodology 70(1), 53–71.
  • Mitchell and Beauchamp (1988) Mitchell, T. J. and J. J. Beauchamp (1988). Bayesian variable selection in linear regression. Journal of the American Statistical Association 83(404), 1023–1032.
  • Murphy (2007) Murphy, K. P. (2007). Conjugate bayesian analysis of the gaussian distribution.
  • Ning et al. (2020) Ning, B., S. Jeong, and S. Ghosal (2020). Bayesian linear regression for multivariate responses under group sparsity. Bernoulli 26(3), 2353–2382.
  • Opper and Archambeau (2009) Opper, M. and C. Archambeau (2009, 03). The variational gaussian approximation revisited. Neural Computation 21(3), 786–792.
  • Ormerod et al. (2017) Ormerod, J. T., C. You, and S. Müller (2017). A variational bayes approach to variable selection. Electronic Journal of Statistics 11(2), 3549–3594.
  • Pinelis and Sakhanenko (1986) Pinelis, I. F. and A. I. Sakhanenko (1986). Remarks on inequalities for large deviation probabilities. Theory of Probability & Its Applications 30(1), 143–148.
  • Raman et al. (2009) Raman, S. et al. (2009). The bayesian group-lasso for analyzing contingency tables. ACM International Conference Proceeding Series 382, 881–888.
  • Ray and Szabó (2022) Ray, K. and B. Szabó (2022). Variational Bayes for high-dimensional linear regression with sparse priors. J. Amer. Statist. Assoc. 117(539), 1270–1281.
  • Ray et al. (2020) Ray, K., B. Szabo, and G. Clara (2020). Spike and slab variational Bayes for high dimensional logistic regression. In Advances in Neural Information Processing Systems, Volume 33, pp.  14423–14434. Curran Associates, Inc.
  • Ročková and George (2018) Ročková, V. and E. I. George (2018). The Spike-and-Slab LASSO. Journal of the American Statistical Association 113(521), 431–444.
  • Scheetz et al. (2006) Scheetz, T. E. et al. (2006). Regulation of gene expression in the mammalian eye and its relevance to eye disease. Proceedings of the National Academy of Sciences of the United States of America 103(39), 14429–14434.
  • Seeger (1999) Seeger, M. (1999). Bayesian methods for support vector machines and Gaussian processes. Master’s thesis, University of Karlsruhe.
  • Silverman et al. (2016) Silverman, M. G. et al. (2016). Association between lowering LDL-C and cardiovascular risk reduction among different therapeutic interventions: A systematic review and meta-analysis. Journal of the American Medical Association 316(12), 1289–1297.
  • Simon et al. (2013) Simon, N., J. Friedman, T. Hastie, and R. Tibshirani (2013). A sparse-group lasso. Journal of Computational and Graphical Statistics 22(2), 231–245.
  • Wang et al. (2007) Wang, L., G. Chen, and H. Li (2007). Group SCAD regression analysis for microarray time course gene expression data. Bioinformatics 23(12), 1486–1494.
  • Xu and Ghosh (2015) Xu, X. and M. Ghosh (2015). Bayesian variable selection and estimation for group lasso. Bayesian Analysis 10(4), 909–936.
  • Yeo and Burge (2004) Yeo, G. and C. B. Burge (2004). Maximum entropy modeling of short sequence motifs with applications to RNA splicing signals. Journal of Computational Biology 11(2-3), 377–394.
  • Yuan and Lin (2006) Yuan, M. and Y. Lin (2006). Model selection and estimation in regression with grouped variables. Journal of the Royal Statistical Society. Series B: Statistical Methodology 68(1), 49–67.
  • Zhang et al. (2019) Zhang, C. et al. (2019). Advances in Variational Inference. IEEE Transactions on Pattern Analysis and Machine Intelligence 41(8), 2008–2026.

SUPPLEMENTARY MATERIAL

Appendix A Co-ordinate ascent algorithms

Derivations are presented for the variational family 𝒬\mathcal{Q}^{\prime}, noting that 𝒬𝒬\mathcal{Q}\subset\mathcal{Q}^{\prime}, hence the update equations under 𝒬\mathcal{Q} follow directly from those under 𝒬\mathcal{Q}^{\prime}. Recall, the family is given as,

𝒬={Q(μ,Σ,γ)=k=1MQk(μGk,ΣGk,γk):=k=1M[γkN(μGk,ΣGk)+(1γk)δ0]}\mathcal{Q}^{\prime}=\left\{Q^{\prime}(\mu,\Sigma,\gamma)=\bigotimes_{k=1}^{M}Q_{k}^{\prime}(\mu_{G_{k}},\Sigma_{G_{k}},\gamma_{k}):=\bigotimes_{k=1}^{M}\left[\gamma_{k}\ N\left(\mu_{G_{k}},\Sigma_{G_{k}}\right)+(1-\gamma_{k})\delta_{0}\right]\right\} (25)

where Σp×p\Sigma\in\mathbb{R}^{p\times p} is a covariance matrix for which Σij=0\Sigma_{ij}=0, for iGk,jGl,kli\in G_{k},j\in G_{l},k\neq l (i.e. there is independence between groups) and ΣGk=(Σij)i,jGkmk×mk\Sigma_{G_{k}}=(\Sigma_{ij})_{i,j\in G_{k}}\in\mathbb{R}^{m_{k}\times m_{k}} denotes the covariance matrix of the kkth group.

A.1 Gaussian Family

Under the Gaussian family, YiiidN(xiβ,τ2)Y_{i}\overset{\text{iid}}{\sim}N(x_{i}^{\top}\beta,\tau^{2}), and the log\log-likelihood is given by,

(𝒟;β,τ2)=n2log(2πτ2)12τ2yXβ2.\ell(\mathcal{D};\beta,\tau^{2})=-\frac{n}{2}\log(2\pi\tau^{2})-\frac{1}{2\tau^{2}}\|y-X\beta\|^{2}.

Recall, we have chosen to model τ2\tau^{2} by an inverse-Gamma prior which has density,

baΓ(a)xa1exp(bx)\frac{b^{a}}{\Gamma(a)}x^{-a-1}\exp\left(\frac{-b}{x}\right)

where a,b>0a,b>0 and in turn we extend 𝒬\mathcal{Q}^{\prime} to 𝒬τ=𝒬×{Γ1(a,b):a,b>0}\mathcal{Q}^{\prime}_{\tau}=\mathcal{Q}^{\prime}\times\{\Gamma^{-1}(a^{\prime},b^{\prime}):a^{\prime},b>0\}.

Under this variational family, the expectation of the negative log-likelihood is given by,

𝔼Qτ\displaystyle\mathbb{E}_{Q^{\prime}_{\tau}} [(𝒟;β)]=𝔼Qτ[n2log(2πτ2)+12τ2yXβ2]\displaystyle\left[-\ell(\mathcal{D};\beta)\right]=\mathbb{E}_{Q^{\prime}_{\tau}}\left[\frac{n}{2}\log(2\pi\tau^{2})+\frac{1}{2\tau^{2}}\|y-X\beta\|^{2}\right]
=\displaystyle= 𝔼Qτ[n2log(2πτ2)+12τ2(y2+Xβ22y,Xβ)]\displaystyle\ \mathbb{E}_{Q^{\prime}_{\tau}}\left[\frac{n}{2}\log(2\pi\tau^{2})+\frac{1}{2\tau^{2}}\left(\|y\|^{2}+\|X\beta\|^{2}-2\langle y,X\beta\rangle\right)\right]
=\displaystyle= n2(log(2π)+log(b)κ(a))\displaystyle\ \frac{n}{2}(\log(2\pi)+\log(b^{\prime})-\kappa(a^{\prime}))
+\displaystyle+ a2b(y2+(i=1pj=1p(XX)ij𝔼Qτ[βiβj])2k=1Mγky,XGkμGk)\displaystyle\ \frac{a^{\prime}}{2b^{\prime}}\Bigg{(}\|y\|^{2}+\left(\sum_{i=1}^{p}\sum_{j=1}^{p}(X^{\top}X)_{ij}\mathbb{E}_{Q^{\prime}_{\tau}}\left[\beta_{i}\beta_{j}\right]\right)-2\sum_{k=1}^{M}\gamma_{k}\langle y,X_{G_{k}}\mu_{G_{k}}\rangle\Bigg{)} (26)

where the expectation

𝔼Qτ[βiβj]={γk(Σij+μiμj)i,jGkγkγhμiμjiGk,jGh,hk\mathbb{E}_{Q^{\prime}_{\tau}}\left[\beta_{i}\beta_{j}\right]=\begin{cases}\gamma_{k}\left(\Sigma_{ij}+\mu_{i}\mu_{j}\right)&\quad i,j\in G_{k}\\ \gamma_{k}\gamma_{h}\mu_{i}\mu_{j}&\quad i\in G_{k},j\in G_{h},h\neq k\end{cases}

A.2 Binomial Family Re-parameterization

To re-parametrization of ΣGk\Sigma_{G_{k}} we follow the same process as in Section 3.2.1. Formally, we have,

ΣGk=(XGkAtXGk+wI)1\Sigma_{G_{k}}=(X^{\top}_{G_{k}}A_{t}X_{G_{k}}+wI)^{-1} (27)

where ww\in\mathbb{R} is the free parameter to be optimized and At=diag(a(t1),,a(tn))A_{t}=\operatorname{diag}(a(t_{1}),\dots,a(t_{n})). Notably, this result follows due to the quadratic nature of the bound employed. Meaning we are able to re-parameterize ΣGk\Sigma_{G_{k}}.

Appendix B Derivation of Variational Posterior Predictive

To compute the variational posterior predictive distribution for the Binomial and Poisson model, we sample βΠ~\beta\sim\tilde{\Pi} and then sample y|x,𝒟y^{*}|x^{*},\mathcal{D} from the respective distribution with mean given by f(xβ)f(x*^{\top}\beta) where xpx^{*}\in\mathbb{R}^{p} and ff is the link function. To sample from Π~\tilde{\Pi} we sample zkindBernoulli(γk)z_{k}\overset{\text{ind}}{\sim}\text{Bernoulli}(\gamma_{k}) and then βGkindN(μGk,ΣGk)\beta_{G_{k}}\overset{\text{ind}}{\sim}N(\mu_{G_{k}},\Sigma_{G_{k}}) if zk=1z_{k}=1 otherwise we set βGk=0mk\beta_{G_{k}}=0_{m_{k}}.

For the Gaussian family we are able to simplify the process by integrating out the variance τ2\tau^{2}. Following Murphy (2007), substituting ξ=τ2\xi=\tau^{2} and recalling the independence of τ2\tau^{2} and β\beta in our variational family, we have,

p~(y|x,𝒟)\displaystyle\tilde{p}(y^{\ast}|x^{\ast},\mathcal{D}) pR+ξ(a+1/2)1exp((yxβ)2+2b2ξ)𝑑ξ𝑑Π~(β)\displaystyle\ \propto\int_{\mathbb{R}^{p}}\int_{R^{+}}\xi^{-(a^{\prime}+1/2)-1}\exp\left(-\frac{(y^{\ast}-x^{\ast\top}\beta)^{2}+2b^{\prime}}{2\xi}\right)\ d\xi\ d\tilde{\Pi}(\beta)
p((yxβ)22b+1)(a+1/2)𝑑Π~(β)\displaystyle\ \propto\int_{\mathbb{R}^{p}}\left(\frac{(y^{\ast}-x^{\ast\top}\beta)^{2}}{2b^{\prime}}+1\right)^{-(a^{\prime}+1/2)}\ d\tilde{\Pi}(\beta)

Recognizing that the expression within the integral has the same functional form as a generalized tt-distribution, whose density is denoted as t(x;μ,σ2,ν)=Γ((ν+1)/2)/(Γ(ν/2)νπσ)(1+(xμ)2/(νσ2))(ν+1)/2t(x;\mu,\sigma^{2},\nu)=\Gamma((\nu+1)/2)/(\Gamma(\nu/2)\sqrt{\nu\pi}\sigma)(1+(x-\mu)^{2}/(\nu\sigma^{2}))^{-(\nu+1)/2}, yields,

p~(y|x,𝒟)=pt(y;xβ,b/a,2a)𝑑Π~(β)\tilde{p}(y^{\ast}|x^{\ast},\mathcal{D})=\int_{\mathbb{R}^{p}}t(y^{\ast};x^{\ast\top}\beta,b^{\prime}/a^{\prime},2a^{\prime})\ d\tilde{\Pi}(\beta) (28)

As (28) is intractable we instead sample from p(y|x,𝒟)p(y^{\ast}|x^{\ast},\mathcal{D}), by

  1. 1.

    Sampling βΠ~\beta\sim\tilde{\Pi}.

  2. 2.

    Sampling yy^{\ast} from Y=μ+σtνY^{\ast}=\mu+\sigma t_{\nu} where μ=xβ,σ=b/a\mu=x^{\ast\top}\beta,\sigma=\sqrt{b^{\prime}/a^{\prime}} and ν=2a\nu=2a^{\prime}, where tνt_{\nu} denotes a t-distribution with ν\nu degrees of freedom.

Appendix C Gibbs Sampler

We present a Gibbs sampler for the Gaussian family of models, noting that the samplers for the Binomial and Poisson family use the same principles. We begin by considering a slight alteration of the prior given in (2). Formally,

βGkind\displaystyle\beta_{G_{k}}\overset{\text{ind}}{\sim} Ψ(βGk;λ)\displaystyle\ \Psi(\beta_{G_{k}};\lambda) (29)
zk|θkind\displaystyle z_{k}|\theta_{k}\overset{\text{ind}}{\sim} Bernoulli(θk)\displaystyle\ \text{Bernoulli}(\theta_{k})
θkiid\displaystyle\theta_{k}\overset{\text{iid}}{\sim} Beta(a0,b0)\displaystyle\ \text{Beta}(a_{0},b_{0})

for k=1,,Mk=1,\dots,M and τ2=ξiidΓ1(ξ;a,b)\tau^{2}=\xi\overset{\text{iid}}{\sim}\Gamma^{-1}(\xi;a,b). Under this prior writing the likelihood as,

p(𝒟|β,z,ξ)=i=1nϕ(yi;f(k=1MzkxGk,βGk),ξ)p(\mathcal{D}|\beta,z,\xi)=\prod_{i=1}^{n}\phi\left(y_{i};f\left(\sum_{k=1}^{M}z_{k}\langle x_{G_{k}},\beta_{G_{k}}\rangle\right),\xi\right) (30)

where ϕ(;μ,σ2)\phi(\cdot;\mu,\sigma^{2}) is the density of the Normal distribution with mean μ\mu and variance σ2\sigma^{2}, and ff is the link function (which in this case is the identity), yields the posterior

p(β,z,ξ|𝒟)p(𝒟|β,z,ξ)π(ξ)k=1Mπ(βGk)π(zk|θk)π(θk).p(\beta,z,\xi|\mathcal{D})\propto p(\mathcal{D}|\beta,z,\xi)\pi(\xi)\prod_{k=1}^{M}\pi(\beta_{G_{k}})\pi(z_{k}|\theta_{k})\pi(\theta_{k}). (31)

Notably, the posterior is equivalent to our previous formulation.

To sample from (31), we construct a Gibbs sampler as outlined in Algorithm 2. Ignoring the superscript for clarity, the distribution θk|𝒟,β,z,θk,ξ\theta_{k}|\mathcal{D},\beta,z,\theta_{-k},\xi is conditionally independent of 𝒟,β,zk,ξ\mathcal{D},\beta,z_{-k},\xi and θk\theta_{-k}. Therefore, θk\theta_{k} is sampled from θk|zk\theta_{k}|z_{k}, which has a Beta(a0+zk,b0+1zk)\text{Beta}(a_{0}+z_{k},b_{0}+1-z_{k}) distribution. Regarding zj(i)z_{j}^{(i)}, the conditional density

p(zk|𝒟,β,zk,θ,ξ)\displaystyle p(z_{k}|\mathcal{D},\beta,z_{-k},\theta,\xi)\propto p(𝒟|β,zk,zk,θ,ξ)π(zk|β,zk,θ,ξ).\displaystyle\ p(\mathcal{D}|\beta,z_{-k},z_{k},\theta,\xi)\pi(z_{k}|\beta,z_{-k},\theta,\xi).
=\displaystyle= p(𝒟;β,z,ξ)π(zk|θk).\displaystyle\ p(\mathcal{D};\beta,z,\xi)\pi(z_{k}|\theta_{k}). (32)

As zkz_{k} is discrete, evaluating the RHS of (C) for zk=0z_{k}=0 and zk=1z_{k}=1, gives the unnormalised conditional probabilities. Summing gives the normalisation constant and thus we can sample zkz_{k} from a Bernoulli distribution with parameter

pk=p(zk=1|𝒟,β,zk,θ,ξ)p(zk=0|𝒟,β,zk,θ,ξ)+p(zk=1|𝒟,β,zk,θ,ξ).p_{k}=\frac{p(z_{k}=1|\mathcal{D},\beta,z_{-k},\theta,\xi)}{p(z_{k}=0|\mathcal{D},\beta,z_{-k},\theta,\xi)+p(z_{k}=1|\mathcal{D},\beta,z_{-k},\theta,\xi)}. (33)

Finally, regarding βGk,j(i)\beta_{G_{k},j}^{(i)} we use a Metropolis-Hastings within Gibbs step, wherein a proposal βGk,j(i)\beta_{G_{k},j}^{(i)} is sampled from a random-walk proposition kernel KK, where in our implementation K(x|ϑ,ε)=N(x;ϑ,2(101ε)1/2)K(x|\vartheta,\varepsilon)=N(x;\vartheta,\sqrt{2}\left(10^{1-\varepsilon}\right)^{1/2}). The proposal is then accepted with probability AA or rejected with probability 1A1-A, in which case βGk,j(i)βGk,j(i1)\beta_{G_{k},j}^{(i)}\leftarrow\beta_{G_{k},j}^{(i-1)}, where AA is given by,

A=min(1,p(𝒟;βGkc,βGk,j,βGk,j(i),z,ξ)π(βGk,j(i)|βGk,j)p(𝒟;βGkc,βGk,j,βGk,j(i1),z,ξ)π(βGk,j(i1)|βGk,j)K(βj(i1)|βj(i),zk(i))K(βj(i)|βj(i1),zk(i1)))A=\min\left(1,\frac{p(\mathcal{D};\beta_{G_{k}^{c}},\beta_{G_{k},-j},\beta_{G_{k},j}^{(i)},z,\xi)\pi(\beta_{G_{k},j}^{(i)}|\beta_{G_{k},-j})}{p(\mathcal{D};\beta_{G_{k}^{c}},\beta_{G_{k},-j},\beta_{G_{k},j}^{(i-1)},z,\xi)\pi(\beta_{G_{k},j}^{(i-1)}|\beta_{G_{k},-j})}\frac{K(\beta_{j}^{(i-1)}|\beta_{j}^{(i)},z_{k}^{(i)})}{K(\beta_{j}^{(i)}|\beta_{j}^{(i-1)},z_{k}^{(i-1)})}\right)
Algorithm 2 MCMC sampler for the Gaussian family GSpSL regression
Initialize β(0),z(0),θ(0),ξ(0)\beta^{(0)},z^{(0)},\theta^{(0)},\xi^{(0)}
for i=1,,Ni=1,\dots,N 
  for k=1,,Mk=1,\dots,M 
   θk(i)iid.Beta(a0+zk(i1),b0+1zk(i1))\theta^{(i)}_{k}\overset{\text{iid.}}{\sim}\text{Beta}(a_{0}+z_{k}^{(i-1)},b_{0}+1-z_{k}^{(i-1)})   
  for k=1,,Mk=1,\dots,M 
   zk(i)ind.Bernoulli(pk)z^{(i)}_{k}\overset{\text{ind.}}{\sim}\text{Bernoulli}(p_{k})   
  for k=1,,Mk=1,\dots,M 
   for j=1,,mkj=1,\dots,m_{k} 
     βGk,j(i)p(βGk,j(i)|𝒟,z(i),βG1:k1(i),βGk,1:j1(i),βGk,j+1:mk(i1),βGk+1:M(i1),ξ(i1))\beta_{G_{k},j}^{(i)}\sim p(\beta_{G_{k},j}^{(i)}|\mathcal{D},z^{(i)},\beta^{(i)}_{G_{1:k-1}},\beta_{G_{k,1:j-1}}^{(i)},\beta_{G_{k,j+1:m_{k}}}^{(i-1)},\beta_{G_{k+1:M}}^{(i-1)},\xi^{(i-1)})      
  Sample ξ(i)iid.Γ1(a+0.5n,b+0.5yk=1MzkXGkβGk(i)2)\xi^{(i)}\overset{\text{iid.}}{\sim}\Gamma^{-1}(a+0.5n,b+0.5\|y-\sum_{k=1}^{M}z_{k}X_{G_{k}}\beta_{G_{k}}^{(i)}\|^{2})
return {β(i),z(i),θ(i),ξ(i)}i=1N\{\beta^{(i)},z^{(i)},\theta^{(i)},\xi^{(i)}\}_{i=1}^{N}.

Appendix D Numerical Study

D.1 Performance with large groups

The performance of GSVB with large groups is examined. For this we fix (n,p,s)=(1,000,5,000,10)(n,p,s)=(1\!,000,5\!,000,10) and vary the group size to be m=10,20,25,50,100m=10,20,25,50,100. The results, presented in Figure 3 highlight that the performance of the method is excellent for groups of size m=10,20,25m=10,20,25, and begins to suffer when groups are of size m=50m=50 and 100100, particularly in settings 3 and 4 wherein GSVB–B did not run.

Refer to caption Refer to caption   Setting 1     Setting 2     Setting 3     Setting 4

×\times 95+ runs did not converge     GSVB-D     GSVB-B     SSGL

Figure 3: Performance evaluation of GSVB and SSGL for Settings 1–4 with (n,p,s)=(1,000,5,000,10)(n,p,s)=(1\!,000,5\!,000,10) across 100 runs. For each method the white diamond (\diamond) indicates the median of the metric, the thick black line ( ) the interquartile range, and the black line ( ) 1.5 times the interquartile range. Rows 1–5: Gaussian family with increasing group sizes of m=10,20,25,50,100m=10,20,25,50,100.

D.2 Runtime

Runtimes for the experiments presented in Section 5 are presented in Table 3 and Table 4. Notably, the runtime for GSVB over MCMC is orders of magnitudes faster. The runtime of GSVB in comparison to SSGL is slower in Settings 4, but marginally slower in the remaining settings. Note all simulations were ran on AMD EPYC 7742 CPUs (128 core, 1TB RAM).

Method Setting 1 Setting 2 Setting 3 Setting 4 Gaus. s=5 GSVB–D 2.1s (1.6s, 3.0s) 1.8s (1.1s, 2.9s) 3.5s (1.6s, 8.0s) 2.2s (1.6s, 5.2s) GSVB–B 1.8s (1.2s, 2.9s) 2.0s (1.3s, 3.1s) 3.4s (1.7s, 7.3s) 2.6s (1.8s, 6.7s) MCMC 7m 53s (7m 31s, 8m 12s) 7m 55s (7m 34s, 8m 2s) 7m 44s (7m 20s, 7m 48s) 8m 14s (7m 49s, 8m 17s) Gaus. s=10 GSVB–D 2.3s (1.7s, 3.8s) 2.6s (1.9s, 3.9s) 7.4s (2.9s, 9.7s) 4.2s (2.4s, 12.1s) GSVB–B 2.9s (1.8s, 4.5s) 2.6s (2.2s, 3.2s) 9.5s (4.8s, 13.9s) 4.3s (2.7s, 14.0s) MCMC 9m 55s (9m 45s, 10m 8s) 10m 13s (9m 48s, 10m 30s) 9m 19s (9m 24s, 9m 33s) 10m 11s (9m 46s, 10m 27s) Binom. s=3 GSVB–D - 5.5s (3.5s, 9.3s) 7.7s (3.7s, 12.0s) 5.2s (2.7s, 8.0s) GSVB–B - 7.2s (6.0s, 11.9s) 7.8s (5.6s, 14.4s) 6.0s (3.7s, 9.5s) MCMC - 16m 45s (16m, 17m 15s) 16m 45s (15m 45s, 17m) 15m 45s (14m 49s, 16m 30s) Pois. s=2 GSVB–D 3.6s (3.3s, 4.1s) 3.3s (2.9s, 3.7s) 3.1s (2.9s, 5.2s) 3.4s (2.4s, 4.1s) GSVB–B 12.5s (10.0s, 18.0s) 10.1s (7.9s, 14.5s) 15.5s (10.1s, 41.6s) 13.3s (10.5s, 24.0s) MCMC 15m 45s (14m 49s, 16m 45s) 15m 45s (15m 15s, 16m 30s) 16m (14m 59s, 18m) 18m 30s (16m 45s, 19m 45s)

Table 3: Median (5%, 95% quartile) runtimes for numerical experiments presented in Figure 1. Note (n,p)=(200,1000)(n,p)=(200,1000) for the Gaussian family and (n,p)=(400,1000)(n,p)=(400,1000) for the Binomial and Poisson family. Note that under setting 1 for the Binomial family there is perfect separation.

Method Setting 1 Setting 2 Setting 3 Setting 4 Gaus. m=10, s=10 GSVB–D 59.1s (47.2s, 1m 16s) 54.3s (47.5s, 1m 8s) 2m 53s (1m 4s, 6m 35s) 8m 7s (4m 30s, 25m 36s) GSVB–B 57.7s (50.7s, 1m 18s) 57.0s (49.5s, 1m 14s) 2m 51s (1m 7s, 6m 38s) 9m 45s (4m 53s, 25m 40s) SSGL 1m 3s (59.5s, 1m 13s) 50.8s (41.6s, 59.5s) 1m 1s (58.1s, 1m 16s) 1m 24s (1m 11s, 2m 3s) Gaus. m=10, s=20 GSVB–D 3m 7s (3m 46s, 4m 31s) 3m 13s (3m 42s, 4m 39s) 9m 42s (3m 56s, 14m 54s) 20m 48s (6m 38s, 39m 23s) GSVB–B 3m 29s (3m 6s, 4m 20s) 3m 23s (3m 57s, 4m 48s) 12m 18s (4m 59s, 16m 20s) 34m 59s (12m 54s, 45m 2s) SSGL 2m 42s (1m 23s, 2m 3s) 1m 16s (59.2s, 2m 40s) 2m 6s (2m 37s, 2m 30s) 2m 16s (2m 31s, 6m 32s) Binom. m=5, s=5 GSVB–D - 3m 43s (2m 54s, 4m 54s) 2m 26s (2m 34s, 4m 5s) 2m 50s (1m 16s, 3m 30s) GSVB–B - 4m 32s (3m 40s, 5m 8s) 3m 26s (2m 20s, 5m 26s) 2m 25s (2m 47s, 5m 18s) SSGL - 2m 54s (2m 44s, 3m 42s) 3m 33s (2m 57s, 3m 26s) 2m 40s (2m 31s, 2m 3s) Pois. m=5, s=3 GSVB–D 16.3s (9.8s, 21.5s) 11.2s (8.2s, 20.8s) 16.0s (10.4s, 43.0s) 12.9s (8.6s, 48.7s) GSVB–B 2m 26s (2m 59s, 3m 20s) 2m 55s (1m 29s, 3m 20s) 3m 53s (2m 49s, 9m 25s) 3m 33s (2m 4s, 9m 11s) SSGL 4m 11s (3m 49s, 8m 38s) 4m 34s (2m 11s, 13m 4s) 2m 29s (2m 56s, 16m 33s) 2m 32s (1m 25s, 3m 2s)

Table 4: Median (5%, 95% quartile) runtimes for numerical experiments presented in Figure 2. Note (n,p)=(500,5000)(n,p)=(500,5000) for the Gaussian family and (n,p)=(1000,5000)(n,p)=(1000,5000) for the Binomial and Poisson family. Note that under setting 1 for the Binomial family there is perfect separation

Appendix E Real data analysis

E.1 Genetic determinants of Low-density Lipoprotein in mice

Here we analyze a dataset from the Mouse Genome Database (Blake et al., 2021) and consists of p=10,346p=10,\!346 single nucleotide polymorphisms (SNPs) collected from n=1,637n=1,\!637 laboratory mice. Notably, each SNP, xijx_{ij} for j=1,,pj=1,\dots,p and i=1,,ni=1,\dots,n, takes a value of 0, 1 or 2. This value indicates how many copies of the risk allele are present. Alongside the genotype data, a phenotype (response) is also collected. The phenotype we consider is low-density lipoprotein cholesterol (LDL-C), which has been shown to be a major risk factor for conditions like coronary artery disease, heart attacks, and strokes (Silverman et al., 2016).

To pre-process the original dataset SNPs with a rare allele frequency, given by RAFj=12ni=1n(𝕀(xij=1)+2𝕀(xij=2))\text{RAF}_{j}=\frac{1}{2n}\sum_{i=1}^{n}\left(\mathbb{I}(x_{ij}=1)+2\mathbb{I}(x_{ij}=2)\right) for j=1,,pj=1,\dots,p, below the first quartile were discarded. Further, due to the high colinearity, covariates with |corr(xj,xk)|0.97|\operatorname{corr}(x_{j},x_{k})|\geq 0.97 for k>jk>j, j=1,,p1j=1,\dots,p-1, were removed. After pre-processing, 3,3413,\!341 SNPs remained. These were used to construct groups by coding each xij:{0,1,2}{(0,0),(0,1),(1,1)}x_{ij}:\{0,1,2\}\mapsto\{(0,0),(0,1),(1,1)\}, in turn giving groups of size m=2m=2.

To evaluate the methods ten fold cross validation is used. Specifically, methods are fit to the validation set and the test set is used for evaluation. This is done by computing the: (i) MSE between the true and predicted value of the response, given by 1n~yXβ^\frac{1}{\tilde{n}}\|y-X\widehat{\beta}\| where n~\tilde{n} is the size of the test set, (ii) posterior predictive (PP) coverage, which measures the proportion of times the response is included in the 95% PP set, and (iii) the average size of the 95% PP set. In addition, (iv) the number of groups selected, given by k=1M𝕀(γk>0.5)\sum_{k=1}^{M}\mathbb{I}(\gamma_{k}>0.5) is also reported.

Importantly the PP distribution is available for GSVB, and given by, p~(y|x,𝒟)=p×+p(y|β,τ2,x,𝒟)𝑑Π~(β,τ2)\tilde{p}(y^{\ast}|x^{\ast},\mathcal{D})=\int_{\mathbb{R}^{p}\times\mathbb{R}^{+}}p(y^{\ast}|\beta,\tau^{2},x^{\ast},\mathcal{D})\ d\tilde{\Pi}(\beta,\tau^{2}) where xpx^{\ast}\in\mathbb{R}^{p} is a feature vector (see Section B of the Supplementary material for details). In genetics studies, xβx^{\ast\top}\beta, is referred to as the polygenic risk score and is commonly used to evaluate genetic risk. While frequentist or MAP methods offer only point estimates, Figure 4 demonstrates that GSVB can offer a distribution for this score, highlighting the uncertainty associated with the estimate.

Refer to caption GSVB-D:    PP mean:   Mass%:     90%     95%     99%   Observed response GSVB-B:    PP mean:   Mass%:     90%     95%     99%

Figure 4: Genetic determinants of LDL-C in mice, variational posterior predictive distribution for GSVB-D (panels A.1 - D.1) and GSVB-B (panels A.2 - D.2) constructed for four observations in the held out test set (fold 1). Notably the red line ( ) represents the measured LDL-C level, with the adjacent text giving this value (yy^{*}) and the posterior mean (y^\widehat{y}). The shading of the variational posterior indicates where 90%90\%, 95%95\% and 99%99\% of the mass is contained.

Regarding the performance of each method results are presented in Table 5, these highlight that GSVB performs excellently obtaining parsimonious models with a comparable MSE to SSGL. In addition, GSVB provides impressive uncertainty quantification, yielding a coverage of 0.9450.945 and 0.9410.941 for GSVB–D and GSVB–B respectively. The SNPs selected by GSVB–D and GSVB-B are similar and reported in Table 6 of the Supplementary material. When the methods are fit to the full dataset, the SNPs selected by both method are: rs13476241, rs13477968, and rs13483814, with GSVB–B selecting rs13477939 in addition. Notably, rs13477968 corresponds to the Ago3 gene, which has been shown to play a role in activating LDL receptors, which in turn regulate LDL-C (Matsui et al., 2010).

Method MSE Num. selected groups PP. coverage PP. length
GSVB-D 0.012 (0.002) 3.50 (0.527) 0.945 (0.027) 0.425 (0.005)
GSVB-B 0.012 (0.002) 4.00 (0.667) 0.941 (0.029) 0.421 (0.005)
SSGL 0.012 (0.002) 57.40 (5.481) - -
Table 5: Genetic determinants of LDL-C in mice, evaluation of methods on the held out genotype data. Reported is the 10 fold cross validated MSE, the number of selected groups, the posterior predictive coverage, and mean posterior predictive interval length.
RSID (or probe ID) Freq. GSVB-D Freq. GSVB-B
rs13476279 6 6
rs13483823 6 5
CEL.4_130248229 5 3
rs13477968 3 5
rs13476241 4 3
CEL.X_65891570 2 4
rs13477903 2 2
rs3688710 2 1
rs13477939 1 -
rs3157124 1 -
rs13478204 - 3
rs6186902 - 2
rs13483814 - 1
rs4222922 - 1
rs6187266 - 1
UT_4_128.521481 - 1
Table 6: Genetic determinants of LDL-C in mice: SNPs selected by GSVB.

E.2 Bardet-Biedl Syndrome Gene Expression Study

Probe ID Gene Name Freq. GSVB-D Freq. GSVB-B Freq. SSGL
1383183_at 1 1 8
1386237_at 1 1 -
1397359_at 4 2 -
1386552_at - 1 -
1378682_at - 1 -
1385926_at - - 2
1396814_at - - 3
1386811_at Hacd4 1 2 -
1386069_at Sp2 4 2 -
1371209_at RT1-CE5 - 1 -
1391624_at Sec14l3 - 1 -
1383829_at Bbx - 2 -
1389274_at Dcakd - - 1
1373995_at Abcg1 - - 1
1375361_at Snx18 - - 1
1368915_at Kmo - - 7
Table 7: Bardet-Biedl syndrome gene expression study: variables selected by GSVB and SSGL.

E.3 MEMset splice site detection

Refer to caption   Groups selected by more than one method   \bullet  Groups for which βGk=0\|\beta_{G_{k}}\|=0

\bullet  Groups for which β^Gk0\|\widehat{\beta}_{G_{k}}\|\neq 0   95% credible intervals for βGk\|\beta_{G_{k}}\| where available  

Figure 5: MEMset splice site detection, comparison of 2\ell_{2}-norms for each group of coefficients β^Gk\widehat{\beta}_{G_{k}} for k=1,,64k=1,\dots,64. The red dotted line ( ​) shows groups which have been selected by more than one method. The points indicate β^Gk\|\widehat{\beta}_{G_{k}}\| where β^Gk\widehat{\beta}_{G_{k}} is the posterior mean for GSVB, the MAP for SSGL and the MLE estimate under the group LASSO. The points are coloured in black (\bullet) when they are non-zero and grey ( \bullet) otherwise. Finally when available the 95% credible set for βGk\|\beta_{G_{k}}\| is given by the solid black line ( ).

Appendix F Proofs of asymptotic results

F.1 A general class of model selection priors and an overview of the proof

Our theoretical results apply to a wider class of model selection priors than the group spike and slab prior (2) underlying our variational approximation. In this section, we thus consider this more general class of model selection priors Castillo et al. (2015); Ning et al. (2020); Ray and Szabó (2022) defined hierarchically via:

sπM(s)S||S|=sUnifM,sβGkind{Ψλ,mk(βGk),GkS,δ0,GkS,\begin{split}s\sim\pi_{M}(s)&\\ S||S|=s\sim\text{Unif}_{M,s}&\\ \beta_{G_{k}}\sim^{ind}\begin{cases}\Psi_{\lambda,m_{k}}(\beta_{G_{k}}),~~&G_{k}\in S,\\ \delta_{0},&G_{k}\not\in S,\end{cases}\end{split} (34)

where πM\pi_{M} is a prior on {0,1,,M}\{0,1,\dots,M\}, UnifM,s\text{Unif}_{M,s} is the uniform distribution on subsets S{1,,M}S\subseteq\{1,\dots,M\} of size ss and

Ψλ,mk(βGk)=Δmkλmkexp(λβGk2)\Psi_{\lambda,m_{k}}(\beta_{G_{k}})=\Delta_{m_{k}}\lambda^{m_{k}}\exp\left(-\lambda\|\beta_{G_{k}}\|_{2}\right) (35)

is a density on mk\mathbb{R}^{m_{k}} with Δmk=2mkπ(1mk)/2Γ((mk+1)/2)1\Delta_{m_{k}}=2^{-m_{k}}\pi^{(1-m_{k})/2}\Gamma((m_{k}+1)/2)^{-1}. This can be concisely written as

(S,β)πM(|S|)1(M|S|)δ0(Sc)GkSΨλ,mk(βGk).\displaystyle(S,\beta)\mapsto\pi_{M}(|S|)\frac{1}{{M\choose|S|}}\delta_{0}(S^{c})\prod_{G_{k}\in S}\Psi_{\lambda,m_{k}}(\beta_{G_{k}}). (36)

Following Castillo et al. (2015); Ning et al. (2020); Ray and Szabó (2022), we assume as usual that

A1MA3πM(s1)πM(s)A2MA4πM(s1),s=1,,M.\displaystyle A_{1}M^{-A_{3}}\pi_{M}(s-1)\leq\pi_{M}(s)\leq A_{2}M^{-A_{4}}\pi_{M}(s-1),\qquad s=1,\dots,M. (37)

We further recall the assumption (38) made on the scale parameter:

λ¯λ2λ¯,λ¯=XM1/mmax,λ¯=3XlogM.\underline{\lambda}\leq\lambda\leq 2\bar{\lambda},\qquad\qquad\underline{\lambda}=\frac{\|X\|}{M^{1/m_{\text{max}}}},\qquad\qquad\bar{\lambda}=3\|X\|\sqrt{\log M}. (38)

The group spike and slab prior fits within this framework by taking πM=Bin(M,a0a0+b0)\pi_{M}=\text{Bin}(M,\tfrac{a_{0}}{a_{0}+b_{0}}), and hence the following results immediately imply Theorems 1 and 2. Recall the parameter set

ρn,sn={β0p:ϕ(Sβ0)c0,|Sβ0|sn,ϕ~(ρn|Sβ0|)c0}\mathcal{B}_{\rho_{n},s_{n}}=\{\beta_{0}\in\mathbb{R}^{p}:\phi(S_{\beta_{0}})\geq c_{0},~|S_{\beta_{0}}|\leq s_{n},~\widetilde{\phi}(\rho_{n}|S_{\beta_{0}}|)\geq c_{0}\}

defined in (23).

Theorem 3 (Contraction).

Suppose that Assumption (K) holds, the prior satisfies (37)-(38) and sns_{n} satisfies mmaxlogsnKlogMm_{\text{max}}\log s_{n}\leq K^{\prime}\log M for some K>0K^{\prime}>0. Then the variational posterior Π~\tilde{\Pi} based on either the variational family 𝒬\mathcal{Q} in (5) or 𝒬\mathcal{Q}^{\prime} in (6) satisfies, with s0=|Sβ0|s_{0}=|S_{\beta_{0}}|,

supβ0ρn,snEβ0Π~(β:X(ββ0)2H0ρn1/2s0logMϕ¯(ρns0))0\sup_{\beta_{0}\in\mathcal{B}_{\rho_{n},s_{n}}}E_{\beta_{0}}\tilde{\Pi}\left(\beta:\|X(\beta-\beta_{0})\|_{2}\geq\frac{H_{0}\rho_{n}^{1/2}\sqrt{s_{0}\log M}}{\bar{\phi}(\rho_{n}s_{0})}\right)\to 0
supβ0ρn,snEβ0Π~(β:ββ02,1H0ρns0logMXϕ¯(ρns0)2)0,\sup_{\beta_{0}\in\mathcal{B}_{\rho_{n},s_{n}}}E_{\beta_{0}}\widetilde{\Pi}\left(\beta:\|\beta-\beta_{0}\|_{2,1}\geq\frac{H_{0}\rho_{n}s_{0}\sqrt{\log M}}{\|X\|\bar{\phi}(\rho_{n}s_{0})^{2}}\right)\to 0,
supβ0ρn,snEβ0Π~(β:ββ02H0ρn1/2s0logMXϕ~(ρns0)2)0,\sup_{\beta_{0}\in\mathcal{B}_{\rho_{n},s_{n}}}E_{\beta_{0}}\widetilde{\Pi}\left(\beta:\|\beta-\beta_{0}\|_{2}\geq\frac{H_{0}\rho_{n}^{1/2}\sqrt{s_{0}\log M}}{\|X\|\widetilde{\phi}(\rho_{n}s_{0})^{2}}\right)\to 0,

for any ρn\rho_{n}\to\infty (arbitrarily slowly), ρn,sn\mathcal{B}_{\rho_{n},s_{n}} defined in (23) and where H0H_{0} depends only on the prior.

Theorem 4 (Dimension).

Suppose that Assumption (K) holds, the prior satisfies (37)-(38) and sns_{n} satisfies mmaxlogsnKlogMm_{\text{max}}\log s_{n}\leq K^{\prime}\log M for some K>0K^{\prime}>0. Then the variational posterior Π~\tilde{\Pi} based on either the variational family 𝒬\mathcal{Q} in (5) or 𝒬\mathcal{Q}^{\prime} in (6) satisfies

supβ0ρn,snEβ0Π~(β:|Sβ|ρn|Sβ0|)0\sup_{\beta_{0}\in\mathcal{B}_{\rho_{n},s_{n}}}E_{\beta_{0}}\tilde{\Pi}\left(\beta:|S_{\beta}|\geq\rho_{n}|S_{\beta_{0}}|\right)\to 0

for any ρn\rho_{n}\to\infty (arbitrarily slowly) and ρn,sn\mathcal{B}_{\rho_{n},s_{n}} defined in (23).

To prove Theorems 3 and 4, we use the following result which relates the VB probability of sets having exponentially small probability under the true posterior.

Lemma 1 (Theorem 5 of Ray and Szabó (2022)).

Let Bn{B}_{n} be a subset of the parameter space, AnA_{n} be events and QnQ_{n} be distributions for β\beta. If there exists C>0C>0 and δn>0\delta_{n}>0 such that

Eβ0Π(βBnc|Y)1AnCeδn,\displaystyle E_{\beta_{0}}\Pi(\beta\in{B}_{n}^{c}|Y)1_{A_{n}}\leq Ce^{-\delta_{n}},

then

Eβ0Qn(βBnc)1An2δn[Eβ0DKL(QnΠ(|Y))1An+Ceδ/2].E_{\beta_{0}}Q_{n}(\beta\in{B}_{n}^{c})1_{A_{n}}\leq\frac{2}{\delta_{n}}\left[E_{\beta_{0}}D_{\text{KL}}(Q_{n}\|\Pi(\cdot|Y))1_{A_{n}}+Ce^{-\delta/2}\right].

The proof thus reduces to finding events AnA_{n} with Pβ0(An)1P_{\beta_{0}}(A_{n})\to 1 on which:

  1. 1.

    The true posterior places only exponentially small probability outside BnB_{n}, that is Π(Bnc|Y)Ceδn\Pi(B_{n}^{c}|Y)\leq Ce^{-\delta_{n}} for some rate δn\delta_{n}\to\infty,

  2. 2.

    The DKLD_{\text{KL}}-divergence between the VB posterior Π~\tilde{\Pi} and the full posterior is o(δn)o(\delta_{n}).

In our setting, we shall take δn=Cs0logM\delta_{n}=Cs_{0}\log M. Full posterior results are dealt with in Section F.2, the DKLD_{\text{KL}}-divergence in Section F.3 and the proofs of Theorems 3 and 4 are completed in Section F.4.

F.2 Asymptotic theory for the full posterior

We now establish contraction rates for the full computationally expensive posterior distribution, keeping track of the exponential tail bounds needed to apply Lemma 1. While the proofs in this section largely follow those in Castillo et al. Castillo et al. (2015), the precise arguments adapting these results to the group sparse setting are rather technical and hence we provide them for convenience. We first establish a Gaussian tail bound in terms of the group structure.

Lemma 2.

The event

𝒯0={maxk=1,,MXGkT(YXβ0)23XlogM}\mathcal{T}_{0}=\left\{\max_{k=1,\dots,M}\|X_{G_{k}}^{T}(Y-X\beta_{0})\|_{2}\leq 3\|X\|\sqrt{\log M}\right\} (39)

satisfies supβ0pPβ0(𝒯0c)2/M\sup_{\beta_{0}\in\mathbb{R}^{p}}P_{\beta_{0}}(\mathcal{T}_{0}^{c})\leq 2/M.

Proof.

Since YXβ0=εNn(0,In)Y-X\beta_{0}=\varepsilon\sim N_{n}(0,I_{n}) under Pβ0P_{\beta_{0}}, applying a union bound over the group structure yields Pβ0(maxkXGkT(YXβ0)2>t)k=1MPβ0(XGkTε2>t)P_{\beta_{0}}(\max_{k}\|X_{G_{k}}^{T}(Y-X\beta_{0})\|_{2}>t)\leq\sum_{k=1}^{M}P_{\beta_{0}}(\|X_{G_{k}}^{T}\varepsilon\|_{2}>t). Recall that for a multivariate normal WNm(0,Σ)W\sim N_{m}(0,\Sigma), we have P(W2EW2x)2exp(x2/(2EW22))P(\|W\|_{2}-E\|W\|_{2}\geq x)\leq 2\exp(-x^{2}/(2E\|W\|_{2}^{2})) by Corollary 3 of Pinelis and Sakhanenko (1986). Since XGkTεNmk(0,XGkTXGk)X_{G_{k}}^{T}\varepsilon\sim N_{m_{k}}(0,X_{G_{k}}^{T}X_{G_{k}}), this implies

(EXGkTε2)2EXGkTε22=Tr(XGkTXGk)X2,(E\|X_{G_{k}}^{T}\varepsilon\|_{2})^{2}\leq E\|X_{G_{k}}^{T}\varepsilon\|_{2}^{2}=\text{Tr}(X_{G_{k}}^{T}X_{G_{k}})\leq\|X\|^{2},

and hence P(XGkTε2X+x)2exp(x2/(2X2))P(\|X_{G_{k}}^{T}\varepsilon\|_{2}\geq\|X\|+x)\leq 2\exp(-x^{2}/(2\|X\|^{2})) for any x>0x>0. Substituting this bound with x=2XlogMx=2\|X\|\sqrt{\log M} into the above union bound thus yields

Pβ0(maxkXGkT(YXβ0)>3XlogM)2Me2logM=2M1.P_{\beta_{0}}(\max_{k}\|X_{G_{k}}^{T}(Y-X\beta_{0})\|_{\infty}>3\|X\|\sqrt{\log M})\leq 2Me^{-2\log M}=2M^{-1}.

Let n(β)\ell_{n}(\beta) denote the log-likelihood of the Nn(Xβ,In)N_{n}(X\beta,I_{n})-distribution. For any βp\beta\in\mathbb{R}^{p}, we have log-likelihood ratio

Λβ(Y)=en(β)n(β0)=e12X(ββ0)22+(YXβ0)TX(ββ0).\Lambda_{\beta}(Y)=e^{\ell_{n}(\beta)-\ell_{n}(\beta_{0})}=e^{-\frac{1}{2}\|X(\beta-\beta_{0})\|_{2}^{2}+(Y-X\beta_{0})^{T}X(\beta-\beta_{0})}. (40)

The next result establishes an almost sure lower bound on the denominator of the Bayes formula. It follows Lemma 2 of Castillo et al. (2015), but must be adapted to account for the uneven prior normalizing factors coming from the group structure.

Lemma 3.

Suppose Assumption (K) holds and that β0p\beta_{0}\in\mathbb{R}^{p} satisfies mmaxlogs0KlogMm_{\text{max}}\log s_{0}\leq K^{\prime}\log M for K>0K^{\prime}>0. Then it holds that with Pβ0P_{\beta_{0}}-probability one,

Λβ(Y)𝑑Π(β)\displaystyle\int\Lambda_{\beta}(Y)d\Pi(\beta) Cπm(s0)s0!eλβ02,1ecs0logM,\displaystyle\geq C\pi_{m}(s_{0})s_{0}!e^{-\lambda\|\beta_{0}\|_{2,1}}e^{-cs_{0}\log M},

where C,c>0C,c>0 depend only on K,KK,K^{\prime}.

The mild condition mmaxlogs0KlogMm_{\text{max}}\log s_{0}\leq K^{\prime}\log M, which will be assumed throughout, relates the true sparsity with the maximal group size. As in Assumption (K), the constant KK^{\prime} is used to provide uniformity, but this can be ignored at first reading.

Proof.

The bound trivially holds true for s0=0s_{0}=0, hence we assume s01s_{0}\geq 1. Write p0:=GkS0mkp_{0}:=\sum_{G_{k}\in S_{0}}m_{k} to be maximal number of non-zero coefficients in β0\beta_{0}. In an abuse of notation, we shall sometimes interchangeably use βS0\beta_{S_{0}} for both the vector in p0\mathbb{R}^{p_{0}} and the vector in p\mathbb{R}^{p} with the entries in S0cS_{0}^{c} set to zero. Using the form (36) of the prior,

Λβ(Y)𝑑Π(β)\displaystyle\int\Lambda_{\beta}(Y)d\Pi(\beta) πM(s0)(Ms0)p0Λβ(Y)GkS0Ψλ,mk(βGk)dβGk.\displaystyle\geq\frac{\pi_{M}(s_{0})}{{M\choose s_{0}}}\int_{\mathbb{R}^{p_{0}}}\Lambda_{\beta}(Y)\prod_{G_{k}\in S_{0}}\Psi_{\lambda,m_{k}}(\beta_{G_{k}})d\beta_{G_{k}}.

By the change of variable bS0=βS0β0,S0b_{S_{0}}=\beta_{S_{0}}-\beta_{0,S_{0}} and the form of the log-likelihood (40), the last display is lower bounded by

πM(s0)(Ms0)eλβ02,1p0e12XS0bS022+(YXβ0)TXS0bS0GkS0Ψλ,mk(bGk)dbGk.\displaystyle\frac{\pi_{M}(s_{0})}{{M\choose s_{0}}}e^{-\lambda\|\beta_{0}\|_{2,1}}\int_{\mathbb{R}^{p_{0}}}e^{-\frac{1}{2}\|X_{S_{0}}b_{S_{0}}\|_{2}^{2}+(Y-X\beta_{0})^{T}X_{S_{0}}b_{S_{0}}}\prod_{G_{k}\in S_{0}}\Psi_{\lambda,m_{k}}(b_{G_{k}})db_{G_{k}}.

Define the measure μ\mu on p0\mathbb{R}^{p_{0}} by dμ(bS0)=e12XS0bS022GkS0Ψλ,mk(bGk)dbGkd\mu(b_{S_{0}})=e^{-\frac{1}{2}\|X_{S_{0}}b_{S_{0}}\|_{2}^{2}}\prod_{G_{k}\in S_{0}}\Psi_{\lambda,m_{k}}(b_{G_{k}})db_{G_{k}}. Let μ¯=μ/μ(p0)\bar{\mu}=\mu/\mu(\mathbb{R}^{p_{0}}) denote the normalized probability measure with corresponding expectation Eμ¯E_{\bar{\mu}}. Defining Z(bS0)=(YXβ0)TXS0bS0Z(b_{S_{0}})=(Y-X\beta_{0})^{T}X_{S_{0}}b_{S_{0}}, Jensen’s inequality implies Eμ¯eZeEμ¯Z=1E_{\bar{\mu}}e^{Z}\geq e^{E_{\bar{\mu}}Z}=1, since Eμ¯Z=0E_{\bar{\mu}}Z=0 as μ¯\bar{\mu} is a symmetric probability distribution about zero. Thus the last display is lower bounded by πm(s0)eλβ02,1μ(p0)/(Ms0)\pi_{m}(s_{0})e^{-\lambda\|\beta_{0}\|_{2,1}}\mu(\mathbb{R}^{p_{0}})/{M\choose s_{0}}, which equals

πM(s0)(Ms0)eλβ02,1p0e12XS0bS022GkS0Ψλ,mk(bGk)dbGk.\displaystyle\frac{\pi_{M}(s_{0})}{{M\choose s_{0}}}e^{-\lambda\|\beta_{0}\|_{2,1}}\int_{\mathbb{R}^{p_{0}}}e^{-\frac{1}{2}\|X_{S_{0}}b_{S_{0}}\|_{2}^{2}}\prod_{G_{k}\in S_{0}}\Psi_{\lambda,m_{k}}(b_{G_{k}})db_{G_{k}}. (41)

Using the group structure, Xb2=k=1MXGkbGk2k=1MXGk2bGk2Xb1\|Xb\|_{2}=\|\sum_{k=1}^{M}X_{\cdot G_{k}}b_{G_{k}}\|_{2}\leq\sum_{k=1}^{M}\|X_{\cdot G_{k}}\|_{2}\|b_{G_{k}}\|_{2}\leq\|X\|\|b\|_{1} since b2b1\|b\|_{2}\leq\|b\|_{1}. Using the form (35) of the density Ψλ,mk\Psi_{\lambda,m_{k}} and recalling that

βS1r(λ/2)|S|eλβS1𝑑βSeλr(λr)|S|/|S|!\int_{\|\beta_{S}\|_{1}\leq r}(\lambda/2)^{|S|}e^{-\lambda\|\beta_{S}\|_{1}}d\beta_{S}\geq e^{-\lambda r}(\lambda r)^{|S|}/|S|!

by (6.2) in Castillo et al. (2015), the integral in the last display is bounded below by

e1/2XbS011GkS0ΔmkλmkeλβGk1dβGke1/2eλ/Xλp0Xp0p0!GkS0Δmk2mk.\begin{split}&e^{-1/2}\int_{\|X\|\|b_{S_{0}}\|_{1}\leq 1}\prod_{G_{k}\in S_{0}}\Delta_{m_{k}}\lambda^{m_{k}}e^{-\lambda\|\beta_{G_{k}}\|_{1}}d\beta_{G_{k}}\\ &\qquad\geq e^{-1/2}e^{-\lambda/\|X\|}\frac{\lambda^{p_{0}}}{\|X\|^{p_{0}}p_{0}!}\prod_{G_{k}\in S_{0}}\Delta_{m_{k}}2^{m_{k}}.\end{split} (42)

Deviating from Castillo et al. (2015), we must now take careful account of the normalizing constants Δmk\Delta_{m_{k}}.

Recall the form of the normalized constants Δmk=2mkπ(1mk)/2Γ((mk+1)/2)1\Delta_{m_{k}}=2^{-m_{k}}\pi^{(1-m_{k})/2}\Gamma((m_{k}+1)/2)^{-1}. The non-asymptotic upper bound in Stirling’s approximation for the Gamma function gives for z2z\geq 2: Γ(z)2π(z1)(z1e)z1e112(z1).\Gamma(z)\leq\sqrt{2\pi(z-1)}\left(\frac{z-1}{e}\right)^{z-1}e^{\frac{1}{12(z-1)}}. Taking z=(mk+1)/22z=(m_{k}+1)/2\geq 2 for mk3m_{k}\geq 3,

Γ((mk+1)/2)2π(mk1)(mk12e)(mk1)/22e1/2πmkmk/21(2e)(mk1)/2,\displaystyle\Gamma((m_{k}+1)/2)\leq 2\sqrt{\pi(m_{k}-1)}\left(\frac{m_{k}-1}{2e}\right)^{(m_{k}-1)/2}\leq 2e^{-1/2}\sqrt{\pi}m_{k}^{m_{k}/2}\frac{1}{(2e)^{(m_{k}-1)/2}}, (43)

where we have used that (mk1mk)mk/2limx(11x)x/2=e1/2(\frac{m_{k}-1}{m_{k}})^{m_{k}/2}\leq\lim_{x\to\infty}(1-\frac{1}{x})^{x/2}=e^{-1/2} since the function in the limit is strictly increasing on (1,)(1,\infty). One can directly verify that the upper bound (43) also holds for mk=1,2m_{k}=1,2. Using (43),

GkS0Δmk2mk=GkS0π(1mk)/2Γ((mk+1)/2)\displaystyle\prod_{G_{k}\in S_{0}}\Delta_{m_{k}}2^{m_{k}}=\prod_{G_{k}\in S_{0}}\frac{\pi^{(1-m_{k})/2}}{\Gamma((m_{k}+1)/2)} 1(2e1)s0(2πe)p0/2GkS0mkmk/2\displaystyle\geq\frac{1}{(\sqrt{2}e^{-1})^{s_{0}}(2\pi e)^{p_{0}/2}}\prod_{G_{k}\in S_{0}}m_{k}^{-m_{k}/2} cp0mmaxp0/2\displaystyle\geq c^{p_{0}}\ m_{\text{max}}^{-p_{0}/2}

for some universal constant c>0c>0. Using this last display, we lower bound (42) by a constant multiple of

eλ/X(λX)p01p0!cp0mmaxp0/2.\displaystyle e^{-\lambda/\|X\|}\left(\frac{\lambda}{\|X\|}\right)^{p_{0}}\frac{1}{p_{0}!}c^{p_{0}}m_{\text{max}}^{-p_{0}/2}.

Using (38), if λ/X1/2\lambda/\|X\|\leq 1/2, then eλ/X(λ/X)p0e1/2Mp0/mmaxe1/2es0logMe^{-\lambda/\|X\|}(\lambda/\|X\|)^{p_{0}}\geq e^{-1/2}M^{-p_{0}/m_{\text{max}}}\geq e^{-1/2}e^{-s_{0}\log M}, while if λ/X1/2\lambda/\|X\|\geq 1/2, then eλ/X(λ/X)p0e6logM2p0eC(K)s0logMe^{-\lambda/\|X\|}(\lambda/\|X\|)^{p_{0}}\geq e^{-6\sqrt{\log M}}2^{-p_{0}}\geq e^{-C(K)s_{0}\log M} since mmaxKlogMm_{\text{max}}\leq K\log M by assumption. Since also mmaxp0/2/p0!e2p0logp0e2mmaxs0log(mmaxs0)m_{\text{max}}^{-p_{0}/2}/p_{0}!\geq e^{-2p_{0}\log p_{0}}\geq e^{-2m_{\text{max}}s_{0}\log(m_{\text{max}}s_{0})}, the last display is lower bounded by eCs0logMe^{-Cs_{0}\log M} under the lemma’s hypotheses. The result then follows by substituting this lower bound for the integral in (41) and using that (Ms0)Ms0/s0!=es0logM/s0!{M\choose s_{0}}\leq M^{s_{0}}/s_{0}!=e^{s_{0}\log M}/s_{0}!. ∎

The next result follows Theorem 10 of Castillo et al. (2015).

Lemma 4 (Dimension).

Suppose that Assumption (K) holds and the prior satisfies (37) and (38). Further assume M>0M>0 is large enough that logMmax{4mmaxlog4A2,4logA2A2}\log M\geq\max\{\frac{4m_{\text{max}}\log 4}{A_{2}},\frac{4\log A_{2}}{A_{2}}\} and M(4mmax+1/2A2)1/A4M\geq(4^{m_{\text{max}}+1/2}A_{2})^{1/A_{4}}. Then for any β0p\beta_{0}\in\mathbb{R}^{p} such that mmaxlogs0KlogMm_{\text{max}}\log s_{0}\leq K^{\prime}\log M and any L>0L>0,

Eβ0Π(β:|Sβ|(L+1)s0|Y)1𝒯0\displaystyle E_{\beta_{0}}\Pi\left(\left.\beta:|S_{\beta}|\geq(L+1)s_{0}\right|Y\right)1_{\mathcal{T}_{0}}
C(K,K)exp{(c(K,K,A2)+144ϕ(S0)2LA42)s0logM},\displaystyle\qquad\quad\leq C(K,K^{\prime})\exp\left\{\left(c(K,K^{\prime},A_{2})+\frac{144}{\phi(S_{0})^{2}}-\frac{LA_{4}}{2}\right)s_{0}\log M\right\},

where s0=|Sβ0|s_{0}=|S_{\beta_{0}}| and 𝒯0\mathcal{T}_{0} is the event (39).

Proof.

Using Bayes formula with likelihood ratio (40) and Lemma 3, for any measurable set BpB\subset\mathbb{R}^{p},

Π(B|Y)1𝒯0=1𝒯0BΛβ(Y)𝑑Π(β)Λβ(Y)𝑑Π(β)Ceλβ02,1+cs0logMs0!πM(s0)B1𝒯0e12X(ββ0)22+(YXβ0)TX(ββ0)𝑑Π(β)\begin{split}\Pi(B|Y)1_{\mathcal{T}_{0}}&=1_{\mathcal{T}_{0}}\frac{\int_{B}\Lambda_{\beta}(Y)d\Pi(\beta)}{\int\Lambda_{\beta}(Y)d\Pi(\beta)}\\ &\leq C\frac{e^{\lambda\|\beta_{0}\|_{2,1}+cs_{0}\log M}}{s_{0}!\pi_{M}(s_{0})}\int_{B}1_{\mathcal{T}_{0}}e^{-\frac{1}{2}\|X(\beta-\beta_{0})\|_{2}^{2}+(Y-X\beta_{0})^{T}X(\beta-\beta_{0})}d\Pi(\beta)\end{split} (44)

for 𝒯0\mathcal{T}_{0} the event defined in (39). Applying Cauchy-Schwarz on the event 𝒯0\mathcal{T}_{0} gives

|(YXβ0)TX(ββ0)|maxk=1,,MXGkT(YXβ0)2k=1MβGkβ0,Gk2λ¯ββ02,1.|(Y-X\beta_{0})^{T}X(\beta-\beta_{0})|\leq\max_{k=1,\dots,M}\|X_{G_{k}}^{T}(Y-X\beta_{0})\|_{2}\sum_{k=1}^{M}\|\beta_{G_{k}}-\beta_{0,G_{k}}\|_{2}\leq\bar{\lambda}\|\beta-\beta_{0}\|_{2,1}. (45)

Therefore, since (YXβ)TX(ββ0)N(0,X(ββ0)22)(Y-X\beta)^{T}X(\beta-\beta_{0})\sim N(0,\|X(\beta-\beta_{0})\|_{2}^{2}) under Pβ0P_{\beta_{0}}, on 𝒯0\mathcal{T}_{0}, the integrand in the second last display is bounded by

e12X(ββ0)22Eβ0[1𝒯0e(1λ2λ¯)(YXβ0)TX(ββ0)]eλ2ββ02,1\displaystyle e^{-\frac{1}{2}\|X(\beta-\beta_{0})\|_{2}^{2}}E_{\beta_{0}}[1_{\mathcal{T}_{0}}e^{(1-\frac{\lambda}{2\bar{\lambda}})(Y-X\beta_{0})^{T}X(\beta-\beta_{0})}]e^{\frac{\lambda}{2}\|\beta-\beta_{0}\|_{2,1}} e12[1(1λ/(2λ¯))2]X(ββ0)22eλ2ββ02,1\displaystyle\leq e^{-\frac{1}{2}[1-(1-\lambda/(2\bar{\lambda}))^{2}]\|X(\beta-\beta_{0})\|_{2}^{2}}e^{\frac{\lambda}{2}\|\beta-\beta_{0}\|_{2,1}}

and hence

Eβ0Π(B|Y)1𝒯0Ceλβ02,1+cs0logMs0!πM(s0)Be12[1(1λ/(2λ¯))2]X(ββ0)22eλ2ββ02,1𝑑Π(β).\displaystyle E_{\beta_{0}}\Pi(B|Y)1_{\mathcal{T}_{0}}\leq C\frac{e^{\lambda\|\beta_{0}\|_{2,1}+cs_{0}\log M}}{s_{0}!\pi_{M}(s_{0})}\int_{B}e^{-\frac{1}{2}[1-(1-\lambda/(2\bar{\lambda}))^{2}]\|X(\beta-\beta_{0})\|_{2}^{2}}e^{\frac{\lambda}{2}\|\beta-\beta_{0}\|_{2,1}}d\Pi(\beta).

Arguing exactly as on p. 2007-8 in Castillo et al. (2015),

β02,1+12ββ02,118λ¯X(ββ0)22+8s0λ¯X2ϕ(S0)214ββ02,1+β2,1\|\beta_{0}\|_{2,1}+\frac{1}{2}\|\beta-\beta_{0}\|_{2,1}\leq\frac{1}{8\bar{\lambda}}\|X(\beta-\beta_{0})\|_{2}^{2}+\frac{8s_{0}\bar{\lambda}}{\|X\|^{2}\phi(S_{0})^{2}}-\frac{1}{4}\|\beta-\beta_{0}\|_{2,1}+\|\beta\|_{2,1}

and hence

Eβ0Π(B|Y)1𝒯0Cecs0logMs0!πM(s0)e8s0λ¯λX2ϕ(S0)2Beλ4ββ02,1+λβ2,1𝑑Π(β).\displaystyle E_{\beta_{0}}\Pi(B|Y)1_{\mathcal{T}_{0}}\leq C\frac{e^{cs_{0}\log M}}{s_{0}!\pi_{M}(s_{0})}e^{\frac{8s_{0}\bar{\lambda}\lambda}{\|X\|^{2}\phi(S_{0})^{2}}}\int_{B}e^{-\frac{\lambda}{4}\|\beta-\beta_{0}\|_{2,1}+\lambda\|\beta\|_{2,1}}d\Pi(\beta). (46)

Setting now B={β:|Sβ|>R}B=\{\beta:|S_{\beta}|>R\} with Rs0R\geq s_{0}, the integral in the last display is bounded by

S:|S|>RπM(|S|)(M|S|)eλ4ββ02,1GkSΔmkλmkdβGk=S:|S|>RπM(|S|)(M|S|)GkS4mk.\displaystyle\sum_{S:|S|>R}\frac{\pi_{M}(|S|)}{{M\choose|S|}}\int e^{-\frac{\lambda}{4}\|\beta-\beta_{0}\|_{2,1}}\prod_{G_{k}\in S}\Delta_{m_{k}}\lambda^{m_{k}}d\beta_{G_{k}}=\sum_{S:|S|>R}^{\infty}\frac{\pi_{M}(|S|)}{{M\choose|S|}}\prod_{G_{k}\in S}4^{m_{k}}.

Using the prior condition (37), this is then bounded by

s=R+1MπM(s)4mmaxsπM(s0)4mmaxs0(4mmaxA2MA4)R+1s0j=0(4mmaxA2MA4)j.\displaystyle\sum_{s=R+1}^{M}\pi_{M}(s)4^{m_{\text{max}}s}\leq\pi_{M}(s_{0})4^{m_{\text{max}}s_{0}}\left(\frac{4^{m_{\text{max}}}A_{2}}{M^{A_{4}}}\right)^{R+1-s_{0}}\sum_{j=0}^{\infty}\left(\frac{4^{m_{\text{max}}}A_{2}}{M^{A_{4}}}\right)^{j}.

Since 4mmaxA2/MA41/24^{m_{\text{max}}}A_{2}/M^{A_{4}}\leq 1/2 by the lemma hypothesis, the last sum is bounded by 2 and hence the first term in (46) is bounded by

2Cexp{cs0logM+8s0λλ¯X2ϕ(S0)2+mmaxs0log4+(R+1s0)log(4mmaxA2/MA4)},\displaystyle 2C\exp\left\{cs_{0}\log M+\frac{8s_{0}\lambda\bar{\lambda}}{\|X\|^{2}\phi(S_{0})^{2}}+m_{\text{max}}s_{0}\log 4+(R+1-s_{0})\log(4^{m_{\text{max}}}A_{2}/M^{A_{4}})\right\},

where C,c>0C,c>0 depend only on K,K>0K,K^{\prime}>0. Taking R=(L+1)s01R=(L+1)s_{0}-1, using the lemma hypotheses and that λ2λ¯=6XlogM\lambda\leq 2\bar{\lambda}=6\|X\|\sqrt{\log M}, the last display is bounded by

2Cexp{(c+144ϕ(S0)2+(L+1)mmaxlog4logM+LlogA2logMLA4)s0logM}.\displaystyle 2C\exp\left\{\left(c+\frac{144}{\phi(S_{0})^{2}}+\frac{(L+1)m_{\text{max}}\log 4}{\log M}+\frac{L\log A_{2}}{\log M}-LA_{4}\right)s_{0}\log M\right\}.

For M>0M>0 large enough that logMmax{4mmaxlog4A2,2logA2A2}\log M\geq\max\{\frac{4m_{\text{max}}\log 4}{A_{2}},\frac{2\log A_{2}}{A_{2}}\}, this is then bounded by

2Cexp{(c+144ϕ(S0)2+A24LA4/2)s0logM}.\displaystyle 2C\exp\left\{\left(c+\frac{144}{\phi(S_{0})^{2}}+\frac{A_{2}}{4}-LA_{4}/2\right)s_{0}\log M\right\}.

We next obtain a contraction rate for the full posterior underlying the variational approximation. The proof follows that of Theorem 3 of Castillo et al. (2015), modified for the group setting.

Lemma 5 (Contraction).

Suppose that Assumption (K) holds and the prior satisfies (37) and (38). Further assume M>0M>0 is large enough that logMmax{4mmaxlog4A2,4logA2A2}\log M\geq\max\{\frac{4m_{\text{max}}\log 4}{A_{2}},\frac{4\log A_{2}}{A_{2}}\} and M(4mmax+1/2A2)1/A4M\geq(4^{m_{\text{max}}+1/2}A_{2})^{1/A_{4}}. Then there exists a constant H0=H0(A1,A3,A4)>0H_{0}=H_{0}(A_{1},A_{3},A_{4})>0 such that for any β0p\beta_{0}\in\mathbb{R}^{p} such that mmaxlogs0KlogMm_{\text{max}}\log s_{0}\leq K^{\prime}\log M and any L>0L>0,

Eβ0Π(β:X(ββ0)2H0(L+2)s0logMϕ¯((L+2)s0)|Y)1𝒯0\displaystyle E_{\beta_{0}}\Pi\left(\left.\beta:\|X(\beta-\beta_{0})\|_{2}\geq\frac{H_{0}\sqrt{(L+2)s_{0}\log M}}{\bar{\phi}((L+2)s_{0})}\right|Y\right)1_{\mathcal{T}_{0}}
C(K,K)exp{(c(K,K,A2)+144ϕ(S0)2LA42)s0logM},\displaystyle\qquad\leq C(K,K^{\prime})\exp\left\{\left(c(K,K^{\prime},A_{2})+\frac{144}{\phi(S_{0})^{2}}-\frac{LA_{4}}{2}\right)s_{0}\log M\right\},

where s0=|Sβ0|s_{0}=|S_{\beta_{0}}| and 𝒯0\mathcal{T}_{0} is the event (39). Moreover, both

Eβ0Π(β:ββ02,1H0(L+2)s0logMXϕ¯((L+2)s0)2|Y)1𝒯0,E_{\beta_{0}}\Pi\left(\left.\beta:\|\beta-\beta_{0}\|_{2,1}\geq\frac{H_{0}(L+2)s_{0}\sqrt{\log M}}{\|X\|\bar{\phi}((L+2)s_{0})^{2}}\right|Y\right)1_{\mathcal{T}_{0}},
Eβ0Π(β:ββ02H0(L+2)s0logMXϕ~((L+2)s0)2|Y)1𝒯0,E_{\beta_{0}}\Pi\left(\left.\beta:\|\beta-\beta_{0}\|_{2}\geq\frac{H_{0}\sqrt{(L+2)s_{0}\log M}}{\|X\|\widetilde{\phi}((L+2)s_{0})^{2}}\right|Y\right)1_{\mathcal{T}_{0}},

satisfy the same inequality.

Proof.

Consider the set E=EL={βp:|Sβ|(L+1)s0}E=E_{L}=\{\beta\in\mathbb{R}^{p}:|S_{\beta}|\leq(L+1)s_{0}\}, which satisfies

Eβ0Π(Ec|Y)1𝒯0C(K,K)exp{(c(K,K,A2)+144ϕ(S0)2LA42)s0logM}E_{\beta_{0}}\Pi(E^{c}|Y)1_{\mathcal{T}_{0}}\leq C(K,K^{\prime})\exp\left\{\left(c(K,K^{\prime},A_{2})+\frac{144}{\phi(S_{0})^{2}}-\frac{LA_{4}}{2}\right)s_{0}\log M\right\} (47)

by Lemma 4. Recall that λβ02,12λ¯ββ02,1+λβ2,1\lambda\|\beta_{0}\|_{2,1}\leq 2\bar{\lambda}\|\beta-\beta_{0}\|_{2,1}+\lambda\|\beta\|_{2,1} by (38). Using this, for any set BEB\subseteq E and 𝒯0={maxkXGkT(YXβ0)2λ¯}\mathcal{T}_{0}=\{\max_{k}\|X_{G_{k}}^{T}(Y-X\beta_{0})\|_{2}\leq\bar{\lambda}\}, (44) and (45) give

Π(B|Y)1𝒯0Cecs0logMs0!πM(s0)B1𝒯0e12X(ββ0)22+3λ¯ββ02,1+λβ2,1𝑑Π(β).\displaystyle\Pi(B|Y)1_{\mathcal{T}_{0}}\leq C\frac{e^{cs_{0}\log M}}{s_{0}!\pi_{M}(s_{0})}\int_{B}1_{\mathcal{T}_{0}}e^{-\frac{1}{2}\|X(\beta-\beta_{0})\|_{2}^{2}+3\bar{\lambda}\|\beta-\beta_{0}\|_{2,1}+\lambda\|\beta\|_{2,1}}d\Pi(\beta).

Note that for any βE\beta\in E, |Sββ0||Sβ|+s0(L+2)s0=:DLs0|S_{\beta-\beta_{0}}|\leq|S_{\beta}|+s_{0}\leq(L+2)s_{0}=:D_{L}s_{0}. Using Definition 2 of the uniform compatibility ϕ¯(s)\bar{\phi}(s),

(41)λ¯ββ02,1\displaystyle(4-1)\bar{\lambda}\|\beta-\beta_{0}\|_{2,1} 4λ¯X(ββ0)2|Sββ0|1/2Xϕ¯(|Sββ0|)λ¯ββ02,1\displaystyle\leq\frac{4\bar{\lambda}\|X(\beta-\beta_{0})\|_{2}|S_{\beta-\beta_{0}}|^{1/2}}{\|X\|\bar{\phi}(|S_{\beta-\beta_{0}}|)}-\bar{\lambda}\|\beta-\beta_{0}\|_{2,1}
18X(ββ0)22+32λ¯2DLs0X2ϕ¯(DLs0)2λ¯ββ02,1.\displaystyle\leq\frac{1}{8}\|X(\beta-\beta_{0})\|_{2}^{2}+\frac{32\bar{\lambda}^{2}D_{L}s_{0}}{\|X\|^{2}\bar{\phi}(D_{L}s_{0})^{2}}-\bar{\lambda}\|\beta-\beta_{0}\|_{2,1}.

Thus for any BEB\subseteq E,

Π(B|Y)1𝒯0Cecs0logMs0!πM(s0)e32λ¯2DLs0X2ϕ¯(DLs0)2B1𝒯0e38X(ββ0)22λ¯ββ02,1+λβ2,1𝑑Π(β).\displaystyle\Pi(B|Y)1_{\mathcal{T}_{0}}\leq C\frac{e^{cs_{0}\log M}}{s_{0}!\pi_{M}(s_{0})}e^{\frac{32\bar{\lambda}^{2}D_{L}s_{0}}{\|X\|^{2}\bar{\phi}(D_{L}s_{0})^{2}}}\int_{B}1_{\mathcal{T}_{0}}e^{-\frac{3}{8}\|X(\beta-\beta_{0})\|_{2}^{2}-\bar{\lambda}\|\beta-\beta_{0}\|_{2,1}+\lambda\|\beta\|_{2,1}}d\Pi(\beta).

Note that by (37), πM(s0)(A1MA3)s0πM(0)\pi_{M}(s_{0})\geq(A_{1}M^{-A_{3}})^{s_{0}}\pi_{M}(0). For B={βE:X(ββ0)2R}B=\{\beta\in E:\|X(\beta-\beta_{0})\|_{2}\geq R\}, the last display implies

Π(B|Y)1𝒯0Cecs0logMs0!A1s0πM(0)MA3s0e32λ¯2DLs0X2ϕ¯(DLs0)2e3R2/8Beλ¯ββ02,1+λβ2,1𝑑Π(β).\displaystyle\Pi(B|Y)1_{\mathcal{T}_{0}}\leq C\frac{e^{cs_{0}\log M}}{s_{0}!A_{1}^{s_{0}}\pi_{M}(0)}M^{A_{3}s_{0}}e^{\frac{32\bar{\lambda}^{2}D_{L}s_{0}}{\|X\|^{2}\bar{\phi}(D_{L}s_{0})^{2}}}e^{-3R^{2}/8}\int_{B}e^{-\bar{\lambda}\|\beta-\beta_{0}\|_{2,1}+\lambda\|\beta\|_{2,1}}d\Pi(\beta).

The last integral is upper bounded by

S:|S|(L+1)s0πM(|S|)(M|S|)eλ¯ββ02,1GkSΔmkλmkdβGk=s=0(L+1)s0πM(s)GkS(λ/λ¯)mk.\displaystyle\sum_{S:|S|\leq(L+1)s_{0}}\frac{\pi_{M}(|S|)}{{M\choose|S|}}\int e^{-\bar{\lambda}\|\beta-\beta_{0}\|_{2,1}}\prod_{G_{k}\in S}\Delta_{m_{k}}\lambda^{m_{k}}d\beta_{G_{k}}=\sum_{s=0}^{(L+1)s_{0}}\pi_{M}(s)\prod_{G_{k}\in S}(\lambda/\bar{\lambda})^{m_{k}}.

Using (37) and (38), the last display is bounded by πM(0)s=0(A2MA4)s2mmaxsπM(0)s=04s2πM(0)\pi_{M}(0)\sum_{s=0}^{\infty}(A_{2}M^{-A_{4}})^{s}2^{m_{\text{max}}s}\leq\pi_{M}(0)\sum_{s=0}^{\infty}4^{-s}\leq 2\pi_{M}(0), since 2mmaxA2MA42mmax+11/42^{m_{\text{max}}}A_{2}M^{-A_{4}}\leq 2^{-m_{\text{max}}+1}\leq 1/4 by the lemma hypothesis. Setting R2=H02DLs0logM/ϕ¯(DLs0)2R^{2}=H_{0}^{2}D_{L}s_{0}\log M/\bar{\phi}(D_{L}s_{0})^{2} for a constant H0>0H_{0}>0, the second last display is bounded by

C(K,K)exp{(c(K,K)+A3logA1logM+288DLϕ¯(DLs0)2)s0logM38R2}\displaystyle C(K,K^{\prime})\exp\left\{\left(c(K,K^{\prime})+A_{3}-\frac{\log A_{1}}{\log M}+\frac{288D_{L}}{\bar{\phi}(D_{L}s_{0})^{2}}\right)s_{0}\log M-\frac{3}{8}R^{2}\right\}
Cexp{[3H028288cA3|logA1|logM]DLs0logMϕ¯(DLs0)2},\displaystyle\quad\leq C\exp\left\{-\left[\frac{3H_{0}^{2}}{8}-288-c-A_{3}-\frac{|\log A_{1}|}{\log M}\right]\frac{D_{L}s_{0}\log M}{\bar{\phi}(D_{L}s_{0})^{2}}\right\},

where we have used that ϕ¯(DLs0)ϕ¯(1)1\bar{\phi}(D_{L}s_{0})\leq\bar{\phi}(1)\leq 1. Taking H0H_{0} large enough depending on A1,A3,A4A_{1},A_{3},A_{4} and using again that ϕ¯(DLs0)1\bar{\phi}(D_{L}s_{0})\leq 1, the last display can be made smaller than (47), which is thus the dominant term in the tail probability bound.

For the 2,1\|\cdot\|_{2,1}-loss, using Definition 2 of the uniform compatibility, for any βE\beta\in E,

λ¯ββ02,1\displaystyle\bar{\lambda}\|\beta-\beta_{0}\|_{2,1} λ¯X(ββ0)2|Sββ0|1/2Xϕ¯(|Sββ0|)12λ¯2DLs0X2ϕ¯(DLs0)2+12X(ββ0)22.\displaystyle\leq\frac{\bar{\lambda}\|X(\beta-\beta_{0})\|_{2}|S_{\beta-\beta_{0}}|^{1/2}}{\|X\|\bar{\phi}(|S_{\beta-\beta_{0}}|)}\leq\frac{1}{2}\frac{\bar{\lambda}^{2}D_{L}s_{0}}{\|X\|^{2}\bar{\phi}(D_{L}s_{0})^{2}}+\frac{1}{2}\|X(\beta-\beta_{0})\|_{2}^{2}.

The result then follows from the first assertion for the prediction loss X(ββ0)2\|X(\beta-\beta_{0})\|_{2}.

For the 2\|\cdot\|_{2}-loss, note that X(ββ0)2ϕ~(DLs0)Xββ02\|X(\beta-\beta_{0})\|_{2}\geq\widetilde{\phi}(D_{L}s_{0})\|X\|\|\beta-\beta_{0}\|_{2} for any βE\beta\in E. Since ϕ~(s)ϕ¯(s)\widetilde{\phi}(s)\leq\bar{\phi}(s) for all ss by Lemma 1 of Castillo et al. (2015) (which extends to the group setting), the result follows. ∎

F.3 DKLD_{\text{KL}}-divergence between the variational and true posterior

We next bound the KL-divergence between the variational family and the true posterior on the following event:

𝒯1=𝒯1(Γ,ε,κ)=𝒯0{Π(β:|Sβ|>Γ|Y)1/4}{Π(β:ββ02>ε|Y)eκ},\mathcal{T}_{1}=\mathcal{T}_{1}(\Gamma,\varepsilon,\kappa)=\mathcal{T}_{0}\cap\{\Pi(\beta:|S_{\beta}|>\Gamma|Y)\leq 1/4\}\cap\{\Pi(\beta:\|\beta-\beta_{0}\|_{2}>\varepsilon|Y)\leq e^{-\kappa}\}, (48)

where Γ,κ,ε>0\Gamma,\kappa,\varepsilon>0. The proof largely follows Section B.2 of Ray and Szabó (2022), again modified to the group sparse setting, which we produce in full for completeness.

Lemma 6.

If 4e1+ΓlogMκ14e^{1+\Gamma\log M-\kappa}\leq 1, then there exists an element Q𝒬𝒬Q\in\mathcal{Q}\subset\mathcal{Q}^{\prime} of the variational families such that

DKL(Q||Π(|Y))1𝒯1\displaystyle D_{\text{KL}}(Q||\Pi(\cdot|Y))1_{\mathcal{T}_{1}} Γ(logM+mmaxlog1ϕ~(Γ))+λΓϕ~(Γ)2(3s01/2ε+3logMX+mmax1/2X)\displaystyle\leq\Gamma\left(\log M+m_{\text{max}}\log\frac{1}{\widetilde{\phi}(\Gamma)}\right)+\frac{\lambda\Gamma}{\widetilde{\phi}(\Gamma)^{2}}\left(3s_{0}^{1/2}\varepsilon+\frac{3\sqrt{\log M}}{\|X\|}+\frac{m_{\text{max}}^{1/2}}{\|X\|}\right)
+log(4e).\displaystyle\qquad+\log(4e).

If Assumption (K) also holds, then

DKL(Q||Π(|Y))1𝒯1\displaystyle D_{\text{KL}}(Q||\Pi(\cdot|Y))1_{\mathcal{T}_{1}} Γ(K+1)logMlog1ϕ~(Γ)+λΓϕ~(Γ)2(3s01/2ε+(3+K)logMX)\displaystyle\leq\Gamma(K+1)\log M\log\frac{1}{\widetilde{\phi}(\Gamma)}+\frac{\lambda\Gamma}{\widetilde{\phi}(\Gamma)^{2}}\left(3s_{0}^{1/2}\varepsilon+\frac{(3+\sqrt{K})\sqrt{\log M}}{\|X\|}\right)
+log(4e).\displaystyle\quad+\log(4e).
Proof.

The full posterior takes the form

Π(|Y)=S:S{1,,M}q^SΠS(|Y)δSc,\Pi(\cdot|Y)=\sum_{S:S\subseteq\{1,\dots,M\}}\hat{q}_{S}\Pi_{S}(\cdot|Y)\otimes\delta_{S^{c}}, (49)

with weights 0q^S10\leq\hat{q}_{S}\leq 1 satisfying Sq^S=1\sum_{S}\hat{q}_{S}=1 and where ΠS(|Y)\Pi_{S}(\cdot|Y) denotes the posterior for βSGkSmk\beta_{S}\in\mathbb{R}^{\sum_{G_{k}\in S}m_{k}} in the restricted model Y=XSβS+ZY=X_{S}\beta_{S}+Z. Arguing exactly as in the proof of Lemma B.2 of Ray and Szabó (2022), one has that on 𝒯1\mathcal{T}_{1} and for 4e1+ΓlogMκ14e^{1+\Gamma\log M-\kappa}\leq 1, there exists a set S~{G1,,GM}\tilde{S}\subseteq\{G_{1},\dots,G_{M}\} such that

|S~|Γ,β0,Sc2ε,q^S~(2e)1MΓ.|\tilde{S}|\leq\Gamma,\qquad\|\beta_{0,S^{c}}\|_{2}\leq\varepsilon,\qquad\hat{q}_{\tilde{S}}\geq(2e)^{-1}M^{-\Gamma}. (50)

Writing p~=GkS~mk\tilde{p}=\sum_{G_{k}\in\tilde{S}}m_{k}, consider the element of the variational family

Q~=(GkS~Nmk(μGk,DGk))(GkS~cδGk)=:Np~(μS~,DS~)δS~c\tilde{Q}=\left(\bigotimes_{G_{k}\in\tilde{S}}N_{m_{k}}(\mu_{G_{k}},D_{G_{k}})\right)\otimes\left(\bigotimes_{G_{k}\in\tilde{S}^{c}}\delta_{G_{k}}\right)=:N_{\tilde{p}}(\mu_{\tilde{S}},D_{\tilde{S}})\otimes\delta_{\tilde{S}^{c}} (51)

where DGkD_{G_{k}} are diagonal matrices to be defined below. This is the distribution that assigns mass one to the model S~\tilde{S} and then fits an independent normal distribution with diagonal covariance on each of the groups in S~\tilde{S}. Since Q~\tilde{Q} is only absolutely continuous with respect to the q^S~ΠS~(|Y)δS~c\hat{q}_{\tilde{S}}\Pi_{\tilde{S}}(\cdot|Y)\otimes\delta_{\tilde{S}^{c}} component of the posterior (49),

infQ𝒬DKL(QΠ(|Y))DKL(Q~Π(|Y)=EβQ~[logdNp~(μS~,DS~)δS~cq^S~dΠS~(|Y)δS~c(β)]=log1q^S~+DKL(Np~(μS~,DS~)ΠS~(|Y)).\begin{split}\inf_{Q\in\mathcal{Q}}D_{\text{KL}}(Q\|\Pi(\cdot|Y))\leq D_{\text{KL}}(\tilde{Q}\|\Pi(\cdot|Y)&=E_{\beta\sim\tilde{Q}}\left[\log\frac{dN_{\tilde{p}}(\mu_{\tilde{S}},D_{\tilde{S}})\otimes\delta_{\tilde{S}^{c}}}{\hat{q}_{\tilde{S}}d\Pi_{\tilde{S}}(\cdot|Y)\otimes\delta_{\tilde{S}^{c}}}(\beta)\right]\\ &=\log\frac{1}{\hat{q}_{\tilde{S}}}+D_{\text{KL}}(N_{\tilde{p}}(\mu_{\tilde{S}},D_{\tilde{S}})\|\Pi_{\tilde{S}}(\cdot|Y)).\end{split} (52)

On 𝒯1\mathcal{T}_{1}, the first term is bounded by log(2eMΓ)\log(2eM^{\Gamma}) by (50), so that it remains to bound the second term in (52).

Define

μS~=(XS~TXS~)1XS~TY,DS~=diag((DGk)GkS~),ΣS~=(XS~TXS~)1,\mu_{\tilde{S}}=(X_{\tilde{S}}^{T}X_{\tilde{S}})^{-1}X_{\tilde{S}}^{T}Y,\qquad D_{\tilde{S}}=\text{diag}((D_{G_{k}})_{G_{k}\in\tilde{S}}),\qquad\Sigma_{\tilde{S}}=(X_{\tilde{S}}^{T}X_{\tilde{S}})^{-1}, (53)

where DGkmk×mkD_{G_{k}}\in\mathbb{R}^{m_{k}\times m_{k}}, GkS~G_{k}\in\tilde{S}, is the diagonal matrix with entries

(DGk)ii=1(XGkTXGk)ii,i=1,,mk.(D_{G_{k}})_{ii}=\frac{1}{(X_{G_{k}}^{T}X_{G_{k}})_{ii}},\qquad\qquad i=1,\dots,m_{k}.

Writing EμS~,DS~E_{\mu_{\tilde{S}},D_{\tilde{S}}} for the expectation under the distribution βS~Np~(μS~,DS~)\beta_{\tilde{S}}\sim N_{\tilde{p}}(\mu_{\tilde{S}},D_{\tilde{S}}),

DKL(Np~(μS~,DS~)ΠS~(|Y))=EμS~,DS~[logdNp~(μS~,DS~)dNp~(μS~,ΣS~)+logdNp~(μS~,ΣS~)dΠS~(|Y)]=:(I)+(II).\begin{split}D_{\text{KL}}(N_{\tilde{p}}(\mu_{\tilde{S}},D_{\tilde{S}})\|\Pi_{\tilde{S}}(\cdot|Y))&=E_{\mu_{\tilde{S}},D_{\tilde{S}}}\left[\log\frac{dN_{\tilde{p}}(\mu_{\tilde{S}},D_{\tilde{S}})}{dN_{\tilde{p}}(\mu_{\tilde{S}},\Sigma_{\tilde{S}})}+\log\frac{dN_{\tilde{p}}(\mu_{\tilde{S}},\Sigma_{\tilde{S}})}{d\Pi_{\tilde{S}}(\cdot|Y)}\right]\\ &=:(I)+(II).\end{split} (54)

We next deal with each term separately.

Term (I) in (54). Using the formula for the Kullback-Leibler divergence between two multivariate Gaussian distributions,

(I)=DKL(Np~(μS~,DS~)Np~(μS~,ΣS~))=12(Tr(ΣS~1DS~)p~+log(|ΣS~|/|DS~|)),(I)=D_{\text{KL}}(N_{\tilde{p}}(\mu_{\tilde{S}},D_{\tilde{S}})\|N_{\tilde{p}}(\mu_{\tilde{S}},\Sigma_{\tilde{S}}))=\tfrac{1}{2}\left(\text{Tr}(\Sigma_{\tilde{S}}^{-1}D_{\tilde{S}})-\tilde{p}+\log(|\Sigma_{\tilde{S}}|/|D_{\tilde{S}}|)\right),

where |A||A| denotes the determinant of a square matrix AA. Further define the matrix

VS~=diag(((XGkTXGk)1)GkS~),V_{\tilde{S}}=\text{diag}(((X_{G_{k}}^{T}X_{G_{k}})^{-1})_{G_{k}\in\tilde{S}}),

which is a block-diagonalization of ΣS~=(XS~TXS~)1\Sigma_{\tilde{S}}=(X_{\tilde{S}}^{T}X_{\tilde{S}})^{-1}. Let S~={Gk1,,Gks}\tilde{S}=\{G_{k_{1}},\dots,G_{k_{s}}\} for s=|S~|s=|\tilde{S}|. By considering multiplication along the block structure, ΣS~1VS~\Sigma_{\tilde{S}}^{-1}V_{\tilde{S}} has (i,j)th(i,j)^{th}-block equal to (XGkiTXGkj)(XGkjTXGkj)1(X_{G_{k_{i}}}^{T}X_{G_{k_{j}}})(X_{G_{k_{j}}}^{T}X_{G_{k_{j}}})^{-1}, i,j=1,,si,j=1,\dots,s. Furthermore, VS~1DS~V_{\tilde{S}}^{-1}D_{\tilde{S}} is a block-diagonal matrix with (i,i)th(i,i)^{th}-block (XGkiTXGki)DGki(X_{G_{k_{i}}}^{T}X_{G_{k_{i}}})D_{G_{k_{i}}}, i=1,,si=1,\dots,s. Thus, by considering the block-diagonal terms,

Tr(ΣS~1DS~)=Tr(ΣS~1VS~VS~1DS~)\displaystyle\text{Tr}(\Sigma_{\tilde{S}}^{-1}D_{\tilde{S}})=\text{Tr}(\Sigma_{\tilde{S}}^{-1}V_{\tilde{S}}V_{\tilde{S}}^{-1}D_{\tilde{S}}) =Tr(diag((XGkiTXGkDGk)GkS~))\displaystyle=\text{Tr}\left(\text{diag}((X_{G_{k_{i}}}^{T}X_{G_{k}}D_{G_{k}})_{G_{k}\in\tilde{S}})\right)
=GkS~Tr(XGkiTXGkiDGki)=GkS~mk=p~,\displaystyle=\sum_{G_{k}\in\tilde{S}}\text{Tr}\left(X_{G_{k_{i}}}^{T}X_{G_{k_{i}}}D_{G_{k_{i}}}\right)=\sum_{G_{k}\in\tilde{S}}m_{k}=\tilde{p},

and hence (I)=12log(|ΣS~||DS~1|)(I)=\tfrac{1}{2}\log(|\Sigma_{\tilde{S}}||D_{\tilde{S}}^{-1}|). Using (53),

|DS~1|=GkS~iGk(XGkTXGk)iiGkS~X2mk=X2p~.|D_{\tilde{S}}^{-1}|=\prod_{G_{k}\in\tilde{S}}\prod_{i\in G_{k}}(X_{G_{k}}^{T}X_{G_{k}})_{ii}\leq\prod_{G_{k}\in\tilde{S}}\|X\|^{2m_{k}}=\|X\|^{2\tilde{p}}.

Let λmax(A)\lambda_{max}(A) and λmin(A)\lambda_{min}(A) denote the largest and smallest eigenvalues, respectively, of a matrix AA. Arguing as in equation (B.12) in Ray and Szabó (2022), for any S{G1,,GM}S\subseteq\{G_{1},\dots,G_{M}\},

λmin(XSTXS)X2ϕ~(|S|)2.\lambda_{min}(X_{S}^{T}X_{S})\geq\|X\|^{2}\widetilde{\phi}(|S|)^{2}. (55)

Therefore,

|ΣS~|=1/|XS~TXS~|1/λmin(XS~TXS~)p~1/(Xϕ~(|S~|))2p~,|\Sigma_{\tilde{S}}|=1/|X_{\tilde{S}}^{T}X_{\tilde{S}}|\leq 1/\lambda_{\min}(X_{\tilde{S}}^{T}X_{\tilde{S}})^{\tilde{p}}\leq 1/(\|X\|\widetilde{\phi}(|\tilde{S}|))^{2\tilde{p}},

and hence (I)=12log(|ΣS~||DS~1|)p~log(1/ϕ~(|S~|))mmaxΓlog(1/ϕ~(Γ))(I)=\tfrac{1}{2}\log(|\Sigma_{\tilde{S}}||D_{\tilde{S}}^{-1}|)\leq\tilde{p}\log(1/\widetilde{\phi}(|\tilde{S}|))\leq m_{\text{max}}\Gamma\log(1/\widetilde{\phi}(\Gamma)) using (50).

Term (II) in (54). One can check that the Np~(μS~,ΣS~)N_{\tilde{p}}(\mu_{\tilde{S}},\Sigma_{\tilde{S}}) distribution has density function proportional to e12YXS~βS~22e^{-\frac{1}{2}\|Y-X_{\tilde{S}}\beta_{\tilde{S}}\|_{2}^{2}} for βS~p~\beta_{\tilde{S}}\in\mathbb{R}^{\tilde{p}}. Therefore,

(II)\displaystyle(II) =EμS~,DS~[logDΠe12YXS~βS~22λβ0,S~2,1DNe12YXS~βS~22λβS~2,1]\displaystyle=E_{\mu_{\tilde{S}},D_{\tilde{S}}}\left[\log\frac{D_{\Pi}e^{-\frac{1}{2}\|Y-X_{\tilde{S}}\beta_{\tilde{S}}\|_{2}^{2}-\lambda\|\beta_{0,\tilde{S}}\|_{2,1}}}{D_{N}e^{-\frac{1}{2}\|Y-X_{\tilde{S}}\beta_{\tilde{S}}\|_{2}^{2}-\lambda\|\beta_{\tilde{S}}\|_{2,1}}}\right]
=log(DΠ/DN)+λEμS~,DS~(βS~2,1β0,S~2,1),\displaystyle=\log(D_{\Pi}/D_{N})+\lambda E_{\mu_{\tilde{S}},D_{\tilde{S}}}(\|\beta_{\tilde{S}}\|_{2,1}-\|\beta_{0,\tilde{S}}\|_{2,1}),

where DΠ=p~e12YXS~βS~22λβS~2,1𝑑βS~D_{\Pi}=\int_{\mathbb{R}^{\tilde{p}}}e^{-\frac{1}{2}\|Y-X_{\tilde{S}}\beta_{\tilde{S}}\|_{2}^{2}-\lambda\|\beta_{\tilde{S}}\|_{2,1}}d\beta_{\tilde{S}} and DN=p~e12YXS~βS~22λβ0,S~2,1𝑑βS~D_{N}=\int_{\mathbb{R}^{\tilde{p}}}e^{-\frac{1}{2}\|Y-X_{\tilde{S}}\beta_{\tilde{S}}\|_{2}^{2}-\lambda\|\beta_{0,\tilde{S}}\|_{2,1}}d\beta_{\tilde{S}} are the normalizing constants for the densities. Arguing exactly as in the proof of Lemma B.2 of Ray and Szabó (2022), one can show that on the event 𝒯1\mathcal{T}_{1} is holds that log(DΠ/DN)2λΓ1/2ε+log2\log(D_{\Pi}/D_{N})\leq 2\lambda\Gamma^{1/2}\varepsilon+\log 2 if 4e1+ΓlogMκ14e^{1+\Gamma\log M-\kappa}\leq 1. On 𝒯1\mathcal{T}_{1}, the second term in the last display is bounded by

λEμS~,DS~βS~β0,S~2,1λ|S~|1/2EμS~,DS~βS~β0,S~2λΓ1/2(μS~β0,S~2+E0,DS~βS~2)\begin{split}\lambda E_{\mu_{\tilde{S}},D_{\tilde{S}}}\|\beta_{\tilde{S}}-\beta_{0,\tilde{S}}\|_{2,1}&\leq\lambda|\tilde{S}|^{1/2}E_{\mu_{\tilde{S}},D_{\tilde{S}}}\|\beta_{\tilde{S}}-\beta_{0,\tilde{S}}\|_{2}\\ &\leq\lambda\Gamma^{1/2}\left(\|\mu_{\tilde{S}}-\beta_{0,\tilde{S}}\|_{2}+E_{0,D_{\tilde{S}}}\|\beta_{\tilde{S}}\|_{2}\right)\end{split} (56)

using Cauchy-Schwarz.

Under Pβ0P_{\beta_{0}}, so that Y=dXβ0+εY=^{d}X\beta_{0}+\varepsilon, and using (53),

μS~β0,S~2(XS~TXS~)1XS~TXS~cβ0,S~c2+(XS~TXS~)1XS~Tε2.\displaystyle\|\mu_{\tilde{S}}-\beta_{0,\tilde{S}}\|_{2}\leq\|(X_{\tilde{S}}^{T}X_{\tilde{S}})^{-1}X_{\tilde{S}}^{T}X_{\tilde{S}^{c}}\beta_{0,\tilde{S}^{c}}\|_{2}+\|(X_{\tilde{S}}^{T}X_{\tilde{S}})^{-1}X_{\tilde{S}}^{T}\varepsilon\|_{2}.

Arguing for this term as in Lemma B.2 of Ray and Szabó (2022), one can show that the first term is bounded by Γ1/2s01/2ε/ϕ~(|S~|)2\Gamma^{1/2}s_{0}^{1/2}\varepsilon/\widetilde{\phi}(|\tilde{S}|)^{2} on 𝒯1\mathcal{T}_{1}. Using that the 2\ell_{2}-operator norm of (XS~TXS~)1(X_{\tilde{S}}^{T}X_{\tilde{S}})^{-1} is bounded by 1/(Xϕ~(|S~|))21/(\|X\|\widetilde{\phi}(|\tilde{S}|))^{2} by (55), the second term in the last display is bounded by XS~Tε2/(X2ϕ~(|S~|)2)\|X_{\tilde{S}}^{T}\varepsilon\|_{2}/(\|X\|^{2}\widetilde{\phi}(|\tilde{S}|)^{2}). But on 𝒯1𝒯0\mathcal{T}_{1}\subset\mathcal{T}_{0},

XS~Tε22=GkS~XGkT(YXβ0)229|S~|X2logM.\|X_{\tilde{S}}^{T}\varepsilon\|_{2}^{2}=\sum_{G_{k}\in\tilde{S}}\|X_{G_{k}}^{T}(Y-X\beta_{0})\|_{2}^{2}\leq 9|\tilde{S}|\|X\|^{2}\log M.

Combining the above bounds thus yields

μS~β0,S~2Γ1/2s01/2εϕ~(Γ)2+3Γ1/2logMXϕ~(|S~|)2\displaystyle\|\mu_{\tilde{S}}-\beta_{0,\tilde{S}}\|_{2}\leq\frac{\Gamma^{1/2}s_{0}^{1/2}\varepsilon}{\widetilde{\phi}(\Gamma)^{2}}+\frac{3\Gamma^{1/2}\sqrt{\log M}}{\|X\|\widetilde{\phi}(|\tilde{S}|)^{2}}

on 𝒯1\mathcal{T}_{1}, thereby controlling the first term in (56).

It remains only to bound the second term in (56). Let eie_{i} denote the ithi^{th} unit vector in mk\mathbb{R}^{m_{k}}, i=1,,mki=1,\dots,m_{k}, and let e¯i\bar{e}_{i} denote its extension to p~\mathbb{R}^{\tilde{p}} with unit entry in the ithi^{th} coordinate of group GkG_{k}. Then (XGkTXGk)ii=XGkei22=Xe¯i22X2ϕ~(1)2(X_{G_{k}}^{T}X_{G_{k}})_{ii}=\|X_{G_{k}}e_{i}\|_{2}^{2}=\|X\bar{e}_{i}\|_{2}^{2}\geq\|X\|^{2}\widetilde{\phi}(1)^{2} for i=1,,mki=1,\dots,m_{k}. Therefore,

E0,DS~βS~22=Tr(DS~)=GkS~iGk1(XS~TXS~)iip~X2ϕ~(1)2.\displaystyle E_{0,D_{\tilde{S}}}\|\beta_{\tilde{S}}\|_{2}^{2}=\text{Tr}(D_{\tilde{S}})=\sum_{G_{k}\in\tilde{S}}\sum_{i\in G_{k}}\frac{1}{(X_{\tilde{S}}^{T}X_{\tilde{S}})_{ii}}\leq\frac{\tilde{p}}{\|X\|^{2}\widetilde{\phi}(1)^{2}}.

Putting together of all the above bounds yields

(II)2λΓ1/2ε+log2+λΓϕ~(Γ)2(s01/2ε+3logMX+mmax1/2X),(II)\leq 2\lambda\Gamma^{1/2}\varepsilon+\log 2+\frac{\lambda\Gamma}{\widetilde{\phi}(\Gamma)^{2}}\left(s_{0}^{1/2}\varepsilon+\frac{3\sqrt{\log M}}{\|X\|}+\frac{m_{\text{max}}^{1/2}}{\|X\|}\right),

using that p~mmax|S~|\tilde{p}\leq m_{\text{max}}|\tilde{S}|, |S~|Γ|\tilde{S}|\leq\Gamma and ϕ~(|S~|)ϕ~(1)1\widetilde{\phi}(|\tilde{S}|)\leq\widetilde{\phi}(1)\leq 1 for S~\tilde{S}\neq\emptyset.

Substituting the bounds for (I)(I) and (II)(II) just derived into (52) and (54) then gives the first result. The second result follows from using that mmaxKlogMm_{\text{max}}\leq K\log M under Assumption (K). ∎

F.4 Proofs of Theorems 3 and 4

To complete the proofs, we apply Lemma 1 with the event 𝒯1\mathcal{T}_{1} defined in (48) for suitably chosen constants Γ,ε,κ>0\Gamma,\varepsilon,\kappa>0.

Lemma 7.

Suppose that Assumption (K) holds, the prior satisfies (37)-(38) and sns_{n} satisfies mmaxlogsnKlogMm_{\text{max}}\log s_{n}\leq K^{\prime}\log M for some K>0K^{\prime}>0. Then

infβ0p:ϕ(Sβ0)c0|Sβ0|snPβ0(𝒯1(Γs0,εs0,κs0))1\inf_{\begin{subarray}{c}\beta_{0}\in\mathbb{R}^{p}:\phi(S_{\beta_{0}})\geq c_{0}\\ |S_{\beta_{0}}|\leq s_{n}\end{subarray}}P_{\beta_{0}}(\mathcal{T}_{1}(\Gamma_{s_{0}},\varepsilon_{s_{0}},\kappa_{s_{0}}))\to 1

as nn\to\infty, where

Γs0=CΓs0,εs0=Cεs0logMXϕ~(Cϕs0),κs0=(CΓs0+1)logM,\Gamma_{s_{0}}=C_{\Gamma}s_{0},\qquad\varepsilon_{s_{0}}=C_{\varepsilon}\frac{\sqrt{s_{0}\log M}}{\|X\|\widetilde{\phi}(C_{\phi}s_{0})},\qquad\kappa_{s_{0}}=(C_{\Gamma}s_{0}+1)\log M, (57)

and the constants have dependence CΓ=CΓ(K,K,A2,A4,c0)C_{\Gamma}=C_{\Gamma}(K,K^{\prime},A_{2},A_{4},c_{0}), Cε=Cε(K,K,A1A4,c0)C_{\varepsilon}=C_{\varepsilon}(K,K^{\prime},A_{1}-A_{4},c_{0}) and Cϕ=Cϕ(K,K,A2,A4,c0)C_{\phi}=C_{\phi}(K,K^{\prime},A_{2},A_{4},c_{0}). Moreover, 4e1+Γs0logMκs014e^{1+\Gamma_{s_{0}}\log M-\kappa_{s_{0}}}\leq 1 for M>0M>0 large enough.

Proof.

We consider each of the sets in 𝒯1\mathcal{T}_{1} individually. In what follows, C=C(K,K)C=C(K,K^{\prime}) and c=c(K,K,A2)c=c(K,K^{\prime},A_{2}) will be constants that may change line-by-line but which will not depend on other parameters.

Applying Lemma 4 with L=2A4(1+logC+c+144/c02)L=\frac{2}{A_{4}}(1+\log C+c+144/c_{0}^{2}), where C,cC,c are the constants in that lemma, we have Eβ0Π(β:|Sβ|>(L+1)s0|Y)1𝒯0es0logM1/4E_{\beta_{0}}\Pi(\beta:|S_{\beta}|>(L+1)s_{0}|Y)1_{\mathcal{T}_{0}}\leq e^{-s_{0}\log M}\leq 1/4 for MM large enough. Setting CΓ=L+1C_{\Gamma}=L+1, the second event in 𝒯1\mathcal{T}_{1} is thus a subset of 𝒯0\mathcal{T}_{0} for Γ=Γs0\Gamma=\Gamma_{s_{0}} and all β0\beta_{0} in the infimum in the lemma.

For the third event in 𝒯1\mathcal{T}_{1}, we apply Lemma 5 with L=2A4(1+logC+c+144/c02+CΓ)L=\frac{2}{A_{4}}(1+\log C+c+144/c_{0}^{2}+C_{\Gamma}), where now C,cC,c are the constants in Lemma 5, to obtain

Eβ0Π(β:ββ02H0(L+2)s0logMXϕ~((L+2)s0)2|Y)1𝒯0eκs0.E_{\beta_{0}}\Pi\left(\left.\beta:\|\beta-\beta_{0}\|_{2}\geq\frac{H_{0}\sqrt{(L+2)s_{0}\log M}}{\|X\|\widetilde{\phi}((L+2)s_{0})^{2}}\right|Y\right)1_{\mathcal{T}_{0}}\leq e^{-\kappa_{s_{0}}}.

Setting Cε=H0L+2C_{\varepsilon}=H_{0}\sqrt{L+2} and Cϕ=L+2C_{\phi}=L+2 shows that the third event in 𝒯1\mathcal{T}_{1} is also contained in 𝒯0\mathcal{T}_{0} for these choices and all β0\beta_{0} in the infimum in the lemma. It thus suffices to control the probability of 𝒯0\mathcal{T}_{0}, with tends to one uniformly in β0p\beta_{0}\in\mathbb{R}^{p} by Lemma 2. Lastly, note that 4e1+Γs0logMκs0=4e1logM14e^{1+\Gamma_{s_{0}}\log M-\kappa_{s_{0}}}=4e^{1-\log M}\leq 1 for MM large enough. ∎

Proof of Theorem 3.

Write n=ρn,sn={β0p:ϕ(Sβ0)c0,|Sβ0|sn,ϕ~(ρn|Sβ0|)c0}\mathcal{B}_{n}=\mathcal{B}_{\rho_{n},s_{n}}=\{\beta_{0}\in\mathbb{R}^{p}:\phi(S_{\beta_{0}})\geq c_{0},~|S_{\beta_{0}}|\leq s_{n},~\widetilde{\phi}(\rho_{n}|S_{\beta_{0}}|)\geq c_{0}\} for the parameter set and let

Ωn={β:X(ββ0)2H0ρn1/2s0logMϕ¯(ρns0)}\displaystyle\Omega_{n}=\left\{\beta:\|X(\beta-\beta_{0})\|_{2}\leq\frac{H_{0}\rho_{n}^{1/2}\sqrt{s_{0}\log M}}{\bar{\phi}(\rho_{n}s_{0})}\right\}

for H0=H0(A1,A3,A4)>0H_{0}=H_{0}(A_{1},A_{3},A_{4})>0 the constant in Lemma 5. Let 𝒯1=𝒯1=𝒯1(Γs0,εs0,κs0)\mathcal{T}_{1}=\mathcal{T}_{1}=\mathcal{T}_{1}(\Gamma_{s_{0}},\varepsilon_{s_{0}},\kappa_{s_{0}}) be the event (48) with parameters (57), so that Lemma 7 gives Eβ0Π~(Ωnc)=Eβ0Π~(Ωnc)1𝒯1+o(1)E_{\beta_{0}}\tilde{\Pi}(\Omega_{n}^{c})=E_{\beta_{0}}\tilde{\Pi}(\Omega_{n}^{c})1_{\mathcal{T}_{1}}+o(1), uniformly over β0n\beta_{0}\in\mathcal{B}_{n}.

Applying Lemma 5 with L+2=ρnL+2=\rho_{n}\to\infty gives that for nn large enough, Eβ0Π(Ωnc|Y)1𝒯0Cecρns0logME_{\beta_{0}}\Pi(\Omega_{n}^{c}|Y)1_{\mathcal{T}_{0}}\leq Ce^{-c\rho_{n}s_{0}\log M}, where C,cC,c depend only on K,K,A2,A4,c0K,K^{\prime},A_{2},A_{4},c_{0}. We can then use Lemma 1 with event An=𝒯1𝒯0A_{n}=\mathcal{T}_{1}\subset\mathcal{T}_{0} and δn=cρns0logM\delta_{n}=c\rho_{n}s_{0}\log M to obtain

Eβ0Π~(Ωnc)1𝒯1\displaystyle E_{\beta_{0}}\tilde{\Pi}(\Omega_{n}^{c})1_{\mathcal{T}_{1}} 2cρns0logM[DKL(Π~Π(|Y))1𝒯1+Ccρns0logM].\displaystyle\leq\frac{2}{c\rho_{n}s_{0}\log M}\left[D_{\text{KL}}(\tilde{\Pi}\|\Pi(\cdot|Y))1_{\mathcal{T}_{1}}+C^{-c\rho_{n}s_{0}\log M}\right].

Since ρns0logM\rho_{n}s_{0}\log M\to\infty and Π~\tilde{\Pi} is by definition the KL-minimizer of the variational family to the posterior, we obtain that for any Q𝒬Q\in\mathcal{Q},

Eβ0Π~(Ωnc)\displaystyle E_{\beta_{0}}\tilde{\Pi}(\Omega_{n}^{c}) 2cρns0logMDKL(QΠ(|Y))1𝒯1+o(1),\displaystyle\leq\frac{2}{c\rho_{n}s_{0}\log M}D_{\text{KL}}(Q\|\Pi(\cdot|Y))1_{\mathcal{T}_{1}}+o(1),

where the o(1)o(1) term is uniform over β0n\beta_{0}\in\mathcal{B}_{n}. But by Lemma 6, there exists an element Q𝒬Q\in\mathcal{Q} of the variational family such that

DKL(QΠ(|Y))1𝒯1C(log1ϕ~(CΓs0)+1ϕ~(CΓs0)2ϕ~(Cϕs0))s0logM,D_{\text{KL}}(Q\|\Pi(\cdot|Y))1_{\mathcal{T}_{1}}\leq C\left(\log\frac{1}{\widetilde{\phi}(C_{\Gamma}s_{0})}+\frac{1}{\widetilde{\phi}(C_{\Gamma}s_{0})^{2}\widetilde{\phi}(C_{\phi}s_{0})}\right)s_{0}\log M,

where CC is uniform over β0n\beta_{0}\in\mathcal{B}_{n} and CΓ,CϕC_{\Gamma},C_{\phi} have dependences specified in Lemma 7. But since CΓ,CϕC_{\Gamma},C_{\phi} are bounded under the hypothesis of the present theorem and ρn\rho_{n}\to\infty, we have ρnCΓ,Cϕ\rho_{n}\geq C_{\Gamma},C_{\phi} for nn large enough and hence ϕ~(CΓs0),ϕ~(Cϕs0)ϕ~(ρns0)c0\widetilde{\phi}(C_{\Gamma}s_{0}),\widetilde{\phi}(C_{\phi}s_{0})\geq\widetilde{\phi}(\rho_{n}s_{0})\geq c_{0}. Using this and the last two displays, we thus have

supβ0nEβ0Π~(Ωnc)C/ρn+o(1)=o(1)\sup_{\beta_{0}\in\mathcal{B}_{n}}E_{\beta_{0}}\tilde{\Pi}(\Omega_{n}^{c})\leq C/\rho_{n}+o(1)=o(1)

as nn\to\infty, thereby establishing the first assertion. The second two assertions follow similarly for using the corresponding results in Lemma 5. ∎

Proof of Theorem 4.

The proof follows similarly to the proof of Theorem 3, using Lemma 4 to control the probability of the posterior set instead of Lemma 5. ∎