This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Do Vendi Scores Converge with Finite Samples?
Truncated Vendi Score for Finite-Sample Convergence Guarantees

Azim Ospanov Department of Computer Science and Engineering, The Chinese University of Hong Kong Farzan Farnia Department of Computer Science and Engineering, The Chinese University of Hong Kong
Abstract

Evaluating the diversity of generative models without reference data poses methodological challenges. The reference-free Vendi [Friedman and Dieng, 2023] and RKE [Jalali et al., 2023] scores address this by quantifying the diversity of generated data using matrix-based entropy measures. Among these two, the Vendi score is typically computed via the eigendecomposition of an n×nn\times n kernel matrix constructed from n generated samples. However, the prohibitive computational cost of eigendecomposition for large nn often limits the number of samples used to fewer than 20,000. In this paper, we investigate the statistical convergence of the Vendi and RKE scores under restricted sample sizes. We numerically demonstrate that, in general, the Vendi score computed with standard sample sizes below 20,000 may not converge to its asymptotic value under infinite sampling. To address this, we introduce the tt-truncated Vendi score by truncating the eigenspectrum of the kernel matrix, which is provably guaranteed to converge to its population limit with n=𝒪(t)n=\mathcal{O}(t) samples. We further show that existing Nyström and FKEA approximation methods converge to the asymptotic limit of the truncated Vendi score. In contrast to the Vendi score, we prove that the RKE score enjoys universal convergence guarantees across all kernel functions. We conduct several numerical experiments to illustrate the concentration of Nyström and FKEA computed Vendi scores around the truncated Vendi score, and we analyze how the truncated Vendi and RKE scores correlate with the diversity of image and text data. The code is available at https://github.com/aziksh-ospanov/truncated-vendi.

1 Introduction

Refer to caption
Figure 1: Statistical convergence of Vendi and RKE scores for different sample sizes on ImageNet data: (Left plots) finite-dimension cosine similarity kernel (Right plots) infinite dimension Gaussian kernel with bandwidth σ=30\sigma=30. The RKE and truncated Vendi scores converged with below 20000 samples, but the Vendi score with Gaussian kernel did not converge.

The increasing use of generative artificial intelligence has underscored the need for accurate evaluation of generative models. In practice, users often have access to multiple generative models trained with different training datasets and algorithms, requiring evaluation methods to identify the most suitable model. The feasibility of a model evaluation approach depends on factors such as the required generated sample size, computational cost, and the availability of reference data. Recent studies on evaluating generative models have introduced assessment methods that relax the requirements on data and computational resources.

Specifically, to enable the evaluation of generative models without reference data, the recent literature has focused on reference-free evaluation scores that remain applicable in the absence of reference samples. The Vendi score [Friedman and Dieng, 2023] is one such reference-free metric that quantifies the diversity of generated data using the entropy of a kernel similarity matrix formulated for the generated samples. Given the sorted eigenvalues λ1λn\lambda_{1}\geq\cdots\geq\lambda_{n} of the normalized matrix 1nK\frac{1}{n}K111In general, we consider the trace-normalized kernel matrix 1Tr(K)K\frac{1}{\mathrm{Tr}(K)}K, which given x:k(x,x)=1\forall x:\>k(x,x)=1, reduces to 1nK\frac{1}{n}K. for the kernel similarity matrix K=[k(xi,xj)]1i,jnK=\bigl{[}k(x_{i},x_{j})\bigr{]}_{1\leq i,j\leq n} of nn generated samples x1,,xnx_{1},\ldots,x_{n}, the definition of (order-1) Vendi score is as:

Vendi(x1,,xn):=exp(i=1nλilog1λi)\mathrm{Vendi}(x_{1},\ldots,x_{n})\,:=\,\exp\Bigl{(}\,\sum_{i=1}^{n}\lambda_{i}\log\frac{1}{\lambda_{i}}\,\Bigr{)} (1)

Following conventional definitions in information theory, the Vendi score corresponds to the exponential of the Von Neumann entropy of normalized kernel matrix 1nK\frac{1}{n}K. More generally, Jalali et al. [2023] define the Rényi Kernel Entropy (RKE) score by applying order-2 Rényi entropy to this matrix, which reduces to the inverse-squared Frobenius norm of the normalized kernel matrix:

RKE(x1,,xn):=11nKF2\mathrm{RKE}(x_{1},\ldots,x_{n})\,:=\frac{1}{\Bigl{\|}\frac{1}{n}K\Bigr{\|}^{2}_{F}} (2)

Although the Vendi and RKE scores do not require reference samples, their computational cost increases rapidly with the number of generated samples nn. Specifically, calculating the Vendi score for the n×nn\times n kernel matrix KK generally involves an eigendecomposition of KK, requiring O(n3)O(n^{3}) computations. Therefore, the computational load of Vendi score becomes substantial for a large sample size nn, and the Vendi score is typically evaluated for sample sizes limited to 20,000. In other words, the Vendi score, as defined in Equation (2), would be computationally infeasible to compute with standard processors for sample sizes greater than a few tens of thousands.

Following the above discussion, a key question that arises is whether the Vendi score estimated from restricted sample sizes (i.e. n20000n\leq 20000) has converged to its asymptotic value with infinite samples, which we call the population Vendi. However, the statistical convergence of the Vendi score has not been thoroughly investigated in the literature. In this work, we study the statistical convergence of the Vendi and RKE diversity scores and aim to analyze the concentration of the estimated scores from a limited number of generated samples n20000n\lessapprox 20000.

1.1 Our Results on Vendi’s Convergence

We discuss the answer to the Vendi convergence question for two types of kernel functions: 1) kernel functions with a finite feature dimension, e.g. the cosine similarity and polynomial kernels, 2) kernel functions with an infinite feature map such as Gaussian (RBF) kernels. For kernel functions with a finite feature dimension dd, we theoretically and numerically show that a sample size n=O(d)n=O(d) is sufficient to guarantee convergence to the population Vendi (asymptotic value when nn\rightarrow\infty). For example, the left plot in Figure 1 shows that in the case of the cosine similarity kernel, the Vendi score on nn randomly selected ImageNet [Deng et al., 2009] samples has almost converged as the sample size reaches 5000, where the dimension dd (using standard DINOv2 embedding [Oquab et al., 2023]) is 768.

In contrast, our numerical results for kernel functions with an infinite feature map demonstrate that for standard datasets, a sample size bounded by 20,000 could be insufficient for convergence of the Vendi score. For example, the right plot of Figure 1 shows the evolution of the Vendi score with the Gaussian kernel on ImageNet data, and the score continues to grow at a significant rate with 20,000 samples222The heavy computational cost prohibits an empirical evaluation of the sample size required for Vendi’s convergence..

Observing the difference between Vendi score convergence for finite and infinite-dimension kernel functions, a natural question is how to extend the definition of Vendi score from finite to infinite dimension case such that the diversity score would statistically converge in both scenarios. We attempt to address the question by introducing an alternative Vendi statistic, which we call the tt-truncated Vendi score. The tt-truncated Vendi score is defined using only the top-tt λ1λt\lambda_{1}\geq\cdots\geq\lambda_{t} eigenvalues of the kernel matrix, where tt is an integer hyperparameter. This modified score is defined as

Truncated-Vendi(t)(x1,,xn)=exp(i=1tλitrunclog1λitrunc)\mathrm{Truncated}\text{-}\mathrm{Vendi}^{(t)}(x_{1},\ldots,x_{n})=\exp\Bigl{(}\sum_{i=1}^{t}{\lambda^{\scriptscriptstyle\text{trunc}}_{i}}\log\frac{1}{\lambda^{\scriptscriptstyle\text{trunc}}_{i}}\Bigr{)}

where we shift each of the top-tt eigenvalue λitrunc=λi+c\lambda^{\scriptscriptstyle\text{trunc}}_{i}=\lambda_{i}+c by the same constant c=(1i=1tλi)/tc=\bigl{(}1-\sum_{i=1}^{t}\lambda_{i}\bigr{)}/t to ensure they add up to 11 and provide a valid probability model. Observe that for a finite kernel dimension dd satisfying dtd\leq t, the truncated and original Vendi scores take the same value, because the truncation will have no impact on the eigenvalues. On the other hand, under an infinite kernel dimension, the two scores may take different values.

As a main theoretical result, we prove that a sample size n=O(t)n=O(t) is always enough to estimate the tt-truncated population Vendi from nn empirical samples, regardless of the finiteness of the kernel feature dimension. This result shows that the tt-truncated Vendi score provides a statistically converging extension of the Vendi score from the finite kernel dimension to the infinite dimension case. To connect the defined tt-truncated Vendi score to existing computation methods for the original Vendi score, we show that the existing computationally-efficient methods for computing the Vendi score can be viewed as approximations of our defined tt-truncated Vendi. Specifically, we show that the Nyström method in [Friedman and Dieng, 2023] and the FKEA method proposed by Ospanov et al. [2024b] provide an estimate of the tt-truncated Vendi.

Refer to caption
Figure 2: Computation of the proposed tt-truncated Vendi score. The kernel similarity matrix eigenspectrum is truncated, and the mass of the truncated tail (excluding the top-tt eigenvalues) is uniformly redistributed among the top-tt eigenvalues.

1.2 Our Results on RKE’s Convergence

For the RKE score, we prove a universal convergence guarantee that holds for every kernel function. The theoretical guarantee shows that the RKE score, and more generally every order-α\alpha entropy score with α2\alpha\geq 2, will converge to its population value within O(1n)O(\frac{1}{\sqrt{n}}) error for nn samples. Our theoretical guarantee also transfers to the truncated version of the RKE score. However, note that the truncation of the eigenspectrum becomes unnecessary in the RKE case, since the score enjoys universal convergence guarantees. Figure 1 shows that using both the cosine-similarity and Gaussian kernel functions, the RKE score nearly converges to its limit value with less than 10000 samples.

Finally, we present the findings of several numerical experiments to validate our theoretical results on the convergence of Vendi, truncated Vendi, and RKE scores. Our numerical results on standard image, text, and video datasets and generative models indicate that in the case of a finite-dimension kernel map, the Vendi score can converge to its asymptotic limit, in which case, as we explained earlier, the Vendi score is identical to the truncated Vendi. On the other hand, in the case of infinite-dimension Gaussian kernel functions, we numerically observe the growth of the score beyond n=n=10,000. Our numerical results further confirm that the scores computed by Nyström method in [Friedman and Dieng, 2023] and the FKEA method [Ospanov et al., 2024b] provide tight estimations of the population truncated Vendi. The following summarizes this work’s contributions:

  • Analyzing the statistical convergence of Vendi and RKE diversity scores under restricted sample sizes n2×104n\lessapprox 2\times 10^{4},

  • Providing numerical evidence on the Vendi score’s lack of convergence for infinite-dimensional kernel functions, e.g. the Gaussian (RBF) kernel,

  • Introducing the truncated Vendi score as a statistically converging extension of the Vendi score from finite to infinite dimension kernel functions,

  • Demonstrating the universal convergence of the RKE diversity score across all kernel functions.

2 Related Works

Diversity evaluation for generative models Diversity evaluation in generative models can be categorized into two primary types: reference-based and reference-free methods. Reference-based approaches rely on a predefined dataset to assess the diversity of generated data. Metrics such as FID [Heusel et al., 2018], KID and distributed KID [Bińkowski et al., 2018, Wang et al., 2023] measure the distance between the generated data and the reference, while Recall [Sajjadi et al., 2018, Kynkäänniemi et al., 2019] and Coverage [Naeem et al., 2020] evaluate the extent to which the generative model captures existing modes in the reference dataset. Pillutla et al. [2021, 2023] propose MAUVE metric that uses information divergences in a quantized embedding space to measure the gap between generated data and reference distribution. In contrast, the reference-free metrics, Vendi [Friedman and Dieng, 2023] and RKE [Jalali et al., 2023], assign diversity scores based on the eigenvalues of a kernel similarity matrix of the generated data. Jalali et al. [2023] interpret the approach as identifying modes and their frequencies within the generated data followed by entropy calculation for the frequency parameters. The Vendi and RKE scores have been further extended to quantify the diversity of conditional prompt-based generative AI models [Ospanov et al., 2024a, Jalali et al., 2024] and to select generative models in online settings [Rezaei et al., 2024, Hu et al., 2024, 2025]. Also, [Zhang et al., 2024, 2025, Jalali et al., 2025, Gong et al., 2025, Wu et al., 2025] extend the entropic kernel-based scores to measure novelty and embedding dissimilarity. In our work, we specifically focus on the statistical convergence of the vanilla Vendi and RKE scores.

Statistical convergence analysis of kernel matrices’ eigenvalues. The convergence analysis of the eigenvalues of kernel matrices has been studied by several related works. Shawe-Taylor et al. [2005] provide a concentration bound for the eigenvalues of a kernel matrix. We note that the bounds in [Shawe-Taylor et al., 2005] use the expectation of eigenvalues 𝔼m[𝝀^(S)]\mathbb{E}_{m}[\hat{\boldsymbol{\lambda}}(S)] for a random dataset S=(𝐱1,,𝐱m)S=(\mathbf{x}_{1},\dots,\mathbf{x}_{m}) of fixed size mm as the center vector in the concentration analysis. However, since eigenvalues are non-linear functions of a matrix, this concentration center vector 𝔼m[𝝀^(S)]\mathbb{E}_{m}[\hat{\boldsymbol{\lambda}}(S)] does not match the eigenvalues of the asymptotic kernel matrix as the sample size approaches to infinity. On the other hand, our convergence analysis focuses on the asymptotic eigenvalues with an infinite sample size, which determines the limit value of Vendi scores. In another related work, Bach [2022] discusses a convergence result for the Von-Neumann entropy of kernel matrix. While this result proves a non-asymptotic guarantee on the convergence of the entropy function, the bound may not guarantee convergence at standard sample sizes for computing Vendi scores (less than 1000010000 in practice). In our work, we aim to provide convergence guarantees for the finite-dimension and generally truncated Vendi scores with restricted sample sizes.

Efficient computation of matrix-based entropy. Several strategies have been proposed in the literature to reduce the computational complexity of matrix-based entropy calculations, which involve the computation of matrix eigenvalues—a process that scales cubically with the size of the dataset. Dong et al. [2023] propose an efficient algorithm for approximating matrix-based Renyi’s entropy of arbitrary order α\alpha, which achieves a reduction in computational complexity down to O(n2sm)O(n^{2}sm) with s,mns,m\ll n. Additionally, kernel matrices can be approximated using low-rank techniques such as incomplete Cholesky decomposition [Fine and Scheinberg, 2001, Bach and Jordan, 2002] or CUR matrix decompositions [Mahoney and Drineas, 2009], which provide substantial computational savings. Pasarkar and Dieng [2024] suggest to leverage Nyström method [Williams and Seeger, 2000] with mm components, which results in O(nm2)O(nm^{2}) computational complexity. Further reduction in complexity is possible using Random Fourier Features, as suggested by Ospanov et al. [2024b], which allows the computation to scale linearly with O(n)O(n) as a function of the dataset size. This work focuses on the latter two methods and the population quantities estimated by them.

Impact of embedding spaces on diversity evaluation. In our image-related experiments, we used the DinoV2 embedding [Oquab et al., 2023], as Stein et al. [2023] demonstrate the alignment of this embedding with human evaluations. We note that the kernel function in the Vendi score can be similarly applied to other embeddings, including the standard InceptionV3[Szegedy et al., 2016] and CLIP embeddings [Radford et al., 2021] as suggested by Kynkäänniemi et al. [2022].

3 Preliminaries

Consider a generative model 𝒢\mathcal{G} that generates samples from a probability distribution PXP_{X}. To conduct a reference-free evaluation of the model, we suppose the evaluator has access to nn independently generated samples from PXP_{X}, denoted by x1,,xn𝒳x_{1},\ldots,x_{n}\in\mathcal{X}. The assessment task is to estimate the diversity of generative model 𝒢\mathcal{G} by measuring the variety of the observed generated data, x1,xnx_{1},\ldots x_{n}. In the following subsections, we will discuss kernel functions and their application to define the Vendi and RKE diversity scores.

3.1 Kernel Functions and Matrices

Following the standard definition, k:𝒳×𝒳k:\mathcal{X}\times\mathcal{X}\rightarrow\mathbb{R} is called a kernel function if for every integer nn\in\mathbb{N} and inputs x1,,xn𝒳x_{1},\ldots,x_{n}\in\mathcal{X}, the following kernel similarity matrix Kn×nK\in\mathbb{R}^{n\times n} is positive semi-definite (PSD):

K=[k(x1,x1)k(x1,xn)k(xn,x1)k(xn,xn)]K=\begin{bmatrix}k(x_{1},x_{1})&\cdots&k(x_{1},x_{n})\\ \vdots&\ddots&\vdots\\ k(x_{n},x_{1})&\cdots&k(x_{n},x_{n})\end{bmatrix} (3)

Aronszajn’s Theorem [Aronszajn, 1950] shows that this definition is equivalent to the existence of a feature map ϕ:𝒳d\phi:\mathcal{X}\rightarrow\mathbb{R}^{d} such that for every x,x𝒳x,x^{\prime}\in\mathcal{X} we have the following where ,\langle\cdot,\cdot\rangle denotes the standard inner product in the d\mathbb{R}^{d} space:

k(x,x)=ϕ(x),ϕ(x)k(x,x^{\prime})\,=\,\bigl{\langle}\phi(x),\phi(x^{\prime})\bigr{\rangle} (4)

In this work, we study the evaluation using two types of kernel functions: 1) finite-dimension kernels where dimension dd is finite, 2) infinite-dimension kernels where there is no feature map satisfying (4) with a finite dd value. A standard example of a finite-dimension kernel is the cosine similarity function where ϕcosine(x)=x/x2\phi_{\text{cosine}}(x)=x/\|x\|_{2}. Also, a widely-used infinite-dimension kernel is the Gaussian (RBF) kernel with bandwidth parameter σ>0\sigma>0 defined as

kGaussian(σ)(x,x):=exp(xx222σ2)k_{\text{\rm Gaussian}(\sigma)}\bigl{(}x,x^{\prime}\bigr{)}\,:=\,\exp\Bigl{(}-\frac{\bigl{\|}x-x^{\prime}\bigr{\|}^{2}_{2}}{2\sigma^{2}}\Bigr{)} (5)

Both the mentioned kernel examples belong to normalized kernels which require k(x,x)=1k(x,x)=1 for every xx, i.e., the feature map ϕ(x)\phi(x) has unit Euclidean norm for every xx. Given a normalized kernel function, the non-negative eigenvalues of the normalized kernel matrix 1nK\frac{1}{n}K for nn points x1,xnx_{1},\ldots x_{n} will sum up to 11, i.e., they form a probability model.

3.2 Matrix-based Entropy Functions and Vendi Score

For a PSD matrix Ad×dA\in\mathbb{R}^{d\times d} with unit trace Tr(A)=1\mathrm{Tr}(A)=1, AA’s eigenvalues form a probability model. The order-α\alpha Renyi entropy of matrix AA is defined using the order-α\alpha entropy of its eigenvalues as

Hα(A):=11αlog(i=1dλiα)H_{\alpha}(A)\,:=\,\frac{1}{1-\alpha}\log\Bigl{(}\sum_{i=1}^{d}\lambda^{\alpha}_{i}\Bigr{)} (6)

For the special case α=2\alpha=2, one can consider the Frobenius norm F\|\cdot\|_{F} and apply the identity AF2=i=1dλi2\bigl{\|}A\bigr{\|}_{F}^{2}=\sum_{i=1}^{d}\lambda^{2}_{i} to show H2(A)=log(1/AF2)H_{2}(A)=\log\bigl{(}1/\bigl{\|}A\bigr{\|}_{F}^{2}\bigr{)}. Moreover, for α=1\alpha=1, the above definition reduces to the Shannon entropy of the eigenvalues as H1(A):=i=1dλilog(1/λi)H_{1}(A)\,:=\,\sum_{i=1}^{d}\lambda_{i}\log({1}/{\lambda_{i}}) [Rényi, 1961].

Jalali et al. [2023] applies the above definition for order α=2\alpha=2 to the normalized kernel similarity matrix 1nK\frac{1}{n}K to define the RKE diversity score (called RKE mode count), which reduces to

RKE(x1,,xn):=exp(H2(1nK))=1nKF2\mathrm{RKE}(x_{1},\ldots,x_{n})\,:=\,\exp\Bigl{(}H_{2}\bigl{(}\frac{1}{n}K\bigr{)}\Bigr{)}=\,\Bigl{\|}\frac{1}{n}K\Bigr{\|}^{-2}_{F} (7)

For a general entropy order α\alpha, [Friedman and Dieng, 2023, Pasarkar and Dieng, 2024] apply the matrix-based entropy definition to the normalized kernel matrix 1nK\frac{1}{n}K and define the order-α\alpha Vendi score for samples x1,,xnx_{1},\ldots,x_{n} as

Vendiα(x1,,xn):=exp(Hα(1nK))\mathrm{Vendi}_{\alpha}\bigl{(}x_{1},\ldots,x_{n}\bigr{)}\,:=\,\exp\Bigl{(}H_{\alpha}\bigl{(}\frac{1}{n}K\bigr{)}\Bigr{)} (8)

Specifically, for order α=1\alpha=1, the above definition results in the standard (order-1) Vendi score in Equation (2).

3.3 Statistical Analysis of Vendi Score

To derive the population limits of Vendi and RKE scores under infinite sampling, which we call population Vendi and population RKE, respectively, we review the following discussion from [Bach, 2022, Jalali et al., 2023]. First, note that the normalized kernel matrix 1nK\frac{1}{n}K, whose eigenvalues are used in the definition of Vendi score, can be written as:

1nK=1nΦΦ\frac{1}{n}K=\frac{1}{n}\Phi\Phi^{\top} (9)

where Φn×d\Phi\in\mathbb{R}^{n\times d} is an n×dn\times d matrix whose rows are the feature presentations of samples, i.e., ϕ(x1),,ϕ(xn)\phi(x_{1}),\ldots,\phi(x_{n}). Therefore, the normalized kernel matrix 1nK\frac{1}{n}K shares the same non-zero eigenvalues with 1nΦΦ\frac{1}{n}\Phi^{\top}\Phi, where the multiplication order is flipped. Note that 1nΦΦ\frac{1}{n}\Phi^{\top}\Phi is equal to the empirical kernel covariance matrix C^X\widehat{C}_{X} defined as:

C^X:=1ni=1nϕ(xi)ϕ(xi)=1nΦΦ.\widehat{C}_{X}:=\frac{1}{n}\sum_{i=1}^{n}\phi(x_{i})\phi(x_{i})^{\top}=\frac{1}{n}\Phi^{\top}\Phi.

As a result, the empirical covariance matrix C^X=1nΦΦ\widehat{C}_{X}=\frac{1}{n}\Phi^{\top}\Phi and kernel matrix 1nK=1nΦΦ\frac{1}{n}K=\frac{1}{n}\Phi\Phi^{\top} share the same non-zero eigenvalues and therefore have the same matrix-based entropy value for every order α\alpha: Hα(1nK)=Hα(C^X)H_{\alpha}(\frac{1}{n}K)=H_{\alpha}(\widehat{C}_{X}). Therefore, if we consider the population kernel covariance matrix C~X=𝔼xPX[ϕ(x)ϕ(x)]\widetilde{C}_{X}=\mathbb{E}_{x\sim P_{X}}\bigl{[}\phi(x)\phi(x)^{\top}\bigr{]}, we can define the population Vendi score as follows.

Definition 1.

Given data distribution PXP_{X}, we define the order-α\alpha population Vendi, Vendiα(PX){\mathrm{Vendi}}_{\alpha}(P_{X}), using the matrix-based entropy of the population kernel covariance matrix C~X=𝔼xPX[ϕ(x)ϕ(x)]\widetilde{C}_{X}=\mathbb{E}_{x\sim P_{X}}\bigl{[}\phi(x)\phi(x)^{\top}\bigr{]} as

Vendiα(PX):=exp(Hα(C~X)){\mathrm{Vendi}}_{\alpha}(P_{X})\,:=\,\exp\Bigl{(}H_{\alpha}(\widetilde{C}_{X})\Bigr{)} (10)

Note that the population RKE score is identical to the population Vendi2\mathrm{Vendi}_{2}, since RKE and Vendi2\mathrm{Vendi}_{2} are the same.

4 Statistical Convergence of Vendi and RKE Scores

Given the definitions of the Vendi score and the population Vendi, a relevant question is how many samples are required to accurately estimate the population Vendi using the Vendi score. To address this question, we first prove the following concentration bound on the vector of ordered eigenvalues [λ1,,λn][\lambda_{1},\ldots,\lambda_{n}] of the kernel matrix for a normalized kernel function.

Theorem 1.

Consider a normalized kernel function kk satisfying k(x,x)=1k(x,x)=1 for every x𝒳x\in\mathcal{X}. Let 𝛌^n\widehat{\boldsymbol{\lambda}}_{n} be the vector of sorted eigenvalues of the normalized kernel matrix 1nK\frac{1}{n}K for nn independent samples x1,,xnPXx_{1},\ldots,x_{n}\sim P_{X}. If we define 𝛌~\widetilde{\boldsymbol{\lambda}} as the vector of sorted eigenvalues of underlying covariance matrix C~X\widetilde{C}_{X}, then if n2+8log(1/δ)n\geq 2+8\log(1/\delta), the following inequality holds with probability at least 1δ1-\delta:

𝝀^n𝝀~232log(2/δ)n\bigl{\|}\widehat{\boldsymbol{\lambda}}_{n}-\widetilde{\boldsymbol{\lambda}}\bigr{\|}_{2}\,\leq\,\sqrt{\frac{32\log\bigl{(}2/\delta\bigr{)}}{n}}

Note that in calculating the subtraction 𝛌^n𝛌~\widehat{\boldsymbol{\lambda}}_{n}-\widetilde{\boldsymbol{\lambda}}, we add |dn||d-n| zero entries to the lower-dimension vector, if the dimension of vectors 𝛌^n\widehat{\boldsymbol{\lambda}}_{n} and 𝛌~\widetilde{\boldsymbol{\lambda}} do not match.

Proof. We defer the proof to the Appendix.

Theorem 1 results in the following corollary on a dimension-free convergence guarantee for every Vendiα\mathrm{Vendi}_{\alpha} score with order α2\alpha\geq 2, including the RKE score (i.e. Vendi2\mathrm{Vendi}_{2}).

Corollary 1.

In the setting of Theorem 1, for every α2\alpha\geq 2 and n2+8log(1/δ)n\geq 2+8\log(1/\delta), the following bound holds with probability at least 1δ1-\delta:

|Vendiα(x1,,xn)1ααVendiα(PX)1αα|\displaystyle\Bigl{|}\mathrm{Vendi}_{\alpha}\bigl{(}x_{1},\ldots,x_{n}\bigr{)}^{\frac{1-\alpha}{\alpha}}-\mathrm{Vendi}_{\alpha}\bigl{(}P_{X}\bigr{)}^{\frac{1-\alpha}{\alpha}}\Bigr{|}\leq 32log2δn\displaystyle\sqrt{\frac{32\log\frac{2}{\delta}}{n}}

Notably, for α=2\alpha=2, we arrive at the following bound on the gap between the empirical and population RKE scores:

|RKE(x1,,xn)1/2RKE(PX)1/2|32log2δn\displaystyle\Bigl{|}\mathrm{RKE}\bigl{(}x_{1},\ldots,x_{n}\bigr{)}^{-1/2}-\mathrm{RKE}\bigl{(}P_{X}\bigr{)}^{-1/2}\Bigr{|}\leq\sqrt{\frac{32\log\frac{2}{\delta}}{n}}

Proof. We defer the proof to the Appendix. Therefore, the bound in Corollary 1 holds regardless of the dimension of kernel feature map, indicating that the RKE score enjoys a universal convergence guarantee across all kernel functions. Next, we show that Theorem 1 implies the following corollary on a dimension-dependent convergence guarantee for order-α\alpha Vendi score with 1α<21\leq\alpha<2, including standard (order-11) Vendi score.

Corollary 2.

In the setting of Theorem 1, consider a finite dimension kernel map where we suppose dim(ϕ)=d<\mathrm{dim}(\phi)=d<\infty. (a) For α=1\alpha=1, assuming n32e2log(2/δ)n\geq 32e^{2}\log(2/\delta), the following bound holds with probability at least 1δ1-\delta:

|log(Vendi1(x1,,xn))log(Vendi1(PX))|\displaystyle\Bigl{|}\,\log\bigl{(}\mathrm{Vendi}_{1}\bigl{(}x_{1},\ldots,x_{n}\bigr{)}\bigr{)}-\log\bigl{(}\mathrm{Vendi}_{1}\bigl{(}P_{X}\bigr{)}\bigr{)}\,\Bigr{|}
\displaystyle\leq\> 8dlog(2/δ)nlog(nd32log(2/δ)).\displaystyle\sqrt{\frac{8d\log\bigl{(}2/\delta\bigr{)}}{n}}\log\Bigl{(}\frac{nd}{32\log(2/\delta)}\Bigr{)}.

(b) For every 1<α<21<\alpha<2 and n2+8log(1/δ)n\geq 2+8\log(1/\delta), the following bound holds with probability at least 1δ1-\delta:

|Vendiα(x1,,xn)1ααVendiα(PX)1αα|\displaystyle\Bigl{|}\,\mathrm{Vendi}_{\alpha}\bigl{(}x_{1},\ldots,x_{n}\bigr{)}^{\frac{1-\alpha}{\alpha}}-\mathrm{Vendi}_{\alpha}\bigl{(}P_{X}\bigr{)}^{\frac{1-\alpha}{\alpha}}\,\Bigr{|}
\displaystyle\leq\> 32d2αlog(2/δ)n\displaystyle\sqrt{\frac{32d^{2-\alpha}\log\bigl{(}2/\delta\bigr{)}}{n}}

Proof. We defer the proof to the Appendix.

Therefore, assuming a finite feature map d<d<\infty and given an entropy order 1α<21\leq\alpha<2, the above results indicate the convergence of the Vendi score to the underlying population Vendi given n=O(d2α)n=O(d^{2-\alpha}) samples. Observe that this result is consistent with our numerical observations of the convergence of Vendi score using the finite-dimension cosine similarity kernel in Figure 1.

5 Truncated Vendi Score and its Estimation via Proxy Kernels

Corollaries 1, 2 demonstrate that if the Vendi score order α\alpha is greater than 22 or the kernel feature map dimension dd is finite, then the Vendi score can converge to the population Vendi with n=O(d)n=O(d) samples. However, the theoretical results do not apply to an order 1α<21\leq\alpha<2 when the kernel map dimension is infinite, e.g. the original order-1 Vendi score [Friedman and Dieng, 2023] with a Gaussian kernel. Our numerical observations indicate that a standard sample size below 20000 could be insufficient for the convergence of order-1 Vendi score (Figure 1). To address this gap, here we define the truncated Vendi score by truncating the eigenspectrum of the kernel matrix, and then show that the existing kernel approximation algorithms for Vendi score concentrate around this modified Vendi score.

Definition 2.

Consider the normalized kernel matrix 1nK\frac{1}{n}K of samples x1,,xnx_{1},\ldots,x_{n}. Then, for an integer parameter t1t\geq 1, consider the top-tt eigenvalues of 1nK\frac{1}{n}K: λ1λ2λt\lambda_{1}\geq\lambda_{2}\geq\cdots\geq\lambda_{t}. Define St=i=1tλiS_{t}=\sum_{i=1}^{t}\lambda_{i} and consider the truncated probability sequence [λ1trunc,,λttrunc][{\lambda}^{\scriptscriptstyle\text{trunc}}_{1},\ldots,{\lambda}^{\scriptscriptstyle\text{trunc}}_{t}]:

λitrunc=λi+1Sttfor i=1,,t{\lambda}^{\scriptscriptstyle\text{trunc}}_{i}=\lambda_{i}+\frac{1-S_{t}}{t}\quad\text{for $\;i=1,\ldots,t$}

We define the order-α\alpha tt-truncated Vendi score as

Vendiα(t)(x1,,xn):=exp(11αlog(i=1tλitruncα))\mathrm{Vendi}_{\alpha}^{(t)}(x_{1},\ldots,x_{n}):=\exp\Bigl{(}\frac{1}{1-\alpha}\log\Bigl{(}\sum_{i=1}^{t}{\lambda}^{{\scriptscriptstyle\text{trunc}}^{\large\alpha}}_{i}\Bigr{)}\Bigr{)}

Notably, for order α=1\alpha=1, the tt-truncated Vendi score is:

Vendi1(t)(x1,,xn):=exp(i=1tλitrunclog1λitrunc)\mathrm{Vendi}_{1}^{(t)}(x_{1},\ldots,x_{n}):=\exp\Bigl{(}\sum_{i=1}^{t}{\lambda}^{\scriptscriptstyle\text{trunc}}_{i}\log\frac{1}{{\lambda}^{\scriptscriptstyle\text{trunc}}_{i}}\Bigr{)}
Remark 1.

The above definition of tt-truncated Vendi score leads to the definition of tt-truncated population Vendi Vendiα(t)(PX)\mathrm{Vendi}_{\alpha}^{(t)}(P_{X}), where the mentioned truncation process is applied to the eigenspectrum of the population kernel covariance matrix C~X\widetilde{C}_{X}. Note that the truncated Vendi score is a statistic and function of random samples x1,,xnx_{1},\ldots,x_{n}, whereas the truncated population Vendi is deterministic and a characteristic of the population distribution PXP_{X}.

According to Definition 2, we find the probability model with the minimum 2\ell_{2}-norm difference from the tt-dimensional vector [λ1,,λt][\lambda_{1},\ldots,\lambda_{t}] including only the top-tt eigenvalues. Then, we use the order-α\alpha entropy of the probability model to define the order-α\alpha tt-truncated population Vendi. Our next result shows that this population quantity can be estimated using n=O(t)n=O(t) samples by tt-truncated Vendi score for every kernel function.

Theorem 2.

Consider the setting in Theorem 1. Then, for every n2+8log(1/δ)n\geq 2+8\log(1/\delta), the difference between the tt-truncated population Vendi and the empirical tt-truncated Vendi score of samples x1,,xnx_{1},\ldots,x_{n} is bounded with probability at least 1δ1-\delta:

|Vendiα(t)(x1,,xn)1ααVendiα(t)(PX)1αα|\displaystyle\Bigl{|}\,\mathrm{Vendi}^{(t)}_{\alpha}\bigl{(}x_{1},\ldots,x_{n}\bigr{)}^{\frac{1-\alpha}{\alpha}}-\mathrm{Vendi}^{(t)}_{\alpha}\bigl{(}P_{X}\bigr{)}^{\frac{1-\alpha}{\alpha}}\,\Bigr{|}
\displaystyle\leq\> 32max{1,t2α}log(2/δ)n\displaystyle\sqrt{\frac{32\max\{1,t^{2-\alpha}\}\log\bigl{(}2/\delta\bigr{)}}{n}}

Proof. We defer the proof to the Appendix. As implied by Theorem 2, the tt-truncated population Vendi can be estimated using O(t)O(t) samples, i.e. the truncation parameter tt plays the role of the bounded dimension of a finite-dimension kernel map. Our next theorem shows that the Nyström method [Friedman and Dieng, 2023] and the FKEA method [Ospanov et al., 2024b] for reducing the computational costs of Vendi scores have a bounded difference with the truncated population Vendi.

Theorem 3.

Consider the setting of Theorem 1. (a) Assume that the kernel function is shift-invariant and the FKEA method with tt random Fourier features is used to approximate the Vendi score. Then, for every δ\delta satisfying n2+8log(1/δ)n\geq 2+8\log(1/\delta), with probability at least 1δ1-\delta:

|FKEA-Vendiα(t)(x1,,xn)1ααVendiα(t)(PX)1αα|\displaystyle\Bigl{|}\mathrm{FKEA}\text{-}\mathrm{Vendi}^{(t)}_{\alpha}\bigl{(}x_{1},\ldots,x_{n}\bigr{)}^{\frac{1-\alpha}{\alpha}}-\mathrm{Vendi}^{(t)}_{\alpha}\bigl{(}P_{X}\bigr{)}^{\frac{1-\alpha}{\alpha}}\Bigr{|}
128max{1,t2α}log(3/δ)min{n,t}\displaystyle\,\leq\sqrt{\frac{128\max\{1,t^{2-\alpha}\}\log\bigl{(}3/\delta\bigr{)}}{\min\{n,t\}}}

(b) Assume that the Nyström method is applied with parameter tt for approximating the kernel function. Then, if for some r1r\geq 1, the kernel matrix KK’s rrth-largest eigenvalue satisfies λrτ{\lambda}_{r}\leq\tau and trτlog(n)t\geq r\tau\log(n), the following holds with probability at least 1δ2n31-\delta-2n^{-3}:

|Nystrom-Vendiα(t)(x1,,xn)1ααVendiα(t)(PX)1αα|\displaystyle\Bigl{|}\mathrm{Nystrom}\text{-}\mathrm{Vendi}^{(t)}_{\alpha}\bigl{(}x_{1},\ldots,x_{n}\bigr{)}^{\frac{1-\alpha}{\alpha}}-\mathrm{Vendi}^{(t)}_{\alpha}\bigl{(}P_{X}\bigr{)}^{\frac{1-\alpha}{\alpha}}\Bigr{|}
𝒪(max{1,t2α}log(2/δ)tτ2log(n)2n)\displaystyle\ \leq\mathcal{O}\Bigl{(}\sqrt{\frac{\max\{1,t^{2-\alpha}\}\log\bigl{(}2/\delta\bigr{)}t\tau^{2}\log(n)^{2}}{n}}\Bigr{)}

Proof. We defer the proof to the Appendix.

6 Numerical Results

Refer to caption
Figure 3: Statistical convergence of Vendi score for different sample sizes on FFHQ[Karras et al., 2019] data: (Left plot) finite-dimension cosine similarity kernel (Right plot) infinite dimension Gaussian kernel with bandwidth σ=35\sigma=35. DINOv2 embedding (dimension 768) is used in computing the scores.
Refer to caption
Figure 4: Statistical convergence of Vendi score for different sample sizes on Synthetic Countries data: (Left plot) finite-dimension cosine similarity kernel (Right plot) infinite dimension Gaussian kernel with bandwidth σ=0.6\sigma=0.6. text-embedding-3-large embedding (dimension 3072) is used in computing the scores.
Refer to caption
Figure 5: Statistical convergence of Vendi score for different sample sizes on Kinetics400[Kay et al., 2017] data: (Left plot) finite-dimension cosine similarity kernel (Right plot) infinite dimension Gaussian kernel with bandwidth σ=4.0\sigma=4.0. I3D embedding (dimension 1024) is used in computing the scores.
Refer to caption
Figure 6: Diversity evaluation of Vendi scores on truncated StyleGAN3 generated FFHQ dataset with varying truncation coefficient ψ\psi. Fixed sample size n=n=20k is used for estimating the scores.
Refer to caption
Figure 7: Diversity evaluation of Vendi scores on truncated StyleGAN-XL generated ImageNet dataset with varying truncation coefficient ψ\psi. Fixed sample size n=n=20k is used for estimating the scores.

We evaluated the convergence of the Vendi score, the truncated Vendi score, and the proxy Vendi scores using the Nyström method and FKEA in our numerical experiments. We provide a comparative analysis of these scores across different data types and models, including image, text, and video. In our experiments, we considered the cosine similarity kernel as a standard kernel function with a finite-dimension map and the Gaussian (RBF) kernel as a kernel function with an infinite-dimension feature map. In the experiments with Gaussian kernels, we matched the kernel bandwidth parameter with those chosen by [Jalali et al., 2023, Ospanov et al., 2024b] for the same datasets. We used 20,000 number of samples per score computation, consistent with standard practice in the literature. To investigate how computation-cutting methods compare to each other, in the experiments we matched the truncation parameter tt of our defined tt-truncated Vendi score with the Nyström method’s hyperparameter on the number of randomly selected rows of kernel matrix and the FKEA’s hyperparameter of the number of random Fourier features. The Vendi and FKEA implementations were adopted from the corresponding references’ GitHub webpages, while the Nyström method was adopted from the scikit-learn Python package.

6.1 Convergence Analysis of Vendi Scores

To assess the convergence of the discussed Vendi scores, we conducted experiments on four datasets including ImageNet and FFHQ [Karras et al., 2019] image datasets, a synthetic text dataset with 400k paragraphs generated by GPT-4 about 100 randomly selected countries, and the Kinetics video dataset [Kay et al., 2017]. Our results, presented in Figures 3, 4, and 5, show that for the finite-dimension cosine similarity kernel the Vendi score converges rapidly to the underlying value and the proxy versions including truncated and Nyström Vendi scores were almost identical to the original Vendi score. This observation is consistent with our theoretical results on the convergence of Vendi scores under finite-dimension kernel maps. On the other hand, in the case of infinite dimension Gaussian kernel, we observed that the Vendi1\mathrm{Vendi}_{1} score did not converge using 20k samples and the score value kept growing with a considerable rate. However, the tt-truncated Vendi score with t=10000t=10000 converged to its underlying statistic shortly after 10000 samples were used. Consistent with our theoretical result, the proxy Nyström and FKEA estimated scores with their rank hyperparameter matched with tt also converged to the limit of the truncated Vendi scores. The numerical results show the connection between the truncated Vendi score and the existing kernel methods for approximating the Vendi score.

6.2 Correlation between the truncated Vendi score and diversity of data

We performed experiments to test the correlation between the truncated Vendi score and the ground-truth diversity of data. To do this, we applied the truncation technique to the FFHQ-based StyleGAN3 [Karras et al., 2021] model and the ImageNet-based StyleGAN-XL [Sauer et al., 2022] model and simulated generative models with different underlying diversity by varying the truncation technique. Considering the Gaussian kernel, we estimated the tt-truncated Vendi score with t=10000t=10000 by averaging the estimated tt-truncated Vendi scores over 55 independent datasets of size 20k where the score seemed to converge to its underlying value. Figures 67 show how the estimated statistic correlates with the truncation parameter for order-α\alpha Vendi scores with α=1, 1.5,2\alpha=1,\,1.5,2. In all these experiments, the estimated truncated Vendi score correlated with the underlying diversity of the models. In addition, we plot the proxy Nyström and FKEA proxy Vendi values computed using 20000 samples which remain close to the estimated tt-truncated statistic. These empirical results suggest that the estimated tt-truncated Vendi score with Gaussian kernel can be used to evaluate the diversity of generated data. Also, the Nyström and FKEA methods were both computationally efficient in estimating the truncated Vendi score from limited generated data. We defer the presentation of the additional numerical results on the convergence of Vendi scores with different orders, kernel functions and embedding spaces to the Appendix.

7 Conclusion

In this work, we investigated the statistical convergence behavior of Vendi diversity scores estimated from empirical samples. We highlighted that, due to the high computational complexity of the score for datasets larger than a few tens of thousands of generated data points, the score is often calculated using sample sizes below 10,000. We demonstrated that such restricted sample sizes do not pose a problem for statistical convergence as long as the kernel feature dimension is bounded. However, our numerical results showed a lack of convergence to the population Vendi when using an infinite-dimensional kernel map, such as the Gaussian kernel. To address this gap, we introduced the truncated population Vendi as an alternative target quantity for diversity evaluation. We showed that existing Nyström and FKEA methods for approximating Vendi scores concentrate around this truncated population Vendi. An interesting future direction is to explore the relationship between other kernel approximation techniques and the truncated population Vendi. Also, a comprehensive analysis of the computational-statistical trade-offs involved in estimating the Vendi score is another relevant future direction.

Acknowledgements.
This work is partially supported by a grant from the Research Grants Council of the Hong Kong Special Administrative Region, China, Project 14209920, and is partially supported by CUHK Direct Research Grants with CUHK Project No. 4055164 and 4937054. Finally, the authors thank the anonymous reviewers for their thoughtful feedback and constructive suggestions.

References

  • Aronszajn [1950] N. Aronszajn. Theory of reproducing kernels. Transactions of the American Mathematical Society, 68(3):337–404, 1950. 10.2307/1990404.
  • Bach [2022] Francis Bach. Information Theory with Kernel Methods, August 2022. URL http://arxiv.org/abs/2202.08545. arXiv:2202.08545 [cs, math, stat].
  • Bach and Jordan [2002] Francis R Bach and Michael I Jordan. Kernel independent component analysis. In Journal of Machine Learning Research, volume 3, pages 1–48, 2002.
  • Bińkowski et al. [2018] Mikołaj Bińkowski, Danica J Sutherland, Michael Arbel, and Arthur Gretton. Demystifying mmd gans. In International Conference on Learning Representations, 2018.
  • Caron et al. [2020] Mathilde Caron, Ishan Misra, Julien Mairal, Priya Goyal, Piotr Bojanowski, and Armand Joulin. Unsupervised Learning of Visual Features by Contrasting Cluster Assignments. In Advances in Neural Information Processing Systems, volume 33, pages 9912–9924. Curran Associates, Inc., 2020. URL https://proceedings.neurips.cc/paper/2020/hash/70feb62b69f16e0238f741fab228fec2-Abstract.html.
  • Deng et al. [2009] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition (CVPR), pages 248–255. IEEE, 2009.
  • Dong et al. [2023] Yuxin Dong, Tieliang Gong, Shujian Yu, and Chen Li. Optimal randomized approximations for matrix-based rényi’s entropy. IEEE Transactions on Information Theory, 2023.
  • Fine and Scheinberg [2001] Shai Fine and Katya Scheinberg. Efficient svm training using low-rank kernel representations. In Journal of Machine Learning Research (JMLR), pages 243–250, 2001.
  • Friedman and Dieng [2023] Dan Friedman and Adji Bousso Dieng. The vendi score: A diversity evaluation metric for machine learning. In Transactions on Machine Learning Research, 2023.
  • Gong et al. [2025] Shizhan Gong, Yankai Jiang, Qi Dou, and Farzan Farnia. Kernel-based unsupervised embedding alignment for enhanced visual representation in vision-language models. arXiv preprint arXiv:2506.02557, 2025.
  • Gross [2011] David Gross. Recovering low-rank matrices from few coefficients in any basis. IEEE Transactions on Information Theory, 57(3):1548–1566, 2011.
  • Heusel et al. [2018] Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. GANs trained by a two time-scale update rule converge to a local nash equilibrium. In Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2018.
  • Hu et al. [2024] Xiaoyan Hu, Ho-fung Leung, and Farzan Farnia. An online learning approach to prompt-based selection of generative models. arXiv preprint arXiv:2410.13287, 2024.
  • Hu et al. [2025] Xiaoyan Hu, Ho-fung Leung, and Farzan Farnia. A multi-armed bandit approach to online selection and evaluation of generative models. In International Conference on Artificial Intelligence and Statistics, pages 1864–1872. PMLR, 2025.
  • Jalali et al. [2023] Mohammad Jalali, Cheuk Ting Li, and Farzan Farnia. An information-theoretic evaluation of generative models in learning multi-modal distributions. In Advances in Neural Information Processing Systems, volume 36, pages 9931–9943, 2023.
  • Jalali et al. [2024] Mohammad Jalali, Azim Ospanov, Amin Gohari, and Farzan Farnia. Conditional vendi score: An information-theoretic approach to diversity evaluation of prompt-based generative models. arXiv preprint arXiv:2411.02817, 2024.
  • Jalali et al. [2025] Mohammad Jalali, Bahar Dibaei Nia, and Farzan Farnia. Towards an explainable comparison and alignment of feature embeddings. arXiv preprint arXiv:2506.06231, 2025.
  • Karras et al. [2019] Tero Karras, Samuli Laine, and Timo Aila. A style-based generator architecture for generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 4401–4410, 2019.
  • Karras et al. [2021] Tero Karras, Miika Aittala, Samuli Laine, Erik Härkönen, Janne Hellsten, Jaakko Lehtinen, and Timo Aila. Alias-free generative adversarial networks. In Advances in Neural Information Processing Systems, volume 34, pages 852–863. Curran Associates, Inc., 2021.
  • Kay et al. [2017] Will Kay, Joao Carreira, Karen Simonyan, Brian Zhang, Chloe Hillier, Sudheendra Vijayanarasimhan, Fabio Viola, Tim Green, Trevor Back, Paul Natsev, Mustafa Suleyman, and Andrew Zisserman. The kinetics human action video dataset, 2017.
  • Kohler and Lucchi [2017] Jonas Moritz Kohler and Aurelien Lucchi. Sub-sampled cubic regularization for non-convex optimization. In International Conference on Machine Learning, pages 1895–1904. PMLR, 2017.
  • Kynkäänniemi et al. [2019] Tuomas Kynkäänniemi, Tero Karras, Samuli Laine, Jaakko Lehtinen, and Timo Aila. Improved precision and recall metric for assessing generative models. In Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc., 2019.
  • Kynkäänniemi et al. [2022] Tuomas Kynkäänniemi, Tero Karras, Miika Aittala, Timo Aila, and Jaakko Lehtinen. The Role of ImageNet Classes in Fréchet Inception Distance. September 2022. URL https://openreview.net/forum?id=4oXTQ6m_ws8.
  • Mahoney and Drineas [2009] Michael W. Mahoney and Petros Drineas. Cur matrix decompositions for improved data analysis. In Proceedings of the National Academy of Sciences, volume 106, pages 697–702, 2009.
  • Naeem et al. [2020] Muhammad Ferjad Naeem, Seong Joon Oh, Youngjung Uh, Yunjey Choi, and Jaejun Yoo. Reliable fidelity and diversity metrics for generative models. In Proceedings of the 37th International Conference on Machine Learning, volume 119 of ICML’20, pages 7176–7185. JMLR.org, 2020.
  • Oquab et al. [2023] Maxime Oquab, Timothée Darcet, Théo Moutakanni, Huy V. Vo, Marc Szafraniec, Vasil Khalidov, Pierre Fernandez, Daniel Haziza, Francisco Massa, Alaaeldin El-Nouby, Mido Assran, Nicolas Ballas, Wojciech Galuba, Russell Howes, Po-Yao Huang, Shang-Wen Li, Ishan Misra, Michael Rabbat, Vasu Sharma, Gabriel Synnaeve, Hu Xu, Herve Jegou, Julien Mairal, Patrick Labatut, Armand Joulin, and Piotr Bojanowski. DINOv2: Learning robust visual features without supervision. In Transactions on Machine Learning Research, 2023. URL https://openreview.net/forum?id=a68SUt6zFt.
  • Ospanov et al. [2024a] Azim Ospanov, Mohammad Jalali, and Farzan Farnia. Dissecting clip: Decomposition with a schur complement-based approach. arXiv preprint arXiv:2412.18645, 2024a.
  • Ospanov et al. [2024b] Azim Ospanov, Jingwei Zhang, Mohammad Jalali, Xuenan Cao, Andrej Bogdanov, and Farzan Farnia. Towards a scalable reference-free evaluation of generative models. In Advances in Neural Information Processing Systems, volume 38, 2024b.
  • Pasarkar and Dieng [2024] Amey Pasarkar and Adji Bousso Dieng. Cousins of the vendi score: A family of similarity-based diversity metrics for science and machine learning. In International Conference on Artificial Intelligence and Statistics. PMLR, 2024.
  • Pillutla et al. [2021] Krishna Pillutla, Swabha Swayamdipta, Rowan Zellers, John Thickstun, Sean Welleck, Yejin Choi, and Zaid Harchaoui. MAUVE: Measuring the gap between neural text and human text using divergence frontiers. In A. Beygelzimer, Y. Dauphin, P. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, 2021. URL https://openreview.net/forum?id=Tqx7nJp7PR.
  • Pillutla et al. [2023] Krishna Pillutla, Lang Liu, John Thickstun, Sean Welleck, Swabha Swayamdipta, Rowan Zellers, Sewoong Oh, Yejin Choi, and Zaid Harchaoui. MAUVE Scores for Generative Models: Theory and Practice. JMLR, 2023.
  • Radford et al. [2021] Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, and Ilya Sutskever. Learning Transferable Visual Models From Natural Language Supervision. In International Conference on Machine Learning, pages 8748–8763. arXiv, February 2021. 10.48550/arXiv.2103.00020. URL http://arxiv.org/abs/2103.00020. arXiv:2103.00020 [cs].
  • Rezaei et al. [2024] Parham Rezaei, Farzan Farnia, and Cheuk Ting Li. Be more diverse than the most diverse: Online selection of diverse mixtures of generative models. arXiv preprint arXiv:2412.17622, 2024.
  • Rényi [1961] Alfréd Rényi. On measures of entropy and information. In Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability, Volume 1: Contributions to the Theory of Statistics, pages 547–561. University of California Press, 1961.
  • Sajjadi et al. [2018] Mehdi S. M. Sajjadi, Olivier Bachem, Mario Lucic, Olivier Bousquet, and Sylvain Gelly. Assessing generative models via precision and recall. In Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc., 2018.
  • Sauer et al. [2022] Axel Sauer, Katja Schwarz, and Andreas Geiger. Stylegan-xl: Scaling stylegan to large diverse datasets. In ACM SIGGRAPH 2022 Conference Proceedings, volume abs/2201.00273, 2022. URL https://arxiv.org/abs/2201.00273.
  • Shawe-Taylor et al. [2005] J. Shawe-Taylor, C.K.I. Williams, N. Cristianini, and J. Kandola. On the eigenspectrum of the gram matrix and the generalization error of kernel-pca. IEEE Transactions on Information Theory, 51(7):2510–2522, 2005. 10.1109/TIT.2005.850052.
  • Stein et al. [2023] George Stein, Jesse Cresswell, Rasa Hosseinzadeh, Yi Sui, Brendan Ross, Valentin Villecroze, Zhaoyan Liu, Anthony L Caterini, Eric Taylor, and Gabriel Loaiza-Ganem. Exposing flaws of generative model evaluation metrics and their unfair treatment of diffusion models. In A. Oh, T. Naumann, A. Globerson, K. Saenko, M. Hardt, and S. Levine, editors, Advances in Neural Information Processing Systems, volume 36, pages 3732–3784. Curran Associates, Inc., 2023.
  • Szegedy et al. [2016] Christian Szegedy, Vincent Vanhoucke, Sergey Ioffe, Jon Shlens, and Zbigniew Wojna. Rethinking the inception architecture for computer vision. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016.
  • Wang et al. [2023] Zixiao Wang, Farzan Farnia, Zhenghao Lin, Yunheng Shen, and Bei Yu. On the distributed evaluation of generative models. arXiv preprint arXiv:2310.11714, 2023.
  • Williams and Seeger [2000] Christopher Williams and Matthias Seeger. Using the nyström method to speed up kernel machines. In Advances in neural information processing systems, pages 682–688, 2000.
  • Wu et al. [2025] Youqi Wu, Jingwei Zhang, and Farzan Farnia. Fusing cross-modal and uni-modal representations: A kronecker product approach, 2025. URL https://arxiv.org/abs/2506.08645.
  • Xu et al. [2015] Zenglin Xu, Rong Jin, Bin Shen, and Shenghuo Zhu. Nystrom approximation for sparse kernel methods: Theoretical analysis and empirical evaluation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 29, 2015.
  • Zhang et al. [2024] Jingwei Zhang, Cheuk Ting Li, and Farzan Farnia. An interpretable evaluation of entropy-based novelty of generative models. arXiv preprint arXiv:2402.17287, 2024.
  • Zhang et al. [2025] Jingwei Zhang, Mohammad Jalali, Cheuk Ting Li, and Farzan Farnia. Unveiling differences in generative models: A scalable differential clustering approach. In Proceedings of the Computer Vision and Pattern Recognition Conference, pages 8269–8278, 2025.

Supplementary Material

Appendix A Proofs

A.1 Proof of Theorem 1

To prove the theorem, we will use the following lemma followed from [Gross, 2011, Kohler and Lucchi, 2017].

Lemma 1 (Vector Bernstein Inequality [Gross, 2011, Kohler and Lucchi, 2017]).

Suppose that z1,,znz_{1},\ldots,z_{n} are independent and identically distributed random vectors with zero mean 𝔼[zi]=𝟎\mathbb{E}[z_{i}]=\mathbf{0} and bounded 2\ell_{2}-norm zi2c\|z_{i}\|_{2}\leq c. Then, for every 0ϵc0\leq\epsilon\leq c, the following holds

(1ni=1nzi2ϵ)exp(nϵ28c2+14)\mathbb{P}\biggl{(}\,\Bigl{\|}\frac{1}{n}\sum_{i=1}^{n}z_{i}\Bigr{\|}_{2}\geq\epsilon\,\biggr{)}\,\leq\,\exp\Bigl{(}-\frac{n\epsilon^{2}}{8c^{2}}+\frac{1}{4}\Bigr{)}

We apply the above Vector Bernstein Inequality to the random vectors ϕ(x1)ϕ(x1),,ϕ(x1)ϕ(x1)\phi(x_{1})\otimes\phi(x_{1}),\ldots,\phi(x_{1})\otimes\phi(x_{1}) where \otimes denotes the Kronecker product. To do this, we define vector vi=ϕ(xi)ϕ(xi)𝔼xP[ϕ(x)ϕ(x)]v_{i}=\phi(x_{i})\otimes\phi(x_{i})-\mathbb{E}_{x\sim P}\bigl{[}\phi(x)\otimes\phi(x)\bigr{]} for every ii. Note that viv_{i} is, by definition, a zero-mean vector and also for every xx we have the following for the normalized kernel function kk:

ϕ(x)ϕ(x)22=ϕ(x)22ϕ(x)22=k(x,x)k(x,x)=1\bigl{\|}\phi(x)\otimes\phi(x)\bigr{\|}^{2}_{2}=\bigl{\|}\phi(x)\bigr{\|}^{2}_{2}\cdot\bigl{\|}\phi(x)\bigr{\|}^{2}_{2}=k(x,x)\cdot k(x,x)=1

Then, the triangle inequality implies that

vi2ϕ(xi)ϕ(xi)2+𝔼xP[ϕ(x)ϕ(x)]2ϕ(xi)ϕ(xi)2+𝔼xP[ϕ(x)ϕ(x)2]=2\bigl{\|}v_{i}\bigr{\|}_{2}\leq\bigl{\|}\phi(x_{i})\otimes\phi(x_{i})\bigr{\|}_{2}+\bigl{\|}\mathbb{E}_{x\sim P}\bigl{[}\phi(x)\otimes\phi(x)\bigr{]}\bigr{\|}_{2}\leq\bigl{\|}\phi(x_{i})\otimes\phi(x_{i})\bigr{\|}_{2}+\mathbb{E}_{x\sim P}\bigl{[}\bigl{\|}\phi(x)\otimes\phi(x)\bigr{\|}_{2}\bigr{]}=2

As a result, the Vector Bernstein Inequality leads to the following for every 0ϵ20\leq\epsilon\leq 2:

(1ni=1nϕ(xi)ϕ(xi)𝔼xP[ϕ(x)ϕ(x)]2ϵ)exp(8nϵ232)\mathbb{P}\Bigl{(}\Bigl{\|}\frac{1}{n}\sum_{i=1}^{n}\phi(x_{i})\otimes\phi(x_{i})-\mathbb{E}_{x\sim P}\bigl{[}\phi(x)\otimes\phi(x)\bigr{]}\Bigr{\|}_{2}\geq\epsilon\Bigr{)}\,\leq\,\exp\Bigl{(}\frac{8-n\epsilon^{2}}{32}\Bigr{)}

On the other hand, note that ϕ(x)ϕ(x)\phi(x)\otimes\phi(x) is the vectorized version of rank-1 ϕ(x)ϕ(x)\phi(x)\phi(x)^{\top}, which shows that the above inequality is equivalent to the following where HS\|\cdot\|_{\mathrm{HS}} denotes the Hilbert-Schmidt norm, which will simplify to the Frobenius norm in the finite dimension case,

(1ni=1n[ϕ(xi)ϕ(xi)]𝔼xP[ϕ(x)ϕ(x)]HSϵ)exp(8nϵ232)\displaystyle\mathbb{P}\Bigl{(}\Bigl{\|}\frac{1}{n}\sum_{i=1}^{n}\bigl{[}\phi(x_{i})\phi(x_{i})^{\top}\bigr{]}-\mathbb{E}_{x\sim P}\bigl{[}\phi(x)\phi(x)^{\top}\bigr{]}\Bigr{\|}_{\mathrm{HS}}\geq\epsilon\Bigr{)}\,\leq\,\exp\Bigl{(}\frac{8-n\epsilon^{2}}{32}\Bigr{)}
\displaystyle\Longrightarrow\;\; (CXC~XHSϵ)exp(8nϵ232)\displaystyle\mathbb{P}\Bigl{(}\Bigl{\|}C_{X}-\widetilde{C}_{X}\Bigr{\|}_{\mathrm{HS}}\geq\epsilon\Bigr{)}\,\leq\,\exp\Bigl{(}\frac{8-n\epsilon^{2}}{32}\Bigr{)}

Subsequently, we can apply the Hoffman-Wielandt inequality which shows that for the sorted eigenvalue vectors of CXC_{X} (denoted by 𝝀^n\widehat{\boldsymbol{\lambda}}_{n} in the theorem) and C~X\widetilde{C}_{X} (denoted by 𝝀~\widetilde{\boldsymbol{\lambda}} in the theorem) we will have 𝝀^n𝝀~2CXC~XHS\|\widehat{\boldsymbol{\lambda}}_{n}-\widetilde{\boldsymbol{\lambda}}\|_{2}\leq\|C_{X}-\widetilde{C}_{X}\|_{\mathrm{HS}}, which together with the previous inequality leads to

(𝝀^n𝝀~2ϵ)exp(8nϵ232)\displaystyle\mathbb{P}\Bigl{(}\bigl{\|}\widehat{\boldsymbol{\lambda}}_{n}-\widetilde{\boldsymbol{\lambda}}\bigr{\|}_{2}\geq\epsilon\Bigr{)}\,\leq\,\exp\Bigl{(}\frac{8-n\epsilon^{2}}{32}\Bigr{)}

If we define δ=exp((8nϵ2)/32)\delta=\exp\bigl{(}(8-n\epsilon^{2})/32\bigr{)} that implies ϵ32log(2/δ)n\epsilon\leq\sqrt{\frac{32\log(2/\delta)}{n}}, we obtain the following for every δexp((2n)/8)\delta\geq\exp\bigl{(}(2-n)/8\bigr{)} (since we suppose 0ϵ20\leq\epsilon\leq 2)

(𝝀^n𝝀~232log(2/δ)n)δ\displaystyle\mathbb{P}\Bigl{(}\bigl{\|}\widehat{\boldsymbol{\lambda}}_{n}-\widetilde{\boldsymbol{\lambda}}\bigr{\|}_{2}\geq\sqrt{\frac{32\log(2/\delta)}{n}}\Bigr{)}\,\leq\,\delta
\displaystyle\Longrightarrow\;\; (𝝀^n𝝀~232log(2/δ)n) 1δ\displaystyle\mathbb{P}\Bigl{(}\bigl{\|}\widehat{\boldsymbol{\lambda}}_{n}-\widetilde{\boldsymbol{\lambda}}\bigr{\|}_{2}\leq\sqrt{\frac{32\log(2/\delta)}{n}}\Bigr{)}\,\geq\,1-\delta

which completes the proof.

A.2 Proof of Corollary 2

The case of α=1\alpha=1. We show that Theorem 1 on the concentration of the eigenvalues 𝝀=[λ1,,λd]\boldsymbol{\lambda}=[\lambda_{1},\ldots,\lambda_{d}] will further imply a concentration bound for the logarithm of Vendi-1 score. In the case of Vendi1\mathrm{Vendi}_{1} (when α1+\alpha\rightarrow 1^{+}), the concentration bound will be formed for the logarithm of the Vendi score, i.e. the Von-Neumann entropy (denoted as HαH_{\alpha}):

H1(CX):=H1(𝝀)=i=1dλ~ilog1λ~iH_{1}(C_{X}):=H_{1}(\boldsymbol{\lambda})=\sum_{i=1}^{d}\widetilde{\lambda}_{i}\log\frac{1}{\widetilde{\lambda}_{i}}

Theorem 1 shows that 𝝀^n𝝀~232log(2/δ)n\|\widehat{\boldsymbol{\lambda}}_{n}-\widetilde{\boldsymbol{\lambda}}\|_{2}\leq\sqrt{\frac{32\log(2/\delta)}{n}} with probability 1δ1-\delta. To convert this concentration bound to a bound on the order-1 entropy (for Vendi-1 score) difference H1(C^n)H1(CX)H_{1}({\widehat{C}_{n}})-H_{1}(C_{X}), we leverage the following two lemmas:

Lemma 2.

For every 0α,β10\leq\alpha,\beta\leq 1 such that |βα|1e|\beta-\alpha|\leq\frac{1}{e}, we have

|αlog1αβlog1β||βα|log1|βα|\Bigl{|}\alpha\log\frac{1}{\alpha}-\beta\log\frac{1}{\beta}\Bigr{|}\,\leq\,|\beta-\alpha|\log\frac{1}{|\beta-\alpha|}

Proof. Let c=|αβ|c=|\alpha-\beta|, where c[0,1e]c\in[0,\frac{1}{e}]. Defining g(z)=zlog(1z)g(z)=z\log(\frac{1}{z}), the first-order optimality condition g(z)=log(z)1=0g^{\prime}(z)=-\log(z)-1=0 yields 1e\frac{1}{e} as the local maximum of g(z)g(z). Therefore, there are three cases of placement of α\alpha and β\beta on the interval [0,1][0,1]: α\alpha and β\beta appear before maximum point, after maximum point or maximum point is between α\alpha and β\beta. We show that regardless of the placement of α\alpha and β\beta, the above inequality remains true.

  • Case 1: α,β[0,1e]\alpha,\beta\in[0,\frac{1}{e}]. Note that g′′(z)=1zg^{\prime\prime}(z)=-\frac{1}{z}. Since the second-order derivative is negative and the function gg is monotonically increasing within the interval [0,1e][0,\frac{1}{e}], the gap between g(α)g(\alpha) and g(β)g(\beta) is maximized when α=0\alpha^{*}=0 and β=cα=c\beta^{*}=c-\alpha^{*}=c. This directly leads to the desired bound as follows:

    |αlog1αβlog1β||αlog1αβlog1β|=|0log0clog1c|clog1c\Bigl{|}\alpha\log\frac{1}{\alpha}-\beta\log\frac{1}{\beta}\Bigr{|}\,\leq\,\Bigl{|}\alpha^{*}\log\frac{1}{\alpha^{*}}-\beta^{*}\log\frac{1}{\beta^{*}}\Bigr{|}\,=\,\bigl{|}0\log 0-c\log\frac{1}{c}\bigr{|}\,\leq\,c\log\frac{1}{c}

    Here, we use the standard limit 0log0=00\log 0=0.

  • Case 2: α,β[1e,1]\alpha,\beta\in[\frac{1}{e},1]. In this case, we note that gg is concave yet decreasing over [1e,1][\frac{1}{e},1], and so the gap between g(α)g(\alpha) and g(β)g(\beta) will be maximized when α=1c\alpha^{*}=1-c and β=1\beta^{*}=1. This leads to:

    |αlog1αβlog1β||αlog1αβlog1β|=(1c)log1(1c)clog1c\Bigl{|}\alpha\log\frac{1}{\alpha}-\beta\log\frac{1}{\beta}\Bigr{|}\,\leq\,\Bigl{|}\alpha^{*}\log\frac{1}{\alpha^{*}}-\beta^{*}\log\frac{1}{\beta^{*}}\Bigr{|}\,=\,(1-c)\log\frac{1}{(1-c)}\leq c\log\frac{1}{c}

    where the last inequality holds because c[0,1e]c\in[0,\frac{1}{e}], and if we define the function h(c)=clog1c(1c)log11ch(c)=c\log\frac{1}{c}-(1-c)\log\frac{1}{1-c}, then we have h(c)=log1c(1c)2h^{\prime}(c)=\log\frac{1}{c(1-c)}-2, which is positive over c[0,c0]c\in[0,c_{0}] (e2<c0<e1e^{-2}<c_{0}<e^{-1} is where c0(1c0)=e2c_{0}(1-c_{0})=e^{-2}), and then negative over [c0,1e][c_{0},\frac{1}{e}], and hence h(c)min{h(0),h(1/e)}=0h(c)\geq\min\{h(0),h(1/e)\}=0 for every c[0,1/e]c\in[0,1/e].

  • Case 3: α[0,1e)\alpha\in[0,\frac{1}{e}) and β(1e,1]\beta\in(\frac{1}{e},1]. When α\alpha and β\beta lie on the opposite ends from the maximum point, the inequality becomes:

    |αlog1αβlog1β|max{|(1/e)log11/eβlog1β|,|αlog1α(1/e)log11/e|}clog1c\bigl{|}\alpha\log\frac{1}{\alpha}-\beta\log\frac{1}{\beta}\bigr{|}\leq\mathrm{max}\Bigr{\{}\bigl{|}(1/e)\log\frac{1}{1/e}-\beta\log\frac{1}{\beta}\bigr{|},\bigr{|}\alpha\log\frac{1}{\alpha}-(1/e)\log\frac{1}{1/e}\bigl{|}\Bigl{\}}\leq c\log\frac{1}{c}

    since we pick the side with the largest difference, this difference is upper bounded by either Case 1 or Case 2 because max{|1eβ|,|α1e|}<c\mathrm{max}\{|\frac{1}{e}-\beta|,|\alpha-\frac{1}{e}|\}<c. Therefore, this case is upper-bounded by clog1cc\log\frac{1}{c}.

All the three cases of placement of α\alpha and β\beta are upper-bounded by clog1cc\log\frac{1}{c}; Therefore, the claim holds.

Lemma 3.

If 𝐮2ϵ\|\mathbf{u}\|_{2}\leq\epsilon for dd-dimensional vector 𝐮𝟎\mathbf{u}\geq\mathbf{0} where ϵ1e\epsilon\leq\frac{1}{e}, then we have

i=1duilog1uiϵdlogdϵ\sum_{i=1}^{d}u_{i}\log\frac{1}{u_{i}}\leq\epsilon\sqrt{d}\log\frac{\sqrt{d}}{\epsilon}

Proof. We prove the above inequality using the KKT conditions for the following maximization problem, representing a convex optimization problem,

max𝐮di=1duilog(1ui)\displaystyle\max_{\mathbf{u}\in\mathbb{R}^{d}}\qquad\;\;\>\sum_{i=1}^{d}u_{i}\log(\frac{1}{u_{i}})
subject toui0,for alli\displaystyle\text{\rm subject to}\quad u_{i}\geq 0,\;\;\text{\rm for all}\>i
i=1dui2ϵ2(equivalent to 𝐮2ϵ)\displaystyle\qquad\qquad\quad\sum_{i=1}^{d}u_{i}^{2}\leq\epsilon^{2}\;\;\ (\text{equivalent to }\|\mathbf{u}\|_{2}\leq\epsilon)

In a concave maximization problem subject to convex constraints, any point that satisfies the KKT conditions is guaranteed to be a global optimum. Let us pick the following solution 𝐮=ϵd𝟏\mathbf{u}^{*}=\frac{\epsilon}{\sqrt{d}}\mathbf{1} and slack variables λ=d2ϵ(log(dϵ)1)\lambda^{*}=\frac{\sqrt{d}}{2\epsilon}\bigr{(}\log(\frac{\sqrt{d}}{\epsilon})-1\bigl{)}, iμi=0\forall_{i}\;\mu_{i}^{*}=0. The Lagrangian of the above problem:

L(𝐮,λ,μ1,,μd)=i=1duilog(1ui)+λ(ϵ2i=1dui2)i=1dμiuiL(\mathbf{u},\lambda,\mu_{1},\dots,\mu_{d})=\sum_{i=1}^{d}u_{i}\log(\frac{1}{u_{i}})+\lambda(\epsilon^{2}-\sum_{i=1}^{d}u_{i}^{2})-\sum_{i=1}^{d}\mu_{i}u_{i}
  • Primal Feasibility. The solution 𝐮\mathbf{u^{*}} satisfies the primal feasibility, since ϵ2i=1d(ϵd)2=ϵ2dϵ2d=0\epsilon^{2}-\sum_{i=1}^{d}(\frac{\epsilon}{\sqrt{d}})^{2}=\epsilon^{2}-d\frac{\epsilon^{2}}{d}=0 and ϵd0\frac{\epsilon}{\sqrt{d}}\geq 0.

  • Dual Feasibility. λ0\lambda^{*}\geq 0 is feasible because of the assumption ϵ1e\epsilon\leq\frac{1}{e} implying that dϵe\frac{\sqrt{d}}{\epsilon}\geq e for every integer dimension d1d\geq 1. Note that this implies λ=d2ϵ(log(dϵ)1)0\lambda^{*}=\frac{\sqrt{d}}{2\epsilon}\bigr{(}\log(\frac{\sqrt{d}}{\epsilon})-1\bigl{)}\geq 0.

  • Complementary Slackness. Since λ(ϵ2i=1d(ϵd)2)=λ0=0\lambda^{*}\bigr{(}\epsilon^{2}-\sum_{i=1}^{d}(\frac{\epsilon}{\sqrt{d}})^{2}\bigl{)}=\lambda^{*}\cdot 0=0, the condition is satisfied.

  • Stationarity. The condition is satisfied as follows:

    uiL(𝐮)\displaystyle\frac{\partial}{\partial u_{i}}L(\mathbf{u^{*}}) =log(ui)12λui+μi=log(ϵd)12d2ϵ(log(ϵd)1)ϵd=0\displaystyle=-\log(u_{i}^{*})-1-2\lambda^{*}u_{i}^{*}+\mu_{i}^{*}=-\log(\frac{\epsilon}{\sqrt{d}})-1-2\cdot\frac{\sqrt{d}}{2\epsilon}\bigr{(}-\log(\frac{\epsilon}{\sqrt{d}})-1\bigl{)}\cdot\frac{\epsilon}{\sqrt{d}}=0

Since all KKT conditions are satisfied and sufficient for global optimality, 𝐮=ϵd𝟏\mathbf{u}^{*}=\frac{\epsilon}{\sqrt{d}}\mathbf{1} is a global optimum of the specified concave maximization problem. We note that this result is also implied by the Schur-concavity property of entropy. Following this result, the specified objective is upper-bounded as follows:

i=1duilog1uiϵdlogdϵ\sum_{i=1}^{d}u_{i}\log\frac{1}{u_{i}}\leq\epsilon\sqrt{d}\log\frac{\sqrt{d}}{\epsilon}

Therefore, the lemma’s proof is complete.

Following the above lemmas, knowing that λ^nλ~232log(2/δ)n\|\widehat{\lambda}_{n}-\widetilde{\lambda}\|_{2}\leq\sqrt{\frac{32\log(2/\delta)}{n}} from Theorem 1 and using the assumption n32e2log(2/δ)236.5log(2/δ)n\geq 32e^{2}\log(2/\delta)\approx 236.5\log(2/\delta) that ensures the upper-bound satisfies 32log(2/δ)n1e\sqrt{\frac{32\log(2/\delta)}{n}}\leq\frac{1}{e}, we can apply the above two lemmas to show that with probability 1δ1-\delta:

|H1(C^n)H1(CX)|=|H1(λ^n)H1(λ~)|8dlog(2/δ)nlog(nd32log(2/δ))\Bigl{|}H_{1}({\widehat{C}_{n}})-H_{1}(C_{X})\Bigr{|}=\Bigl{|}H_{1}(\widehat{\lambda}_{n})-H_{1}(\widetilde{\lambda})\Bigr{|}\leq\sqrt{\frac{8d\log(2/\delta)}{n}}\log\Bigl{(}\frac{nd}{32\log(2/\delta)}\Bigr{)}

Note that under a kernel function with finite dimension dd, the above bound will be 𝒪(dnlog(nd))\mathcal{O}\Bigl{(}\sqrt{\frac{d}{n}}\log\bigl{(}nd\bigr{)}\Bigr{)}.

The case of 1<α<21<\alpha<2. Note that the inequality vαd2α2v2\|v\|_{\alpha}\leq d^{\frac{2-\alpha}{2}}\|v\|_{2} holds for every dd-dimensional vector vdv\in\mathbb{R}^{d}. Therefore, we can repeat the proof of Corollary 1 to show the following for every 1<α<21<\alpha<2

|Vendiα(x1,,xn)1ααVendiα(Px)1αα|\displaystyle\Bigl{|}\mathrm{Vendi}_{\alpha}(x_{1},\ldots,x_{n})^{\frac{1-\alpha}{\alpha}}-\mathrm{Vendi}_{\alpha}(P_{x})^{\frac{1-\alpha}{\alpha}}\Bigr{|}\, =|𝝀^nα𝝀~α|\displaystyle=\,\Bigl{|}\bigl{\|}\widehat{\boldsymbol{\lambda}}_{n}\bigr{\|}_{\alpha}-\bigl{\|}\widetilde{\boldsymbol{\lambda}}\bigr{\|}_{\alpha}\Bigr{|}
𝝀^n𝝀~α\displaystyle\leq\,\bigl{\|}\widehat{\boldsymbol{\lambda}}_{n}-\widetilde{\boldsymbol{\lambda}}\bigr{\|}_{\alpha}
d2α2𝝀^n𝝀~2.\displaystyle\leq\,d^{\frac{2-\alpha}{2}}\bigl{\|}\widehat{\boldsymbol{\lambda}}_{n}-\widetilde{\boldsymbol{\lambda}}\bigr{\|}_{2}.

Consequently, Theorem 1 impies that for every 1α<21\leq\alpha<2 and δexp((2n)/8)\delta\geq\exp((2-n)/8), the following holds with probability at least 1δ1-\delta

|Vendiα(x1,,xn)1ααVendiα(Px)1αα|d2α232log(2/δ)n=32d2αlog(2/δ)n\Bigl{|}\mathrm{Vendi}_{\alpha}(x_{1},\ldots,x_{n})^{\frac{1-\alpha}{\alpha}}-\mathrm{Vendi}_{\alpha}(P_{x})^{\frac{1-\alpha}{\alpha}}\Bigr{|}\,\leq\,d^{\frac{2-\alpha}{2}}\sqrt{\frac{32\log(2/\delta)}{n}}\,=\,\sqrt{\frac{32d^{2-\alpha}\log(2/\delta)}{n}}

A.3 Proof of Corollary 1

Considering the α\alpha-norm definition 𝐯α=(i=1d|vi|α)1/α\|\mathbf{v}\|_{\alpha}=\bigl{(}\sum_{i=1}^{d}|v_{i}|^{\alpha}\bigr{)}^{1/\alpha}, we can rewrite the order-α\alpha Vendi definition as

Vendiα(x1,,xn)=𝝀^nαα1αVendiα(x1,,xn)1αα=𝝀^nα\mathrm{Vendi}_{\alpha}(x_{1},\ldots,x_{n})=\bigl{\|}\widehat{\boldsymbol{\lambda}}_{n}\bigr{\|}^{\frac{\alpha}{1-\alpha}}_{\alpha}\quad\Longleftrightarrow\quad\mathrm{Vendi}_{\alpha}(x_{1},\ldots,x_{n})^{\frac{1-\alpha}{\alpha}}=\bigl{\|}\widehat{\boldsymbol{\lambda}}_{n}\bigr{\|}_{\alpha}

where 𝝀^n\widehat{\boldsymbol{\lambda}}_{n} is defined in Theorem 1. Similarly, given the definition of 𝝀~\widetilde{\boldsymbol{\lambda}} we can write

Vendiα(Px)1αα=𝝀~α\mathrm{Vendi}_{\alpha}(P_{x})^{\frac{1-\alpha}{\alpha}}=\bigl{\|}\widetilde{\boldsymbol{\lambda}}\bigr{\|}_{\alpha}

Therefore, for every α2\alpha\geq 2, the following hold due to the triangle inequality:

|Vendiα(x1,,xn)1ααVendiα(Px)1αα|\displaystyle\Bigl{|}\mathrm{Vendi}_{\alpha}(x_{1},\ldots,x_{n})^{\frac{1-\alpha}{\alpha}}-\mathrm{Vendi}_{\alpha}(P_{x})^{\frac{1-\alpha}{\alpha}}\Bigr{|}\, =|𝝀^nα𝝀~α|\displaystyle=\,\Bigl{|}\bigl{\|}\widehat{\boldsymbol{\lambda}}_{n}\bigr{\|}_{\alpha}-\bigl{\|}\widetilde{\boldsymbol{\lambda}}\bigr{\|}_{\alpha}\Bigr{|}
𝝀^n𝝀~α\displaystyle\leq\,\bigl{\|}\widehat{\boldsymbol{\lambda}}_{n}-\widetilde{\boldsymbol{\lambda}}\bigr{\|}_{\alpha}
𝝀^n𝝀~2.\displaystyle\leq\,\bigl{\|}\widehat{\boldsymbol{\lambda}}_{n}-\widetilde{\boldsymbol{\lambda}}\bigr{\|}_{2}.

As a result, Theorem 1 shows that for every α2\alpha\geq 2 and δexp((2n)/8)\delta\geq\exp((2-n)/8), the following holds with probability at least 1δ1-\delta

|Vendiα(x1,,xn)1ααVendiα(Px)1αα|32log(2/δ)n\Bigl{|}\mathrm{Vendi}_{\alpha}(x_{1},\ldots,x_{n})^{\frac{1-\alpha}{\alpha}}-\mathrm{Vendi}_{\alpha}(P_{x})^{\frac{1-\alpha}{\alpha}}\Bigr{|}\,\leq\,\sqrt{\frac{32\log(2/\delta)}{n}}

A.4 Proof of Theorem 2

We begin by proving the following lemma showing that the eigenvalues used in the definition of the tt-truncated Vendi score are the projection of the original eigenvalues onto a tt-dimensional probability simplex.

Lemma 4.

Consider 𝐯[0,1]d\mathbf{v}\in[0,1]^{d} that satisfies 𝟏𝐯=1\mathbf{1}^{\top}\mathbf{v}=1. i.e., the sum of 𝐯\mathbf{v}’s entries equals 11. Given integer 1td1\leq t\leq d, define vector 𝐯(t)[0,1]d\mathbf{v}^{(t)}\in[0,1]^{d} whose last dtd-t entries are 0, i.e., vi(t)=0v^{(t)}_{i}=0 for t+1idt+1\leq i\leq d, and its first tt entries are defined as vj(t)=vj+1Sttv^{(t)}_{j}=v_{j}+\frac{1-S_{t}}{t} where St=v1++vtS_{t}=v_{1}+\cdots+v_{t}. Then, 𝐯(t)\mathbf{v}^{(t)} is the projection of 𝐯\mathbf{v} onto the following simplex set and has the minimum 2\ell_{2}-norm distance to this set

Δt:={𝐮[0,1]d:vi=0for allt+1id,i=1tvi=1}.\Delta_{t}:=\Bigl{\{}\mathbf{u}\in[0,1]^{d}:\;v_{i}=0\>\text{\rm for all}\>t+1\leq i\leq d,\;\;\;\sum_{i=1}^{t}v_{i}=1\Bigr{\}}.

Proof. To prove the lemma, first note that 𝐯(t)Δt\mathbf{v}^{(t)}\in\Delta_{t}, i.e. its first tt entries are non-negative and add up to 11, and also its last dtd-t entries are zero. Then, consider the projection problem discussed in the lemma:

min𝐮ti=1t(uivi)2\displaystyle\min_{\mathbf{u}\in\mathbb{R}^{t}}\qquad\;\;\>\sum_{i=1}^{t}\bigl{(}u_{i}-v_{i}\bigr{)}^{2}
subject toui0,for alli\displaystyle\text{\rm subject to}\quad u_{i}\geq 0,\;\;\text{\rm for all}\>i
i=1tui=1\displaystyle\qquad\qquad\quad\sum_{i=1}^{t}u_{i}=1

Then, since we know from the assumptions that vi0v_{i}\geq 0 and i=1tvi1\sum_{i=1}^{t}v_{i}\leq 1, the discussed 𝐮t\mathbf{u}^{*}\in\mathbb{R}^{t} where ui=vi+(1St)/tu^{*}_{i}=v_{i}+(1-S_{t})/t together with Lagrangian coefficients μi=0\mu_{i}=0 (for inequality constraint ui0u_{i}\geq 0) and λ=(1St)/t\lambda=(1-S_{t})/t (for equality constraint) satisfy the KKT conditions. The primal and dual feasibility conditions as well as the complementary slackness clearly hold for these selection of primal and dual variables. Also, the KKT stationarity condition is satisfied as for every ii we have uiviλμi=0u^{*}_{i}-v_{i}-\lambda-\mu_{i}=0. Since the optimization problem is a convex optimization task with affine constraints, the KKT conditions are sufficient for optimaility which proves the lemma.

Based on the above lemma, the eigenvalues 𝝀^n(t)\widehat{\boldsymbol{\lambda}}_{n}^{(t)} used to calculate the tt-truncated Vendi score Vendiα(t)(x1,,xn)\mathrm{Vendi}^{(t)}_{\alpha}(x_{1},\ldots,x_{n}) are the projections of the top-tt eigenvalues in 𝝀^n\widehat{\boldsymbol{\lambda}}_{n} for the original score Vendiα(x1,,xn)\mathrm{Vendi}_{\alpha}(x_{1},\ldots,x_{n}) onto the tt-simplex subset of d\mathbb{R}^{d} according to the 2\ell_{2}-norm. Similarly, the eigenvalues 𝝀~n(t)\widetilde{\boldsymbol{\lambda}}_{n}^{(t)} used to calculate the tt-truncated population Vendi Vendiα(t)(PX)\mathrm{Vendi}^{(t)}_{\alpha}(P_{X}) are the projections of the top-tt eigenvalues in 𝝀~\widetilde{\boldsymbol{\lambda}} for the original population Vendi Vendiα(Px)\mathrm{Vendi}_{\alpha}(P_{x}) onto the tt-simplex subset of d\mathbb{R}^{d}.

Since 2\ell_{2}-norm is a Hilbert space norm and the tt-simplex subset Δt\Delta_{t} is a convex set, we know from the convex analysis that the 2\ell_{2}-distance between the projected points 𝝀^n(t)\widehat{\boldsymbol{\lambda}}_{n}^{(t)} and 𝝀~(t)\widetilde{\boldsymbol{\lambda}}^{(t)} is upper-bounded by the 2\ell_{2}-distance between the original points 𝝀^n\widehat{\boldsymbol{\lambda}}_{n} and 𝝀~\widetilde{\boldsymbol{\lambda}}. As a result, Theorem 1 implies that

(𝝀^n𝝀~232log(2/δ)n) 1δ\displaystyle\mathbb{P}\Bigl{(}\bigl{\|}\widehat{\boldsymbol{\lambda}}_{n}-\widetilde{\boldsymbol{\lambda}}\bigr{\|}_{2}\leq\sqrt{\frac{32\log(2/\delta)}{n}}\Bigr{)}\,\geq\,1-\delta
\displaystyle\Longrightarrow\;\; (𝝀^n(t)𝝀~(t)232log(2/δ)n) 1δ\displaystyle\mathbb{P}\Bigl{(}\bigl{\|}\widehat{\boldsymbol{\lambda}}^{(t)}_{n}-\widetilde{\boldsymbol{\lambda}}^{(t)}\bigr{\|}_{2}\leq\sqrt{\frac{32\log(2/\delta)}{n}}\Bigr{)}\,\geq\,1-\delta

However, note that the eigenvalue vectors 𝝀^n(t)\widehat{\boldsymbol{\lambda}}^{(t)}_{n} and 𝝀~(t)\widetilde{\boldsymbol{\lambda}}^{(t)} can be analyzed in a bounded tt-dimensional space as their entries after index t+1t+1 are zero. Therefore, we can apply the proof of Corollary 2 to show that for every 1α<21\leq\alpha<2 and δexp((2n)/8)\delta\geq\exp((2-n)/8), the following holds with probability at least 1δ1-\delta

|Vendiα(x1,,xn)1ααVendiα(Px)1αα|32t2αlog(2/δ)n\Bigl{|}\mathrm{Vendi}_{\alpha}(x_{1},\ldots,x_{n})^{\frac{1-\alpha}{\alpha}}-\mathrm{Vendi}_{\alpha}(P_{x})^{\frac{1-\alpha}{\alpha}}\Bigr{|}\,\leq\,\sqrt{\frac{32t^{2-\alpha}\log(2/\delta)}{n}}

To extend the result to a general α>1\alpha>1, we reach the following inequality covering the above result as well as the result of Corollary 1 in one inequality

|Vendiα(x1,,xn)1ααVendiα(Px)1αα|32max{1,t2α}log(2/δ)n\Bigl{|}\mathrm{Vendi}_{\alpha}(x_{1},\ldots,x_{n})^{\frac{1-\alpha}{\alpha}}-\mathrm{Vendi}_{\alpha}(P_{x})^{\frac{1-\alpha}{\alpha}}\Bigr{|}\,\leq\,\sqrt{\frac{32\max\{1,t^{2-\alpha}\}\log(2/\delta)}{n}}

A.5 Proof of Theorem 3

Proof of Part (a). As defined by Ospanov et al. [2024b], the FKEA method uses the eigenvalues of tt random Fourier frequencies ω1,,ωt\omega_{1},\ldots,\omega_{t} where for each ωi\omega_{i} they consider two features cos(ωix)\cos(\omega_{i}^{\top}x) and sin(ωix)\sin(\omega_{i}^{\top}x). Following the definitions, it can be seen that k(x,x)=𝔼ωpω[cos(ω(xx))]k(x,x^{\prime})=\mathbb{E}_{\omega\sim p_{\omega}}\bigl{[}\cos(\omega^{\top}(x-x^{\prime}))\bigr{]} which is approximated by FKEA as 1ti=1tcos(ωi(xx))\frac{1}{t}\sum_{i=1}^{t}\cos(\omega_{i}^{\top}(x-x^{\prime})). Therefore, if we define kernel matrix KiK_{i} as the kernel matrix for ki(x,x)=cos(ωi(xx))k_{i}(x,x^{\prime})=\cos(\omega_{i}^{\top}(x-x^{\prime})), then we will have

1nKFKEA(t)=1ti=1t1nKi\frac{1}{n}K^{\mathrm{FKEA}(t)}=\frac{1}{t}\sum_{i=1}^{t}\frac{1}{n}K_{i}

where 𝔼ωipω[1nKi]=1nK\mathbb{E}_{\omega_{i}\sim p_{\omega}}[\frac{1}{n}K_{i}]=\frac{1}{n}K.

On the other hand, we note that 1nKHS1\|\frac{1}{n}K\|_{\mathrm{HS}}\leq 1 holds as the kernel function is normalized and hence |k(x,x)|1|k(x,x^{\prime})|\leq 1. Since the Frobenius norm is the 2\ell_{2}-norm of the vectorized version of the matrix, we can apply Vector Bernstein inequality in Lemma 1 to show that for every 0ϵ20\leq\epsilon\leq 2:

(1ti=1t[1nKi]1nKFϵ)exp(8tϵ232)\displaystyle\mathbb{P}\Bigl{(}\Bigl{\|}\frac{1}{t}\sum_{i=1}^{t}\bigl{[}\frac{1}{n}K_{i}\bigr{]}-\frac{1}{n}K\Bigr{\|}_{F}\geq\epsilon\Bigr{)}\,\leq\,\exp\Bigl{(}\frac{8-t\epsilon^{2}}{32}\Bigr{)}
\displaystyle\Longrightarrow\;\; (1nKFKEA(t)1nKFϵ)exp(8tϵ232)\displaystyle\mathbb{P}\Bigl{(}\Bigl{\|}\frac{1}{n}K^{\mathrm{FKEA}(t)}-\frac{1}{n}K\Bigr{\|}_{F}\geq\epsilon\Bigr{)}\,\leq\,\exp\Bigl{(}\frac{8-t\epsilon^{2}}{32}\Bigr{)}

Then, we apply the Hoffman-Wielandt inequality to show that for the sorted eigenvalue vectors of 1nK\frac{1}{n}K (denoted by 𝝀^n\widehat{\boldsymbol{\lambda}}_{n}) and 1nKFKEA(t)\frac{1}{n}K^{\mathrm{FKEA}(t)} (denoted by 𝝀FKEA(t){\boldsymbol{\lambda}}^{\mathrm{FKEA}(t)}) we will have 𝝀^n𝝀FKEA(t)21nKFKEA(t)1nKHS\|\widehat{\boldsymbol{\lambda}}_{n}-{\boldsymbol{\lambda}}^{\mathrm{FKEA}(t)}\|_{2}\leq\|\frac{1}{n}K^{\mathrm{FKEA}(t)}-\frac{1}{n}K\|_{\mathrm{HS}}, which together with the previous inequality leads to

(𝝀^n𝝀FKEA(t)2ϵ)exp(8tϵ232)\displaystyle\mathbb{P}\Bigl{(}\Bigl{\|}\widehat{\boldsymbol{\lambda}}_{n}-{\boldsymbol{\lambda}}^{\mathrm{FKEA}(t)}\Bigr{\|}_{2}\geq\epsilon\Bigr{)}\,\leq\,\exp\Bigl{(}\frac{8-t\epsilon^{2}}{32}\Bigr{)}

Furthermore, as we shown in the proof of Theorem 1 for every 0γ20\leq\gamma\leq 2

(𝝀^n𝝀~2γ)exp(8nγ232)\displaystyle\mathbb{P}\Bigl{(}\bigl{\|}\widehat{\boldsymbol{\lambda}}_{n}-\widetilde{\boldsymbol{\lambda}}\bigr{\|}_{2}\geq\gamma\Bigr{)}\,\leq\,\exp\Bigl{(}\frac{8-n\gamma^{2}}{32}\Bigr{)}

which, by applying the union bound for γ=ϵ\gamma=\epsilon, together with the previous inequality shows that

(𝝀~𝝀FKEA(t)22ϵ)\displaystyle\mathbb{P}\Bigl{(}\Bigl{\|}\widetilde{\boldsymbol{\lambda}}-{\boldsymbol{\lambda}}^{\mathrm{FKEA}(t)}\Bigr{\|}_{2}\geq 2\epsilon\Bigr{)}\, exp(8tϵ232)+exp(8nϵ232)\displaystyle\leq\,\exp\Bigl{(}\frac{8-t\epsilon^{2}}{32}\Bigr{)}+\exp\Bigl{(}\frac{8-n\epsilon^{2}}{32}\Bigr{)}
 2exp(8min{n,t}ϵ232)\displaystyle\leq\,2\exp\Bigl{(}\frac{8-\min\{n,t\}\epsilon^{2}}{32}\Bigr{)}

Therefore, Lemma 4 implies that

(𝝀~(t)𝝀FKEA(t)2ϵ)\displaystyle\mathbb{P}\Bigl{(}\Bigl{\|}\widetilde{\boldsymbol{\lambda}}^{(t)}-{\boldsymbol{\lambda}}^{\mathrm{FKEA}(t)}\Bigr{\|}_{2}\geq\epsilon\Bigr{)}\,  2exp(32min{n,t}ϵ2128)\displaystyle\leq\,2\exp\Bigl{(}\frac{32-\min\{n,t\}\epsilon^{2}}{128}\Bigr{)}

If we define δ=2exp(32min{n,t}ϵ2128)\delta=2\exp\bigl{(}\frac{32-\min\{n,t\}\epsilon^{2}}{128}\bigr{)}, implying that ϵ128log(3/δ)min{n,t}\epsilon\leq\sqrt{\frac{128\log(3/\delta)}{\min\{n,t\}}}, then the above inequality shows that

(𝝀~(t)𝝀FKEA(t)2128log(3/δ)min{n,t})\displaystyle\mathbb{P}\Bigl{(}\Bigl{\|}\widetilde{\boldsymbol{\lambda}}^{(t)}-{\boldsymbol{\lambda}}^{\mathrm{FKEA}(t)}\Bigr{\|}_{2}\leq\sqrt{\frac{128\log(3/\delta)}{\min\{n,t\}}}\Bigr{)}\, 1δ\displaystyle\geq 1-\delta

Therefore, if we follow the same steps of the proof of Theorem 2, we can show

|FKEA-Vendiα(t)(x1,,xn)1ααVendiα(t)(Px)1αα|128max{1,t2α}log(3/δ)min{n,t}\Bigl{|}\mathrm{FKEA}\text{-}\mathrm{Vendi}^{(t)}_{\alpha}(x_{1},\ldots,x_{n})^{\frac{1-\alpha}{\alpha}}-\mathrm{Vendi}^{(t)}_{\alpha}(P_{x})^{\frac{1-\alpha}{\alpha}}\Bigr{|}\,\leq\,\sqrt{\frac{128\max\{1,t^{2-\alpha}\}\log(3/\delta)}{\min\{n,t\}}}

Proof of Part (b). To show this theorem, we use Theorem 3 from [Xu et al., 2015], which shows that if the rrth largest eigenvalue of the kernel matrix 1nK\frac{1}{n}K satisfies λrτn\lambda_{r}\leq\frac{\tau}{n}, then given tCrlog(n)t\geq Cr\log(n) (CC is a universal constant), the following spectral norm bound will hold with probability 12n31-\frac{2}{n^{3}}:

1nK1nKNystrom(t)sp𝒪(τlog(n)nt).\bigl{\|}\frac{1}{n}K-\frac{1}{n}K^{\mathrm{Nystrom(t)}}\bigr{\|}_{sp}\leq\mathcal{O}\Bigl{(}\frac{\tau\log(n)}{\sqrt{nt}}\Bigr{)}.

Therefore, Weyl’s inequality implies the following for the vector of sorted eigenvalues of 1nK\frac{1}{n}K, i.e. 𝝀^n\widehat{\boldsymbol{\lambda}}_{n}, and that of 1nKNystrom(t)\frac{1}{n}K^{\mathrm{Nystrom(t)}}, i.e., 𝝀Nystrom(t){\boldsymbol{\lambda}}^{\mathrm{Nystrom(t)}},

𝝀^n𝝀Nystrom(t)𝒪(τlog(n)nt).\bigl{\|}\widehat{\boldsymbol{\lambda}}_{n}-{\boldsymbol{\lambda}}^{\mathrm{Nystrom(t)}}\bigr{\|}_{\infty}\leq\mathcal{O}\Bigl{(}\frac{\tau\log(n)}{\sqrt{nt}}\Bigr{)}.

As a result, considering the subvectors 𝝀^n[1:t]\widehat{\boldsymbol{\lambda}}_{n}[1:t] and 𝝀Nystrom(t)[1:t]{\boldsymbol{\lambda}}^{\mathrm{Nystrom(t)}}[1:t] with the first tt entries of the vectors, we will have:

𝝀^n[1:t]𝝀Nystrom(t)[1:t]𝒪(τlog(n)nt)𝝀^n[1:t]𝝀Nystrom(t)[1:t]2𝒪(τlog(n)tn)\bigl{\|}\widehat{\boldsymbol{\lambda}}_{n}[1:t]-{\boldsymbol{\lambda}}^{\mathrm{Nystrom(t)}}[1:t]\bigr{\|}_{\infty}\leq\mathcal{O}\Bigl{(}\frac{\tau\log(n)}{\sqrt{nt}}\Bigr{)}\quad\Longrightarrow\quad\bigl{\|}\widehat{\boldsymbol{\lambda}}_{n}[1:t]-{\boldsymbol{\lambda}}^{\mathrm{Nystrom(t)}}[1:t]\bigr{\|}_{2}\leq\mathcal{O}\Bigl{(}\tau\log(n)\sqrt{\frac{t}{n}}\Bigr{)}

Noting that the non-zero entries of 𝝀Nystrom(t){\boldsymbol{\lambda}}^{\mathrm{Nystrom(t)}} are all included in the first-tt elements, we can apply Lemma 4 which shows that with probability 12n31-2n^{-3} we have

𝝀^n(t)𝝀Nystrom(t)2𝒪(τlog(n)tn)\Bigl{\|}\widehat{\boldsymbol{\lambda}}^{(t)}_{n}-{\boldsymbol{\lambda}}^{\mathrm{Nystrom(t)}}\Bigr{\|}_{2}\leq\mathcal{O}\Bigl{(}\tau\log(n)\sqrt{\frac{t}{n}}\Bigr{)}

Also, in the proof of Theorem 2, we showed that

(𝝀^n(t)𝝀~(t)232log(2/δ)n) 1δ\displaystyle\mathbb{P}\Bigl{(}\bigl{\|}\widehat{\boldsymbol{\lambda}}^{(t)}_{n}-\widetilde{\boldsymbol{\lambda}}^{(t)}\bigr{\|}_{2}\leq\sqrt{\frac{32\log(2/\delta)}{n}}\Bigr{)}\,\geq\,1-\delta

Combining the above inequalities using a union bound, shows that with probability at least 1δ2n31-\delta-2n^{-3} we have

𝝀Nystrom(t)𝝀~(t)2\displaystyle\Bigl{\|}{\boldsymbol{\lambda}}^{\mathrm{Nystrom(t)}}-\widetilde{\boldsymbol{\lambda}}^{(t)}\Bigr{\|}_{2}\> 𝝀Nystrom(t)𝝀^n(t)2+𝝀^n(t)𝝀~(t)2\displaystyle\leq\>\Bigl{\|}{\boldsymbol{\lambda}}^{\mathrm{Nystrom(t)}}-\widehat{\boldsymbol{\lambda}}^{(t)}_{n}\Bigr{\|}_{2}+\Bigl{\|}\widehat{\boldsymbol{\lambda}}^{(t)}_{n}-\widetilde{\boldsymbol{\lambda}}^{(t)}\Bigr{\|}_{2}
32log(2/δ)n+𝒪(τlog(n)tn)\displaystyle\leq\>\sqrt{\frac{32\log(2/\delta)}{n}}+\mathcal{O}\Bigl{(}\tau\log(n)\sqrt{\frac{t}{n}}\Bigr{)}
=𝒪(log(2/δ)+tlog(n)2τ2n)\displaystyle=\>\mathcal{O}\Bigl{(}\sqrt{\frac{\log(2/\delta)+t\log(n)^{2}\tau^{2}}{n}}\Bigr{)}

Hence, repeating the final steps in the proof of Theorem 2, we can prove

|Nystrom-Vendiα(t)(x1,,xn)1ααVendiα(t)(Px)1αα|𝒪(max{t2α,1}(log(2/δ)+tlog(n)2τ2)n)\Bigl{|}\mathrm{Nystrom}\text{-}\mathrm{Vendi}^{(t)}_{\alpha}(x_{1},\ldots,x_{n})^{\frac{1-\alpha}{\alpha}}-\mathrm{Vendi}^{(t)}_{\alpha}(P_{x})^{\frac{1-\alpha}{\alpha}}\Bigr{|}\,\leq\,\mathcal{O}\Bigl{(}\sqrt{\frac{\max\{t^{2-\alpha},1\}\bigl{(}\log(2/\delta)+t\log(n)^{2}\tau^{2}\bigr{)}}{n}}\Bigr{)}

Appendix B Additional Numerical Results

Refer to caption
Figure 8: Diversity evaluation of Vendi scores on StyleGAN-XL generated ImageNet dataset with varying truncation parameter ψ\psi. The setting is based on DinoV2 embedding and bandwidth σ=30\sigma=30

In this section, we present supplementary results concerning the evaluation of diversity and the convergence behavior of different variants of the Vendi score. We extend the convergence experiments discussed in the main text to include the truncated StyleGAN3-t FFHQ dataset (Figure 12) and the StyleGAN-XL ImageNet dataset (Figure 13). Furthermore, we demonstrate that the truncated Vendi statistic effectively captures the diversity characteristics across various data modalities. Specifically, we conducted similar experiments as shown in Figures 7 and 6 on text data (Figure 9) and video data (Figure 11), showcasing the applicability of the metric across different domains.

Refer to caption
Figure 9: Diversity evaluation of Vendi scores on synthetic text dataset about 100 countries generated by GPT-4 with varying number of countries. The setting is based on text-embedding-3-large embedding and bandwidth σ=0.5\sigma=0.5
Refer to caption
Figure 10: The diagram outlining an intuition behind a kernel bandwidth σ\sigma selection in diversity evaluation.
Refer to caption
Figure 11: Diversity evaluation of Vendi scores on Kinetics400 dataset with varying number of classes. The setting is based on I3D embedding and bandwidth σ=4.0\sigma=4.0

We observe in Figure 12 that the convergence behavior is illustrated across various values of ψ\psi. The results indicate that, for a fixed bandwidth σ\sigma, the truncated, Nyström, and FKEA variants of the Vendi score converge to the truncated Vendi statistic. As demonstrated in Figure 6 of the main text, this truncated Vendi statistic effectively captures the diversity characteristics inherent in the underlying dataset.

We note that in presence of incremental changes to the diversity of the dataset, finite-dimensional kernels, such as cosine similarity kernel, remain relatively constant. This effect is illustrated in Figure 13, where increase in truncation factor ψ\psi results in incremental change in diversity. This is one of the cases where infinite-dimensional kernel maps with a sensitivity (bandwidth) parameter σ\sigma are useful in controlling how responsive the method should be to the change in diversity.

Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 12: Statistical convergence of Vendi score for different sample sizes on StyleGAN3 generated FFHQ data at various truncation factors ψ\psi: (Left plot) finite-dimension cosine similarity kernel (Right plot) infinite dimension Gaussian kernel with bandwidth σ=35\sigma=35. DinoV2 embedding (dimension 768) is used in computing the scores.
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 13: Statistical convergence of Vendi score for different sample sizes on StyleGAN-XL generated ImageNet data: (Left plot) finite-dimension cosine similarity kernel (Right plot) infinite dimension Gaussian kernel with bandwidth σ=40\sigma=40. DinoV2 embedding (dimension 768) is used in computing the scores.
Table 1: Statistical convergence of diversity scores for different sample size on DALL-E 3 generated MSCOCO data
nn VENDI-1.0 RKE Vendi-t FKEA-Vendi Nystrom-Vendi Recall Coverage
2000 239.91 13.47 239.91 228.69 239.91 0.76 0.86
4000 315.35 13.51 315.35 280.68 315.35 0.81 0.87
6000 357.15 13.56 346.27 310.9 345.49 0.83 0.91
8000 392.36 13.56 354.8 329.56 357.41 0.87 0.91
Table 2: Statistical convergence of diversity scores for different sample size on SDXL generated MSCOCO data
nn VENDI-1.0 RKE Vendi-t FKEA-Vendi Nystrom-Vendi Recall Coverage
2000 187.17 10.65 187.17 173.06 187.18 0.78 0.85
4000 236.49 10.7 236.49 222.78 236.08 0.82 0.87
6000 264.82 10.7 258.21 236.37 257.34 0.86 0.87
8000 289.08 10.71 265.84 251.59 266.23 0.86 0.86
10000 304.44 10.72 267.39 256.24 268.34 0.86 0.87
Table 3: Compilation time (in seconds) of different Vendi scores with increasing sample size
Metric samples nn
10000 20000 30000 40000 50000 60000 70000
Vendi 97s 631s 1868s - - - -
FKEA-Vendi 19s 36s 53s 71s 88s 105s 124s
Nystrom-Vendi 31s 44s 78s 91s 112s 136s 164s

B.1 Bandwidth σ\sigma Selection

In our experiments, we select the Gaussian kernel bandwidth, σ\sigma, to ensure that the Vendi metric effectively distinguishes the inherent modes within the dataset. The kernel bandwidth directly controls the sensitivity of the metric to the underlying data clusters. As illustrated in Figure 10, varying σ\sigma significantly impacts the diversity computation on the ImageNet dataset. A smaller bandwidth (e.g., σ=20,30\sigma=20,30) results in the metric treating redundant samples as distinct modes, artificially inflating the number of clusters, which in turn slows down the convergence of the metric. On the other hand, large bandwidth results in instant convergence of the metric, i.e. in σ=60\sigma=60 n=100n=100 and n=1000n=1000 have almost the same amount of diversity.

Appendix C Selection of Embedding space

To show that proposed truncated Vendi score remains feasible under arbitrary embedding selection, we conducted experiments from Figures 6 and 7. Figures 14, 15, 16 and 17 extend the results to CLIP Radford et al. [2021] and SWaV Caron et al. [2020] embeddings. These experiments demonstrate that FKEA, Nyström and tt-truncated Vendi correlate with increasing diversity of the evaluated dataset. We emphasize that proposed statistic remains feasible under arbitrary embedding space that is capable of mapping image samples into a latent space.

Refer to caption
Figure 14: Diversity evaluation of Vendi scores on ImageNet dataset with varying number of classes based on CLIP embedding and bandwidth σ=5.0\sigma=5.0
Refer to caption
Figure 15: Diversity evaluation of Vendi scores on truncated StyleGAN3 generated FFHQ with varying truncation coefficient ψ\psi based on CLIP embedding and bandwidth σ=5.0\sigma=5.0
Refer to caption
Figure 16: Diversity evaluation of Vendi scores on ImageNet dataset with varying number of classes based on SWaV embedding and bandwidth σ=1.0\sigma=1.0
Refer to caption
Figure 17: Diversity evaluation of Vendi scores on truncated StyleGAN3 generated FFHQ with varying truncation coefficient ψ\psi based on SwAV embedding and bandwidth σ=1.0\sigma=1.0