This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

\AtBeginShipoutNext\AtBeginShipoutUpperLeft

MIT-CTP 5692

A Maximum Entropy Principle in Deep Thermalization
and in Hilbert-Space Ergodicity

Daniel K. Mark Center for Theoretical Physics, Massachusetts Institute of Technology, Cambridge, MA 02139, USA    Federica Surace California Institute of Technology, Pasadena, CA 91125, USA    Andreas Elben California Institute of Technology, Pasadena, CA 91125, USA    Adam L. Shaw California Institute of Technology, Pasadena, CA 91125, USA    Joonhee Choi Department of Electrical Engineering, Stanford University, Stanford, CA, USA    Gil Refael California Institute of Technology, Pasadena, CA 91125, USA    Manuel Endres California Institute of Technology, Pasadena, CA 91125, USA    Soonwon Choi soonwon@mit.edu Center for Theoretical Physics, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
(August 12, 2025)
Abstract

We report universal statistical properties displayed by ensembles of pure states that naturally emerge in quantum many-body systems. Specifically, two classes of state ensembles are considered: those formed by i) the temporal trajectory of a quantum state under unitary evolution or ii) the quantum states of small subsystems obtained by partial, local projective measurements performed on their complements. These cases respectively exemplify the phenomena of “Hilbert-space ergodicity” and “deep thermalization.” In both cases, the resultant ensembles are defined by a simple principle: the distributions of pure states have maximum entropy, subject to constraints such as energy conservation, and effective constraints imposed by thermalization. We present and numerically verify quantifiable signatures of this principle by deriving explicit formulae for all statistical moments of the ensembles; proving the necessary and sufficient conditions for such universality under widely-accepted assumptions; and describing their measurable consequences in experiments. We further discuss information-theoretic implications of the universality: our ensembles have maximal information content while being maximally difficult to interrogate, establishing that generic quantum state ensembles that occur in nature hide (scramble) information as strongly as possible. Our results generalize the notions of Hilbert-space ergodicity to time-independent Hamiltonian dynamics and deep thermalization from infinite to finite effective temperature. Our work presents new perspectives to characterize and understand universal behaviors of quantum dynamics using statistical and information theoretic tools.

I Introduction

Refer to caption
Figure 1: Overview of this work. In this work, we study ensembles of states generated by natural dynamics. (a) Traditional experiments on quantum systems can only measure the average properties of such ensembles, described by a density matrix ρ\rho (red cross in Bloch sphere). (b) In contrast, modern experimental capabilities are able to generate and investigate ensembles of labeled quantum states, and study their higher statistical properties. We specifically study two ensembles of states in this work: (c) the temporal ensemble of states generated by taking an initial state |Ψ0\ket{\Psi_{0}} and evolving it to different times under a (fixed) Hamiltonian HH, and (d) the projected ensemble of states {|Ψ(zB)}\{\ket{\Psi(z_{B})}\} generated by projective measurement (performed on a large subsystem) of a single state |Ψ\ket{\Psi} resulting in many different outcomes zBz_{B}. These ensembles, respectively, exemplify the phenomena of Hilbert-space ergodicity and deep thermalization. (e) Our key finding is that these ensembles of states have maximum entropy, subject to constraints arising from dynamics. For example, quantum thermodynamics constrains projected ensembles to have a fixed first moment ρ\rho. While many ensembles of states share the same first moment ρ\rho (red crosses), the (unnormalized) projected ensemble is the one with the maximum ensemble entropy (bottom).

The second law of thermodynamics is a foundational concept in statistical physics with far reaching implications across many areas of studies, from quantum gravity to information theory and computation [1, 2, 3, 4, 5]. At its core is the idea of a maximum entropy principle [6, 7, 8, 9, 10]; that a macroscopic system in equilibrium is described by the state with maximum entropy, subject to physical constraints such as global energy and particle number conservation. It is remarkable that such a simple principle enables highly accurate predictions about a large system without knowledge of the vast majority of its degrees of freedom.

In quantum systems, the maximum entropy principle [11, 12] is often justified by one of the following two conditions relevant to many conventional experiments [Fig. 1(a)]. First, one only has access to local or spatially averaged observables [13, 14, 15, 16]. Second, one can only measure observables averaged over long time intervals or at late, equilibrium times without fine-grained temporal resolution [17, 15, 16]. In both cases, measurement outcomes become largely agnostic to microscopic details of the initial state or system dynamics and are well-described by thermal density matrices. It is this inability to obtain detailed, microscopic information from traditional experiments that render quantum states to be in a probabilistic ensemble and ultimately empower the maximum entropy principle to make accurate and practical predictions of accessible observables.

Modern quantum experiments challenge such traditional settings [Fig. 1(b)]. Quantum devices built using atomic or superconducting qubit platforms have realized and probed diverse phenomena ranging from non-equilibrium dynamics [18, 19, 20, 21, 22] to exotic phases of matter [23, 24, 25, 26, 27]. These experiments have the capability to fully interrogate the entire quantum system, enabling the measurement of arbitrary, non-local multi-particle correlation functions at the fundamental quantum projection limit. They are also able to deterministically evolve quantum many-body states with a high repetition rate, enabling the study of quantum dynamics with an unprecedented level of accuracy and details. Existing maximum entropy principles were not designed to capture such microscopic measurement outcomes, which are non-local and time-resolved. Therefore, we set out to develop a new approach that may provide an effective statistical model for such data. Here, we report the discovery of a generalized maximum entropy principle for quantum many-body systems that is well aligned with modern capabilities of experiments.

We specifically study two settings, the temporal and projected ensembles. The former considers the trajectory of quantum states under time evolution, while the latter considers ensembles of states generated by partial projective measurements performed on the system [Fig. 1(c,d)]. In both settings, the average states, or first moments, of the ensembles are described by density matrices ρ\rho which have been extensively studied in the literature [17, 13, 14, 12, 15, 16]. Here, certain universal phenomena apply, such as the fact that the reduced density matrices of small subsystems are well described by thermal Gibbs states [14, 15], or dephasing in the energy-eigenbasis, which leads to the so-called diagonal ensemble as an effective description of equilibrium properties [12, 17, 13]. However, the temporal and projected ensembles of generic systems have not been studied beyond their first moments. Our results extend this universality to higher moments, finding that the first moment uniquely fixes the higher moments.

We find a new type of maximum entropy principle that describes ensembles of states in the above settings [Fig. 1(e)]. The entropy in question is an ensemble entropy, which quantifies how the states in an ensemble are spread out over the relevant Hilbert space. This is distinct from the conventional von Neumann entropy, which only depends on the average state (density matrix) of the ensemble. In broad strokes, our main finding is that in many physical situations, the ensembles of pure states maximize the ensemble entropy up to physical constraints such as energy conservation or local thermalization, thus establishing a new maximum entropy principle. Special cases of our ensembles of states have been discussed in the literature under the names of “deep thermalization” [28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38] and “Hilbert space ergodicity” [39, 40] respectively, in particular at infinite effective temperature. Here we present a unified principle that holds under general conditions.

Ultimately, our work is a theory of fluctuations: the variations in observables over states in an ensemble, which go fundamentally beyond ensemble-averaged observables that are typically studied. In companion experimental work [41], we use insights gained from this work to predict and verify universal forms of fluctuations in a variety of quantities, interpolating from global to local, and in closed as well as open system dynamics.

Our findings have several implications. First, we confirm that Hamiltonian dynamics uniformly explores its available Hilbert space. Specifically, the temporal trace of a many-body quantum state satisfies the maximum entropy principle under the constraint of energy conservation. For sufficiently long evolution, the trajectory of the state is only constrained by the population of the energy eigenstates: all other degrees of freedom are randomly distributed with maximum entropy. This improves our understanding of thermalization as a dynamical phenomenon [42, 43] and its relationship to static notions of thermal equilibrium [44, 45]. While thermalization has traditionally concerned the dynamics of local quantities —e.g. how such quantities reach their equilibrium values— here we study the dynamics of the global state, confirming and formalizing our intuition about ergodicity in Hilbert space [46, 39]. We find necessary and sufficient conditions for Hilbert-space ergodicity and find that curiously, just as in classical dynamics, a system can be ergodic without being chaotic.

Second, we generalize notions of deep thermalization [28, 29, 30] away from infinite temperature. Deep thermalization refers to the observation that in generic many-body dynamics, not only does the density matrix of a local subsystem reach thermal equilibrium, so do the higher moments of the projected ensemble. Additionally, it has been pointed out that the higher moments may take a longer time than lower moments to converge, implying the possibility of sustained thermalization (in non-local quantities) even when local observables have converged [30, 31, 36]. The equilibrium values of such higher moments have considerable structure. At infinite effective temperature, the projected ensemble [28, 29] has been found to approach the Haar ensemble [47, 48], a paradigmatic uniform random ensemble of quantum states. We extend this phenomenology to finite temperature, identifying finite temperature generalizations of the Haar ensemble with maximum entropy which describe our ensembles of states. Our result should be taken as complementary to, but distinct from conventional theories of quantum thermalization such as the eigenstate thermalization hypothesis (ETH) [12, 15]. It does not rely on the ETH, but when used in conjuction, implies that all higher moments of the projected ensemble reach thermal equilbrium with universal, Gibbs-like structure.

Third, our approach predicts and, where relevant, confirms the existence of a universality in natural many-body states, revealed in their higher moments. Not only do mean quantities captured by density matrices ρ\rho exhibit universal thermal behavior, so do higher-order moments such as the variance and skewness, as dictated by our maximum entropy principle. Higher order quantities beyond local observables have received recent attention in non-equilibrium dynamics [49, 50], and our results suggest at the broader applicability of higher moments — the maximum entropy principle holds not only at the level of density matrices [51, 52], but may also hold at the level of many-body wavefunctions [53, 54, 55]. This possibility has been recently investigated for energy eigenstates in Ref. [55]. Our results support such an approach based on the maximum entropy principle, and suggest that they may be more widely applicable; here we primarily apply them to time-evolved many-body states.

Finally, we draw further connections between thermalization and quantum information theory. Ensembles of quantum states are natural objects for quantum communication: fundamental results such as the Holevo bound investigate the use of such ensembles to transmit information [56]. Our results indicate that ensembles of states that result from many-body dynamics are both maximally difficult to compress, distinguish, or use for information transmission, providing a new perspective on the nature of information scrambling and hiding in many-body dynamics [57, 58, 59, 60, 61].

The maximum entropy principle has repeatedly exceeded expectations in its ability to describe complex, interacting systems [6, 7, 8, 9, 10]. Most results have so far been restricted to classical or single-particle quantum physics. Our work represents a first step to adapting this principle to the modern setting of many-body physics. While we were able to firmly establish this principle in certain settings, we find that it may apply more generally in many-body physics. Indeed, see Ref. [62] for a recent application of this principle to many-body physics to full-counting statistics in systems with a U(1)U(1) symmetry. These developments hint at broader applications of the maximum entropy principle in many-body physics.

II Summary of results

Ensemble Statistical description Maximum entropy principle Measurable signature Information theoretic properties
Temporal ensemble Random phase ensemble (Sec. IV) Maximum entropy with fixed energy populations {|cE|2}\{|c_{E}|^{2}\} Porter Thomas (PT) dist. in outcome probabilities pz(t)p_{z}(t) (Thm. 3). Constant mutual information [Eq. (43)]
I(Z;T)(1γ)/ln2I(Z;T)\approx(1-\gamma)/\ln 2
Projected ensemble:
Energy non-revealing basis
Scrooge ensemble (Sec. V) Maximum entropy with fixed first moment ρA\rho_{A} PT dist. in joint probabilities p(oA,zB)p(o_{A},z_{B}) [Eq. (56)] Minimal mutual information [Eq. (61)]
I(OA;ZB)Q(ρ)I(O_{A};Z_{B})\approx Q(\rho)
General measurement basis Generalized Scrooge ensemble (Sec. VI) Collection of maximum-entropy ensembles with fixed first moments ρd(zB)\rho_{d}(z_{B}) PT dist. in normalized probabilities poA(zB)/pd,oA(zB)p_{o_{A}}(z_{B})/p_{d,o_{A}}(z_{B}) [Eq. (72)] Minimal interaction information [Eq. (76)]
I(OA;ZB;T)𝔼zBQ(ρ¯(zB))I(O_{A};Z_{B};T)\approx\mathbb{E}_{z_{B}}Q(\bar{\rho}(z_{B}))
Table 1: Summary of key results in this work. We describe the key ensembles considered in this work: the temporal ensemble and the projected ensembles generated by measurements in a basis {zB}\{z_{B}\} under two different cases: when {zB}\{z_{B}\} is energy non-revealing and when it is general. Their statistical descriptions are probability distributions which have maximum entropy under appropriate constraints. We finally discuss their measurable signatures and information theoretic properties.

In this work, we study ensembles of quantum states. These ensembles \mathcal{E} are collections of pure quantum states |Ψj\ket{\Psi_{j}}, labeled by an index jj, with an associated probability distribution pjp_{j}, that is, ={pj,|Ψj}\mathcal{E}=\{p_{j},\ket{\Psi_{j}}\}. Quantum state ensembles can be discrete (i.e. the index jj takes discrete values, such as a discrete set of measurement outcomes) or continuous (the index jj takes continuous values, such as time). In situations where the label is not important, we simply write ={P(Ψ),|Ψ}\mathcal{E}=\{P(\Psi),\ket{\Psi}\}.

Quantum state ensembles have traditionally been studied in quantum information for tasks such as quantum communication [63]. However, until recently they have not been necessary in the context of many-body physics. Under traditional measurements, observations made on an ensemble of states only depend on the first moment ρ(1)𝑑ΨP(Ψ)|ΨΨ|\rho^{(1)}_{\mathcal{E}}\equiv\int d\Psi P(\Psi)|\Psi\rangle\!\langle\Psi|. This is because without the ability to determine the index jj, measurements of an observable OO would yield its expected value, averaged over the ensemble, i.e. 𝑑ΨP(Ψ)Ψ|O|Ψ=tr(Oρ(1))\int d\Psi P(\Psi)\langle\Psi|O|\Psi\rangle=\text{tr}(O\rho^{(1)}_{\mathcal{E}}). Therefore, in this setting, one can only operate at the level of mixed states ρ(1)\rho^{(1)}_{\mathcal{E}} [17, 13, 14, 12, 15, 16].

Modern experiments challenge this assumption and may involve full knowledge of the ensemble \mathcal{E}. In modern quantum devices, one can determine the index jj in a variety of settings. For example, in the ensembles we study, jj will be the evolution time, which can be precisely controlled, or jj will be the outcome of a projective measurement which is recorded by the experimenter. When the index jj can be determined, we can study a wider range of properties, such as the fluctuations of the expectation values Ψj|O|Ψj\langle\Psi_{j}|O|\Psi_{j}\rangle. The simplest such quantity is the variance 𝑑ΨP(Ψ)Ψ|O|Ψ2[𝑑ΨP(Ψ)Ψ|O|Ψ]2\int d\Psi P(\Psi)\langle\Psi|O|\Psi\rangle^{2}-\big{[}\int d\Psi P(\Psi)\langle\Psi|O|\Psi\rangle\big{]}^{2}. This quantity depends on more than the first moment ρ(1)\rho^{(1)}_{\mathcal{E}}. Specifically, it depends on the second moment ρ(2)𝑑ΨP(Ψ)|ΨΨ|2\rho^{(2)}_{\mathcal{E}}\equiv\int d\Psi P(\Psi)|\Psi\rangle\!\langle\Psi|^{\otimes 2} since the first term is equal to 𝑑ΨP(Ψ)Ψ|O|Ψ2=tr(ρ(2)O2)\int d\Psi P(\Psi)\langle\Psi|O|\Psi\rangle^{2}=\text{tr}(\rho^{(2)}_{\mathcal{E}}O^{\otimes 2}). In general, the statistical properties of quantum state ensembles can be systematically characterized through their moments

ρ(k)𝑑ΨP(Ψ)|ΨΨ|k,\displaystyle\rho^{(k)}_{\mathcal{E}}\equiv\int d\Psi P(\Psi)\outerproduct{\Psi}{\Psi}^{\otimes k}, (1)

for positive integer kk111For discrete ensembles, the integral reduces to a sum over all states in the ensemble, see Sec. III..

Recent work has illustrated that rich physical phenomena are encoded in the higher moments of state ensembles. As a concrete, but non-exhaustive example, higher moment quantities are necessary to detect the measurement-induced phase transition in monitored quantum circuits [64, 65, 66, 67]. In this setting, an ensemble of pure states is generated by the different outcomes of mid-circuit measurements. A phase transition is observed in quantities such as the bipartite entanglement entropy averaged over states in the ensemble. Their average kk-th Rényi entropies can be computed from the kk-th moment of the ensemble, through the quantity 𝑑ΨP(Ψ)trA(trB(|ΨΨ|)k)\int d\Psi P(\Psi)\text{tr}_{A}(\text{tr}_{B}(|\Psi\rangle\!\langle\Psi|)^{k}), where AA and BB are a bipartition of the system.

Without reference to a specific observable, the fluctuations of quantum state ensembles can be studied by examining how the states are distributed in Hilbert space. We quantify this spread through an ensemble entropy, which is the Shannon entropy of the probability distribution P(Ψ)P(\Psi). For continuous ensembles, this requires the choice of a reference distribution, in our case, the Haar ensemble. Thus, for continuous quantum state ensembles, the ensemble entropy is given as the (negative) Kullback-Liebler (KL) divergence between P(Ψ)P(\Psi) and the Haar ensemble

Ent()\displaystyle\text{Ent}(\mathcal{E}) 𝒟(Haar)\displaystyle\equiv-\mathcal{D}(\mathcal{E}\|\text{Haar}) (2)
𝑑ΨP(Ψ)log2(P(Ψ)PHaar(Ψ)).\displaystyle\equiv-\int d\Psi~P(\Psi)\log_{2}\left(\frac{P(\Psi)}{P_{\text{Haar}}(\Psi)}\right)~.

We find that the ensembles we study are statistically described by universal continuous222Even when the ensembles are finite, such as the projected ensemble, we treat them as finite samples from an underlying continuous ensemble. state ensembles with maximum ensemble entropy, up to constraints imposed by energy conservation or effective constraints imposed by thermalization. We note that the ensemble entropy differs from the traditionally considered von Neumann entropy S(ρ(1))=tr[ρ(1)log2ρ(1)]S(\rho^{(1)}_{\mathcal{E}})=-\text{tr}[\rho^{(1)}_{\mathcal{E}}\log_{2}\rho^{(1)}_{\mathcal{E}}] which only depends on the average state ρ(1)\rho^{(1)}_{\mathcal{E}}. Instead, Ent()\text{Ent}(\mathcal{E}) depends on the distribution of the individual states P(Ψ)P(\Psi), which is determined by all moments of \mathcal{E}, and has distinct information-theoretic properties (Fig. 2). We also remark that while similar notions have been studied in the past, e.g. in Refs. [53, 54], this work explicitly shows the link between this entropy measure and ensembles of states obtained from dynamics, the first to do so to our knowledge.

In the following, we demonstrate this result by focusing on two specific settings —temporal and projected ensembles— and their relationship. Finally, we discuss implications of our results, summarized in Table 1.

II.1 Temporal ensembles are random phase ensembles

The first ensemble of states we study is the temporal ensemble, consisting of quantum states generated by time evolution of an initial state |Ψ0\ket{\Psi_{0}} under a Hamiltonian HH [Fig. 3(a)], with probabilities being uniform on the time interval t[τ/2,τ/2]t\in[-\tau/2,\tau/2].

Temp.(τ)\displaystyle\mathcal{E}_{\text{Temp.}}(\tau) {(1/τ,|Ψ(t))}t=τ/2τ/2\displaystyle\equiv\left\{\left(1/\tau,\ket{\Psi(t)}\right)\right\}_{t=-\tau/2}^{\tau/2} (3)
{(1/τ,exp(iHt)|Ψ0)}t=τ/2τ/2.\displaystyle\equiv\left\{(1/\tau,\exp(-iHt)\ket{\Psi_{0}})\right\}_{t=-\tau/2}^{\tau/2}~.

This ensemble consists of the trajectory of a quantum state as it evolves under HH. The statistical properties of the trajectory quantify its degree of ergodicity in Hilbert space, as our results will make clear.

Conventionally, quantum ergodicity is formulated in terms of the energy eigenstates of a system [68, 69, 70]. While this applies to time-independent Hamiltonian evolution or Floquet dynamics, it is not applicable to quantum dynamics in general. Furthermore, the notion of ergodicity is fundamentally dynamical in nature, and its definition in terms of static quantities such as energy eigenstates illustrates a tension in our understanding of quantum ergodicity. Recent works [39, 40] challenge this formulation. Specifically, they show that the ergodicity of certain time-dependent Hamiltonian dynamics cannot be formulated using conventional approaches based on the statistical properties of (quasi)-energy eigenstates, but instead can be characterized by the statistical properties of their trajectories. The central idea of Refs. [39, 40] is to test if a time-evolved state uniformly visits all ambient points in the Hilbert space, a property dubbed (complete) Hilbert-space ergodicity.

Our work generalizes the notion of Hilbert-space ergodicity [39] to systems with energy conservation. The presence of energy conservation explicitly disallows ergodicity over all of Hilbert space. Nevertheless, we find that many-body dynamics under time-independent Hamiltonian evolution is still Hilbert-space ergodic, in the sense that the evolved states uniformly explore their available Hilbert space, which is constrained due to energy conservation. In order to understand the effects of energy conservation, consider the simple case of a single qubit initially polarized in the xx-direction, subject to a magnetic field along the zz-axis. Over time, the qubit periodically traces out an equatorial circle on the Bloch sphere, instead of uniformly covering the sphere. In contrast, quantum dynamics without such conservation laws, specifically a quasiperiodically driven system, can uniformly explore Hilbert space [39].

As in classical dynamics [71], we find that a quantum many-body system may be Hilbert-space ergodic without being quantum chaotic (i.e. being integrable). We show that Hilbert-space ergodicity holds as long as the system satisfies the no-resonance conditions (Definition 3), colloquially the absence of higher order resonances in the spectrum (Definition 3). This condition is widely assumed to be true in chaotic systems [69, 17, 72, 13], and we numerically observe that it is also true in certain integrable systems such as the XXZ model, but not in others such as the transverse field Ising model.

To prove our claim of Hilbert-space ergodicity, we show that in the τ\tau\rightarrow\infty infinite interval limit, the temporal ensemble is equal to the random phase ensemble [46, 73]. This is the ensemble of states with fixed amplitudes |cE||E|Ψ0||c_{E}|\equiv|\langle E|\Psi_{0}\rangle| in some basis {|E}\{|E\rangle\}, here the energy eigenbasis, and with complex phases arg(cE)\text{arg}(c_{E}) independently random from the uniform distribution Unif([0,2π))\text{Unif}([0,2\pi)). In a time-evolved state, the amplitudes |cE|2|c_{E}|^{2} are fixed by the initial state and Hamiltonian, and the phases are the only degrees of freedom.

Theorem 1 (informal): The infinite-time temporal ensemble obtained by evolving an initial state is equal to a random phase ensemble if the Hamiltonian HH satisfies all kk-th no-resonance conditions. Conversely, for almost every initial state, if the infinite-time temporal ensemble is equal to a random phase ensemble, the Hamiltonian HH satisfies all kk-th no-resonance conditions.

The above Theorem is somewhat of a folklore result [72, 70, 46], often stated without proof. Here we provide a full statement and proof, for several reasons. Firstly, Theorem 1 has a natural interpretation in terms of a maximum entropy principle. The random phase ensemble has maximum ensemble entropy [Eq. (2)] on this restricted Hilbert space, and in Sec. IV.2, we quantify the rate of convergence of the temporal ensemble to the random phase ensemble with increasing interval τ\tau. Next, this leads to a novel type of universality in the many-body setting, which we outline below.

Unlike the above example of a two-level system, the exponentially larger Hilbert space in many-body systems leads to behaviour much more akin to ergodicity in the full Hilbert space. Specifically, we show that the trajectory appears pseudorandom in the following sense: in the limit of large Hilbert space dimension DD, the temporal ensemble is statistically equal to the Haar ensemble, deformed as follows

|Ψ(t)stat.Dρd|ϕ,|ϕHaar,\ket{\Psi(t)}\overset{\text{stat.}}{\approx}\sqrt{D\rho_{d}}\ket{\phi}~,~\ket{\phi}\sim\text{Haar}, (4)

where ρdE|EE|Ψ0Ψ0|EE|\rho_{d}\equiv\sum_{E}|E\rangle\!\langle E|\Psi_{0}\rangle\!\langle\Psi_{0}|E\rangle\!\langle E| is the diagonal ensemble. By “statistically equal,” we mean that Eq. (4) is not true on the level of wavefunctions, but the moments ρ(k)\rho^{(k)} of the ensembles are very close. To show this, we develope a simple analytic expression for the moments of the temporal ensemble that holds in the DD\rightarrow\infty limit (Sec. IV.1). One consequence of this pseudorandomness is our finding that the measurement probabilities, upon appropriate rescaling, universally follow the Porter-Thomas distribution (below, and Sec. IV.3), a hallmark of random quantum states which had been observed in quantum dynamics without conservation laws [74]. We discuss information-theoretic aspects of this result and also prove an analog of the ergodic theorem, a result in the theory of dynamical systems [75], where we relate the statistics defined over time and configuration space (discussed below and in Sec. IV.5).

II.2 Projected ensembles are (generalized) Scrooge ensembles

Not only do temporal ensembles satisfy the maximum entropy principle, so do projected ensembles. Projected ensembles, introduced in a modern context by Refs. [28, 29] (and previously studied in Ref. [76]), are ensembles of states obtained from a single, larger many-body state |Ψ\ket{\Psi}. To do so, we partition the system into two parts AA and BB, such that the Hilbert space factorizes (for simplicity of discussion) as =AB\mathcal{H}=\mathcal{H}_{A}\otimes\mathcal{H}_{B}. We then measure BB in a basis {|zB}\{\ket{z_{B}}\}. Each measurement outcome |zB\ket{z_{B}} projects |Ψ\ket{\Psi} into a distinct pure state |Ψ(zB)A\ket{\Psi(z_{B})}\in\mathcal{H}_{A} [Fig. 1(d)], defined as

|Ψ(zB)(𝕀AzB|)|Ψ/p(zB),\ket{\Psi(z_{B})}\equiv(\mathbb{I}_{A}\otimes\bra{z_{B}})\ket{\Psi}/\sqrt{p(z_{B})}~, (5)

where the normalization constant is the probability of obtaining the outcome zBz_{B}, p(zB)Ψ|(𝕀A|zBzB|)|Ψp(z_{B})\equiv\bra{\Psi}(\mathbb{I}_{A}\otimes\outerproduct{z_{B}}{z_{B}})\ket{\Psi}. This defines the projected ensemble:

Proj.={p(zB),|Ψ(zB)}zB.\displaystyle\mathcal{E}_{\text{Proj.}}=\{p(z_{B}),\ket{\Psi(z_{B})}\}_{z_{B}}~. (6)

For concreteness, we primarily study global states |Ψ=exp(iHt)|Ψ0\ket{\Psi}=\exp(-iHt)\ket{\Psi_{0}} obtained from time-evolution by a generic, ergodic many-body Hamiltonian HH, which rapidly reach thermal equilibrium at a temperature set by the energy of the initial state [51, 52]. We will also study eigenstates of HH that exhibit similar thermal properties.

In Refs. [28, 29], it was found that when the state |Ψ\ket{\Psi} is obtained from time-evolution under a chaotic many-body Hamiltonian at infinite effective temperature, the projected ensemble of small subsystems is very well-approximated by the Haar ensemble. This phenomenon is known as deep thermalization [30]. This has since been rigorously shown in a variety of settings, including deep random circuits [29, 36], dual-unitary models [77, 30, 31, 32, 78], free fermion models [33], or Hamiltonian dynamics with certain assumptions [34]. Various aspects including the rate of deep thermalization [30, 31, 78, 36], the effect of symmetries [35, 37], and the presence of magic in the global state [38] have been investigated. The majority of what we know so far is restricted to the setting of infinite temperature dynamics. In this work, we set out to study the projected ensembles formed in finite temperature dynamics.

At finite temperature, projected ensembles are not close to the Haar ensemble. To see this, note that the first moment of the projected ensemble ρ(1)zBp(zB)|Ψ(zB)Ψ(zB)|=trB[|ΨΨ|]\rho^{(1)}_{\mathcal{E}}\equiv\sum_{z_{B}}p(z_{B})\outerproduct*{\Psi(z_{B})}{\Psi(z_{B})}=\text{tr}_{B}[\outerproduct{\Psi}{\Psi}] is precisely equal to the reduced density matrix (RDM). In turn, only at infinite temperature is the RDM equal to the first moment of the Haar ensemble, the maximally mixed state 𝕀A/DA\mathbb{I}_{A}/D_{A}. This follows from a standard result in quantum thermalization [69] that the reduced density matrix ρA\rho_{A} is well described by a Gibbs state, i.e. ρAexp(βHA)\rho_{A}\propto\exp(-\beta H_{A}), where HAH_{A} is the Hamiltonian truncated to the subsystem AA, and β1\beta^{-1} is an effective temperature. It has remained an open question as to how to describe these projected ensembles at finite temperature.

In this work, we find that a maximum entropy principle still applies: under an “energy non-revealing” condition on the measurement basis {|zB}\{|z_{B}\rangle\} which we shall discuss below, the projected ensembles are described by the ensembles that have maximum entropy, over all ensembles with a given first moment. These ensembles have been previously studied and were termed Scrooge ensembles [79, 80, 81] for their unique information-theoretic properties333As a technical point, Scrooge ensembles only have maximum entropy when they are regarded as ensembles of unnormalized states [80], see Sec. III.. Each density matrix ρ\rho uniquely defines a Scrooge ensemble Scrooge(ρ)\text{Scrooge}(\rho), and our result is that at finite temperatures, the projected ensemble is the Scrooge ensemble corresponding to the appropriate thermal RDM.

Claim 1 (Corollary of Theorem 4 below, informal): The projected ensemble is statistically described by the Scrooge ensemble, if the measurement basis {|zB}\{\ket{z_{B}}\} is energy non-revealing.

The Scrooge ensemble may be essentially understood as a deformation of the Haar ensemble such that its first moment equals ρ\rho [79] [Fig. 4(b)]. Specifically, we can form the Scrooge ensemble by first sampling Haar-random states |ψHaar\ket{\psi}\sim\text{Haar}, then deforming them by a given density matrix ρ\rho, i.e.

|ΦScrooge(ρ)|Φρ|ψ,|ψHaar.\ket{\Phi}\sim\text{Scrooge}(\rho)~\Longleftrightarrow~\ket{\Phi}\propto\sqrt{\rho}\ket{\psi},~\ket{\psi}\sim\text{Haar}~. (7)

The probability p(Φ)p(\Phi) is further weighted by the norm ψ|ρ|ψ\bra{\psi}\rho\ket{\psi}, leading to the distribution of states

Scrooge(ρ)={p(Φ)=Dψ|ρ|ψ,|Φ=ρ|ψψ|ρ|ψ1/2}|ψHaar.\displaystyle\mathcal{E}_{\text{Scrooge}}(\rho)=\left\{\begin{aligned} &p(\Phi)=D\bra{\psi}\rho\ket{\psi},\\ &\ket{\Phi}=\frac{\sqrt{\rho}\ket{\psi}}{\bra{\psi}\rho\ket{\psi}^{1/2}}\end{aligned}\right\}_{\ket{\psi}\sim\text{Haar}}. (8)

In Sec. V.3, we give closed form expressions for the resultant probability distribution function p(Φ)p(\Phi), as well as its kk-th moments.

Scrooge ensembles do not always describe the projected ensemble. They only do so in the special case when the measurement basis zBz_{B} is uncorrelated with the energy, defined precisely in Sec. VI.4. Under these circumstances, we can treat the projected states |Ψ(zB)\ket{\Psi(z_{B})} as independent, identical samples from a distribution. However, this is not always true: when zBz_{B} is correlated with the subsystem energy, it is also correlated with the projected state |Ψ(zB)\ket*{\Psi(z_{B})}, such that the energies EAΨ(zB)|HA|Ψ(zB)E_{A}\equiv\langle\Psi(z_{B})|H_{A}|\Psi(z_{B})\rangle and EBzB|HB|zBE_{B}\equiv\langle z_{B}|H_{B}|z_{B}\rangle are anticorrelated. Under these general conditions, we introduce the generalized Scrooge ensemble to describe the projected ensemble (Section VI). Each projected state |Ψ(zB)|\Psi(z_{B})\rangle should be treated as a sample from a different Scrooge ensemble for each zBz_{B}, i.e. where the first moment is not the RDM ρA\rho_{A}, but is a zBz_{B}-dependent mixed state ρ¯(zB)\bar{\rho}(z_{B}), which we may determine by the time-average of the projected state, 𝔼t[|Ψ(zB,t)Ψ(zB,t)|]\mathbb{E}_{t}[|\Psi(z_{B},t)\rangle\!\langle\Psi(z_{B},t)|]. Therefore, the generalized Scrooge ensemble is more accurately a collection of ensembles, where the projected ensemble is comprised of exactly one sample from each ensemble, but we will abuse notation and refer to it as an ensemble. This is our second key result.

Theorem 4 (informal): At sufficiently long times tt,444While our proof requires times that are exponential in system size, we find from numerical experiments that we only require times that grow weakly (at most polynomially) with system size. the projected ensemble generated by a time-evolved state |Ψ(t)\ket{\Psi(t)} is approximately equal to the generalized Scrooge ensemble, which depends on the measurement basis {zB}\{z_{B}\}, Hamiltonian HH, and initial state |Ψ0\ket{\Psi_{0}}, if the Hamiltonian HH satisfies all kk-th no-resonance conditions. The converse direction holds for almost every initial state |Ψ0\ket{\Psi_{0}}.

When the measurement basis {zB}\{z_{B}\} is uncorrelated with the global energy, the generalized Scrooge ensemble reduces to the Scrooge ensemble (Claim 1). Specifically, this happens when the mixed state ρ¯(zB)\bar{\rho}(z_{B}) is independent of zBz_{B} (up to a rescaling), giving a concrete definition for “energy non-revealing” bases.

We prove this result using our knowledge of the temporal ensemble. We show that the temporal trajectory of the projected state |Ψ(zB,t)|\Psi(z_{B},t)\rangle is a Scrooge ensemble (here, because of the subsystem projection, there are no hard constraints on the amplitudes, as there were in the case of the global state). We then utilize our ergodic theorem (Sec. IV.5) to translate our results for a distribution over time to a distribution over the outcomes zBz_{B}.

While Theorem 4 concerns the time-evolved states |Ψ(t)\ket{\Psi(t)}, we numerically observe that the generalized Scrooge ensemble also describes the projected ensembles obtained from energy eigenstates, suggesting that our maximum entropy principle holds for a broader class of many-body states.

In projected ensembles, the maximum entropy principle sheds light on the nature of correlations between system and bath. Specifically, the projected ensemble reveals correlations between the subsystem AA and its complement BB, which we quantify in terms of the mutual information. We find that the nontrivial correlations between them are minimal and basis independent. Here, nontrivial correlations mean any amount of correlation (mutual information) that is beyond the thermal correlations that remain after infinite-time averaging. In other words, any meaningful correlations between AA and BB originate only from energy conservation, and other symmetries if they exist. Below and in Sec. V.8 and VI.3, we discuss consequences of these correlations for tasks such as communication and compression.

II.3 Implications and discussion

Below, we summarize several broader points that we found applicable to both the temporal and projected ensemble.

Ergodic theorem — Remarkably, the temporal and projected ensembles are closely related. We prove that a version of the ergodic theorem holds in many-body dynamics. This enables us to translate statements about distributions over time (temporal ensemble) into statements about distributions over measurement outcomes, at fixed, typical points in time. In particular, we use this to prove Theorem 4, a statement about the projected ensembles at typical points in time.

Operational meaning of ensemble entropy — In this work, we use the ensemble entropy to define our maximum entropy principle. We also identify its operational interpretation: the ensemble entropy quantifies the information required to classically represent an ensemble of states. Our maximally entropic ensembles are therefore the most difficult to compress, formalizing our intuition that the states generated by natural, chaotic many-body dynamics lack overall structure that can be exploited for compression.

Specifically, the ensemble entropy quantifies the number of classical bits required to store an ensemble of states up to numerical precision ϵ\epsilon [Fig. 2(a)]. This should be contrasted with the minimum number of qubits required to store the ensemble in quantum memory, which was established by Schumacher to depend only on the first moment ρ\rho_{\mathcal{E}} of the ensemble through the von Neumann entropy S(ρ)S(\rho_{\mathcal{E}}) [82].

Information-theoretic consequences — Not only are our ensembles of states maximally difficult to classically store, they are also maximally difficult to distinguish by measurement, as quantified by their accessible information [79, 63]. The accessible information quantifies the maximum rate of classical information that can be transmitted by sending states from the ensemble over a quantum channel (in terms of bits of information per channel use) [Fig. 2(b)]. It is upper-bounded by the well-known Holevo bound [56], and lower-bounded by Josza, Robb, and Wooters [79]. They found that the Scrooge ensemble achieves this lower bound, dubbed the subentropy Q(ρ)Q(\rho) [Eq. (61)]. They dubbed the ensemble the “Scrooge ensemble” because it is uniquely “stingy” in the amount of information revealed under measurement. Therefore, the projected ensemble not only has maximum ensemble entropy, it also minimizes the amount of information that can be extracted. These information theoretic properties are directly measurable [Fig. 4(f) and Fig. 5(d)] but have not yet been observed in experiment.

In the case of temporal ensembles, the accessible information quantifies the amount of information that can be gained about the evolution time of the state. We find that it takes on a universal value (1γ)/ln20.61(1-\gamma)/\ln 2\approx 0.61, where γ\gamma is the Euler-Mascheroni constant [Fig. 3(d)].

Refer to caption
Figure 2: Information theoretic implications of our work. Our results indicate that two tasks are maximally difficult for our ensemble of states: (a) Information compression of an ensemble of states \mathcal{E}. The ensemble entropy quantifies the task of classically representing MM states from \mathcal{E} up to precision ϵ\epsilon (denoted here by a fine grid partitioning the Hilbert space into 2Nϵ2^{N_{\epsilon}} elements). This task requires M[Nϵ+Ent()]M\cdot[N_{\epsilon}+\text{Ent}(\mathcal{E})] classical bits. This classical storage task differs from the analogous quantum task: storing the same MM states in quantum memory only requires MS(ρ)MS(\rho_{\mathcal{E}}) qubits, where the S(ρ)S(\rho_{\mathcal{E}}) is the von Neumann entropy of the average state ρ\rho_{\mathcal{E}}. (b) We also consider a task originally studied by Holevo [56]: sending classical information by sending states from \mathcal{E} over a quantum channel. This channel capacity is given by the accessible information Acc()\text{Acc}(\mathcal{E}), which our ensembles of states uniquely minimize.

Measurable signature: universal Porter-Thomas distribution — Our maximum entropy principle selects distributions of states that are “closest” to the Haar ensemble, satisfying certain constraints. The relationship between our ensembles of states to the Haar ensemble gives rise to a measurable signature of our maximum entropy principle: the Porter-Thomas (PT) distribution [83]. Originally identified in nuclear physics settings, the PT distribution has received renewed attention in many-body physics. This distribution represents the distribution of overlaps between Haar random states |ψ\ket{\psi} with a fixed (but arbitrary) state |Φ\ket{\Phi}, i.e. the probability p(Φ)|Φ|ψ|2p(\Phi)\equiv|\langle\Phi|\psi\rangle|^{2} of measuring |Φ\ket{\Phi} from a Haar-random state |ψ\ket{\psi} [84]. Not only is the PT distribution a measurable signature of Haar-random states, we find that it is also present in many natural many-body states. In companion work [41], we experimentally and theoretically investigate the emergence of the Porter-Thomas distribution in global quantities, its relationship to local quantities, and how the Porter-Thomas distribution is modified in the presence of noise, using our results to learn about the noise in an experiment.

II.4 Organization of paper

The rest of the paper is organized as follows. In Section III, we provide an overview of the central object of our work — ensembles of pure states and their information-theoretic and measurable properties. We then proceed to study our two ensembles of states: the temporal ensemble (Section IV) and the projected ensemble in a special and a general setting (Sections V and VI). Finally, in Section VII we discuss the PT distribution in the properties of many-body eigenstates.

III Ensembles of pure states

Ensembles of pure states generated by quantum many-body dynamics are the central objects of study in this work. In this section, we provide an overview of ensembles of pure states and their properties. After defining normalized and unnormalized ensembles of states, we introduce two quintessential ensembles: the Haar and Gaussian ensembles, which we use as the reference distributions for our ensemble entropy. We then discuss the quantities we use to analyze our ensembles of states: the kk-th statistical moments, ensemble entropy, mutual information, and PT distribution.

Ensembles of pure states are collections of pure states in a Hilbert space of dimension DD, weighted by a probability distribution.

Definition 1.

An ensemble of pure states is a probability distribution of states |Ψj\ket{\Psi_{j}} weighted by probabilities pjp_{j}:

={pj,|Ψj}\mathcal{E}=\{p_{j},\ket{\Psi_{j}}\} (9)

when the pjp_{j}’s are omitted, we take the states |Ψj\ket{\Psi_{j}} to be uniformly distributed over indices jj 555We ignore subtleties about the global phase of the states |Ψj|\Psi_{j}\rangle, and assume that the states are gauge-fixed to, for example, have real positive values of the first entry u1|Ψj\langle u_{1}|\Psi_{j}\rangle in some basis {|ui}\{|u_{i}\rangle\}..

Our distributions may be continuous or discrete. In the continuous case, the distribution is a measure over normalized wavefunctions, which we can specify in terms of its components Ψzz|Ψ\Psi_{z}\equiv\innerproduct{z}{\Psi}, subject to the constraint that z|Ψz|2=1\sum_{z}\absolutevalue{\Psi_{z}}^{2}=1:

P(Ψ)dΨ=P({Ψz})δ(Ψ21)zd2Ψz,P(\Psi)d\Psi=\int P(\{\Psi_{z}\})\delta(\|\Psi\|^{2}-1)\prod_{z}d^{2}\Psi_{z}~, (10)

where zd2Ψz\prod_{z}d^{2}\Psi_{z} is the Euclidean measure on D\mathbb{C}^{\otimes D}.

Unnormalized ensembles of states — We will also find it useful to define distributions of unnormalized states. In this case, the distribution is simply a measure over DD-dimensional complex vectors, defined in terms of its components Ψ~zz|Ψ~\tilde{\Psi}_{z}\equiv\innerproduct*{z}{\tilde{\Psi}}. For consistency, we will use tildes throughout this work to denote unnormalized states |Ψ~\ket*{\tilde{\Psi}}.

An unnormalized ensemble can be mapped to a normalized one by normalizing each states. This results in the probability distribution P(Ψ)dΨ=𝑑Ψ~P(Ψ~)δ(|Ψ~/Ψ~|Ψ)P(\Psi)d\Psi=\int d\tilde{\Psi}P(\tilde{\Psi})\delta(\ket*{\tilde{\Psi}}\!/\|\tilde{\Psi}\|-\ket{\Psi}). For each state |Ψ\ket{\Psi}, the normalized ensemble averages over the distribution of the norm Ψ~\|\tilde{\Psi}\| to give the measure P(Ψ)P(\Psi). Therefore, the unnormalized ensemble contains more information than the normalized one: specifically information about fluctuations of the norm Ψ~\|\tilde{\Psi}\|.

III.1 Gaussian and Haar random ensembles of states

The ensemble entropy, which quantifies our maximum entropy principle, is defined with reference to the Haar (or Gaussian) ensemble. These ensembles are unique in being invariant under arbitrary unitary transformations and have been the subject of considerable study as paradigmatic distributions of random quantum states [48].

Random unnormalized states: the Gaussian ensemble — The Gaussian ensemble is a probability distribution over unnormalized states, where each component Ψ~zz|Ψ~\tilde{\Psi}_{z}\equiv\innerproduct*{z}{\tilde{\Psi}} is an independent complex Gaussian variable.

PGauss(Ψ~)dΨ~=zDπexp[D|Ψ~z|2]d2Ψ~z.\displaystyle P_{\text{Gauss}}(\tilde{\Psi})d\tilde{\Psi}=\prod_{z}\frac{D}{\pi}\exp[-D|\tilde{\Psi}_{z}|^{2}]d^{2}\tilde{\Psi}_{z}~. (11)

The normalization above is chosen so that the average norm is 𝔼[Ψ~|Ψ~]=1\mathbb{E}[\innerproduct*{\tilde{\Psi}}{\tilde{\Psi}}]=1.

Random normalized states: the Haar ensemble — The Haar ensemble is the unique distribution of normalized states that is invariant under any unitary transformation UU. It is naturally related to the Gaussian ensemble: one can sample from the Haar ensemble by sampling states |Ψ\ket{\Psi} from the Gaussian ensemble, then normalizing them. This induces the measure

PHaar(Ψ)dΨ=PGauss(Ψ~)δ(|Ψ|Ψ~/Ψ~)𝑑Ψ~.\displaystyle P_{\text{Haar}}(\Psi)d\Psi=\int P_{\text{Gauss}}(\tilde{\Psi})\delta(\ket{\Psi}-\ket*{\tilde{\Psi}}/\|\tilde{\Psi}\|)d\tilde{\Psi}~. (12)

The Haar ensemble has been widely studied in quantum information science and utilized for a myriad of applications, see e.g. Ref. [48]. In particular, the moments of the Haar ensemble play a key role, which we now discuss.

III.2 Statistical kk-th moments

An ensemble of states contains a large amount of information, beyond that of its density matrix. In this work, we systematically study and quantify ensembles of states by their higher-order moments.

Higher order quantities have received less attention in many-body physics because they have been inaccessible to traditional experiments, requiring a large number of controlled repetitions in order to access higher order statistical properties. Modern quantum devices address these limitations with both high repetition rates and a high degree of controllability [28, 85].

In this work, we find that not only do the average properties (first moments) thermally equilibrate, so too do the higher moments of our natural ensembles of states. We first define the kk-th moments.

The kk-th moment of an ensemble \mathcal{E} is, for discrete ensembles

ρ(k)\displaystyle\rho^{(k)}_{\mathcal{E}} jpj|ΨjΨj|k,\displaystyle\equiv\sum_{j}p_{j}\outerproduct{\Psi_{j}}{\Psi_{j}}^{\otimes k}, (13)

for continuous ensembles is

ρ(k)𝑑ΨP(Ψ)|ΨΨ|k,\rho^{(k)}_{\mathcal{E}}\equiv\int d\Psi P(\Psi)\outerproduct{\Psi}{\Psi}^{\otimes k}, (14)

and for unnormalized ensembles we have (denoting the moments with tildes)

ρ~(k)=𝑑Ψ~P(Ψ~)|Ψ~Ψ~|k.\tilde{\rho}^{(k)}_{\mathcal{E}}=\int d\tilde{\Psi}P(\tilde{\Psi})\outerproduct*{\tilde{\Psi}}{\tilde{\Psi}}^{\otimes k}. (15)

The kk-th moment of an ensemble can be used to describe the expectation value of any linear operator O(k):kkO^{(k)}\!:\!\mathcal{H}^{\otimes k}\rightarrow\mathcal{H}^{\otimes k} which acts on kk-copies of the Hilbert space. The ensemble average of O(k)O^{(k)} can be obtained from the kk-th moment: 𝔼|Ψ[Ψ|kO(k)|Ψk]=tr[O(k)ρ(k)]\mathbb{E}_{\mathcal{\ket*{\Psi}\sim E}}[\bra{\Psi}^{\otimes k}O^{(k)}\ket{\Psi}^{\otimes k}]=\text{tr}[O^{(k)}\rho^{(k)}_{\mathcal{E}}]. For example, O(k)=OkO^{(k)}=O^{\otimes k} describes the kk-th moment of the expectation values 𝔼[Ψ|O|Ψk]\mathbb{E}_{\mathcal{E}}[\bra{\Psi}O\ket{\Psi}^{k}]. Here, we use 𝔼[f(Ψ)]\mathbb{E}_{\mathcal{E}}[f(\Psi)] to denote the ensemble average 𝑑ΨP(Ψ)f(Ψ)\int d\Psi P(\Psi)f(\Psi).

The kk-th moment can be thought of as a generalization of the density matrix. Namely, the first moment ρ(1)=jpj|ΨjΨj|\rho^{(1)}=\sum_{j}p_{j}\outerproduct{\Psi_{j}}{\Psi_{j}} is the conventional density matrix, and describes the average state of the ensemble. All higher moments ρ(k)\rho^{(k)}_{\mathcal{E}} are likewise positive semi-definite operators and, for ensembles of normalized states, have unit trace.

Moments of Haar ensemble — The higher moments of the Gaussian and Haar ensembles are particularly simple. The moments of the Haar ensemble have analytical expression [48]:

ρHaar(k)\displaystyle\rho^{(k)}_{\text{Haar}} Haar𝑑Ψ|ΨΨ|k=σSkPerm(σ)i=0k1(D+i),\displaystyle\equiv\int_{\text{Haar}}d\Psi\outerproduct{\Psi}{\Psi}^{\otimes k}=\frac{\sum_{\sigma\in S_{k}}\text{Perm}(\sigma)}{\prod_{i=0}^{k-1}(D+i)}~, (16)

with similar expression for the moments of the Gaussian ensemble. Here, Perm(σ)\text{Perm}(\sigma) is the operator that permutes kk states by the permutation σSk\sigma\in S_{k}, i.e. Perm(σ)[|Ψ1|Ψk]=|Ψσ(1)|Ψσ(k)\text{Perm}(\sigma)[\ket*{\Psi_{1}}\otimes\cdots\otimes\ket*{\Psi_{k}}]=\ket*{\Psi_{\sigma(1)}}\otimes\cdots\otimes\ket*{\Psi_{\sigma(k)}}. This result follows from the Schur-Weyl duality [86]. Its simplicity enables the design of useful quantum algorithms such as classical shadow tomography [87, 88], as well as to analytical solutions of models of quantum dynamics such as random quantum circuits [64]. We will find that our ensembles of states have kk-th moments with similar structure as the Haar kk-th moments, enabling phenomena such as the emergence of the Porter-Thomas distribution.

Trace distance of higher moments — Higher moments also enable the systematic and quantitative comparison of two ensembles of states. Specifically, we compare their kk-th moments through their trace distance.

ΔHaar(k)\displaystyle\Delta^{(k)}_{\text{Haar}} 12ρ(k)ρHaar(k),\displaystyle\equiv\frac{1}{2}\|\rho^{(k)}_{\mathcal{E}}-\rho^{(k)}_{\text{Haar}}\|_{*}~, (17)

where the trace (or nuclear) norm \|\cdot\|_{*} of a matrix is the sum of the absolute values of its eigenvalues. In particular, the trace distance sets an upper bound on how well any kk-copy observable O(k)O^{(k)} can distinguish two ensembles of states [63].

Refs. [28, 29] found that with increasing system size, projected ensembles of infinite effective temperature states — either obtained from quench dynamics of simple initial states or chosen from high energy eigenstates — have closer kk-th moments to the Haar moment (in other words, projected ensembles form better approximate kk-designs). Analogously, we find that the kk-th moments of the projected and (generalized) Scrooge ensembles converge in trace distance with increasing system size [Fig. 4(b) and Fig. 5(b)].

III.3 Ensemble entropy

The central claim of our work is that natural many-body dynamics generates ensembles of states with maximum entropy. The entropy in question, the ensemble entropy, differs from the conventional von Neumann entropy S(ρ)tr[ρlog(ρ)]S(\rho)\equiv-\tr[\rho\log(\rho)]. This is not suitable as an entropy of an ensemble of states, since it only depends on the first moment ρ\rho, and there are many ensembles of states which share the same average state ρ\rho. Instead, our ensemble entropy is the Shannon entropy of the distribution of states over Hilbert space. For continuous distributions, this entropy must be defined relative to a reference distribution.

Definition 2.

The ensemble entropy Ent()\text{Ent}(\mathcal{E}) is the negative of the Kullback-Leibler (KL) divergence between \mathcal{E} and the reference measure. For normalized ensembles, the reference distribution is the Haar measure [Eq. (2)]

Ent()\displaystyle\text{Ent}(\mathcal{E}) 𝒟(Haar)\displaystyle\equiv-\mathcal{D}(\mathcal{E}\|\text{Haar}) (18)
=𝑑ΨP(Ψ)log2(P(Ψ)PHaar(Ψ)),\displaystyle=-\int d\Psi~P(\Psi)\log_{2}\left(\frac{P(\Psi)}{P_{\text{Haar}}(\Psi)}\right)~,

while for unnormalized ensembles, the reference distribution is the Gaussian measure

Ent()\displaystyle\text{Ent}(\mathcal{E}) 𝒟(Gauss)\displaystyle\equiv-\mathcal{D}(\mathcal{E}\|\text{Gauss}) (19)
=𝑑Ψ~P(Ψ~)log2(P(Ψ~)PGauss(Ψ~)).\displaystyle=-\int d\tilde{\Psi}~P(\tilde{\Psi})\log_{2}\left(\frac{P(\tilde{\Psi})}{P_{\text{Gauss}}(\tilde{\Psi})}\right)~.

The ensemble entropy is non-positive, achieving zero only when the two distributions are equal. See Refs. [53, 54] for similar attempts to quantify the entropy of state ensembles.

Connection to von Neumann entropy — While the ensemble entropy may seem an unconventional notion of entropy, in fact it can be related to the von Neumann entropy of higher moments of the ensemble. For finite ensembles ={pj,|Ψj}\mathcal{E}=\{p_{j},|\Psi_{j}\rangle\}, the quantity

limkS(ρ(k)).\text{lim}_{k\rightarrow\infty}S(\rho^{(k)}_{\mathcal{E}})~. (20)

converges to the Shannon entropy jpjlog2pj-\sum_{j}p_{j}\log_{2}p_{j}, which in turn converges to the ensemble entropy (plus a divergent constant), when the discrete ensemble is appropriately selected (see below). This follows from the fact that with increasing number of copies kk, two states |ψk,|ϕk\ket{\psi}^{\otimes k},\ket{\phi}^{\otimes k} become increasingly distinguishable (orthogonal), since |ϕ|ψ|2k|\innerproduct{\phi}{\psi}|^{2k} quickly vanishes when |ϕ|ψ|1|\innerproduct{\phi}{\psi}|\neq 1. Therefore, for any discrete ensemble {pj,|Ψj}\{p_{j},\ket{\Psi_{j}}\}, the quantities pjp_{j} and |Ψjk\ket{\Psi_{j}}^{\otimes k} converge to the eigenvalues and eigenvectors of ρ(k)\rho^{(k)} respectively. A similar argument can be made for continuous distributions, and we obtain Ent()=limkS(ρ(k))+C\text{Ent}(\mathcal{E})=\text{lim}_{k\rightarrow\infty}S(\rho^{(k)}_{\mathcal{E}})+C, with a divergent constant CC.

Ensemble entropy and information compression — The ensemble entropy has a simple operational meaning: it quantifies the minimum number of bits required to store the states in the ensemble in classical memory, up to a fixed precision ϵ\epsilon. If storing a single such state requires NϵN_{\epsilon} classical bits, we find that storing a large number of states MM drawn from the ensemble \mathcal{E} requires M[Nϵ+Ent()]M[N_{\epsilon}+\text{Ent}(\mathcal{E})] classical bits (note that Ent()0\text{Ent}(\mathcal{E})\leq 0), as illustrated in Fig. 2(a). Our maximum entropy principle indicates that the required storage for many-body states generated by time evolution is maximal.

To see this, note that storing states up to accuracy ϵ\epsilon amounts to partitioning the Hilbert space in a uniform ϵ\epsilon-net, i.e. into a discrete set of points which are a distance ϵ\epsilon apart [3]. To store the states in the ensemble, each state in the continuous Hilbert space is assigned to one of these discrete points. Storing these states thus amounts to storing these labels — they may either be stored directly or they may be compressed according to how often the labels appear. In the former case, each label requires NϵN_{\epsilon} bits, if there are 2Nϵ2^{N_{\epsilon}} points in the net. In the latter, the optimal rate of compression is given by the relative entropy between the discrete probabilities on this ϵ\epsilon-net and the uniform distribution [3]. In the limit of small ϵ\epsilon, the discrete and continuous distributions converge, and therefore the overall amount of compression achieved is equal to the ensemble entropy Ent()\text{Ent}(\mathcal{E}) per state, and the overall number of bits required to store MM states is M[Nϵ+Ent()]M[N_{\epsilon}+\text{Ent}(\mathcal{E})].

Finally, we note that this task is different from the task of storing an ensemble of states in quantum memory, i.e. compressing the states into a smaller number of qubits. This can be done by Schumacher compression [82], with compression rate given by the von Neumann entropy of the first moment S(ρ)S(\rho_{\mathcal{E}}). In contrast, the task of classical approximate storage depends on all moments of the ensemble.

III.4 Mutual information

In addition to the ensemble entropy, we study another information-theoretic property: the mutual information, which quantifies the information-theoretic correlations between two variables. We find that not only is the ensemble entropy maximized in our ensembles, the mutual information is also minimized. These two quantities are a priori unrelated, and lead to distinct operational consequences.

The mutual information I(X;Z)I(X;Z) between random variables XX and ZZ is defined as

I(X;Z)=H(X)H(X|Z)\displaystyle I(X;Z)=H(X)-H(X|Z) (21)
=xp(x)log2[p(x)]+zp(z)xp(x|z)log2[p(x|z)].\displaystyle=-\sum_{x}p(x)\log_{2}[p(x)]+\sum_{z}p(z)\sum_{x}p(x|z)\log_{2}[p(x|z)]~.

This can be interpreted as the information gain (reduction in entropy) of the variable XX from measurements of the variable ZZ. In this work, we take the logarithm to base 2, which quantifies the mutual information in units of bits.

The mutual information is relevant in several information theoretic tasks, including quantifying the maximum rate of information transmission through a noisy classical channel [79], where the variables XX and ZZ are here the channel input and output. Below, we focus on the significance of mutual information to a quantum communication task.

Holevo studied the use of a noiseless quantum channel to transmit classical information [56]. Given an ensemble of (possibly mixed) states ={px,ρx}\mathcal{E}=\{p_{x},\rho_{x}\}, Holevo considered a task where one party (Alice) transmits a classical message {x1,x2,}\{x_{1},x_{2},\dots\} to another (Bob), by sending a sequence of states {ρx1,ρx2,}\{\rho_{x_{1}},\rho_{x_{2}},\dots\}. Bob decodes this message using measurements ZZ on the received states. When the quantum states |ψxi\ket*{\psi_{x_{i}}} are not orthogonal, the symbols xix_{i} cannot be unambiguously determined: the maximum classical transmission rate is therefore given by the accessible information

Acc()supZI(X;Z),\text{Acc}(\mathcal{E})\equiv\text{sup}_{Z}I(X;Z), (22)

the mutual information maximized over complete measurement bases ZZ.

Holevo provided an upper bound for the accessible information: the Holevo-χ\chi quantity χ()S(ρ)xpxS(ρx)\chi(\mathcal{E})\equiv S(\rho_{\mathcal{E}})-\sum_{x}p_{x}S(\rho_{x}) [56]. In an ensemble of pure states, the Holevo-χ\chi quantity is equal to the von Neumann entropy χ()=S(ρ)\chi(\mathcal{E})=S(\rho_{\mathcal{E}}). For a given first moment ρ\rho_{\mathcal{E}}, the Holevo bound is saturated when the ensemble of states is formed using the eigenbasis of ρ\rho_{\mathcal{E}}, i.e. ={λj,|j}\mathcal{E}=\{\lambda_{j},\ket{j}\}.

Josza, Robb and Wooters [79] subsequently established a lower bound for the accessible information, dubbed the subentropy Q(ρ)Q(\rho_{\mathcal{E}}). That is, for any ensemble of pure states \mathcal{E} with a given first moment ρ\rho_{\mathcal{E}}, the accessible information is bounded by:

Q(ρ)Acc()S(ρ).\displaystyle Q(\rho_{\mathcal{E}})\leq\text{Acc}(\mathcal{E})\leq S(\rho_{\mathcal{E}})~. (23)

Like the von Neumann entropy, the subentropy only depends on the first moment ρ\rho_{\mathcal{E}}. For any ensemble \mathcal{E}, the subentropy is defined as the mutual information I(X;Z)I(X;Z) averaged over all complete projective measurement bases ZZ. This is equivalent to averaging over unitaries UU which rotate measurement bases and has expression [79, 89]

Q(ρ)j(ijλjλjλi)log2λj,Q(\rho_{\mathcal{E}})\equiv-\sum_{j}\left(\prod_{i\neq j}\frac{\lambda_{j}}{\lambda_{j}-\lambda_{i}}\right)\log_{2}\lambda_{j}~, (24)

where λj\lambda_{j} are the eigenvalues of ρ\rho_{\mathcal{E}}. More recently, the subentropy has received attention in studies of the dynamics of information in many-body systems [61].

Josza, Robb and Wooters established that the Scrooge ensemble saturates this lower bound (Section V[79]). Scrooge ensembles are therefore ensembles with minimum accessible information (or, the most information “stingy” ensembles), leading to their name.

(Generalized) Scrooge ensembles describe projected ensembles, therefore our projected ensembles have minimal accessible information. That is, in natural many-body systems, it is maximally hard to predict the measurement outcome of one subsystem from measurements on the other.

III.5 Porter-Thomas distribution

Finally, we turn to a measurable signature of our maximum entropy principle, the Porter-Thomas (PT) distribution. Originally studied in nuclear physics [83], the PT distribution has received renewed attention due to its presence in Haar-random many-body states, approximated by the output of deep random quantum circuits [84].

Concretely, a positive random variable XX follows the Porter-Thomas distribution with mean μ\mu [i.e. XPT(μ)X\sim\text{PT}(\mu)] if its probability distribution function is

Pr[X=x]dx\displaystyle\text{Pr}[X=x]dx =exp(x/μ)dx/μ\displaystyle=\exp(-x/\mu)dx/\mu (25)

In particular, the overlap of states |Ψ\ket{\Psi} drawn from the Haar random distribution to any fixed state |Φ\ket{\Phi} follows a PT-distribution with μ=1/D\mu=1/D [74]

Pr|ΨHaar[|Ψ|Φ|2=x]dx=exp(Dx)dDx.\text{Pr}_{\ket{\Psi}\sim\text{Haar}}[\absolutevalue{\innerproduct{\Psi}{\Phi}}^{2}=x]dx=\exp(-Dx)dDx~. (26)

Notably, this is a statement about the probability distribution of probabilities |Ψ|Φ|2|\langle\Psi|\Phi\rangle|^{2}. While relatively unknown in physical settings, this has received attention in the statistical literature under several names, including the fingerprint and histogram of the histogram [90, 91]. The fingerprint is sufficient to describe the properties of a distribution that do not depend on its labels, most notably its Shannon entropy.

The probability-of-probabilities can be studied in modern quantum experiments. For example, the PT distribution expected from Haar-random states has been predicted and subsequently experimentally verified in deep random quantum circuits [84, 74] and many-body states at infinite temperature [29]. Even though the ensembles we study in this work are not the Haar ensemble, we find that the maximum entropy principle leads to the same PT distribution, when observables are appropriately normalized. In companion work, we experimentally and theoretically study this object in detail [41].

IV Temporal ensembles

The most natural setting of quantum dynamics is time evolution under a time-independent Hamiltonian. It is an important endeavor to understand the phenomena that can arise in this setting. Despite the fact that it is simple to state, relatively few tools are available to analyze the dynamics under an interacting many-body Hamiltonian, and one must often turn to more analytically tractable models of quantum dynamics such as random quantum circuits [64].

In this section, we show universal statistical properties of the temporal ensemble — the time-trace of a state evolving under Hamiltonian dynamics. Intuitively, Hamiltonian dynamics conserves energy, and its trajectory in Hilbert space cannot be unconstrained. In fact, the Schrödinger equation conserves the population |cE|2|c_{E}|^{2} of each energy eigenstate |E\ket{E}, where |cE|2|Ψ0|E|2|c_{E}|^{2}\equiv|\langle\Psi_{0}|E\rangle|^{2} for an initial state |Ψ0\ket{\Psi_{0}}. This imposes a total of DD (the Hilbert space dimension) constraints, and the only remaining degrees of freedom are the complex phases arg(E|Ψ(t))\text{arg}(\langle E|\Psi(t)\rangle). One might conjecture that these phases are uniformly and independently random, satisfying the maximum entropy principle.

Indeed, this expectation is correct: the temporal ensemble is (statistically) described by the random phase ensemble, the ensemble of states with fixed magnitudes and independent, uniformly random phases in the energy eigenbasis. This has been a folklore result in several works, see e.g. Refs. [72, 70, 46], however, it has to our knowledge not been presented with the same level of rigor as in this work. Here, we rigorously state our result with a necessary and sufficient condition — the kk-th no-resonance condition — and discuss its corollaries.

The first moments of the temporal ensemble are heavily used in the quantum thermalization literature. Seminal works in quantum chaos and thermalization introduced the diagonal ensemble ρdE|cE|2|EE|\rho_{d}\equiv\sum_{E}|c_{E}|^{2}\outerproduct*{E}{E} to characterize observables at thermal equilibrium [51, 52]. However, this only captures the average values of observables and does not address higher moments such as their variance over time. Previous work either bounds [13, 17] or predicts the dependence of fluctuations with the total Hilbert space dimension [69, 92, 93], but here we provide approximate formulae for all moments. We apply our result on the temporal ensemble to a novel class of observables: global projective measurements, discovering universal behaviour in its higher moments and information-theoretic properties.

We first formalize the above intuitive expectation and present the sketch of our proof. We define the temporal ensemble as the set of |Ψ(t)\ket{\Psi(t)} explored by a fixed initial state |Ψ0\ket{\Psi_{0}} time-evolving under a Hamiltonian HH over a time interval t[τ/2,τ/2]t\in[-\tau/2,\tau/2].

Temp.(τ)={exp(iHt)|Ψ0}t=τ/2τ/2\displaystyle\mathcal{E}_{\text{Temp.}}(\tau)=\{\exp(-iHt)\ket{\Psi_{0}}\}_{t=-\tau/2}^{\tau/2} (27)

When τ\tau\rightarrow\infty, we equate this to the random phase ensemble, the distribution of states with fixed magnitudes in a given basis |j|Ψ|=|cj||j|Ψ0|\absolutevalue{\innerproduct{j}{\Psi}}=\absolutevalue{c_{j}}\equiv\absolutevalue{\innerproduct{j}{\Psi_{0}}}, with i.i.d. random phases ϕjUnif([0,2π))\phi_{j}\sim\text{Unif}([0,2\pi)).

Rand. Phase={j=1D|cj|eiϕj|j|ϕjUnif([0,2π))},\displaystyle\mathcal{E}_{\text{Rand.~Phase}}=\Big{\{}\sum_{j=1}^{D}\absolutevalue{c_{j}}e^{i\phi_{j}}\ket{j}~\Big{|}~\phi_{j}\sim\text{Unif}([0,2\pi))\Big{\}},

This ensemble was studied in Ref. [46], which characterizes properties such as the entanglement of its typical states. Ref. [73] develops a graphical calculus to perform the kk-copy averages we discuss in this work, although we will not need its full generality here (see also Ref. [94, 95]).

Our result, which equates the temporal ensemble (in the limit τ\tau\rightarrow\infty) with the random phase ensemble, relies on the following assumption:

Definition 3 (kk-th no-resonance condition).

A Hamiltonian HH satisfies the kk-th no-resonance condition [17, 72, 96, 97] if for any two sets of kk eigenvalues {Eαj}j=1k\{E_{\alpha_{j}}\}_{j=1}^{k} and {Eβj}j=1k\{E_{\beta_{j}}\}_{j=1}^{k} the equation

Eα1+Eα2++Eαk=Eβ1+Eβ2++EβkE_{\alpha_{1}}+E_{\alpha_{2}}+\cdots+E_{\alpha_{k}}=E_{\beta_{1}}+E_{\beta_{2}}+\cdots+E_{\beta_{k}}~ (28)

is true if and only if the sets of indices (α1,α2,,αk)(\alpha_{1},\alpha_{2},\dots,\alpha_{k}) and (β1,β2,,βk)(\beta_{1},\beta_{2},\dots,\beta_{k}) are equal up to reordering.

In other words, the two sets of indices must be related by some permutation σSk\sigma\in S_{k}, βj=ασ(j)\beta_{j}=\alpha_{\sigma(j)}. It is easy to see that the kk-th no-resonance condition implies the kk^{\prime}-th no-resonance conditions, for k<kk^{\prime}<k. The no-resonance conditions are considered generically satisfied and are used in works such as Refs. [69, 17, 72, 13]: it is believed that if the condition does not hold in a given ergodic system, any small perturbation will generically break any kk-th resonances [96]. We note recent work [98] that studies near-violations of the kk-th no-resonance condition, concluding that they are generically small and bounding their effects on certain quantities.

The no-resonance condition for k=1k=1 simply states that the energy eigenspectrum has no degeneracies. However, this may not be satisfied in generic Hamiltonians. The presence of non-Abelian symmetries ensures that there are degenerate multiplets of states. Strictly speaking, this violates all no-resonance conditions. However, for the purposes of the temporal ensemble, we may disregard these degeneracies: the initial state |Ψ0\ket{\Psi_{0}} projects the degenerate eigenspace onto a single eigenstate. Our conclusions below hold as long as the spectrum of HH, with degeneracies removed, satisfies the kk-th no-resonance conditions [94]. We call this the no-resonance condition modulo degeneracies.

With the above definitions, we are ready to state our result equating the temporal and random phase ensembles.

Theorem 1.

Given an initial state |Ψ0\ket{\Psi_{0}} and a Hamiltonian HH, the infinite-time temporal ensemble is equal to the random phase ensemble:

limτTemp.(τ)=Rand. Phase,\lim_{\tau\rightarrow\infty}\mathcal{E}_{\text{Temp.}}(\tau)=\mathcal{E}_{\text{Rand.~Phase}}~, (29)

with fixed energy populations |cE|2=|Ψ0|E|2|c_{E}|^{2}=\absolutevalue{\innerproduct{\Psi_{0}}{E}}^{2}, if HH satisfies all kk-th no-resonance conditions modulo degeneracies. The converse holds as long as |Ψ0|\Psi_{0}\rangle has non-zero population on the eigenspace of every non-degenerate eigenvalue EE, i.e. for almost every |Ψ0|\Psi_{0}\rangle666If there is a symmetry and |Ψ0|\Psi_{0}\rangle is supported only on every eigenstate in a symmetry sector, we may conclude that the kk-th no-resonance conditions modulo degeneracies hold for the eigenvalues in that sector..

Proof sketch. Here we sketch the idea of the proof, deferring its details to Appendix A. The general idea is to evaluate the kk-th moment ρtemp.(k)\rho_{\text{temp.}}^{(k)} of the temporal ensemble, and show that it matches the moments of the random phase ensemble. The evaluation of ρtemp.(k)\rho_{\text{temp.}}^{(k)} is possible for an infinite interval τ=\tau=\infty based on the no-resonance condition Eq. (28), and the corresponding moments of the random phase ensemble can also be explicitly evaluated. We note that having matching moments does not always imply that two ensembles are equal to one another. Under certain conditions such as the (complex) Carleman condition [99], this equivalence can be made. The evaluated moments satisfy the complex Carleman condition, completing our proof. The converse direction can be easily shown by noting that if the kk-th no-resonance condition is not satisfied, the moment ρtemp.(k)\rho_{\text{temp.}}^{(k)} will have additional terms that are not present in the random phase ensemble (as long as all coefficients cEc_{E} are non-zero) and hence the temporal and random phase ensembles will differ. ∎

Ensemble entropy — The ensemble entropy is always divergent in any distribution of states with fixed magnitudes, since the constrained space is a sub-dimensional manifold of the full Hilbert space. This can be intuitively understood from the following property of the KL divergence 𝒟(Haar)\mathcal{D}(\mathcal{E}\|\text{Haar}). Assume we are given samples from \mathcal{E} and promised that the underlying ensemble is either \mathcal{E} or the Haar ensemble. The KL divergence quantifies how many samples one needs to correctly conclude that the ensemble is \mathcal{E} and not the Haar ensemble [3]. In our case, we can rule out the Haar ensemble even with a single wavefunction |Ψ(t)|\Psi(t)\rangle from the temporal ensemble. This is because the state |Ψ(t)|\Psi(t)\rangle has the fixed amplitudes |E|Ψ(t)|=|cE||\langle E|\Psi(t)\rangle|=|c_{E}|, which occurs with probability zero in the Haar ensemble. Therefore, given complete knowledge of the state |Ψ(t)|\Psi(t)\rangle, the constraints {|cE|}\{|c_{E}|\}, and the promise that the underlying ensemble is either \mathcal{E} or the Haar ensemble, this discrimination task can be done infinitely quickly, and hence the KL divergence is infinitely large. While artificial in practice, this operational meaning for the KL divergence accounts for its divergence in the temporal ensemble.

However, the maximum entropy principle still holds in a certain sense: we can separate the entropy into contributions from the phase and magnitude degrees of freedom; the contribution from the phase is maximized by independent, uniformly distributed phases (Appendix C). We find a constant, non-divergent contribution

Ent(Temp.)\displaystyle\text{Ent}(\mathcal{E}_{\text{Temp.}}) 𝒟(Temp. Haar)\displaystyle\equiv-\mathcal{D}(\mathcal{E}_{\text{Temp.}}\|\text{ Haar}) (30)
=Elog(D|cE|2)+limδ0D(log(δ/2)1).\displaystyle=\sum_{E}\log(D\absolutevalue{c_{E}}^{2})+\lim_{\delta\rightarrow 0}D\left(\log(\delta/2)-1\right)~.

We interpret this expression as proportional to the volume E|cE|\prod_{E}|c_{E}| of the constrained space. Remarkably, this expression will reappear in the ensemble entropy of the projected ensemble (Section V.7).

IV.1 Asymptotic product form

In this work, we shall be interested in the moments of the random phase ensemble. Here, we provide a simplified, approximate form of the kk-th moments, which enables their further analysis.

Theorem 2 (Asymptotic product form).

In the limit of small purity tr(ρd2)\tr(\rho_{d}^{2}), the random phase ensemble has the product form:

ρTemp.(k)=ρdkσSkPerm(σ)+O(tr(ρd2))\displaystyle\rho^{(k)}_{\text{Temp.}}=\rho_{d}^{\otimes k}\sum_{\sigma\in S_{k}}\text{Perm}(\sigma)+O(\tr(\rho_{d}^{2})) (31)

Here, ρdE|cE|2|EE|\rho_{d}\equiv\sum_{E}\absolutevalue{c_{E}}^{2}\outerproduct{E}{E} is the diagonal ensemble, a common construction in the thermalization literature [51, 52], and its purity tr(ρd2)\tr(\rho_{d}^{2}) is typically exponentially small in total system size.

Proof sketch: We provide the full proof of Theorem 2 in Appendix D, and only state the key idea, which is the following: the kk-th moment of the temporal ensemble can be written as a sum over all lists of labels of energy eigenstates {αj}\{\alpha_{j}\} and over all permutations of {αj}\{\alpha_{j}\} (Theorem 1). Crucially, {αj}\{\alpha_{j}\} is allowed to have repeated elements. Only if no elements are repeated will the number of unique permutations of {αj}\{\alpha_{j}\} be k!k!. Using the same counting also for lists with repeated elements, the expression for the moments considerably simplifies.

ρTemp.(k)\displaystyle\rho^{(k)}_{\text{Temp.}} σSk{αj}j=1kcαjcσ(αj)|EαjEσ(αj)|\displaystyle\approx\sum_{\sigma\in S_{k}}\sum_{\{\alpha_{j}\}}\bigotimes_{j=1}^{k}c_{\alpha_{j}}c^{*}_{\sigma(\alpha_{j})}\outerproduct{E_{\alpha_{j}}}{E_{\sigma(\alpha_{j})}} (32)
=ρdkσSkPerm(σ)\displaystyle=\rho_{d}^{\otimes k}\sum_{\sigma\in S_{k}}\text{Perm}(\sigma) (33)

Eq. (32) is not exact because it over-counts permutations when {αj}\{\alpha_{j}\} contains repeated elements: there are additional correction terms to compensate for this over-counting. Careful consideration gives that the trace norm of all correction terms can be bounded by O(tr(ρd2))O(\tr(\rho_{d}^{2})), with a kk-dependent prefactor (Appendix D). ∎

Theorem 2 explicitly demonstrates how the higher order moments of the temporal ensemble depend on the first moment ρd\rho_{d}. Specifically, the moments of the temporal ensemble have a structural resemblance to the moments of the Haar ensemble [Eq. (16)], up to a correction which is typically exponentially small, formalizing our intuition of maximally entropic ensembles as “distorted” Haar ensembles. This asymptotic behavior will also be common to projected ensembles.

IV.2 Finite-time temporal ensembles

Theorem 1 studies the properties of temporal ensembles over infinite time intervals. Here, we investigate how temporal ensembles over finite time intervals converge to their infinite-interval values. We analytically study its convergence rate in random-matrix models and present numerical evidence that the above convergence rate also holds for realistic Hamiltonians. Specifically, we study the convergence of the finite time kk-th moment ρTemp.(k)(τ)\rho^{(k)}_{\text{Temp.}}(\tau) to its infinite-time limit ρTemp.(k)ρTemp.(k)(τ=)\rho^{(k)}_{\text{Temp.}}\equiv\rho^{(k)}_{\text{Temp.}}(\tau=\infty). This has expression:

ρTemp.(k)(τ)1ττ/2τ/2𝑑t|Ψ(t)Ψ(t)|k\displaystyle\rho^{(k)}_{\text{Temp.}}(\tau)\equiv\frac{1}{\tau}\int_{-\tau/2}^{\tau/2}dt\outerproduct{\Psi(t)}{\Psi(t)}^{\otimes k} (34)
={αj},{βj}sinc(j(EαjEβj)τ2)[jcαjcβj|EαjEβj|].\displaystyle=\!\!\sum_{\{\alpha_{j}\},\{\beta_{j}\}}\!\!\text{sinc}\Big{(}\sum_{j}(E_{\alpha_{j}}-E_{\beta_{j}})\frac{\tau}{2}\Big{)}\Bigg{[}\bigotimes_{j}c_{\alpha_{j}}c_{\beta_{j}}^{*}\outerproduct*{E_{\alpha_{j}}}{E_{\beta_{j}}}\Bigg{]}.

We consider the convergence both in terms of the trace norm and Frobenius norm, with the latter being analytically easier to study. For the first moment, i.e. k=1k=1, we numerically observe that the distance in Frobenius and trace norms both decay as τ1\tau^{-1} (see App. F). This holds for a generic chaotic model, the mixed field Ising model (the main model we study in this work, Sec. IV.4) as well as for Hamiltonians sampled at random from the Gausian Unitary Ensemble (GUE). We attribute this asymptotic 1/τ1/\tau scaling to the 1/x1/x asymptotic behavior of the sinc(x)\text{sinc}(x) function in Eq. (34). In Appendix F, we support these numerical findings by analytical calculations in random matrix theory, showing that the average of the squared Frobenius norm decays as τ2\tau^{-2}. This suggests that the average Frobenius norm decays as τ1\tau^{-1}, consistent with our numerical observations.

Interestingly, for higher order moments k=2,3k=2,3, we observe different behavior. For the mixed field Ising model as well as individual samples from the GUE, the Frobenius distance decays at intermediate times as τ1/2\tau^{-1/2} before a crossover time, after which it decays as τ1\tau^{-1} [Fig. 3(b)]. This crossover happens in times which are exponentially long in system size, which we attribute to the presence of a minimum distance of gaps EiEj+EkElE_{i}-E_{j}+E_{k}-E_{l} (for k=2k=2) for each individual Hamiltonian. While such minimum distance is exponentially small in system size, it has a finite, nonzero value Δmin\Delta_{\min} for finite-sized systems. Only at times much longer than the inverse of this minimum gap does the asymptotic 1/(Δminτ)1/(\Delta_{\min}\tau) behavior become clear.

Notably, we find that for the first moment, k=1k=1, the GUE ensemble mean closely resembles the behavior of individual instances and decays as τ1\tau^{-1}. In contrast, for higher moments, k=2k=2, the behavior of individual samples and the sample mean differ. With an increasing number of samples, the crossover is less pronounced. We relate these findings to statistical properties of the eigenvalues of GUE Hamiltonians. It is well known that the eigenvalues exhibit level repulsion, i.e. the probability to find two distinct eigenvalues close to each other EiEj0E_{i}-E_{j}\approx 0 is vanishing. In contrast, as suggested by Ref. [98] and the analytical and numerical results presented in Appendix F, we find that the gaps EiEj+EkElE_{i}-E_{j}+E_{k}-E_{l} of eigenvalues exhibit only weak repulsion, leading to the crossover timescale growing with the size of the sample.

Refer to caption
Figure 3: Temporal ensembles from natural dynamics. (a) The temporal ensemble is the collection of states formed by time-evolving a given state |Ψ\ket{\Psi} by a Hamiltonian HH, for various times tt. (b) Temporal ensembles formed over a finite interval of width τ\tau approach the infinite-time ensemble as τ\tau increases: the distance (here, the Frobenius norm) between their kk-th moment decreases as 1/τ1/21/\tau^{1/2}, before decreasing as τ1\tau^{-1}. Data shown here for k=2k=2, for a state evolving under the mixed field Ising model (MFIM) in red, and a Gaussian Unitary Ensemble (GUE)-random Hamiltonian in blue. (c) Signatures of the temporal ensemble can be measured through the fluctuations of quantities such as the outcome probability p(z,t)|z|Ψ(t)|2p(z,t)\equiv\absolutevalue{\innerproduct{z}{\Psi(t)}}^{2} (normalized by a factor of DD). Their histogram of values over time reveals a universal, Porter-Thomas (exponential) distribution. Histograms and time traces for two values of zz shown in red and blue. (d) The mutual information I(Z;T)I(Z;T) as a function of the time interval τ\tau of the ensemble. I(Z;T)I(Z;T) quickly saturates to the theoretical value of (1γ)/ln2(1-\gamma)/\ln 2, agreeing with the predicted (1γπ/(2σHτ))/ln2(1-\gamma-\sqrt{\pi}/(2\sigma_{H}\tau))/\ln 2 [inset, dashed, Eq. (44)]. In (c,d), data is obtained from quench evolution in the MFIM, for the initial state in Eq. (37) with θ=0.6\theta=0.6, and the units of time are determined by the O(1)O(1) Hamiltonian parameters of Eq. (36).

IV.3 Porter-Thomas distribution as measurable signature of temporal ensembles

An immediate consequence of Theorem 2 is the emergence of the Porter-Thomas distribution (Sec. III.5) in the probabilities of measuring a given outcome zz, as a function of time. This result was originally presented in Ref. [94]. For clarity, we denote this distribution as pz(t)p(z,t)z|Ψ(t)Ψ(t)|zp_{z}(t)\equiv p(z,t)\equiv\innerproduct{z}{\Psi(t)}\!\innerproduct{\Psi(t)}{z} 777The subscript zz indicates that the measurement outcome zz is being fixed and the distribution pz(t)p_{z}(t) is taken over time tt., which can be experimentally estimated by repeated measurements of the state |Ψ(t)\ket{\Psi(t)}.

Unlike local observables, whose expectation values quickly equilibrate in time— O(t)\langle O(t)\rangle has exponentially small relative fluctuations about its equilibrium value 𝔼t[O(t)]\mathbb{E}_{t}[\langle O(t)\rangle][17]— global observables such as the projector |zz||z\rangle\langle z| never equilibrate. Theorem 3 indicates that they have O(1)O(1) relative fluctuations which follow a universal, exponential distribution. We utilized this result in Ref. [94] to propose a protocol for many-body benchmarking in analog quantum simulators. Here, we will use the PT distribution as a signature of our maximum entropy principle; it will be a common signature among the temporal ensemble, projected ensemble and beyond.

Theorem 3.

Given an initial state |Ψ0\ket{\Psi_{0}}, a Hamiltonian HH which satisfies the kk-th no-resonance condition, and a fixed state |z\ket{z}, the outcome distribution pz(t)p_{z}(t) follows an approximate Porter-Thomas (or, exponential) distribution over time, with mean pavg(z)𝔼t[pz(t)]p_{\text{avg}}(z)\equiv\mathbb{E}_{t}[p_{z}(t)]. Specifically, the kk-th moments satisfy:

𝔼t[pz(t)k]\displaystyle\mathbb{E}_{t}[p_{z}(t)^{k}] =pavg(z)k[k!+O(Dβ1)]\displaystyle=p_{\text{avg}}(z)^{k}\left[k!+O(D_{\beta}^{-1})\right] (35)

where Dβ1z,E|z|E|4|cE|4/pavg(z)D_{\beta}^{-1}\equiv\sum_{z,E}\absolutevalue{\innerproduct{z}{E}}^{4}|c_{E}|^{4}/p_{\text{avg}}(z) is the inverse effective dimension, which captures the effective size of the Hilbert space explored during quench evolution that is accessible by measurements in {|z}\{\ket{z}\}. Our notation β\beta associates DβD_{\beta} with the temperature of the initial state, which to a large extent controls the size of the Hilbert space explored.

Proof.

Its kk-th moment can be easily obtained from the kk-th moment of the temporal ensemble, which gives the desired 𝔼t[pz(t)k]z|kρTemp.(k)|zkk!pavg(z)k\mathbb{E}_{t}[p_{z}(t)^{k}]\equiv\bra{z}^{\otimes k}\rho^{(k)}_{\text{Temp.}}\ket{z}^{\otimes k}\approx k!~p_{\text{avg}}(z)^{k}. This relation follows from Theorem 2. We conclude that pz(t)p_{z}(t) follows a Porter-Thomas distribution888More accurately, we bound the distance between the kk-th moments of pz(t)p_{z}(t) and the Porter-Thomas distribution. We do not currently have a means to directly bound the distance between distributions, such as the total variation distance., with a mean value equal to pavg(z)𝔼t[pz(t)]=z|ρd|zp_{\text{avg}}(z)\equiv\mathbb{E}_{t}[p_{z}(t)]=\langle z|\rho_{d}|z\rangle. This is a simple illustration of how the first moment pavg(z)p_{\text{avg}}(z) uniquely determines all moments of the distribution pz(t)p_{z}(t). This relation is only approximate because of the correction terms in Theorem 2. We are able to (loosely) bound the effects of these correction terms on the kk-th moments of p(z,t)p(z,t) in terms of the parameter Dβ1D_{\beta}^{-1}. This is not a straightforward application of the bound provided in Theorem 2, which is too weak for our purposes. We refer the reader to Ref. [94] for a detailed proof, or to Appendix I for a very similar proof. ∎

IV.4 Numerical model

We verify our theory predictions via explicit exact numerical simulations of interacting quantum many-body systems. As an illustration, we work with a paradigmatic model of many-body quantum chaos, the one-dimensional mixed field Ising model (MFIM) with open boundary conditions.

HMFIM=j=1NhxXj+hyYj+j=1N1JXjXj+1,H_{\text{MFIM}}=\sum_{j=1}^{N}h_{x}X_{j}+h_{y}Y_{j}+\sum_{j=1}^{N-1}JX_{j}X_{j+1}~, (36)

where XjX_{j} and YjY_{j} are Pauli matrices on site jj. We use parameter values established in Ref. [100]: (hx,hy,J)=(0.890,0.9045,1)(h_{x},h_{y},J)=(0.890,0.9045,1). Unless otherwise stated, we use this model in this and future sections for our numerical simulations.

In both the projected and temporal ensemble, we study the time-evolution of the initial state

|Ψ0=[exp(iθX/2)|0]N,\ket*{\Psi_{0}}=\left[\text{exp}(i\theta X/2)\ket{0}\right]^{\otimes N}~, (37)

where the parameter θ\theta is used to tune the state from infinite temperature (θ=0\theta=0) to positive or negative temperature. Unless otherwise stated, we present data for θ=0.6\theta=0.6, which has energy density E/N0.51E/N\approx 0.51, far from the center of the spectrum.

Here, we illustrate the content of Theorem 3 in Fig. 3(c), where we plot the probabilities pz(t)p_{z}(t) for two choices of zz, which exhibit large fluctuations over time, about different average values pavg(z)p_{\text{avg}}(z). The histograms of {pz(t)}\{p_{z}(t)\} show good agreement with the PT distributions pz(t)PT[pavg(z)]p_{z}(t)\sim\text{PT}[p_{\text{avg}}(z)] even over a relatively short time interval, confirming our analysis.

IV.5 Ergodicity

Theorem 3 is a statement about the distribution of p(z,t)p(z,t) over time (denoted pz(t)p_{z}(t) to make explicit that the variable zz is held constant). We can make an analogous statement about a distribution over zz, holding tt constant. At a fixed, late time tt, the quantities p~t(z)p(z,t)/pavg(z)\tilde{p}_{t}(z)\equiv p(z,t)/p_{\text{avg}}(z) are approximately distributed according to the PT distribution with mean 1.

p~t(z)PT(1),\tilde{p}_{t}(z)\sim\text{PT}(1)~, (38)

This was originally derived in Ref. [94] by showing that the weighted kk-th moment m(k)(t)zpavg(z)p~t(z)km^{(k)}(t)\equiv\sum_{z}p_{\text{avg}}(z)\tilde{p}_{t}(z)^{k} for a typical, late time tt satisfies:

m(k)(t)=k!+O(Dβ1/2).m^{(k)}(t)=k!+O(D_{\beta}^{-1/2})~. (39)

We interpret this equivalence of distributions over time and over outcomes zz as a kind of “ergodicity,” in the sense of the ergodic theorem in the study of classical dynamical systems [75], which states that in an ergodic system, the average over time is equal to the “spatial” average over configurations. Such an equivalence will be essential to our characterization of projected ensembles in Section VI.

IV.6 Mutual information: building a (bad) “clock”

The temporal ensemble has unique information theoretic properties. In particular, we find near-universal behavior in the mutual information between the random variables ZZ and TT which respectively represent the measurement outcomes zz and evolution times tt.

To illustrate the operational meaning of the mutual information, let us consider the following task. We are given several copies of the time-evolved state |Ψ(t)\ket{\Psi(t)}. We have knowledge of the initial state of the system as well as its Hamiltonian, but not the duration of the time evolution tt, which is assumed to be uniformly distributed on some large interval t[t0,t0+τ]t\in[t_{0},t_{0}+\tau]. Given multiple single-copy measurements performed in an optimal basis, how much information about tt can we learn?

Before performing detailed analysis, we can make a prediction based on our maximum entropy principle: since a temporal ensemble is maximally entropic, we expect that the amount of information must be small no matter which basis the measurements are performed. Indeed, our results imply that the (near) optimality is achieved by any generic measurement basis including the conventional configuration basis |zz|\outerproduct{z}{z}. Furthermore the information gain per measurement takes the universal value (1γ)/ln20.6099(1-\gamma)/\ln 2\approx 0.6099, independent of the Hamiltonian, initial state, measurement basis, and system size.

In order to formally define the mutual information I(Z;T)I(Z;T), we must make the following definitions. We let time be a uniformly distributed variable over a interval of (long but arbitrary) length τ\tau. We then define a joint probability distribution over measurement outcomes zz and times tt, by P(z,t)=pz(t)/τP(z,t)=p_{z}(t)/\tau, which is normalized: τ/2τ/2𝑑tzP(z,t)=1\int_{-\tau/2}^{\tau/2}dt\sum_{z}P(z,t)=1. We shall also need the marginal distributions P(t)zP(z,t)=1/τP(t)\equiv\sum_{z}P(z,t)=1/\tau, P(z)𝑑tP(z,t)=pavg(z)P(z)\equiv\int dtP(z,t)=p_{\text{avg}}(z) and the conditional distribution P(t|z)=P(t,z)/P(z)=p~z(t)/τP(t|z)=P(t,z)/P(z)=\tilde{p}_{z}(t)/\tau.

Then, the mutual information can be computed as

I(Z;T)=H(T)H(T|Z),\displaystyle I(Z;T)=H(T)-H(T|Z)~, (40)
=𝑑tP(t)log2(P(t))\displaystyle=-\int dtP(t)\log_{2}(P(t))
+zP(z)τ/2τ/2𝑑tP(t|z)log2P(t|z),\displaystyle~~~~+\sum_{z}P(z)\int_{-\tau/2}^{\tau/2}dtP(t|z)\log_{2}P(t|z)~, (41)
zpavg(z)0𝑑p~Pr(p~)p~log2p~+O(Dβ1)\displaystyle\approx\sum_{z}p_{\text{avg}}(z)\int_{0}^{\infty}d\tilde{p}~\text{Pr}(\tilde{p})\tilde{p}\log_{2}\tilde{p}+O(D_{\beta}^{-1}) (42)
(1γ)/ln2.\displaystyle\approx(1-\gamma)/\ln 2~. (43)

where in Eq. (42) we have used Theorem 3 to replace the integral over time with an integral over p~\tilde{p}, which follows a PT distribution, i.e. with probability density Pr(p~)dp~=exp(p~)dp~\text{Pr}(\tilde{p})d\tilde{p}=\exp(-\tilde{p})d\tilde{p}.

It is remarkable that this universal value does not depend on system sizes, and, in particular, does not vanish with any parameters. This implies that, despite the fact that ergodic dynamics hides information about time tt as strongly as possible, a finite amount (0.6099 bits) of information still cannot be concealed, in the large τ\tau limit. In other words, our result implies that a certain level of temporal fluctuations in observables is inevitable in the ergodic dynamics of generic pure states, putting the lower bound on the ability of unitary dynamics to hide temporal information. We also note that estimating the evolution duration tt is equivalent to determining an overall strength scaling factor of a many-body Hamiltonian by evolving to a fixed time tt, and hence this “clock” may equivalently be regarded as a sensor for the overall Hamiltonian strength.

IV.6.1 Finite-time mutual information

We note that the mutual information discussed above should be distinguished from more conventional metrics for clocks such as sensitivity. The sensitivity at very small intervals Δt\Delta t (i.e. in the limit of many measurements) depends instead on the energy uncertainty σH(ψ0|H2|ψ0ψ0|H|ψ02)1/2\sigma_{H}\equiv\left(\langle\psi_{0}|H^{2}|\psi_{0}\rangle-\langle\psi_{0}|H|\psi_{0}\rangle^{2}\right)^{1/2}, which is not universal. The mutual information instead describes how quickly a large interval τ\tau can be refined by measurements.

Our universal result above is valid in the limit of large time intervals τ\tau, over which p(z,t)p(z,t) follows the approximate PT distribution. At shorter intervals (but still at sufficiently late times tt), the mutual information I(Z;T)I(Z;T) takes on a smaller value, due to p(z,t)p(z,t) having a non-zero correlation time. This is illustrated in Fig. 3(d). In fact, I(Z;T)I(Z;T) approaches its late-time value as a power law:

I(Z;T)(1γπ2σHτ)/ln2,I(Z;T)\approx\left(1-\gamma-\frac{\sqrt{\pi}}{2\sigma_{H}\tau}\right)/\ln 2~, (44)

where σH=(Ψ0|H2|Ψ0Ψ0|H|Ψ02)1/2\sigma_{H}=(\bra{\Psi_{0}}H^{2}\ket{\Psi_{0}}-\bra{\Psi_{0}}H\ket{\Psi_{0}}^{2})^{1/2} is the uncertainty in energy of the initial state. This is illustrated in the inset of Fig. 3(d).

We present the full derivation in Appendix G. In brief, we obtain this result by estimating the autocorrelation time of p(z,t)p(z,t), using approximations of the many-body spectrum. We find that the correlation time is inversely proportional to the energy uncertainty of the state σH\sigma_{H}. If we interrogate p(z,t)p(z,t) when t is an integer multiple of the correlation time, we get an effective number σHτ\sigma_{H}\tau of points in times for which p(z,t)p(z,t) is distinct. This gives a mutual information that approaches the value 1γ1-\gamma at the rate (σHτ)1(\sigma_{H}\tau)^{-1}. The coefficient π/2\sqrt{\pi}/2 follows from a detailed calculation. This result complements Section IV.2, providing a convergence timescale of finite-time temporal ensembles from the perspective of the mutual information. We leave to future work the analysis of the ultimate sensitivity at very small intervals τπ2σH(1γ)\tau\ll\frac{\sqrt{\pi}}{2\sigma_{H}(1-\gamma)}, at which point Eq. (44) breaks down.

Our results in this section describe universal, statistical features of the trajectory of a many-body state under chaotic many-body dynamics. They confirm an intuitive, folklore heuristic that chaotic many-body dynamics resembles random unitary dynamics. Here, we find signatures of Haar random states such as the PT distribution in the time-evolved many-body states. We attribute this quasi-randomness to our maximum-entropy principle. This translates to our finding that these temporal trajectories can be used for certain protocols such as benchmarking and sensing. Furthermore, their performance in these protocols is universal and independent of details such as the Hamiltonian, initial state, and measurement basis.

V Projected Ensembles

Refer to caption
Figure 4: Projected ensembles from natural dynamics. (a) The projected ensemble is formed by performing projective measurements on a large subsystem BB of a many-body state, here obtained from time-evolution of an initial state. (b) In this setting, the projected ensemble is described by the Scrooge ensemble, which is a unique probability distribution of states (color gradient) for each value of the first moment (red cross, in our setting equal to the reduced density matrix). (c) At constant energy density E/N0.5E/N\approx 0.5, the trace distance between the second moments of the projected and Scrooge ensembles decreases with quench evolution time before saturating to an equilibrium value which depends on system size. We time-evolve a product state with the mixed-field Ising model (MFIM) on a one-dimensional open chain of length NN [Eq. (36)], from N=12N=12 (lighter) to N=22N=22 (darker) and fix the subsystem AA to be four sites near the center of the chain. (d) The equilibrium trace distance decreases exponentially with system size for both second (blue) and third moment (green). (e) A signature of the Scrooge ensemble is the emergence of the Porter-Thomas distribution in measurement outcome distributions; here we plot a histogram of the normalized outcome distributions DpzA(zB)Dp_{z_{A}}(z_{B}), with fixed zAz_{A} and different values of zBz_{B}. Histograms for three different values of zAz_{A} (blue, red, purple) follow PT distributions with different means D𝔼zB[pzA(zB)]D\mathbb{E}_{z_{B}}[p_{z_{A}}(z_{B})]. (f) The mutual information I(OA;ZB|T)I(O_{A};Z_{B}|T) (conditioned at fixed times) is an information-theoretic signature of the Scrooge ensemble: its value is independent of the basis OAO_{A}, and is equal to the subentropy Q[ρA]Q[\rho_{A}] [Eq. (61)] (black dashed). Here we plot I(OA;ZB|T)I(O_{A};Z_{B}|T) for three different observables OA={X4,Y4,Z4}O_{A}=\{X^{\otimes 4},Y^{\otimes 4},Z^{\otimes 4}\} (red, green and blue respectively), against the energy density E/NE/N of the state, controlled by the polarization angle θ\theta of the initial state [Eq. (37)].

Next, we turn to the projected ensemble, which is the second setting of our maximum entropy principle. When a many-body state is bipartitioned into a small system AA and its complement BB, the large subsystem BB acts like a “thermal bath” to the smaller subsystem AA. This observation is instrumental to the emergence of thermal equilibration in a closed quantum system. Nevertheless, the extent to which subsystem BB truly acts as a bath is not fully characterized, owing to the complexity of the overall quantum many-body state. Projected ensembles offer a lens into the nature of correlations between subsystem and bath. Without the information of the bath, the subsystem is described by the reduced density matrix ρA=trB(|ΨΨ|)\rho_{A}=\tr_{B}(\outerproduct{\Psi}{\Psi}). In order to go further, it is necessary to obtain information about the bath. In projected ensembles, this is done by projecting the bath onto definite states |zB|z_{B}\rangle, typically by performing measurements in a product basis. This in turn projects |Ψ\ket{\Psi} into a pure state |Ψ(zB)A\ket*{\Psi(z_{B})}\in\mathcal{H}_{A}. Taken over all measurement outcomes, a single wavefunction |Ψ\ket{\Psi} generates a large ensemble of states {|Ψ(zB)}\{\ket{\Psi(z_{B})}\}. By studying the statistical properties of this ensemble of states, as well as any correlations between |Ψ(zB)\ket{\Psi(z_{B})} and zBz_{B}, the projected ensemble construction provides a new way to study system-bath correlations.

Our maximum entropy principle states the following: there are correlations between |Ψ(zB)\ket{\Psi(z_{B})} and zBz_{B} which are “thermal” in the sense that they arise from (and only from) energy conservation. Furthermore, there are random fluctuations over these thermal correlations. In Section VI, we find that these fluctuations are universal and follow a maximum entropy principle.

In this section, we study a special case. In certain measurement bases, the correlations between |Ψ(zB)\ket*{\Psi(z_{B})} and zBz_{B} vanish. Nevertheless, effects of energy conservation remain: a mixture of all projected states equals ρA\rho_{A}, which is, up to subleading corrections, a thermal Gibbs state with temperature set by the bath [69]. Therefore, while the projected states |Ψ(zB)\ket*{\Psi(z_{B})} are effectively identical independent random samples from a distribution of states, the underlying distribution is not Haar-random, but has first moment ρA\rho_{A}. In this section, we find that this distribution has maximum entropy under this first moment constraint, a distribution that has been previously studied under the name “Scrooge ensemble” [79, 80, 76].

Explicitly, the projected ensemble is defined as the set of states obtained by projectively measuring a pure state |Ψ\ket{\Psi} on a large subsystem BB [Fig. 4(a)]. Specifically, the subsystem BB is measured, giving a random outcome zBz_{B}, which occurs with probability p(zB)p(z_{B}):

p(zB)Ψ|(𝕀A|zBzB|)|Ψ.p(z_{B})\equiv\bra{\Psi}\left(\mathbb{I}_{A}\otimes\outerproduct{z_{B}}{z_{B}}\right)\ket{\Psi}~. (45)

This measurement outcome projects the initial state |Ψ\ket{\Psi} onto a state |Ψ(zB)\ket{\Psi(z_{B})} in the Hilbert space of the remaining subsystem, denoted AA:

|Ψ(zB)(𝕀AzB|)|Ψ/p(zB)A\ket{\Psi(z_{B})}\equiv\left(\mathbb{I}_{A}\otimes\bra{z_{B}}\right)\ket{\Psi}/\sqrt{p(z_{B})}\in\mathcal{H}_{A} (46)

The projected ensemble is the set of states

Proj.={p(zB),|Ψ(zB)}\mathcal{E}_{\text{Proj.}}=\{p(z_{B}),\ket{\Psi(z_{B})}\} (47)

While this object was first introduced in different contexts, e.g. in Ref. [76], the projected ensemble was later studied by some of the authors in Refs. [28, 29] as a means of quantifying the notion of randomness in a single many-body state (as opposed to ensembles of them), by studying its projected ensemble.

V.1 Results at infinite temperature: approximate kk-designs

In Ref. [29], it was found that a wide class of many-body states induce projected ensembles that are statistically close to the Haar ensemble. This closeness was measured by the trace distance ΔHaar(k)\Delta^{(k)}_{\text{Haar}} between the kk-th moments of the projected ensemble and the Haar ensemble.

Specifically, it was proven that the projected ensembles are approximate kk-designs when the global states are typical states drawn from the Haar ensemble or from a state kk^{\prime}-design (with k>kk^{\prime}>k). This was also numerically demonstrated in natural many-body states, obtained from either (a) time-evolution of an initial state or (b) the eigenstates of a chaotic many-body Hamiltonian, close to infinite temperature, defined as states |Ψ0\ket{\Psi_{0}} that satisfy Ψ0|H|Ψ0=tr(H)/D\bra{\Psi_{0}}H\ket{\Psi_{0}}=\tr(H)/D. Subsequent work has rigorously established the emergence of the projected kk-design in specific settings, such as in dual-unitary models [77, 30, 31, 32, 78], free fermion models [33], or under assumptions of the reduced density matrix [34]. Further work has studied the effects of symmetry [35, 37], the effect of quantum magic in the global state [38] on the projected ensemble, and deep thermalization timescales [30, 31, 36].

In the above studies, it was important that the initial state |Ψ0\ket{\Psi_{0}} be at infinite temperature. This is because the first moment of the projected ensemble is the reduced density matrix ρAtrB[|ΨΨ|]\rho_{A}\equiv\tr_{B}[\outerproduct{\Psi}{\Psi}]. For thermal states, ρA\rho_{A} is close to a Gibbs state [69]; only at infinite temperature is this equal to the first moment of the Haar ensemble: the maximally mixed state 𝕀/DA\mathbb{I}/D_{A}. Therefore, away from infinite temperature, the projected ensembles of such states cannot equal the Haar ensemble. It has remained an open question as to which ensemble describes the finite-temperature projected ensemble.

V.2 Projected ensembles at finite temperature: Scrooge ensemble

Our main result of this section is that under appropriate conditions, the projected ensemble of a finite temperature state is described, by the Scrooge ensemble [79, 80, 76]. Introduced in Ref. [79], any density matrix ρ\rho has a corresponding Scrooge ensemble, which we denote Scrooge[ρ]\text{Scrooge}[\rho] [Figure 4(b)]. This is most easily understood as a “ρ\rho-distortion” of the Haar ensemble, defined by the probability distribution. The state |Ψ\ket{\Psi} has a higher probability when it has a larger overlap with the principal axes of ρ\rho. We can sample from the Scrooge ensemble by “distorting” wavefunctions sampled from the Haar ensemble, as described in Section II. We can also describe the Scrooge ensemble in terms of its distribution function [80].

PScr(Ψ)dΨ\displaystyle P_{\text{Scr}}(\Psi)d\Psi
=D!2πDdetρΨ|ρ1|ΨD1PHaar(Ψ)dΨ\displaystyle=\frac{D!}{2~\pi^{D}\det\rho}\bra{\Psi}\rho^{-1}\ket{\Psi}^{-D-1}P_{\text{Haar}}(\Psi)d\Psi (48)
=D!2πDm=1Dλm(m=1D|Ψm|2λm)D1PHaar(Ψ)dΨ,\displaystyle=\frac{D!}{2~\pi^{D}\prod_{m=1}^{D}\lambda_{m}}\left(\sum_{m=1}^{D}\frac{\absolutevalue{\Psi_{m}}^{2}}{\lambda_{m}}\right)^{-D-1}~P_{\text{Haar}}(\Psi)d\Psi~,

where {λm,|m}\{\lambda_{m},\ket{m}\} is the eigensystem of ρ\rho, and Ψmm|Ψ\Psi_{m}\equiv\innerproduct{m}{\Psi}. In the context of the projected ensemble, the first moment ρ\rho is the reduced density matrix ρA\rho_{A} of |Ψ\ket{\Psi}.

In the ideal limit in which the measurement basis is a Haar-random basis on BB, Ref. [76] proves that the resulting projected ensemble is the Scrooge ensemble.

However, this limit of Haar-random measurements is not realistic to our setting, in which the measurement basis is typically a complete basis of unentangled product states. This may have qualitatively different behavior from the highly entangled Haar-random basis states. Furthermore, in realistic systems, the measurement basis may be strongly correlated with the Hamiltonian realized in the system.

It will turn out that the first condition does not affect the behavior of the projected ensemble, but the second condition does. While we provide a precise definition in Section VI, “correlations between measurement basis and the Hamiltonian” may be taken to be the distribution of expectation values zB|HB|zB\bra{z_{B}}H_{B}\ket{z_{B}}, where HBH_{B} are the terms of the Hamiltonian HH restricted onto the large subsystem BB. We find that the projected states |Ψ(zB)\ket*{\Psi(z_{B})} are correlated with the energy zB|HB|zB\bra{z_{B}}H_{B}\ket{z_{B}}. If there is a large spread of values zB|HB|zB\bra{z_{B}}H_{B}\ket{z_{B}}, the projected states are distinguishable. Conversely, if zB|HB|zB\bra{z_{B}}H_{B}\ket{z_{B}} is independent of zBz_{B}, the projected states are independent of their label zBz_{B}, and the projected ensemble can be regarded as a collection of identical samples from the same distribution. In brief, we claim:

Claim 1.

The projected ensemble is statistically described by the Scrooge ensemble, if the measurement basis {|zB}\{\ket{z_{B}}\} is energy non-revealing.

When such correlations are not negligible, the projected ensemble is described by a generalized Scrooge ensemble, discussed in Section VI. There, we provide a proof of our general claim for the projected ensembles of time-evolved states, based on our results about the temporal ensemble in Section IV.

Our conventional understanding of many-body dynamics is that on sufficiently small subsystems, a many-body state will rapidly reach thermal equilibrium, i.e. its reduced density matrix will be close to that of a Gibbs state at an appropriate temperature [69]. As with the temporal ensemble, our results imply that not only is the first moment of the projected ensemble (the reduced density matrix) a universal function of the temperature, so are all higher statistical moments.

V.3 Moments of the Scrooge ensemble

We directly test our claim by comparing the kk-th moments of the projected ensemble to those of the Scrooge ensemble. In order to do so, here we provide explicit expressions for the kk-th moment of the Scrooge ensemble, which have previously not been available in the literature.

Our expression for the kk-th moment is in terms of its matrix elements in (kk-copies of) the eigenbasis {λm,|m}\{\lambda_{m},\ket{m}\} of ρ\rho. We find that the nonzero elements of the Scrooge kk-th moment ρScr.(k)\rho^{(k)}_{\text{Scr.}} are those which permute the kk copies. For a basis element |𝒎|m1,,mk\ket{\boldsymbol{m}}\equiv\ket{m_{1},\dots,m_{k}}, 𝒎|ρScr.(k)|𝒎\langle\boldsymbol{m}|\rho^{(k)}_{\text{Scr.}}|\boldsymbol{m^{\prime}}\rangle is non-zero only |𝒎\ket{\boldsymbol{m^{\prime}}} is a permutation of |𝒎\ket{\boldsymbol{m}}, i.e. |𝒎=|σ(𝒎),σSk\ket{\boldsymbol{m^{\prime}}}=\ket{\sigma(\boldsymbol{m})},~\sigma\in S_{k}. Furthermore, the matrix elements only depend on 𝒎\boldsymbol{m} and are independent of the permutation σ\sigma.

ρScr.(k)=𝒎𝒎=σ(𝒎)ρScr.,𝒎(k)|𝒎𝒎|.\rho^{(k)}_{\text{Scr.}}=\sum_{\boldsymbol{m}}\sum_{\boldsymbol{m^{\prime}}=\sigma(\boldsymbol{m})}\rho^{(k)}_{\text{Scr.},\boldsymbol{m}}\outerproduct{\boldsymbol{m}}{\boldsymbol{m^{\prime}}}~. (49)

This structure is similar to the kk-th moments of the Haar ensemble [Eq. (16)]. However, unlike in the Haar ensemble, the coefficients ρScr.,𝒎(k)\rho^{(k)}_{\text{Scr.},\boldsymbol{m}} are nontrivial and are given by

ρScr.,𝒎(k)=(i=1Dλi1)kΛkμm1μmk|μi=λi1,\rho^{(k)}_{\text{Scr.},\boldsymbol{m}}=\left(\prod_{i=1}^{D}\lambda^{-1}_{i}\right)\frac{\partial^{k}\Lambda_{k}}{\partial\mu_{m_{1}}\dots\partial\mu_{m_{k}}}{\Bigg{|}}_{\mu_{i}=\lambda_{i}^{-1}}~, (50)

where

Λk(μ1,,μD)\displaystyle\Lambda_{k}(\mu_{1},\dots,\mu_{D}) j=1Dμjk2lnμjij(μjμi).\displaystyle\equiv\sum_{j=1}^{D}\frac{\mu^{k-2}_{j}\ln\mu_{j}}{\prod_{i\neq j}(\mu_{j}-\mu_{i})}. (51)

We derive this result in Appendix H. While this expression appears complicated, it reduces to a simple form in the limit that ρ\rho is a sufficiently mixed state. We illustrate this for k=2k=2. When mλm2n1\sum_{m}\lambda_{m}^{2n}\ll 1 for every n>1n>1, Eq. (50) simplifies to the product form (c.f. Theorem 2)

ρScr.(2)ρ2(𝕀+𝕊),\displaystyle\rho^{(2)}_{\text{Scr.}}\approx\rho^{\otimes 2}(\mathbb{I}+\mathbb{S})~, (52)

where 𝕀\mathbb{I} and 𝕊\mathbb{S} are the identity and swap operators permuting two copies of the Hilbert space 2\mathcal{H}^{\otimes 2}.

V.4 Numerical evidence: trace distance between kk-th moments

We numerically support our claim by identifying the projected ensemble with the Scrooge ensemble. Detailed information about the model of our numerical simulations is provided in Section IV.4. In Figure 4(c,d), we compare the k=2,3k=2,3 moments of the Scrooge ensemble ρScr.(k)\rho^{(k)}_{\text{Scr.}} with those of the projected ensemble

ρProj.(k)zBp(zB)|Ψ(zB)Ψ(zB)|k,\rho^{(k)}_{\text{Proj.}}\equiv\sum_{z_{B}}p(z_{B})|\Psi(z_{B})\rangle\langle\Psi(z_{B})|^{\otimes k}~, (53)

obtained from a time-evolved state |Ψ(t)\ket{\Psi(t)} under the MFIM. As a function of quench evolution time tt, the trace distance Δ(k)\Delta^{(k)} decreases as a power law with tt [29], before saturating at a value which depends on system size, illustrated for k=2k=2 in Figure 4(c). This saturation value decreases exponentially with increasing system size for both k=2k=2 and k=3k=3 [Fig. 4(d)].

V.5 Unnormalized ensembles of states

It is mathematically more convenient to treat the projected ensemble as an ensemble of the unnormalized states |Ψ~(zB)(𝕀AzB|)|Ψ=p(zB)|Ψ(zB)\ket*{\tilde{\Psi}(z_{B})}\equiv(\mathbb{I}_{A}\otimes\bra{z_{B}})\ket{\Psi}=\sqrt{p(z_{B})}|\Psi(z_{B})\rangle, such that the projection operation |ΨΨ||Ψ~(zB)Ψ~(zB)|\outerproduct{\Psi}{\Psi}\rightarrow\outerproduct*{\tilde{\Psi}(z_{B})}{\tilde{\Psi}(z_{B})} acts linearly on the state |ΨΨ|\outerproduct{\Psi}{\Psi}. The kk-th moment of the ensemble of unnormalized states [Eq. (15), reproduced below], differs from that of the ensemble of normalized states [Eq. (14)].

ρ~(k)\displaystyle\tilde{\rho}^{(k)} 1DBzB|Ψ~(zB,t)Ψ~(zB,t)|k𝑑Ψ~|Ψ~Ψ~|k.\displaystyle\equiv\frac{1}{D_{B}}\sum_{z_{B}}\outerproduct*{\tilde{\Psi}(z_{B},t)}{\tilde{\Psi}(z_{B},t)}^{\otimes k}\equiv\int d\tilde{\Psi}\outerproduct*{\tilde{\Psi}}{\tilde{\Psi}}^{\otimes k}~. (54)

The primary reason for considering the unnormalized ensemble moments is that we will be able to derive analytical expressions for the corresponding moments of projected ensembles generated by time-evolved states.

In terms of unnormalized ensembles, our claim is that the projected ensemble is statistically described by the unnormalized-state analog of the Scrooge ensemble; which we refer to as the unnormalized Scrooge ensemble, and denote S~crooge[ρ]\tilde{\text{S}}\text{crooge}[\rho]. This is simply a distortion of the Gaussian ensemble of unnormalized states [Eq. (12), Ref. [80]]. In the eigenbasis of ρ\rho, the probability distribution function factorizes:

PS~cr.(Ψ~)dΨ~\displaystyle P_{\tilde{\text{S}}\text{cr.}}(\tilde{\Psi})d\tilde{\Psi} =m=1D1λmπexp[λm1|Ψ~m|2]d2Ψ~m.\displaystyle=\prod_{m=1}^{D}\frac{1}{\lambda_{m}\pi}\exp[-\lambda_{m}^{-1}|\tilde{\Psi}_{m}|^{2}]d^{2}\tilde{\Psi}_{m}~. (55)

Note that we can map the unnormalized S~crooge[ρ]\tilde{\text{S}}\text{crooge}[\rho] into the normalized ensemble Scrooge[ρ]\text{Scrooge}[\rho] by the procedure described in Sec. III.

Indeed, the kk-th moments of S~crooge[ρ]\tilde{\text{S}}\text{crooge}[\rho] are far simpler than those of Scrooge[ρ]\text{Scrooge}[\rho]. They are exactly of the product form:

ρ~S~cr.(k)=𝑑Ψ~exp[λj1|Ψ~j|2]𝒎,𝒎,𝒎=σ(𝒎)i=1k|Ψ~mi|2|𝒎𝒎|\displaystyle\tilde{\rho}^{(k)}_{\tilde{\text{S}}\text{cr.}}=\int d\tilde{\Psi}~\exp[-\lambda_{j}^{-1}|\tilde{\Psi}_{j}|^{2}]\sum_{\begin{subarray}{c}\boldsymbol{m},\boldsymbol{m^{\prime}},\\ \boldsymbol{m^{\prime}}=\sigma(\boldsymbol{m})\end{subarray}}\prod_{i=1}^{k}|\tilde{\Psi}_{m_{i}}|^{2}\outerproduct{\boldsymbol{m}}{\boldsymbol{m^{\prime}}}
=ρkσSkPerm(σ).\displaystyle=\rho^{\otimes k}\sum_{\sigma\in S_{k}}\text{Perm}(\sigma)~. (56)

V.6 Porter-Thomas distribution in outcome probabilities

The simple kk-th moments of the unnormalized Scrooge ensembles lend themselves to a simple, measurable signature of our claim. We consider the joint probabilities poA(zB)|oA|Ψ~(zB)|2=|oA,zB|Ψ|2p_{o_{A}}(z_{B})\equiv|\langle o_{A}|\tilde{\Psi}(z_{B})|^{2}=|\langle o_{A},z_{B}|\Psi\rangle|^{2}. Eq. (56) predicts that if we fix oAo_{A}, the distribution of probabilities poA(zB)p_{o_{A}}(z_{B}) over the bitstrings zBz_{B} will be a PT distribution poA(zB)PT(oA|ρA|oA)p_{o_{A}}(z_{B})\sim\text{PT}\left(\bra{o_{A}}\rho_{A}\ket{o_{A}}\right). In Fig. 4(e), AA is a subsystem of four qubits. We plot the histogram of poA(zB)p_{o_{A}}(z_{B}) for three different outcomes oA=0000,1010o_{A}=0000,1010 or 0010, all of which follow PT distributions.

In contrast to the joint probability poA(zB)p_{o_{A}}(z_{B}), the distribution of the conditional probabilities p(oA|zB)p(oA,zB)/p(zB)p(o_{A}|z_{B})\equiv p(o_{A},z_{B})/p(z_{B}) is non-universal. While the moments of the joint probability poA(zB)p_{o_{A}}(z_{B}) correspond to the moments of the unnormalized Scrooge ensemble, the moments of the conditional probability p(oA|zB)=|oA|Ψ(zB)|2p(o_{A}|z_{B})=|\langle o_{A}|\Psi(z_{B})\rangle|^{2} are given by the moments of the normalized Scrooge ensemble [Eq. (50)]. The predicted distribution of p(oA|zB)p(o_{A}|z_{B}) is non-universal and depends on the observable OAO_{A}, hence is more challenging to verify, illustrating the conceptual utility of ensembles of unnormalized states.

V.7 Maximally entropic ensembles of unnormalized states

That the Scrooge ensemble describes the projected ensemble is no coincidence: it is known that the unnormalized Scrooge ensemble has maximum entropy, subject to the constraint that its first moment is ρA\rho_{A} [101, 80] (see Ref. [53] for a discussion of maximum-entropy normalized ensembles). Showing this is a simple application of Lagrange multipliers, but we reproduce it here for completeness. We want to solve for the probability distribution PMaxEntP_{\text{MaxEnt}} which minimizes the relative entropy D(Gauss)D\left(\mathcal{E}\|\text{Gauss}\right), subject to the first moment constraint. This optimization may be independently solved in the eigenbasis of ρA\rho_{A}, i.e. PMaxEntP_{\text{MaxEnt}} factorizes in the eigenbasis. In each component, PMaxEntP_{\text{MaxEnt}} has exponential form: PMaxEnt(Ψ~j)/PGauss(Ψ~j)exp(cj|Ψ~j|2)P_{\text{MaxEnt}}(\tilde{\Psi}_{j})/P_{\text{Gauss}}(\tilde{\Psi}_{j})\propto\exp(-c_{j}|\tilde{\Psi}_{j}|^{2}). The unknown values cjc_{j} are solved by enforcing the first moment constraint. With ρA=jλj|jj|\rho_{A}=\sum_{j}\lambda_{j}\outerproduct{j}{j}, we have

λj\displaystyle\lambda_{j} =|Ψ~j|2PMaxEnt(Ψ~)𝑑Ψ~\displaystyle=\int|\tilde{\Psi}_{j}|^{2}P_{\text{MaxEnt}}(\tilde{\Psi})d\tilde{\Psi} (57)
=1Z[kjDD+ck]D(D+cj)2=1D+cj,\displaystyle=\frac{1}{Z}\left[\prod_{k\neq j}\frac{D}{D+c_{k}}\right]\frac{D}{(D+c_{j})^{2}}=\frac{1}{D+c_{j}}~, (58)

where in the second line we have performed the relevant Gaussian integrals, and we note that the normalization constant ZZ satisfies Z=kD/(D+ck)Z=\prod_{k}D/(D+c_{k}). It follows that cj=λj1Dc_{j}=\lambda_{j}^{-1}-D, precisely equal to the unnormalized Scrooge ensemble in Eq. (55). This derivation is fairly general and may apply to multiple settings, e.g., recent work in the context of maximum entropy states under charge conservation [62].

The corresponding maximized value of the ensemble entropy is

Ent()=𝒟(Gauss)\displaystyle\text{Ent}(\mathcal{E})=-\mathcal{D}(\mathcal{E}\|\text{Gauss}) (59)
=j𝑑Ψ~jPS~cr.(Ψ~j)[(λj1D)|Ψ~j|2+log(Dλj)]\displaystyle=\sum_{j}\int d\tilde{\Psi}_{j}P_{\tilde{\text{S}}\text{cr.}}(\tilde{\Psi}_{j})\left[\left(\lambda_{j}^{-1}-D\right)|\tilde{\Psi}_{j}|^{2}+\log(D\lambda_{j})\right]
=jlog(Dλj),\displaystyle=\sum_{j}\log(D\lambda_{j})~, (60)

identical to the (non-divergent part of the) ensemble entropy of the temporal ensemble [Eq. (30)], strongly suggesting a connection between the two ensembles.

V.8 Basis-independent mutual information

The Scrooge ensemble was originally studied for its information theoretic properties: it is the ensemble of states with minimal accessible information (Section III.4[79]. Specifically, Josza, Robb, and Wooters [79] lower bounded the accessible information of any ensemble of states with first moment ρ\rho by the subentropy Q(ρ)Q(\rho) [Eq. (61)], and found that the Scrooge ensemble achieves this lower bound. A non-trivial, testable consequence of their result is that the mutual information I(OA;ZB|T)I(O_{A};Z_{B}|T) is equal to the subentropy, for any complete basis OAO_{A}.

I(OA;ZB|T)Q(ρA).\displaystyle I(O_{A};Z_{B}|T)\approx Q(\rho_{A})~. (61)

Here, the conditional mutual information I(OA;ZB|T)I(O_{A};Z_{B}|T) refers the mutual information between the variables OAO_{A} and ZBZ_{B} of the probability distribution p(oA,zB,t)p(o_{A},z_{B},t) at fixed times tt, i.e. conditioning on the time TT. As a reminder, we denote variables with upper case letters and their values in lower case. Although time will not play a role here (as long as the evolution time is sufficiently long), we primarily adopt this notation anticipating our result in Sec. VI.3, where the dependence on time will be important.

The operational meaning of the mutual information can be understood by considering the following scenario. Suppose Alice and Bob share an entangled quantum many-body state obtained from ergodic Hamiltonian dynamics of duration tt. Bob performs a measurement on his (larger) portion of the state in the computational basis, resulting in a measurement outcome zBz_{B}. Alice performs the measurement of an arbitrary (potentially optimized) local observable leading to an outcome oAo_{A}. How much information do Alice and Bob share in this exercise? Or, equivalently, how much can Alice or Bob know about the measurement outcomes of the other from their own measurement outcomes? Our result implies that they know as little as possible, while some amount of information is inevitably shared due to energy conservation. Furthermore, this feature is universal, independent of the choice of Alice’s observable as long as Bob’s measurement basis realizes a Scrooge ensemble (we discuss the effects of measurement basis below).

We verify this property in Fig. 4(f). We choose three different measurement bases OA{X4,Y4,Z4}O_{A}\in\{X^{\otimes 4},Y^{\otimes 4},Z^{\otimes 4}\}, i.e. the XX, YY and ZZ bases on the subsystem AA. We verify that the mutual information I(OA;ZB|T)I(O_{A};Z_{B}|T) of a time-evolved state |Ψ(t)\ket{\Psi(t)} is independent of OAO_{A} and agrees with the subentropy Q(ρA)Q(\rho_{A}) across a wide range of energy densities of the initial state |Ψ0\ket{\Psi_{0}}, i.e. over a wide range of temperatures of the reduced density matrix ρA\rho_{A}.

Taken together, our results provide a surprising universality in the behavior of projected ensembles, even away from infinite temperature. When the basis ZBZ_{B} is uncorrelated with the projected states |Ψ(zB)\ket{\Psi(z_{B})}, they are statistically described by independent samples from the Scrooge ensemble |Ψ(zB)Scrooge[ρA]\ket{\Psi(z_{B})}\sim\text{Scrooge}[\rho_{A}]. We discussed measurable signatures of the Scrooge ensemble: statistics of the outcomes probabilities poA(zB)p_{o_{A}}(z_{B}), and universal behaviour of the conditional mutual information I(OA;ZB|T)I(O_{A};Z_{B}|T). However, we have not yet provided a derivation for the emergence of the Scrooge ensemble. Furthermore, our results in this section will not hold for generic measurement bases. In the following section, we resolve both issues by deriving a generalized Scrooge ensemble from the temporal ensemble.

VI From temporal to projected ensembles: The effect of measurement basis

As aforementioned, properties of the projected ensemble depend on the measurement basis on BB: our findings in Section V only hold for certain measurement bases. In this section, we seek to extend our description of projected ensembles to general bases. Instead of Scrooge ensembles, we will find that projected ensembles are described by a generalized Scrooge ensemble, which we introduce, that accounts for correlations between zBz_{B} and the projected states |Ψ(zB)\ket{\Psi(z_{B})}.

For projected ensembles generated by from time-evolved states, we derive this result from properties of the temporal ensemble. Our finding is more general than time-evolved states: we numerically find that the projected ensembles arising from energy eigenstates are also well described by generalized Scrooge ensembles. We discuss measurable and information-theoretic signatures of the generalized Scrooge ensemble, which mirror that of the Scrooge ensemble. Finally, we identify conditions under which the generalized Scrooge ensemble reduces to the Scrooge ensemble. These results give a complete description of the projected ensembles generated by time-evolved states, proven for generic Hamiltonian dynamics, and broaden our understanding of deep thermalization.

As an illustrative example, we consider the mixed-field Ising model (MFIM). For convenience, we repeat its Hamiltonian [Eq. (36)] here.

HMFIM=j=1NhxXj+hyYj+j=1N1JXjXj+1,H_{\text{MFIM}}=\sum_{j=1}^{N}h_{x}X_{j}+h_{y}Y_{j}+\sum_{j=1}^{N-1}JX_{j}X_{j+1}~, (62)

The projected ensemble is equal to the Scrooge ensemble when the subsystem BB is measured in the ZBZ_{B} basis. In other bases, e.g. the XBX_{B} basis, there are correlations between the measurement outcomes |xB\ket{x_{B}} and the projected states |Ψ(xB)\ket{\Psi(x_{B})}. In particular, their energies are anti-correlated: measurement outcomes |xB\ket{x_{B}} with high energies xB|HB|xB\bra{x_{B}}H_{B}\ket{x_{B}} are likely to result in projected states with lower energies Ψ(xB)|HA|Ψ(xB)\bra{\Psi(x_{B})}H_{A}\ket{\Psi(x_{B})} (and vice versa), where HAH_{A} and HBH_{B} are the terms of the Hamiltonian restricted to the subsystems AA and BB respectively. Note that the energies we refer to here are local energies on subsystems AA and BB and therefore such energetic correlations are not strict, but exist in a statistical sense, see e.g. Ref. [102]. Under these conditions, one must specify exponentially many probability distributions which take into account the correlations between xBx_{B} and |Ψ(xB)\ket{\Psi(x_{B})} arising from energy or other conservation laws. This is in contrast to what we have discussed in Sec. V, where we could treat the projected states |Ψ(xB)\ket{\Psi(x_{B})} with different labels (measurement outcomes) xBx_{B} as samples from the same probability distribution. To emphasize the effect of measurement bases, in this section we will refer to such measurement outcomes and projected states as xBx_{B} and |Ψ(xB)\ket{\Psi(x_{B})} respectively.

Special case: Magnetization-sectored bases — To gain intuition about the effects of measurement basis, it is illustrative to consider a special case: one where the outcomes can be partitioned into sets (“sectors”), within which there are no further correlations between the outcomes and projected states. For example, the MFIM we consider has a Zeeman field in the YY-axis. When this Zeeman field is sufficiently large, the sector MM of bitstrings |yB\ket{y_{B}} with the same magnetization have approximately the same energy on BB, but have different energies from bitstrings from other sectors. One consequence of this is the following: consider the “sector-resolved reduced density matrix,” defined as

ρA(M)\displaystyle\rho_{A}(M) =yBMp(yB)|Ψ(yB,t)Ψ(yB,t)|yBMp(yB).\displaystyle=\frac{\sum_{y_{B}\in M}p(y_{B})\outerproduct*{\Psi(y_{B},t)}{\Psi(y_{B},t)}}{\sum_{y_{B}\in M}p(y_{B})}~. (63)

These mixed states ρA(M)\rho_{A}(M) show systematic differences across magnetization sectors MM. Accordingly, it is not appropriate to take the projected states from |Ψ(yB,t)|\Psi(y_{B},t)\rangle as identical samples from the same ensemble of states. Rather, projected states corresponding to the same sector MM are drawn from the same ensemble. As we will show below, these are Scrooge ensembles, with different first moments:

yBM,|Ψ(yB)Scrooge[ρA(M)].\forall y_{B}\in M~,~\ket*{\Psi(y_{B})}\sim\text{Scrooge}[\rho_{A}(M)]~. (64)

However, this sector picture is only valid in special cases. Our generalized Scrooge ensemble extends this observation to the general case.

Refer to caption
Figure 5: Generalized Scrooge ensembles describe projected ensembles. (a) When the measurement basis on BB (here denoted XBX_{B}) is correlated with the total energy of the system, the projected states |Ψ(xB)\ket{\Psi(x_{B})} are correlated with the measurement outcome xBx_{B}, here illustrated by differences between outcomes with “low” (orange) and “high” (blue) energies. This leads to deviations from the Scrooge ensemble, and the projected ensemble is instead described by a generalized Scrooge ensemble (Claim 2). (b) We verify this conjecture numerically with the trace distance Δ(3)\Delta^{(3)} between the third moments of the projected ensemble and either the Scrooge ensemble (squares, downward triangle) or the generalized Scrooge ensemble (circles and upward triangles). We consider projected ensembles of states time-evolved (for sufficiently long times) under the MFIM [Eqs. (36-37)] and measured in the XBX_{B} basis. The difference between the two ensembles is more obvious for subsystems of two qubits (squares, circles) than for subsystems of four qubits (upward, downward triangles), due to the larger bath relative to the system size. While the distance to the Scrooge ensemble saturates to a non-zero value, the distance to the generalized Scrooge ensemble continues to decrease. (c) Consistent with the generalized Scrooge ensemble, the histogram (blue) of the rescaled probabilities p~(xB)pxA(xB)/𝔼t[pxA(xB)]\tilde{p}(x_{B})\equiv p_{x_{A}}(x_{B})/\mathbb{E}_{t}[p_{x_{A}}(x_{B})] follows a PT distribution, while the bare probabilities DpxA(xB)Dp_{x_{A}}(x_{B}) (red line) do not, with xA=0000x_{A}=0000 plotted here as a representative. (d) As with the mutual information in Scrooge ensembles, in generalized Scrooge ensembles the interaction information I(OA;XB;T)I(O_{A};X_{B};T) displays a universal behavior, independent of basis OAO_{A}, and agrees with the weighted sum of subentropies [Eq. (76)] (dashed). Inset: in contrast, the mutual information I(OA;XB|T)I(O_{A};X_{B}|T) is no longer independent of the basis OAO_{A}, here illustrated in red, green, and blue respectively by OA={X4,Y4,Z4}O_{A}=\{X^{\otimes 4},Y^{\otimes 4},Z^{\otimes 4}\}, due to time-averaged (thermal) correlations between the measurement bases OAO_{A} and XBX_{B}.

VI.1 Generalized Scrooge Ensemble

Our key idea is to treat each projected state |Ψ(xB)\ket*{\Psi(x_{B})} as a random sample from a different Scrooge ensemble Scrooge[ρ¯(xB)]\text{Scrooge}[\bar{\rho}(x_{B})], where ρ¯(xB)\bar{\rho}(x_{B}) is a density matrix associated to each outcome xBx_{B}. We define the density matrix ρ¯(xB)\bar{\rho}(x_{B}) be the time-averaged state

ρ¯(xB)\displaystyle\bar{\rho}(x_{B}) 𝔼t[|Ψ~(xB)Ψ~(xB)|]/𝔼t[Ψ~(xB)|Ψ~(xB)]\displaystyle\equiv\mathbb{E}_{t}[|\tilde{\Psi}(x_{B})\rangle\langle\tilde{\Psi}(x_{B})|]/\mathbb{E}_{t}[\langle\tilde{\Psi}(x_{B})|\tilde{\Psi}(x_{B})\rangle]
=(𝕀AxB|)ρd(𝕀A|xB)/pd(xB),\displaystyle=\left(\mathbb{I}_{A}\otimes\bra{x_{B}}\right)\rho_{d}\left(\mathbb{I}_{A}\otimes\ket{x_{B}}\right)/p_{d}(x_{B})~, (65)
pd(xB)\displaystyle p_{d}(x_{B}) 𝔼t[Ψ~(xB)|Ψ~(xB)]\displaystyle\equiv\mathbb{E}_{t}[\langle\tilde{\Psi}(x_{B})|\tilde{\Psi}(x_{B})\rangle]
=tr[(𝕀AxB|)ρd(𝕀A|xB)].\displaystyle=\tr\left[\left(\mathbb{I}_{A}\otimes\bra{x_{B}}\right)\rho_{d}\left(\mathbb{I}_{A}\otimes\ket{x_{B}}\right)\right]~. (66)

pd(xB)p_{d}(x_{B}) is the time-averaged probability of measuring the outcome xBx_{B}, and we express both ρ¯(xB)\bar{\rho}(x_{B}) and pd(xB)p_{d}(x_{B}) in terms of the diagonal ensemble ρd𝔼t[|Ψ(t)Ψ(t)|]\rho_{d}\equiv\mathbb{E}_{t}[|\Psi(t)\rangle\langle\Psi(t)|]. We refer to this collection of probability distributions, each sampled from once, as the generalized Scrooge ensemble.

Claim 2.

The states |Ψ(xB)\ket*{\Psi(x_{B})} of the projected ensemble are sampled from the generalized Scrooge ensemble

|Ψ(xB)Scrooge[ρ¯(xB)].\displaystyle\ket*{\Psi(x_{B})}\sim\text{Scrooge}[\bar{\rho}(x_{B})]~. (67)

In order to support this claim, we compare the kk-th moments of the normalized projected ensemble with the predicted kk-th moment

ρGen. Scr.(k)=xBpd(xB)ρScrooge(k)(ρ¯(xB)).\rho^{(k)}_{\text{Gen.~Scr.}}=\sum_{x_{B}}p_{d}(x_{B})\rho^{(k)}_{\text{Scrooge}}(\bar{\rho}(x_{B}))~. (68)

As illustrated in Fig. 5(b), the higher moments of the projected ensemble agree with ρGen. Scr.(k)\rho^{(k)}_{\text{Gen.~Scr.}} and differ from ρScrooge(k)\rho^{(k)}_{\text{Scrooge}} [Eq. (50)], with exponential convergence with increasing system size. Here, we have calculated the average states ρ¯(xB)\bar{\rho}(x_{B}) by numerically time-averaging |Ψ~(xB)Ψ~(xB)|\outerproduct*{\tilde{\Psi}(x_{B})}{\tilde{\Psi}(x_{B})} over a long interval. Having numerically demonstrated our claim, we proceed to derive it with results from the temporal ensemble.

Temporal ensemble of projected states — In order to derive results about projected ensembles, we first establish a result about the temporal ensemble of a projected state. Specifically, we consider the set of unnormalized states |Ψ~(xB,t)\ket*{\tilde{\Psi}(x_{B},t)} for a fixed outcome xBx_{B}. We show that the time trajectory (temporal ensemble) is described by the unnormalized ensemble S~crooge[ρ¯(xB)]\tilde{\text{S}}\text{crooge}[\bar{\rho}(x_{B})].

Lemma 1 (Temporal kk-th moment of the projected state).

Consider the temporal ensemble comprised of the (unnormalized) projected states |Ψ~(xB,t)(𝕀AxB|)|Ψ(t)\ket*{\tilde{\Psi}(x_{B},t)}\equiv(\mathbb{I}_{A}\otimes\bra{x_{B}})\ket{\Psi(t)}. The kk-th moment of this temporal ensemble is equal to the kk-th moment of the corresponding unnormalized Scrooge ensemble, up to small corrections:

ρTemp. Proj.(k)(xB)𝔼t[|Ψ~(xB,t)Ψ~(xB,t)|k]\displaystyle\rho^{(k)}_{\text{Temp. Proj.}}(x_{B})\equiv\mathbb{E}_{t}\left[\outerproduct*{\tilde{\Psi}(x_{B},t)}{\tilde{\Psi}(x_{B},t)}^{\otimes k}\right] (69)
=pd(xB)kρ~S~cr.(k)[ρ¯(xB)]+O(Δβ(xB)pd(xB)k),\displaystyle=p_{d}(x_{B})^{k}\tilde{\rho}^{(k)}_{\tilde{\text{S}}\text{cr.}}[\bar{\rho}(x_{B})]+O(\Delta_{\beta}(x_{B})p_{d}(x_{B})^{k})~,

where Δβ(xB)tr[(𝕀A|xBxB|)2ρd(2)]/pd(xB)2\Delta_{\beta}(x_{B})\equiv\tr[(\mathbb{I}_{A}\otimes\outerproduct{x_{B}}{x_{B}})^{\otimes 2}\rho_{d}^{(2)}]/p_{d}(x_{B})^{2} is a number, typically exponentially small in system size, characterizing the validity of our approximation and where ρd(2)|cE|4|E,EE,E|\rho_{d}^{(2)}\equiv\absolutevalue{c_{E}}^{4}\outerproduct{E,E}{E,E}.

Proof.

The leading term follows from Theorem 2. However, the bound O(Δβ(xB))O(\Delta_{\beta}(x_{B})) does not immediately follow from the bound given in Theorem 2, which is too weak and should be improved for this context.

We outline the proof here, with full details in Appendix I. We use an intermediate step in the proof of Theorem 2 in Appendix D. This step expresses the difference ρTemp. Proj.(k)(xB)pd(xB)kρ~S~cr.(k)[ρ¯(xB)]\rho^{(k)}_{\text{Temp. Proj.}}(x_{B})-p_{d}(x_{B})^{k}\tilde{\rho}^{(k)}_{\tilde{\text{S}}\text{cr.}}[\bar{\rho}(x_{B})] in terms of lists of energies (E1,,Ek)(E_{1},\dots,E_{k})— those with at least one repeated element. This allows us to replace two copies of the diagonal matrix ρd2E1,E2|cE1|2|cE2|2|E1,E2E1,E2|\rho_{d}^{\otimes 2}\equiv\sum_{E_{1},E_{2}}\absolutevalue{c_{E_{1}}}^{2}\absolutevalue{c_{E_{2}}}^{2}\outerproduct{E_{1},E_{2}}{E_{1},E_{2}} with the term ρd(2)E|cE|4|E,EE,E|\rho_{d}^{(2)}\equiv\sum_{E}\absolutevalue{c_{E}}^{4}\outerproduct{E,E}{E,E}. The relative size of the correction term is then Δβ(xB)\Delta_{\beta}(x_{B}), which is typically exponentially small in total system size (Appendix J). ∎

Projected ensemble of time-evolved states — While Lemma 1 is interesting in its own right, we use our notion of ergodicity (Sec. IV.5) to turn it into a statement about the projected ensemble at fixed (but generically late) points in time.

Theorem 4.

At typical (late) times tt, the weighted kk-th moment ρ~Proj.(k)(t)\tilde{\rho}^{(k)}_{\text{Proj.}}(t) of the unnormalized projected ensemble is equal to the kk-th moment of the generalized Scrooge ensemble:

ρ~Proj.(k)(t)\displaystyle\tilde{\rho}^{(k)}_{\text{Proj.}}(t) xB|Ψ~(xB,t)Ψ~(xB,t)|kpd(xB)k1\displaystyle\equiv\sum_{x_{B}}\frac{\outerproduct*{\tilde{\Psi}(x_{B},t)}{\tilde{\Psi}(x_{B},t)}^{\otimes k}}{p_{d}(x_{B})^{k-1}} (70)
=xBpd(xB)ρ~S~cr.(k)[ρ¯(xB)]+O(Δβ1/2),\displaystyle=\sum_{x_{B}}p_{d}(x_{B})\tilde{\rho}^{(k)}_{\tilde{\text{S}}\text{cr.}}[\bar{\rho}(x_{B})]+O(\Delta^{1/2}_{\beta})~, (71)

where ΔβxBpd(xB)Δβ(xB)\Delta_{\beta}\equiv\sum_{x_{B}}p_{d}(x_{B})\Delta_{\beta}(x_{B}), the weighted mean of Δβ(xB)\Delta_{\beta}(x_{B}).

Proof.

From Lemma 1, the time-averaged kk-th moment 𝔼t[ρ~Proj.(k)(t)]\mathbb{E}_{t}[\tilde{\rho}^{(k)}_{\text{Proj.}}(t)] is approximately equal to the weighted sum xBpd(xB)ρ~S~cr.(k)[ρ¯(xB)]\sum_{x_{B}}p_{d}(x_{B})\tilde{\rho}^{(k)}_{\tilde{\text{S}}\text{cr.}}[\bar{\rho}(x_{B})]. We then show in Appendix J that the variance over time vart[ρ~Proj.(k)(t)]\text{var}_{t}[\tilde{\rho}^{(k)}_{\text{Proj.}}(t)] is exponentially small, and therefore the projected kk-th moment ρ~Proj.(k)(t)\tilde{\rho}^{(k)}_{\text{Proj.}}(t) is close to its temporal average. The temporal fluctuations are of size vart[ρ~Proj.(k)(t)]=O(Δβ)\|\text{var}_{t}[\tilde{\rho}^{(k)}_{\text{Proj.}}(t)]\|=O(\Delta_{\beta}). While the most general bound on Δβ\Delta_{\beta} that we can prove is Δβ1\Delta_{\beta}\leq 1, we argue in Appendix J that it is typically exponentially small. ∎

Finally, we note that the kk-th moment described in Eq. (188) is a slightly different moment than both the moments of the normalized and unnormalized projected ensembles. Each projected state is weighted by the factor pd(xB)1kp_{d}(x_{B})^{1-k} in order to have a quantity that can be bounded by a small quantity (Δβ\Delta_{\beta}). The norm Ψ~(xB)|Ψ~(xB)\innerproduct*{\tilde{\Psi}(x_{B})}{\tilde{\Psi}(x_{B})} fluctuates about the value pd(xB)p_{d}(x_{B}). These fluctuations are not universal999More precisely, the norm is a random variable which is a weighted sum of PT variables, where the weights are the eigenvalues of ρ¯(xB)\bar{\rho}(x_{B}). The resulting distribution is known as a hypoexponential distribution and is discussed extensively in companion work [41] but depends on the spectrum of ρ¯(xB)\bar{\rho}(x_{B}).. Accordingly, the average value of Ψ~(xB)|Ψ~(xB)k/pd(xB)k\innerproduct*{\tilde{\Psi}(x_{B})}{\tilde{\Psi}(x_{B})}^{k}/p_{d}(x_{B})^{k} is equal to the kk-th moment of this distribution, which is larger than 1. With increasing size of the subsystem AA, these values concentrate around 1 and therefore, our weighted kk-th moment in Eq. (188) converges to the normalized ensemble moment.

VI.2 Porter-Thomas distribution in generalized Scrooge ensembles

In Sections IV.3 and V.6, we discussed the Porter-Thomas distribution as a measurable signature of the temporal and Scrooge ensembles respectively. Perhaps unsurprisingly, the PT distribution is also present in the generalized Scrooge ensemble, but only after suitable data processing. In Sec. V.6, we established that in the Scrooge ensemble, the joint probabilities poA(zB)p_{o_{A}}(z_{B}) follow a PT distribution with mean 𝔼zB[poA(zB)]oA|ρA|oA/DB\mathbb{E}_{z_{B}}[p_{o_{A}}(z_{B})]\equiv\bra{o_{A}}\rho_{A}\ket{o_{A}}/D_{B}. In the generalized Scrooge ensemble, since each projected state |Ψ~(xB)\ket*{\tilde{\Psi}(x_{B})} is sampled from a different Scrooge ensemble, we must normalize the overlaps poA(xB)p_{o_{A}}(x_{B}) by their respective time-averaged values 𝔼t[poA(xB)]=oA|ρ¯(xB)|oA\mathbb{E}_{t}[p_{o_{A}}(x_{B})]=\bra{o_{A}}\bar{\rho}(x_{B})\ket{o_{A}} (see Sec. IV.5). Having done so,

p(oA,zB)/𝔼t[p(oA,zB)]PT(1),p(o_{A},z_{B})/\mathbb{E}_{t}[p(o_{A},z_{B})]\sim\text{PT}(1)~, (72)

as illustrated in Fig. 5(c).

VI.3 Interaction information

We next turn to the information-theoretic properties of the generalized Scrooge ensemble. Specifically, we consider the mutual information I(OA;XB)I(O_{A};X_{B}) between any basis {oA}\{o_{A}\} on AA and projective measurements {|xB}\{\ket{x_{B}}\}. Unlike in the Scrooge ensemble, there are additional correlations between xBx_{B} and the projected states |Ψ(xB)\ket*{\Psi(x_{B})}. With the generalized Scrooge ensemble, we find that the mutual information I(OA;XB|T)I(O_{A};X_{B}|T) (i.e. the mutual information of the distribution p(oAzB,t)p(o_{A}z_{B},t) at fixed, late times) can be written as follows:

I(OA;XB|T)I(OA;XB)+xBpd(xB)Q[ρ¯(xB)],I(O_{A};X_{B}|T)\approx I(O_{A};X_{B})+\sum_{x_{B}}p_{d}(x_{B})Q\left[\bar{\rho}(x_{B})\right]~, (73)

where the first term I(OA;XB)I(O_{A};X_{B}) refers to correlations that exist even in the time-averaged distribution 𝔼t[p(oAzB,t)]\mathbb{E}_{t}[p(o_{A}z_{B},t)]. These correlations are non-universal and are thermal in nature, arising from energy correlations between the bases {|oA}\{\ket{o_{A}}\} and {|xB}\{\ket{x_{B}}\}. In contrast, the second term captures the correlations associated with fluctuations in time. These fluctuations in time are universal and are expressed as a weighted sum of subentropies Q[ρ¯(xB)]Q\left[\bar{\rho}(x_{B})\right]. Eq. (73) follows from the following:

I(OA;XB|T)I(OA;XB)\displaystyle I(O_{A};X_{B}|T)-I(O_{A};X_{B})
=H(OA|T)H(OA|XB,T)H(OA)+H(OA|XB)\displaystyle=H(O_{A}|T)-H(O_{A}|X_{B},T)-H(O_{A})+H(O_{A}|X_{B})
H(OA|XB)H(OA|XB,T)\displaystyle\approx H(O_{A}|X_{B})-H(O_{A}|X_{B},T) (74)
xBpd(xB)[H(OA|XB=xB)H(OA|XB=xB,T)]\displaystyle\approx\sum_{x_{B}}p_{d}(x_{B})\left[H(O_{A}|X_{B}=x_{B})-H(O_{A}|X_{B}=x_{B},T)\right] (75)
xBpd(xB)Q[ρ~(xB)].\displaystyle\approx\sum_{x_{B}}p_{d}(x_{B})Q\left[\tilde{\rho}(x_{B})\right]~. (76)

Eq. (74) we have used the fact that H(OA|T)H(OA)H(O_{A}|T)\approx H(O_{A}), because the time-dependent reduced density matrix ρA\rho_{A} is exponentially close to its time averaged value trB[ρd]\tr_{B}[\rho_{d}]. We then approximate H(OA|XB,T)=xBp(xB)H(OA|XB=xB,T)pd(xB)H(OA|XB=xB,T)H(O_{A}|X_{B},T)=\sum_{x_{B}}p(x_{B})H(O_{A}|X_{B}=x_{B},T)\approx p_{d}(x_{B})H(O_{A}|X_{B}=x_{B},T). This is justified because p(xB,t)p(x_{B},t) fluctuates about pd(xB)p_{d}(x_{B}). When summed over in H(OA|XB,T)H(O_{A}|X_{B},T), these fluctuations average away and we can replace p(xB)p(x_{B}) with pd(xB)p_{d}(x_{B}). Eq. (75) is a weighted sum of the mutual information for the conditional distributions p(oA,t|xB)p(o_{A},t|x_{B}). Theorem 4 allows us to equate this mutual information with the subentropies Q[ρ¯(xB)]Q[\bar{\rho}(x_{B})], giving the desired Eq. (73).

The quantity I(OA;XB;T)I(OA;XB|T)I(OA;XB)I(O_{A};X_{B};T)\equiv I(O_{A};X_{B}|T)-I(O_{A};X_{B}) is known as the interaction information, which is a multivariate analog of the mutual information [103]. It is a measure of the correlations that exist due to interaction among all variables, and is symmetric under exchange of all variables, and is also the classical analogue of the (quantum) tripartite mutual information studied in many-body physics [104, 105]. In particular, our result predicts that the interaction information is independent of basis OAO_{A}, and has expression xBpd(xB)Q[ρ¯(xB)]\sum_{x_{B}}p_{d}(x_{B})Q[\bar{\rho}(x_{B})], which we can bound as xBpd(xB)Q[ρ~(xB)]Q[ρA]\sum_{x_{B}}p_{d}(x_{B})Q[\tilde{\rho}(x_{B})]\leq Q[\rho_{A}] using the concavity of the subentropy [89]. The interaction information quantifies the amount of information unique to a time-evolved state, i.e. with time-averaged correlations subtracted. Similar to our results in the temporal and Scrooge ensembles, this information in universal, and is minimized subject to constraints.

We demonstrate this result with numerical data in Fig. 5(d), in which we plot the interaction information for three measurement bases OA{X4,Y4,Z4}O_{A}\in\{X^{\otimes 4},Y^{\otimes 4},Z^{\otimes 4}\}. While the mutual information I(OA;XB|T)I(O_{A};X_{B}|T) depends on the choice of OAO_{A} (inset), the interaction information I(OA;XB;T)I(O_{A};X_{B};T) is independent and is equal to the weighted average subentropy [Eq. (76)].

VI.4 Recovering the Scrooge ensemble from the generalized Scrooge ensemble

The generalized Scrooge ensemble in Theorem 4 specializes to the original Scrooge ensemble when the states ρ¯(xB)\bar{\rho}(x_{B}) are independent of measurement outcome xBx_{B}, i.e. ρ¯(xB)ρA\bar{\rho}(x_{B})\propto\rho_{A}. As presented in Section V, this is approximately true when states evolved by the MFIM are measured in the {|zB}\{\ket{z_{B}}\} basis. The crucial property of the ZBZ_{B} measurement basis is that the energy expectation values E(|z)z|H|zE(\ket{z})\equiv\innerproduct{z}{H|z} and energy variances σH2(|z)z|(HEz)2|z\sigma^{2}_{H}(\ket{z})\equiv\innerproduct{z}{(H-E_{z})^{2}|z} are independent of the bitstring zz. This is an example of “minimal correlation” between measurement basis and energy eigenbasis.

We argue that these two properties are sufficient for the Scrooge ensemble to be recovered approximately for zBz_{B} measurements in the MFIM. In particular, we will show that for any state |ϕAA\ket{\phi_{A}}\in\mathcal{H}_{A}, ϕA|ρd(zB)|ϕAϕA|ρd|ϕA\innerproduct{\phi_{A}|\rho_{d}(z_{B})}{\phi_{A}}\simeq\innerproduct{\phi_{A}|\rho_{d}}{\phi_{A}} up to subleading corrections in the size NN of the system, and therefore ρd(zB)ρd\rho_{d}(z_{B})\simeq\rho_{d}.

From the definition of ρd(zB)\rho_{d}(z_{B}), the expectation value ϕA|ρd(zB)|ϕA\innerproduct{\phi_{A}|\rho_{d}(z_{B})}{\phi_{A}} can be expressed in terms of the eigenstates |E\ket{E} of HH as

ϕA|ρd(zB)|ϕA=E|cE|2|ϕA,zB|E|2\innerproduct{\phi_{A}|\rho_{d}(z_{B})}{\phi_{A}}=\sum_{E}|c_{E}|^{2}|\innerproduct{\phi_{A},z_{B}}{E}|^{2} (77)

We next use the theorem proved in Ref. [106], which states that for large NN, the energy distribution of the product state |ϕA,zB\ket{\phi_{A},z_{B}} converges to a Gaussian centered about E(|ϕA,zB)=ϕA,zB|H|ϕA,zBE(\ket{\phi_{A},z_{B}})=\innerproduct{\phi_{A},z_{B}}{H|\phi_{A},z_{B}} with variance σH2(|ϕA,zB))=ϕA,zB|(HE(|ϕA,zB))2|ϕA,zB\sigma^{2}_{H}(\ket{\phi_{A},z_{B}}))=\innerproduct{\phi_{A},z_{B}}{(H-E(\ket{\phi_{A},z_{B}}))^{2}|\phi_{A},z_{B}}. Furthermore, the mean E(|ϕA,zB)E(\ket{\phi_{A},z_{B}}) and variance σH2(|ϕA,zB))\sigma^{2}_{H}(\ket{\phi_{A},z_{B}})) are independent of zBz_{B} up to subleading corrections. To show this, we write the Hamiltonian as H=HA+HB+HABH=H_{A}+H_{B}+H_{AB}, where HA,HBH_{A},H_{B} only act on subsystems A,BA,B respectively and HABH_{AB} is the remaining term, acting on the boundary between AA and BB. As mentioned above, the expectation value of HBH_{B} and variance are independent of zBz_{B}:

EB(|zB)\displaystyle E_{B}(\ket{z_{B}}) zB|HB|zB=0,\displaystyle\equiv\innerproduct{z_{B}}{H_{B}|z_{B}}=0, (78)
σHB2(|zB)\displaystyle\sigma^{2}_{H_{B}}(\ket{z_{B}}) zB|(HBEB(|zB))2|zB\displaystyle\equiv\innerproduct{z_{B}}{(H_{B}-E_{B}(\ket{z_{B}}))^{2}|z_{B}} (79)
=NB(hx2+hy2)+(NB1)J2.\displaystyle=N_{B}(h_{x}^{2}+h_{y}^{2})+(N_{B}-1)J^{2}~.

We also have ϕA,zB|HAB|ϕA,zB=0\innerproduct{\phi_{A},z_{B}}{H_{AB}|\phi_{A},z_{B}}=0, which gives

E(|ϕA,zB)=EA(|ϕA),\displaystyle E(\ket{\phi_{A},z_{B}})=E_{A}(\ket{\phi_{A}}), (80)
σH2(|ϕA,zB)=σHA2(|ϕA)+σHB2+O(1)\displaystyle\sigma^{2}_{H}(\ket{\phi_{A},z_{B}})=\sigma^{2}_{H_{A}}(\ket{\phi_{A}})+\sigma_{H_{B}}^{2}+O(1) (81)

where the terms O(1)O(1) come from the boundary contributions and hence do not scale with system size. From Eqs. (80) and (81), we find that the energy distribution of |ϕA,zB\ket{\phi_{A},z_{B}}, and consequently the quantity ϕA|ρd(zB)|ϕA\innerproduct{\phi_{A}}{\rho_{d}(z_{B})|\phi_{A}} in Eq. (77) are independent of the measurement outcome zBz_{B} (up to subleading corrections in system size), recovering the Scrooge ensemble from the generalized Scrooge ensemble in the limit of large system sizes.

VI.5 Projected ensembles of energy eigenstates

Refer to caption
Figure 6: Scrooge and generalized Scrooge ensembles from energy eigenstates. (a) We study the projected ensembles arising from energy eigenstates of the MFIM with broken time-reversal symmetry [Eq. (82)], with total size N=12N=12 and a subsystem of size NA=4N_{A}=4 in the middle of the chain. (b) For each eigenstate |E\ket{E}, we construct its reduced density matrix ρA(E)\rho_{A}^{(E)} and corresponding Scrooge ensemble Scrooge[ρA(E)]\text{Scrooge}[\rho_{A}^{(E)}]. The projected ensemble is always close to the Scrooge ensemble prediction, as measured by the trace distance Δ(2)\Delta^{(2)} of their second moments (blue), and is only close to the Haar ensemble at infinite temperature, when E/N0E/N\approx 0 (red). Curiously, and unexpectedly, the generalized Scrooge ensemble appears to be better descriptions at the edges of the spectrum. (c) The trace distances Δ(2)\Delta^{(2)} (blue) and Δ(3)\Delta^{(3)} (green), averaged over the 100 eigenstates closest to the target energy E=0.5NE=0.5N [orange line in (b)] decrease exponentially with system size. (d) For a fixed eigenstate |E\ket{E}, the histogram of normalized measurement probabilities p~pxA(E)(xB)/𝔼E[pxA(E)(xB)]\tilde{p}\equiv p_{x_{A}}^{(E)}(x_{B})/\mathbb{E}_{E}[p_{x_{A}}^{(E)}(x_{B})] follows a PT distribution, while the raw probabilities DpxA(E)(xB)Dp_{x_{A}}^{(E)}(x_{B}) (red line) do not, supporting the presence of the generalized Scrooge ensemble. Here we plot xA=0010x_{A}=0010 as a representative.

Thus far, we have investigated the projected ensembles of time-evolved states |Ψ(t)\ket{\Psi(t)}, and provided analytical arguments based on the temporal ensemble. In fact, the generalized Scrooge ensemble also seems to describe projected ensembles obtained from energy eigenstates, as verified by numerical simulation results illustrated in Fig. 6.

Here, we consider the projected ensembles arising from eigenstates of the MFIM, with additional terms which remove time-reversal symmetry [Eq. (82) and in Appendix L]. We first test our simpler prediction: the emergence of the Scrooge ensemble by measuring in the ZBZ_{B} basis. In Fig. 6(b), the projected ensemble agrees with its respective Scrooge ensemble for energy eigenstates at all energies, but only agrees with the Haar ensemble at infinite temperature, when E/N0E/N\approx 0. Furthermore, the distance between the projected and Scrooge ensemble decreases exponentially with total system size [Fig. 6(c)].

We next test our generalized Scrooge ensemble prediction, by generating the projected ensemble with measurements in the XBX_{B} basis. In this setting, it is difficult to make quantitatively accurate predictions for the generalized Scrooge ensemble, because we do not have a prescription of the average state ρ¯(xB)\bar{\rho}(x_{B}). The expression ρ¯(xB)𝔼t[|Ψ~(xB,t)Ψ~(xB,t)|]\bar{\rho}(x_{B})\propto\mathbb{E}_{t}[\outerproduct*{\tilde{\Psi}(x_{B},t)}{\tilde{\Psi}(x_{B},t)}] in time-evolved states does not apply to energy eigenstates. Nevertheless, motivated by the eigenstate thermalization hypothesis (ETH), we expect that the eigenstates with nearby energy values will generate highly-similar projected ensembles: in particular, the average states ρ¯(xB,E)\bar{\rho}(x_{B},E) should be smooth functions of the total energy. Therefore, we can estimate ρ¯(xB,E)\bar{\rho}(x_{B},E) by averaging the projected states |Ψ~E(xB)Ψ~E(xB)|\outerproduct*{\tilde{\Psi}_{E}(x_{B})}{\tilde{\Psi}_{E}(x_{B})} over energy eigenstates of a sufficiently narrow energy window. While this method is not quantitatively accurate enough to give exponentially-decreasing trace distances (the equivalent of Fig. 6(c), but for the generalized Scrooge ensemble), we can observe the PT distribution in normalized outcome probabilities [Fig. 6(d)]. We first average the probability outcomes pxA(E)(xB)|E|xA,xB|2p_{x_{A}}^{(E)}(x_{B})\equiv\absolutevalue{\innerproduct{E}{x_{A},x_{B}}}^{2} over a small energy window to obtain 𝔼E[pxA(E)(xB)]\mathbb{E}_{E}[p_{x_{A}}^{(E)}(x_{B})] for each value of xAx_{A} and xBx_{B}. The histogram of normalized probabilities p~pxA(E)(xB)/𝔼E[pxA(E)(xB)]\tilde{p}\equiv p_{x_{A}}^{(E)}(x_{B})/\mathbb{E}_{E}[p_{x_{A}}^{(E)}(x_{B})] over all xBx_{B}’s (with fixed xAx_{A}) shows a PT distribution, while the raw probabilities pxA(E)(xB)p_{x_{A}}^{(E)}(x_{B}) do not.

Time-reversal symmetry and the real Scrooge ensemble — Finally, we remark on the role of symmetry. Unlike in time-evolved states, the presence of time-reversal symmetry in a Hamiltonian HH qualitatively affects the projected ensembles from energy eigenstates. Time-reversal symmetry — defined as the existence of a unitary operator UU that maps HH to its complex conjugate HH^{*}: UHU=HUHU^{\dagger}=H^{*} [107] — ensures that all eigenstates of HH have purely real coefficients in a certain basis. If this coincides with the measurement basis (e.g. the XBX_{B} basis in the MFIM), then the projected states are guaranteed to be real vectors. In such cases, the projected ensemble instead follows a real Scrooge ensemble, which is an analogous “ρ\rho-distortion” of the ensemble of real random vectors. The real Scrooge ensemble has corresponding signatures such as a “real PT distribution” [83] which we demonstrate in Appendix L. Here we simply break the time-reversal symmetry by adding appropriate terms:

HMFIM=\displaystyle H^{\prime}_{\text{MFIM}}= j=1NhxXj+hyYj+hzZj\displaystyle\sum_{j=1}^{N}h_{x}X_{j}+h_{y}Y_{j}+h_{z}Z_{j} (82)
+j=1N1JXjXj+1+JYjYj+1,\displaystyle+\sum_{j=1}^{N-1}JX_{j}X_{j+1}+J^{\prime}Y_{j}Y_{j+1}~,

with J=0.4J^{\prime}=0.4, hz=0.5h_{z}=0.5 and the values of hx,hy,Jh_{x},h_{y},J kept the same as the other numerical examples [Eq. (36)].

The presence of time-reversal symmetry was not relevant to the projected ensembles of time-evolved states because the time-evolution naturally breaks time-reversal symmetry. However, time-evolved states can be engineered to be purely real, requiring both time-reversal and particle-hole symmetry and suitably symmetric initial states and measurement bases (Appendix L).

VII Porter-Thomas distributions as signatures of randomness: Higher-order ETH

Refer to caption
Figure 7: Porter-Thomas (PT) distribution in energy populations. We observe the presence of the PT distribution in the energy populations |cE|2|E|Ψ0|2\absolutevalue{c_{E}}^{2}\equiv\absolutevalue{\innerproduct{E}{\Psi_{0}}}^{2} as a consequence of the “higher-order ETH” [96]. In a narrow energy window, the populations are distributed according to a Porter-Thomas distribution, with a smoothly varying mean f(E)f(E). (a) A state at infinite effective temperature |Ψ=|0N\ket{\Psi}=|0\rangle^{\otimes N} has constant mean f(E)f(E) (black dashed line). (b) The histogram of |cE|2\absolutevalue{c_{E}}^{2} over all eigenstates follows a PT distribution. (c) In contrast, for a state |Ψ=(exp(iθX/2)|0)N\ket{\Psi}=(\exp(i\theta X/2)|0\rangle)^{\otimes N} at finite temperature (details below), |cE|2\absolutevalue{c_{E}}^{2} monotonically increases about the smooth function f(E)exp(βE)f(E)\propto\exp(\beta E), with β\beta fitted (dashed line). (d) The histogram of |cE|2/f(E)\absolutevalue{c_{E}}^{2}/f(E) shows a PT distribution, both over the entire spectrum as well as in a narrow energy window [inset, energy window denoted by green band in (c)], while the raw populations D|cE|2D\absolutevalue{c_{E}}^{2} at finite temperature (red line) do not.

In this work we demonstrated instances of the PT distribution emerging from the maximum entropy principle in many-body dynamics: in the temporal and Scrooge ensembles, the PT distribution describes the probabilities of outcome probabilities Pr[p]\text{Pr}[p]. We found that in the generalized Scrooge ensemble, it was necessary to appropriately rescale the outcome probabilities to obtain the PT distribution.

Here, we demonstrate another instance of the PT distribution, hinting at the wider applicability of the maximum entropy principle in many-body physics. Specifically, we find the PT distribution in the distribution of overlaps |cE|2=|Ψ|E|2\absolutevalue{c_{E}}^{2}=\absolutevalue{\innerproduct{\Psi}{E}}^{2} between energy eigenstates |E\ket{E} and a simple physical state |Ψ\ket{\Psi}. This is consistent with a “higher-order” or kk-ETH [96], in which energy eigenstates in a narrow energy window are conjectured to form a state kk-design in an appropriate microcanonical subspace. This is a generalization of the conventional ETH, which is a statement about the the behavior of properties such as the overlaps Ej|O|Ek\bra{E_{j}}O\ket{E_{k}} between energy eigenstates |Ek\ket{E_{k}} and operators OO. Specifically, Ref. [96] defines the kk-ETH to hold for a given kk-copy operator OO if

𝔼t[UtkΠEOΠEUtk]=𝔼U[UkΠEOΠEU,k],\mathbb{E}_{t}[U_{t}^{\otimes k}\Pi_{E}O\Pi_{E}U_{-t}^{\otimes k}]=\mathbb{E}_{U}[U^{\otimes k}\Pi_{E}O\Pi_{E}U^{\dagger,\otimes k}]~, (83)

where ΠE\Pi_{E} is a projector onto a microcanonical subspace MC\mathcal{H}_{\text{MC}}, spanned by the energy eigenstates in a narrow energy window, Utexp(iHt)U_{t}\equiv\exp(-iHt), and the average over UU is over Haar-random unitaries acting on the microcanonical space MC\mathcal{H}_{\text{MC}}. They numerically verify that Eq. (83) approximately holds when O=okO=o^{\otimes k} (for local operators oo) and when O=𝒮AO=\mathcal{S}_{A} is the two-copy swap operator on a subsystem AA, which they used to derive the Page correction in the 2-Rényi entropy.

Here, we take O=|Ψ0Ψ0|kO=\outerproduct{\Psi_{0}}{\Psi_{0}}^{\otimes k}. Assuming that the kk-ETH holds for such operators, Eq. (83) predicts that the eigenstate overlaps |cE|2\absolutevalue{c_{E}}^{2} in a microcanonical window follow a Porter-Thomas distribution, with a mean value f(E)f(E).

|cE|2PT(f(E)).\absolutevalue{c_{E}}^{2}\sim\text{PT}(f(E))~. (84)

We further expect f(E)f(E) to be a smooth function of the energy. By dividing |cE|2\absolutevalue{c_{E}}^{2} by a fitted smooth function f(E)f(E), we expect the rescaled overlap to be distributed according to a PT distribution. We empirically find that f(E)f(E) is often well described by a Boltzmann distribution exp(βE)\exp(-\beta E). In Fig. 7(a-d) we demonstrate the emergence of the PT distribution in the populations over the eigenstates of the time-reversal-broken MFIM HH^{\prime} [Eq. (82)] for N=12N=12 for: an infinite-temperature state |Ψ0=|0N\ket{\Psi_{0}}=|0\rangle^{\otimes N} (a,b), and a finite-temperature state |Ψ0=[exp(iθX/2)|0]N\ket{\Psi_{0}}=[\exp(i\theta X/2)|0\rangle]^{\otimes N} (c,d), with θ=0.3557\theta=0.3557 chosen such that Ψ0|H|Ψ0=0.5N\langle\Psi_{0}|H|\Psi_{0}\rangle=0.5N.

The presence of the PT distribution in the energy populations not only serves as an easily visualized consequence of the kk-ETH, it also accounts for observations in the literature, such as universal ratios of 2 and 3 between the spectral form factor (SFF) and survival probability at late times [108], which follow from the second moment of the complex and real PT distributions. Specifically, the SFF is expressed as |tr[exp(iHt)]|2/D2=E,Eexp(i(EE)t)/D2|\text{tr}\left[\exp(-iHt)\right]|^{2}/D^{2}=\sum_{E,E^{\prime}}\exp(i(E^{\prime}-E)t)/D^{2}. At late times, only the E=EE=E^{\prime} contribution remains and the SFF has a late time value of 1. In contrast, the survival probability of an initial state |Ψ0|\Psi_{0}\rangle has expression |Ψ0|exp(iHt)|Ψ0|2=E,E|cE|2|cE|2exp(i(EE)t)|\langle\Psi_{0}|\exp(-iHt)|\Psi_{0}\rangle|^{2}=\sum_{E,E^{\prime}}|c_{E}|^{2}|c_{E^{\prime}}|^{2}\exp(i(E^{\prime}-E)t). Again, at late times, only the E=EE=E^{\prime} term survives and we obtain E|cE|4\sum_{E}|c_{E}|^{4}. From the discussion above, at infinite effective temperature, this sum is approximately 2/D2/D or 3/D3/D, when HH respectively does not and does have time-reversal symmetry, and |Ψ0|\Psi_{0}\rangle respects these symmetries.

VIII Outlook

In this work, we have demonstrated that a novel type of maximum entropy principle arises in natural chaotic many-body systems. We substantiate our findings by studying two different ensembles of states, each underlying the phenomena of Hilbert space ergodicity and deep thermalization, and find them well-described by ensembles with maximum entropy. Furthermore, we present the operational meaning, information theoretic properties, and simple, experimentally-accessible signatures, i.e., Porter-Thomas distribution in certain rescaled observables, of our claimed universality.

Our work extends our understanding of deep thermalization, most notably away from infinite temperature. We leave it to future work to generalize more features of infinite-temperature projected ensembles, such as the existence of distinct thermalization times for kk-th moments [30, 31, 36] and the dynamics at short times [78].

The universality we establish in this work may be exploited for practical purposes. For example, in Ref. [94], the presence of the PT distribution is leveraged to design a protocol to estimate fidelities in natural many-body systems. In companion experimental work, we explore deviations of the PT distribution in the presence of a bath. By understanding these deviations from universality, we design protocols to learn properties of the bath [41]. Further applications may include generalizing schemes that use aspects of quantum randomness to natural systems, such as randomized measurement protocols (including classical shadow tomography) [87, 88, 109, 95] (with applications to characterizing thermal states [110]), as well as random number generation [111, 112].

Finally, our results should be seen as complementary to and distinct from conventional results such as the ETH. The ETH models some aspects of ergodic many-body systems with random matrix theory, which has been successful in describing striking features such as the level repulsion in eigenvalues. In recent years, there has been a growing body of work scrutinizing as well as generalizing the ETH, motivated by recent interest in quantities such as multi-time correlators. See e.g. Refs. [113, 96, 114, 115]. Our findings in the temporal and projected ensembles as well as our analysis of the kk-ETH contribute to such work; providing a guiding principle to understand the emergence of effective randomness in many-body states. We look forward to such work and anticipate a broader understanding of complex quantum many-body systems.

Acknowledgements

We thank David Huse, Wen Wei Ho, Daniel Ranard and Saúl Pilatowsky-Cameo for insightful discussions on this work. We acknowledge support by the NSF QLCI Award OMA-2016245, the DOE (DE-SC0021951), the Institute for Quantum Information and Matter, an NSF Physics Frontiers Center (NSF Grant PHY-1733907), the Center for Ultracold Atoms, an NSF Physics Frontiers Center (NSF Grant PHY-1734011), the DARPA ONISQ program (W911NF2010021), the AFOSR YIP (FA9550-19-1-0044), Army Research Office MURI program (W911NF2010136), NSF CAREER award 2237244 and NSF CAREER award 1753386. Support is also acknowledged from the U.S. Department of Energy, Office of Science, National Quantum Information Science Research Centers, Quantum Systems Accelerator. AE acknowledges funding by the German National Academy of Sciences Leopoldina under the grant number LPDS 2021-02 and by the Walter Burke Institute for Theoretical Physics at Caltech. FMS acknowledges support provided by the U.S. Department of Energy Office of Science, Office of Advanced Scientific Computing Research, (DE-SC0020290), by Amazon Web Services, AWS Quantum Program, and by the DOE QuantISED program through the theory consortium “Intersections of QIS and Theoretical Particle Physics” at Fermilab. JC acknowledges support from the Terman Faculty Fellowship at Stanford University.

Appendix A Proof of Theorem 1 (The temporal ensemble is the random phase ensemble)

Theorem 1.

Given an initial state |Ψ0\ket{\Psi_{0}} and a Hamiltonian HH, the infinite-time temporal ensemble is equal to the random phase ensemble:

limτTemp.(τ)=Rand. Phase,\lim_{\tau\rightarrow\infty}\mathcal{E}_{\text{Temp.}}(\tau)=\mathcal{E}_{\text{Rand.~Phase}}~, (85)

with fixed energy populations |cE|2=|Ψ0|E|2|c_{E}|^{2}=\absolutevalue{\innerproduct{\Psi_{0}}{E}}^{2}, if and only if HH satisfies all kk-th no-resonance conditions modulo degeneracies.

Proof.

We prove Theorem 1 by matching the kk-th moments of the temporal and random phase ensembles, and showing that they agree for all orders kk. We enumerate the possible terms with multisets.

Definition 4 (Multisets).

We define the ordered multiset {αj}j=1k\{\alpha_{j}\}_{j=1}^{k} as the ordered list of elements αj{1,,D}\alpha_{j}\in\{1,\dots,D\}, with possibly repeated elements. We denote Mk(D)M_{k}(D) the set of all such multisets of length kk with symbols in {1,,D}\{1,\dots,D\}.

We now compute the kk-th moment of the temporal ensemble:

ρTemp.(k)𝔼t[Utk|Ψ0Ψ0|kUtk]\displaystyle\rho^{(k)}_{\text{Temp.}}\equiv\mathbb{E}_{t}\left[U_{t}^{\otimes k}\outerproduct{\Psi_{0}}{\Psi_{0}}^{\otimes k}U_{-t}^{\otimes k}\right] (86)
=𝜶,𝜷Mk(D)[j=1k|EαjEαj|]|Ψ0Ψ0|k[j=1k|EβjEβj|]δ(j=1kEαjj=1kEβj),\displaystyle=\sum_{\boldsymbol{\alpha},\boldsymbol{\beta}\in M_{k}(D)}\left[\bigotimes_{j=1}^{k}\outerproduct{E_{\alpha_{j}}}{E_{\alpha_{j}}}\right]\outerproduct{\Psi_{0}}{\Psi_{0}}^{\otimes k}\left[\bigotimes_{j=1}^{k}\outerproduct{E_{\beta_{j}}}{E_{\beta_{j}}}\right]\delta\left(\sum_{j=1}^{k}E_{\alpha_{j}}-\sum_{j=1}^{k}E_{\beta_{j}}\right), (87)
=𝜶Mk(D)[σπ(𝜶)|Eα1,,EαkEα1,,Eαk||Ψ0Ψ0|k|Eσ(α1),,Eσ(αk)Eσ(α1),,Eσ(αk)|],\displaystyle=\sum_{\boldsymbol{\alpha}\in M_{k}(D)}\bigg{[}\sum_{\sigma\in\pi(\boldsymbol{\alpha})}\outerproduct{E_{\alpha_{1}},\cdots,E_{\alpha_{k}}}{E_{\alpha_{1}},\cdots,E_{\alpha_{k}}}\outerproduct{\Psi_{0}}{\Psi_{0}}^{\otimes k}\outerproduct{E_{\sigma({\alpha_{1}})},\cdots,E_{\sigma({\alpha_{k}})}}{E_{\sigma({\alpha_{1}})},\cdots,E_{\sigma({\alpha_{k}})}}\bigg{]}, (88)
=𝜶Mk(D)σπ(𝜶)j=1k|Eαj|Ψ0|2|EαjEσ(αj)|\displaystyle=\sum_{\boldsymbol{\alpha}\in M_{k}(D)}\sum_{\sigma\in\pi(\boldsymbol{\alpha})}\bigotimes_{j=1}^{k}\absolutevalue{\innerproduct{E_{\alpha_{j}}}{\Psi_{0}}}^{2}\outerproduct{E_{\alpha_{j}}}{E_{\sigma({\alpha_{j}})}} (89)
=ρdk𝜶Mk(D)σπ(𝜶)j=1k|EαjEσ(αj)|,\displaystyle=\rho_{d}^{\otimes k}\sum_{\boldsymbol{\alpha}\in M_{k}(D)}\sum_{\sigma\in\pi(\boldsymbol{\alpha})}\bigotimes_{j=1}^{k}\outerproduct{E_{\alpha_{j}}}{E_{\sigma({\alpha_{j}})}}~, (90)

where Utexp(iHt)U_{t}\equiv\exp(-iHt), and 𝔼t[]=limT1T0Tdt\mathbb{E}_{t}[\cdot]=\lim_{T\rightarrow\infty}\frac{1}{T}\int_{0}^{T}\cdot~dt is the infinite-time average, |E\ket{E} are energy eigenstates of the Hamiltonian, and ρdE|E|Ψ0|2|EE|\rho_{d}\equiv\sum_{E}\absolutevalue{\innerproduct{E}{\Psi_{0}}}^{2}\outerproduct{E}{E} is the diagonal ensemble, prominently featured in literature on many-body thermalization [15]. In going from (86) to (87) we have performed an infinite-time integration to obtain a delta function, and in going to (88), we have used the kk-th no-resonance condition.

We then compute the kk-th moment of the random phase ensemble:

ρRand. phase(k)\displaystyle\rho^{(k)}_{\text{Rand. phase}} =𝜶,𝜷Mk(D)j=1kcαjcβj|αjβj|j(12π02π𝑑θj)exp[i(θαjθβj)]δ(αj),(βj)\displaystyle=\sum_{\boldsymbol{\alpha},\boldsymbol{\beta}\in M_{k}(D)}\bigotimes_{j=1}^{k}c_{\alpha_{j}}c_{\beta_{j}}^{*}\outerproduct*{\alpha_{j}}{\beta_{j}}\underbrace{\prod_{j}\left(\frac{1}{2\pi}\int_{0}^{2\pi}d\theta_{j}\right)\exp[i(\theta_{\alpha_{j}}-\theta_{\beta_{j}})]}_{\delta_{(\alpha_{j}),(\beta_{j})}} (91)
=𝜶Mk(D)σπ(M)j=1k|cαj|2|αjσ(αj)|\displaystyle=\sum_{\boldsymbol{\alpha}\in M_{k}(D)}\sum_{\sigma\in\pi(M)}\bigotimes_{j=1}^{k}\absolutevalue{c_{\alpha_{j}}}^{2}\outerproduct*{\alpha_{j}}{\sigma({\alpha_{j}})} (92)

Therefore, the kk-th moments of the temporal and random phase ensembles are equal. Identifying a distribution from its moments is known as the “moment problem.” For complex multivariate (vector-valued) distributions P(z1,,zD)P(𝒛)P(z_{1},\cdots,z_{D})\equiv P(\boldsymbol{z}), the complex Carleman’s condition gives a sufficient condition to the moment problem. That is, define the moments mj(k)𝑑zP(𝒛)|zj|2km^{(k)}_{j}\equiv\int dzP(\boldsymbol{z})|z_{j}|^{2k}. If the following sum diverges:

j,k=1(mj(k))1/2k=,\forall j~,~\sum_{k=1}^{\infty}\left(m^{(k)}_{j}\right)^{-1/2k}=\infty~, (93)

then the sequence of moments (mj(k))(m^{(k)}_{j}) uniquely defines the probability distribution (Thm. 15.11 of Ref. [99]). In the random phase ensemble, the complex Carleman condition is satisfied. This allows us to identify the temporal and random phase ensembles, if the kk-th no-resonance condition is satisfied.

In the other direction, if the kk-th no-resonance condition is not satisfied, the delta function in Eq. (87) will give additional contributions. Therefore, the kk-th moments of the temporal and random phase ensembles do not agree, giving the desired statement. ∎

Appendix B kk-th twirling identity

The above computation of ρTemp.(k)\rho^{(k)}_{\text{Temp.}} in Appendix A can be generalized into a statement about the twirling channel on any operator AA that acts on kk copies of the Hilbert space. Note, however, that unlike Theorem 1, we now require the full kk-th no-resonance conditions and cannot disregard degeneracies. As an illustration, we state and explicitly prove the result for k=2k=2. This identity originally appeared in the SM of Ref. [94].

Lemma 2 (Second Hamiltonian twirling identity).

If a Hamiltonian HH satisfies the second no-resonance condition, then for any operator AA acting on 2\mathcal{H}^{\otimes 2}, the following identity holds:

𝔼t[Ut2AUt2]=𝒟2[A]𝕀+𝒟2[A𝒮]𝒮E|EE|2A|EE|2,\displaystyle\mathbb{E}_{t}\left[U_{t}^{\otimes 2}AU_{-t}^{\otimes 2}\right]=\mathcal{D}^{\otimes 2}\big{[}A\big{]}\mathbb{I}+\mathcal{D}^{\otimes 2}\big{[}A\mathcal{S}\big{]}\mathcal{S}-\sum_{E}\outerproduct{E}{E}^{\otimes 2}A\outerproduct{E}{E}^{\otimes 2}~, (94)

where Utexp(iHt)U_{t}\equiv\exp(-iHt), and 𝔼t[]=limT1T0Tdt\mathbb{E}_{t}[\cdot]=\lim_{T\rightarrow\infty}\frac{1}{T}\int_{0}^{T}\cdot~dt is the infinite-time average, and |E\ket{E} are energy eigenstates of the Hamiltonian. 𝕀\mathbb{I} and 𝒮\mathcal{S} are the identity and swap operators on 2\mathcal{H}^{\otimes 2}, defined as 𝕀(|Ψ1|Ψ2)=|Ψ1|Ψ2\mathbb{I}(\ket{\Psi_{1}}\otimes\ket{\Psi_{2}})=\ket{\Psi_{1}}\otimes\ket{\Psi_{2}} and 𝒮(|Ψ1|Ψ2)=|Ψ2|Ψ1\mathcal{S}(\ket{\Psi_{1}}\otimes\ket{\Psi_{2}})=\ket{\Psi_{2}}\otimes\ket{\Psi_{1}} for any |Ψ1,|Ψ2\ket{\Psi_{1}},\ket{\Psi_{2}}. Finally, 𝒟2\mathcal{D}^{\otimes 2} is the dephasing channel in the energy eigenbasis: 𝒟2[A]E,E|E,EE,E|A|E,EE,E|\mathcal{D}^{\otimes 2}\big{[}A\big{]}\equiv\sum_{E,E^{\prime}}\outerproduct{E,E^{\prime}}{E,E^{\prime}}A\outerproduct{E,E^{\prime}}{E,E^{\prime}}.

Proof.
𝔼t[Ut2AUt2]=limT1T0T𝑑tUt2AUt2\displaystyle\mathbb{E}_{t}\left[U_{t}^{\otimes 2}AU_{-t}^{\otimes 2}\right]=\lim_{T\rightarrow\infty}\frac{1}{T}\int^{T}_{0}dt~U_{t}^{\otimes 2}AU_{-t}^{\otimes 2} (95)
=E1,E2,E3,E4[limT1T0T𝑑tei(E1E2+E3E4)t]|E1,E3E1,E3|A|E2,E4E2,E4|\displaystyle=\sum_{E_{1},E_{2},E_{3},E_{4}}\left[\lim_{T\rightarrow\infty}\frac{1}{T}\int^{T}_{0}dt~e^{-i(E_{1}-E_{2}+E_{3}-E_{4})t}\right]\outerproduct{E_{1},E_{3}}{E_{1},E_{3}}A\outerproduct{E_{2},E_{4}}{E_{2},E_{4}} (96)
=E1,E2,E3,E4δ(E1E2+E3E4)|E1,E3E1,E3|A|E2,E4E2,E4|\displaystyle=\sum_{E_{1},E_{2},E_{3},E_{4}}\delta(E_{1}-E_{2}+E_{3}-E_{4})~\outerproduct{E_{1},E_{3}}{E_{1},E_{3}}A\outerproduct{E_{2},E_{4}}{E_{2},E_{4}} (97)
=2-NRE,E[|E,EE,E|A|E,EE,E|+|E,EE,E|A|E,EE,E|]E|E,EE,E|A|E,EE,E|,\displaystyle\overset{\text{$2$-NR}}{=}\sum_{E,E^{\prime}}\bigg{[}\outerproduct{E,E^{\prime}}{E,E^{\prime}}A\outerproduct{E,E^{\prime}}{E,E^{\prime}}+\outerproduct{E,E^{\prime}}{E,E^{\prime}}A\outerproduct{E^{\prime},E}{E^{\prime},E}\bigg{]}-\sum_{E}\outerproduct{E,E}{E,E}A\outerproduct{E,E}{E,E}~, (98)

where “2-NR” indicates the use of the second no-resonance condition. The second no-resonance condition states that the delta function δ(E1E2+E3E4)\delta(E_{1}-E_{2}+E_{3}-E_{4}) is non-zero in only two cases: E1=E2,E3=E4E_{1}=E_{2},~E_{3}=E_{4} or E1=E4,E3=E2E_{1}=E_{4},~E_{3}=E_{2}, which give the two terms above. The final term in Eq. (98) corresponds to E1=E2=E3=E4E_{1}=E_{2}=E_{3}=E_{4}. This special case belongs to the above two terms simultaneously. It is double counted in the first two terms, hence is subtracted. ∎

Lemma 3 (kk-th Hamiltonian twirling identity).

If the Hamiltonian satisfies the kk-th no-resonance condition, its kk-th twirling channel acting on any operator AA on kk copies of the Hilbert space k\mathcal{H}^{\otimes k} is:

𝔼t[UtkAUtk]\displaystyle\mathbb{E}_{t}\left[U_{t}^{\otimes k}AU_{-t}^{\otimes k}\right] =M=(E1,,Ek)[σπ(M)|E1,,EkE1,,Ek|A|Eσ(1),,Eσ(k)Eσ(1),,Eσ(k)|],\displaystyle=\sum_{M=(E_{1},\cdots,E_{k})}\bigg{[}\sum_{\sigma\in\pi(M)}\outerproduct{E_{1},\cdots,E_{k}}{E_{1},\cdots,E_{k}}A\outerproduct{E_{\sigma(1)},\cdots,E_{\sigma(k)}}{E_{\sigma(1)},\cdots,E_{\sigma(k)}}\bigg{]}, (99)
=σSk𝒟k[APerm(σ1)]Perm(σ)+E.C.\displaystyle=\sum_{\sigma\in S_{k}}\mathcal{D}^{\otimes k}\big{[}A\text{Perm}(\sigma^{-1})\big{]}\text{Perm}(\sigma)+\mathrm{E.C.} (100)

where Perm(σ)\text{Perm}(\sigma)—with σSk\sigma\in S_{k}—is a permutation operator acting on the kk-copy Hilbert space:

Perm(σ){Ei}|E1,E2,,EkEσ(1),Eσ(2),,Eσ(k)|,\displaystyle\text{Perm}(\sigma)\equiv\sum_{\{E_{i}\}}\outerproduct*{E_{1},E_{2},\dots,E_{k}}{E_{\sigma(1)},E_{\sigma(2)},\dots,E_{\sigma(k)}}~, (101)

and 𝒟k\mathcal{D}^{\otimes k} is the dephasing channel in the energy basis, 𝒟k[A]{Ei}|{Ei}{Ei}|A|{Ei}{Ei}|\mathcal{D}^{\otimes k}\big{[}A\big{]}\equiv\sum_{\{E_{i}\}}\outerproduct*{\{E_{i}\}}{\{E_{i}\}}A\outerproduct*{\{E_{i}\}}{\{E_{i}\}}.

In (99), the sum is taken over multisets MM. If MM has rr unique indices with multiplicities n1,,nrn_{1},\cdots,n_{r} (such that i=1nnr=k\sum_{i=1}^{n}n_{r}=k), its group of permutations π(M)\pi(M) is a subgroup of the symmetric group SkS_{k} with k!/i=1nni!k!/\prod_{i=1}^{n}n_{i}! elements. In (100), we re-write Eq. (99) in terms of permutations, and additional terms E.C. associated with multiple counting (similar to the k=2k=2 case). At this level, it is difficult to provide a useful bound on the size of E.C. However, bounds may be proven in specific contexts, such as Theorem 3. All terms denoted “E.C.” can be explicitly derived from the twirling identities.

The proof of the kk-th twirling identity is analogous to that of the second twirling identity: in this case, there are k!k! permutations relevant for the kk-th no resonance theorem, along with more terms than the simplest one in the k=2k=2 case. As an example, for k=3k=3, in addition to the terms associated with cases where two energies are the same Ei=EjE_{i}=E_{j}, there is a higher order term arising when all three energies are the same: E1=E2=E3E_{1}=E_{2}=E_{3}.

Appendix C Entropy of temporal ensemble

In Section IV, we stated that the ensemble entropy of the temporal ensemble, a negative quantity which is the negative of its relative entropy to the Haar ensemble, is given by [Eq. (30)]

Ent(Temp.)=Elog(D|cE|2)+limδ0D(log(δ/2)1).\displaystyle\text{Ent}(\mathcal{E}_{\text{Temp.}})=\sum_{E}\log(D\absolutevalue{c_{E}}^{2})+\lim_{\delta\rightarrow 0}D\left(\log(\delta/2)-1\right)~. (102)

Here, we derive this result. Note that any ensemble in the constrained space can be described by the joint distribution P({θE})P(\{\theta_{E}\}). We emphasize that the ensemble entropy is the differential entropy of the distribution of states over Hilbert space. The ensemble entropy for any ensemble of states with constrained magnitudes |cE|2\absolutevalue{c_{E}}^{2} is formally negative infinite since they lie on a sub-dimensional (measure zero) manifold of the Hilbert space. However, we can still extract meaningful information from the ensemble entropy as follows. We may take the probability distribution P(Ψ)P(\Psi) on the constrained space as the limit of probability distributions with independent phase and magnitude components: P(Ψ)=P({|ΨE|})P({θE})P(\Psi)=P(\{\absolutevalue{\Psi_{E}}\})P(\{\theta_{E}\}). This separates the entropy into contributions from the phase and magnitude degrees of freedom:

𝒟(Haar)=𝑑ΨP(Ψ)log(P(Ψ)PHaar(Ψ))=Sθ+S|ΨE|2.\displaystyle-\mathcal{D}(\mathcal{E}\|\text{Haar})=-\int d\Psi~P(\Psi)\log\left(\frac{P(\Psi)}{P_{\text{Haar}}(\Psi)}\right)=S_{\theta}+S_{\absolutevalue{\Psi_{E}}^{2}}~. (103)

The phase contribution is

Sθ=E02π𝑑θEP({θE})log[(2π)DP({θE})],S_{\theta}=-\prod_{E}\int_{0}^{2\pi}d\theta_{E}P(\{\theta_{E}\})\log[(2\pi)^{D}P(\{\theta_{E}\})]~, (104)

which is uniquely maximized when the phases are uniformly distributed, taking value 0.

The magnitude component is divergent: However, if we replace the delta function with an integral over the thin shell |ΨE|2[|cE|2(1δ),|cE|2(1+δ)]\absolutevalue{\Psi_{E}}^{2}\in[\absolutevalue{c_{E}}^{2}(1-\delta),\absolutevalue{c_{E}}^{2}(1+\delta)], we obtain a constant contribution in addition to the divergent one.

S|ΨE|2\displaystyle S_{\absolutevalue{\Psi_{E}}^{2}} =limδ0E|cE|2(1δ)|cE|2(1+δ)d|ΨE|22|cE|2δlog(1/(|cE|2δ)2Dexp(D|ΨE|2))=limδ0[E(log(|cE|2))D|cE|2+Dlog(δD/2)]\displaystyle=-\lim_{\delta\rightarrow 0}\prod_{E}\int_{\absolutevalue{c_{E}}^{2}(1-\delta)}^{\absolutevalue{c_{E}}^{2}(1+\delta)}\frac{d\absolutevalue{\Psi_{E}}^{2}}{2\absolutevalue{c_{E}}^{2}\delta}\log(\frac{1/(\absolutevalue{c_{E}}^{2}\delta)}{2D\exp(-D\absolutevalue{\Psi_{E}}^{2})})=\lim_{\delta\rightarrow 0}\left[\sum_{E}(\log(\absolutevalue{c_{E}}^{2}))-D\absolutevalue{c_{E}}^{2}+D\log(\delta D/2)\right]
=Elog(|cE|2)+limδ0D(log(δD/2)1).\displaystyle=\sum_{E}\log(\absolutevalue{c_{E}}^{2})+\lim_{\delta\rightarrow 0}D\left(\log(\delta D/2)-1\right)~. (105)

We interpret the constant contribution to the entropy as the volume of Hilbert (sub)space spanned by the temporal ensemble. This subspace is in fact a hypertorus, with radii given by |cE|{\absolutevalue{c_{E}}}. In the limit of large Hilbert space dimension DD, the volume of this hypertorus is approximately E|cE|\prod_{E}\absolutevalue{c_{E}}. Therefore our measure of entropy is quantitatively equivalent to the volume of the Hilbert space.

Appendix D Proof of Theorem 2 (Asymptotic product formula)

Theorem 2.

The random phase ensemble asymptotically approaches the product form:

ρTemp.(k)=ρdkσSkPerm(σ)+O(tr(ρd2))\displaystyle\rho^{(k)}_{\text{Temp.}}=\rho_{d}^{\otimes k}\sum_{\sigma\in S_{k}}\text{Perm}(\sigma)+O(\tr(\rho_{d}^{2})) (106)

Here, ρdE|cE|2|EE|\rho_{d}\equiv\sum_{E}\absolutevalue{c_{E}}^{2}\outerproduct{E}{E} is the diagonal ensemble, a common construction in the thermalization literature [51, 52], and its purity tr(ρd2)\tr(\rho_{d}^{2}) is typically exponentially small in total system size.

Proof.

We want to show that:

ρTemp.(k)\displaystyle\rho^{(k)}_{\text{Temp.}} =ρdk𝜶Mk(D)σπ(𝜶)j=1k|EαjEσ(αj)|DρdkσSkPerm(σ)=ρdk𝜶Mk(D)σSkj=1k|EαjEσ(αj)|\displaystyle=\rho_{d}^{\otimes k}\sum_{\boldsymbol{\alpha}\in M_{k}(D)}\sum_{\sigma\in\pi(\boldsymbol{\alpha})}\bigotimes_{j=1}^{k}\outerproduct{E_{\alpha_{j}}}{E_{\sigma({\alpha_{j}})}}\overset{D\rightarrow\infty}{\longrightarrow}\rho_{d}^{\otimes k}\sum_{\sigma\in S_{k}}\text{Perm}(\sigma)=\rho_{d}^{\otimes k}\sum_{\boldsymbol{\alpha}\in M_{k}(D)}\sum_{\sigma\in S_{k}}\bigotimes_{j=1}^{k}\outerproduct{E_{\alpha_{j}}}{E_{\sigma({\alpha_{j}})}} (107)

Here, we derive the bound on the error

ρTemp.(k)ρdkσSkPerm(σ)k!2tr(ρd2),\|\rho^{(k)}_{\text{Temp.}}-\rho_{d}^{\otimes k}\sum_{\sigma\in S_{k}}\text{Perm}(\sigma)\|_{*}\leq\frac{k!}{2}\tr(\rho_{d}^{2}), (108)

where \|\cdot\|_{*} is the trace norm, or the sum of absolute values of the eigenvalues. To do so, we note that

ρdkσSkPerm(σ)ρTemp.(k)=ρdk[𝜶Mk(D)(σSkσπ(𝜶))j=1k|EαjEσ(αj)|]\rho_{d}^{\otimes k}\sum_{\sigma\in S_{k}}\text{Perm}(\sigma)-\rho^{(k)}_{\text{Temp.}}=\rho_{d}^{\otimes k}\left[\sum_{\boldsymbol{\alpha}\in M_{k}(D)}\left(\sum_{\sigma\in S_{k}}-\sum_{\sigma\in\pi(\boldsymbol{\alpha})}\right)\bigotimes_{j=1}^{k}\outerproduct{E_{\alpha_{j}}}{E_{\sigma({\alpha_{j}})}}\right] (109)

We then divide the Hilbert space k\mathcal{H}^{\otimes k} into classes ss of multisets 𝜶\boldsymbol{\alpha}, grouped by their degeneracy structure: the number of degenerate indices αj\alpha_{j} and their multiplicity. For example, s=()s=() indicates the class of multisets with no degeneracies, s=(2)s=(2) indicates the presence of 1 degeneracy of multiplicity 2, and s=(2,2)s=(2,2) indicates two degeneracies of multiplicity 2, respectively. The size of the permutation group π(𝜶)\pi(\boldsymbol{\alpha}) is uniquely determined by the class s(𝜶)s(\boldsymbol{\alpha}), and is a divisor of |Sk|=k!|S_{k}|=k!. Therefore,

ρdkσSkPerm(σ)ρTemp.(k)\displaystyle\rho_{d}^{\otimes k}\sum_{\sigma\in S_{k}}\text{Perm}(\sigma)-\rho^{(k)}_{\text{Temp.}} =ρdk[sMk(D)𝜶s(σSkσπ(𝜶))j=1k|EαjEσ(αj)|]\displaystyle=\rho_{d}^{\otimes k}\left[\sum_{s\subset M_{k}(D)}\sum_{\boldsymbol{\alpha}\in s}\left(\sum_{\sigma\in S_{k}}-\sum_{\sigma\in\pi(\boldsymbol{\alpha})}\right)\bigotimes_{j=1}^{k}\outerproduct{E_{\alpha_{j}}}{E_{\sigma({\alpha_{j}})}}\right] (110)
=sMk(D)ρdkPsσSkPerm(σ)\displaystyle=\sum_{s\subset M_{k}(D)}\rho_{d}^{\otimes k}P_{s}\sum_{\sigma\in S_{k}}\text{Perm}(\sigma) (111)

Where PsP_{s} is a projector onto the space of multisets in the class ss. Each term in the sum over ss, ρdkPsσSkPerm(σ)\rho_{d}^{\otimes k}P_{s}\sum_{\sigma\in S_{k}}\text{Perm}(\sigma),is positive semi-definite. We then use the triangle inequality to bound the trace distance, noting that the term from the class s=()s=() gives 0 contribution. In our sum (111) over multisets with at least one degeneracy, we can replace ρdk\rho_{d}^{\otimes k} with ρd(2)ρd(k2)\rho_{d}^{(2)}\otimes\rho_{d}^{\otimes(k-2)}, where ρd(2)E|cE|4|E,EE,E|\rho_{d}^{(2)}\equiv\sum_{E}\absolutevalue{c_{E}}^{4}\outerproduct{E,E}{E,E}. Finally, the projector PsP_{s} only decreases the trace distance and can be omitted.

ρdkσSkPerm(σ)ρTemp.(k)\displaystyle\|\rho_{d}^{\otimes k}\sum_{\sigma\in S_{k}}\text{Perm}(\sigma)-\rho^{(k)}_{\text{Temp.}}\|_{*} s()(ρd(2)ρd(k2))σSkPerm(σ)\displaystyle\leq\sum_{s\neq()}\|\left(\rho_{d}^{(2)}\otimes\rho_{d}^{\otimes(k-2)}\right)\sum_{\sigma\in S_{k}}\text{Perm}(\sigma)\|_{*} (112)
=s()tr[(ρd(2)ρd(k2))σSkPerm(σ)]\displaystyle=\sum_{s\neq()}\tr\left[\left(\rho_{d}^{(2)}\otimes\rho_{d}^{\otimes(k-2)}\right)\sum_{\sigma\in S_{k}}\text{Perm}(\sigma)\right] (113)
s()k!tr(ρd2)=(p(k)1)k!tr(ρd2)k!exp(π2k3)tr(ρd2)\displaystyle\leq\sum_{s\neq()}k!\tr(\rho_{d}^{2})=(p(k)-1)k!\tr(\rho_{d}^{2})\leq k!\exp(\pi\sqrt{\frac{2k}{3}})\tr(\rho_{d}^{2}) (114)

where p(k)p(k) is the partition number of kk (the number of possible classes ss), which is upper bounded by exp(π2k3)\exp(\pi\sqrt{\frac{2k}{3}}) [116]. In going from Eq. (113) to Eq. (114) we have used the fact that

tr[(ρd(2)ρd(k2))σSkPerm(σ)]=tr(ρd2)+(k1)tr(ρd3)+(k12)tr(ρd2)2+k!tr(ρd2),\tr\left[\left(\rho_{d}^{(2)}\otimes\rho_{d}^{\otimes(k-2)}\right)\sum_{\sigma\in S_{k}}\text{Perm}(\sigma)\right]=\tr(\rho_{d}^{2})+(k-1)\tr(\rho_{d}^{3})+\binom{k-1}{2}\tr(\rho_{d}^{2})^{2}+\cdots\leq k!\tr(\rho_{d}^{2})~, (115)

this bound is weak since we expect all terms to be exponentially smaller (in system size) than the first. This implies that the O(k!)O(k!) prefactor is likely not tight in terms of scaling with kk. However, we will typically restrict our attention to small values of kk, and are more interested in the factor of tr(ρd2)\tr(\rho_{d}^{2}). In typical systems, we expect tr(ρd2)D1\tr(\rho_{d}^{2})\sim D^{-1} to exponentially decrease as the size of the many-body system increases. ∎

Appendix E Convergence of finite-time temporal ensembles: Further numerical results

Refer to caption
Figure 8: Convergence of finite time ensembles. (a) Comparison of convergence of first (k=1k=1, squares) and second moment (k=2k=2, circles) in trace (green) and Frobenius distance (blue) for a temporal ensemble generated by a single, randomly drawn GUE Hamiltonian with Hilbert space dimension D=26D=2^{6}. (b) Convergence in Frobenius distance of the first three moments k=1,2,3k=1,2,3 (squares, circles, crosses) for single, randomly drawn GUE Hamiltonian with Hilbert space dimension D=26D=2^{6} (blue) and the MFIM with N=6N=6 spins (red). (c) Ensemble averaged Frobenius distance of the temporal ensembles generated by a randomly drawn GUE Hamiltonians in various Hilbert space dimension D=24,26,28D=2^{4},2^{6},2^{8}. Mean is takes over 10000 (D=24,26D=2^{4},2^{6}) and 2000 random instances (D=28D=2^{8}). Error bars denote two standard errors of the mean, 2000. Inset shows the median of the Frobenius distances of the same data. (d,e) Histogram of gaps and differences of gaps accumulated over 50000 randomly chosen GUE Hamiltonian with Hilbert space dimension D=24D=2^{4}. Indices i,j,k,li,j,k,l take all values in 1,,D1,\dots,D and we excluded the (trivially vanishing) diagonal elements i=ji=j in (d) and i=k,j=li=k,j=l and i=l,j=ki=l,j=k in (e). The insets enlarge the regions close to zero.

Here, we present additional numerical results investigating the convergence of finite time temporal ensembles discussed in Section IV.2.

In Fig. 8(a), we compare convergence of first and second moment of the finite time temporal ensemble in trace and Frobenius distance. We consider a single Hamiltonian drawn at random from the GUE ensemble in a system with Hilbert space dimension d=26d=2^{6}. For both, first and second moment, we observe the same behavior at long times: trace and Frobenius distance decay as τ1\tau^{-1}, where τ\tau is the interval length. In contrast, the intermediate τ1/2\tau^{-1/2} decay of the Frobenius distance and its pronounced cross-over to the late time τ1\tau^{-1} appears to be absent when considering the trace distance. This suggests, that while the early and intermediate time behavior of trace and Frobenius distance can differ, the Frobenius distance is a quantitative proxy for the trace distance at long times.

In Fig. 8(b), we complement the results of the main text by comparing convergence of the kk-th moments of the finite time temporal for different k=1,2,3k=1,2,3. We consider a single Hamiltonian drawn at random from the GUE ensemble in a system with Hilbert space dimension D=26D=2^{6} and the MFIM with N=6N=6 spins. For computational simplicity, we focus on the Frobenius norm. Independent of the value of kk, we observe a late time decay τ1\propto\tau^{-1}. As discussed in the main text [c.f. Eq. (34)], we attribute this to the presence of a finite minimal gap EiEjE_{i}-E_{j} (difference of gaps for k2k\leq 2) for an individual Hamiltonian and the late time decay of the sinc function.

In Fig. 8(c), we investigate the convergence of ensemble-averaged distances. Notably, we find that for the first moment, k=1k=1, the GUE ensemble mean closely resembles the behavior of individual instances and decays as τ1\tau^{-1} (not shown). For higher moments, k=2k=2, we find that, at early times, ensemble average distances decays as τ1/2\tau^{-1/2} whereas at very late times we observe a decay consistent with τ1\tau^{-1}. The cross-over between these two regimes is, however, less sharply pronounced as for single fixed Hamiltonians [as visualized by the ensemble median, representing a typical instance, in inset of Fig. 8(c)]. We relate these findings to statistical properties of the energy eigenlevels of GUE Hamiltonians. It is well known that the statistics of eigenvalues exhibits (strong) level repulsion, i.e., the probability to find two distinct eigenvalues close to each other EiEj0E_{i}-E_{j}\approx 0 is vanishing. In contrast, the numerical results presented Figs. 8(d, e) suggest the gaps EiEj+EkElE_{i}-E_{j}+E_{k}-E_{l} of eigenvalues exhibit only weak repulsion, i.e., the probability of finding two distinct gaps close to each other EiEj+EkEl0E_{i}-E_{j}+E_{k}-E_{l}\approx 0 is suppressed but finite. We leave a detailed investigation of such higher-order gap repulsion and its consequences for the late time decay of the ensemble-averaged Frobenius (and trace) distance to future study (see also Section F.1).

Appendix F Convergence of finite-time temporal ensembles: Analytical results from random matrix theory

Here we prove the τ1\tau^{-1} scaling of the convergence of the finite-time temporal ensemble for Hamiltonians from the Gaussian Unitary Ensemble (GUE), discussed in Section IV.2. We focus on the (ensemble-averaged) squared Frobenius norm F\|\cdot\|_{F} for its analytic tractability. Note that, since we average over random Hamiltonians HH, the initial state can be fixed to be an arbitrary state.

𝔼HGUE[ρTemp.(k)(τ)ρTemp.(k)F2]\displaystyle\mathbb{E}_{H\sim\text{GUE}}\left[\|\rho^{(k)}_{\text{Temp.}}(\tau)-\rho^{(k)}_{\text{Temp.}}\|_{F}^{2}\right] =𝔼HGUE[tr[(ρTemp.(k)(τ))2]2tr[ρTemp.(k)(τ)ρTemp.(k)]+tr[(ρTemp.(k))2]]\displaystyle=\mathbb{E}_{H\sim\text{GUE}}\left[\tr[\left(\rho^{(k)}_{\text{Temp.}}(\tau)\right)^{2}]-2\tr[\rho^{(k)}_{\text{Temp.}}(\tau)\rho^{(k)}_{\text{Temp.}}]+\tr[\left(\rho^{(k)}_{\text{Temp.}}\right)^{2}]\right] (116)
=𝔼HGUE[tr[(ρTemp.(k)(τ))2]tr[(ρTemp.(k))2]]\displaystyle=\mathbb{E}_{H\sim\text{GUE}}\left[\tr[\left(\rho^{(k)}_{\text{Temp.}}(\tau)\right)^{2}]-\tr[\left(\rho^{(k)}_{\text{Temp.}}\right)^{2}]\right] (117)

where we used that the long time limit average moment does not evolve in time.

tr[ρTemp.(k)(τ)ρTemp.(k)]=tr[ττdt2τ(eiHt|Ψ0Ψ0|eiHt)kρTemp.(k)]=Ψ0|kρTemp.(k)|Ψ0k=tr[ρTemp.(k),2].\displaystyle\tr[\rho^{(k)}_{\text{Temp.}}(\tau)\rho^{(k)}_{\text{Temp.}}]=\tr[\int_{-\tau}^{\tau}\frac{\text{d}t}{2\tau}\left(e^{-iHt}\ket{\Psi_{0}}\bra{\Psi_{0}}e^{iHt}\right)^{\otimes k}\rho^{(k)}_{\text{Temp.}}]=\bra{\Psi_{0}}^{\otimes k}\rho^{(k)}_{\text{Temp.}}\ket{\Psi_{0}}^{\otimes k}=\tr[\rho^{(k),2}_{\text{Temp.}}]. (118)

To carry out the average over the GUE, we decompose H=UΛUH=U\Lambda U^{\dagger} with Λ\Lambda a diagonal matrix of eigenvalues of HH and UU the unitary matrix formed by its eigenvectors. Importantly, for HH from GUE, Λ\Lambda and UU are statistically independent and UU is a random unitary matrix distributed according to the Haar measure on the unitary group. We can thus evaluate the average over the eigenvectors UU separately using Weingarten calculus. We find for the first term

𝔼HGUE[tr[ρTemp.(k),2(τ)]]=\displaystyle\mathbb{E}_{H\sim\text{GUE}}\left[\tr[\rho^{(k),2}_{\text{Temp.}}(\tau)]\right]= 14τ202τdt¯02t¯dtdΛP(Λ)dUtr[|Ψ0Ψ0|2kU2k(eiΛte+iΛt)kU,2k]\displaystyle\frac{1}{4\tau^{2}}\int_{0}^{2\tau}\!\!\text{d}\bar{t}\int_{0}^{2\bar{t}}\!\text{d}t^{\prime}\int\!\text{d}\Lambda\,P(\Lambda)\int\!\text{d}U\tr\left[\ket{\Psi_{0}}\bra{\Psi_{0}}^{\otimes 2k}U^{\otimes 2k}\left(e^{-i\Lambda t^{\prime}}\otimes e^{+i\Lambda t^{\prime}}\right)^{\otimes k}U^{{\dagger},\otimes 2k}\right]
=\displaystyle= 14τ202τdt¯02t¯dt1D(D+1)(D+k1)dΛP(Λ)πS2ktr[Wπ(eiΛte+iΛt)k].\displaystyle\frac{1}{4\tau^{2}}\int_{0}^{2\tau}\!\!\text{d}\bar{t}\int_{0}^{2\bar{t}}\!\text{d}t^{\prime}\frac{1}{D(D+1)\cdots(D+k-1)}\int\!\text{d}\Lambda\,P(\Lambda)\sum_{\pi\in S_{2k}}\tr\left[W_{\pi}\left(e^{-i\Lambda t^{\prime}}\otimes e^{+i\Lambda t^{\prime}}\right)^{\otimes k}\right]. (119)

where P(Λ)P(\Lambda) denotes the joint probability distribution of the eigenvalues of Hamiltonians HH from GUE, dU\int\!\text{d}U denotes the Haar integral of the unitary group, WπW_{\pi} with πS2k\pi\in S_{2k} are 2k2k permutation operators acting on the 2k2k-copy space and S2kS_{2k} denotes the symmetric group of order 2k2k. In the case k=1k=1, we can simplify the expressions further. We find

P(Λ)dΛπS2tr[Wπ(eiΛτe+iΛτ)]=1D(D+1)[K(τ)+D]\displaystyle\int\!P(\Lambda)\text{d}\Lambda\sum_{\pi\in S_{2}}\tr\left[W_{\pi}\left(e^{-i\Lambda\tau}\otimes e^{+i\Lambda\tau}\right)\right]=\frac{1}{D(D+1)}\left[K(\tau)+D\right]

where

K(τ)=𝔼HGUE[|tr(eiHτ)|2]=\displaystyle K(\tau)=\mathbb{E}_{H\sim\text{GUE}}\left[\absolutevalue{\tr(e^{-iH\tau})}^{2}\right]= P(Λ)dΛ|tr(eiΛτ)|2\displaystyle\int\!P(\Lambda)\text{d}\Lambda~\absolutevalue{\tr(e^{-i\Lambda\tau})}^{2}

is the spectral form factor averaged over Hamiltonians HH. Furthermore, ρH(1)=E|E|Ψ0|2|EE|\rho_{H}^{(1)}=\sum_{E}\absolutevalue{\innerproduct{E}{\Psi_{0}}}^{2}\outerproduct{E}{E} with {|E}\{\ket{E}\} the eigenvectors of HH. Thus, we have tr[ϱH(1),2]=E|E|Ψ0|4\tr[\varrho_{H}^{(1),2}]=\sum_{E}\absolutevalue{\innerproduct{E}{\Psi_{0}}}^{4}, which averaged over HH from GUE,

𝔼HGUE[E|E|Ψ0|4]=dU[i=1D|i|U|Ψ0|4]=2D+1\displaystyle\mathbb{E}_{H\sim\text{GUE}}\left[\sum_{E}|\innerproduct{E}{\Psi_{0}}|^{4}\right]=\int\!\text{d}U\left[\sum_{i=1}^{D}|\innerproduct{i|U}{\Psi_{0}}|^{4}\right]=\frac{2}{D+1} (120)

Finally, we obtain

𝔼HGUE[ρH(1)(τ)ρH(1)F2]=14τ202τdt¯02t¯dtK(t)DD(D+1)Dτ2+𝒪(τ4)\displaystyle\mathbb{E}_{H\sim\text{GUE}}\left[\|\rho^{(1)}_{H}(\tau)-\rho^{(1)}_{H}\|_{F}^{2}\right]=\frac{1}{4\tau^{2}}\int_{0}^{2\tau}\!\!\text{d}\bar{t}\int_{0}^{2\bar{t}}\!\text{d}t^{\prime}\frac{K(t^{\prime})-D}{D(D+1)}\sim\frac{D}{\tau^{2}}+\mathcal{O}(\tau^{-4}) (121)

Using the improved approximation [117] for the spectral form factor K(τ)K(\tau) and an asymptotic expansion of the occuring Bessel function [118] gives a τ2\tau^{-2} scaling of the squared Frobenius norm. We note that if the standard box approximation is used, Eq. (121) does not converge to 0. As we detail below, an alternative approach is to compute the convergence for the GUE ensemble for a finite dimension DD. The finite DD expressions, in fact, do not require any regularization, and the large DD limit can be safely taken on the final result.

F.1 Convergence for GUE ensemble and finite dimension

Here we show how to study the convergence to the temporal ensemble for Hamiltonians from GUE ensembles of finite dimension DD. These finite DD expressions do not suffer from the problems of the infinite-DD spectral form factor, which has to be properly regularized. Therefore, we can compute the distance from the temporal ensemble for finite DD, and take the DD\rightarrow\infty limit at the end.

We are interested in the average of the following distance for HH belonging to the GUE ensemble:

Tr[(ρTemp.(k)(τ)ρTemp.(k)())2]={αj}{βj}Perm({αj})sinc2(j(EαjEβj)τ)j|cαj|2|cβj|2.\mathrm{Tr}[(\rho^{(k)}_{\mathrm{Temp.}}(\tau)-\rho^{(k)}_{\mathrm{Temp.}}(\infty))^{2}]=\sum_{\{\alpha_{j}\}}\;\sum_{\{\beta_{j}\}\notin\mathrm{Perm}(\{\alpha_{j}\})}\mathrm{sinc}^{2}\left(\sum_{j}(E_{\alpha_{j}}-E_{\beta_{j}})\tau\right)\prod_{j}|c_{\alpha_{j}}|^{2}|c_{\beta_{j}}|^{2}. (122)

Note that the sum is over the sequences of indices {αj},{βj}\{\alpha_{j}\},\{\beta_{j}\} under the condition that the two sequences are not related by a permutation. Let us consider, as a first example, the case k=1k=1. We have:

𝔼HGUE[Tr[(ρTemp.(1)(τ)ρTemp.(1)())2]]=iji𝔼HGUE[sinc2((EiEj)τ)]𝔼HGUE[|ci|2|cj|2],\mathbb{E}_{H\sim\text{GUE}}\left[\mathrm{Tr}[(\rho^{(1)}_{\mathrm{Temp.}}(\tau)-\rho^{(1)}_{\mathrm{Temp.}}(\infty))^{2}]\right]=\sum_{i}\sum_{j\neq i}\mathbb{E}_{H\sim\text{GUE}}\left[\mathrm{sinc}^{2}\left((E_{i}-E_{j})\tau\right)\right]\mathbb{E}_{H\sim\text{GUE}}\left[|c_{i}|^{2}|c_{j}|^{2}\right], (123)

where we used that the eigenvalues and the eigenvectors of HH are independently distributed.

For the eigenvectors we get:

𝔼HGUE[|ci|2|cj|2]=1D(D+1).\mathbb{E}_{H\sim\text{GUE}}\left[|c_{i}|^{2}|c_{j}|^{2}\right]=\frac{1}{D(D+1)}. (124)

The convergence in time depends on the following average over the distribution of the eigenvalues:

𝔼HGUE[sinc2((EiEj)τ)]\displaystyle\mathbb{E}_{H\sim\text{GUE}}\left[\mathrm{sinc}^{2}\left((E_{i}-E_{j})\tau\right)\right] =dE1dEDsinc2[(E1E2)τ]pD(E1,,ED)\displaystyle=\int\mathrm{d}E_{1}\dots\mathrm{d}E_{D}\,\mathrm{sinc}^{2}\left[(E_{1}-E_{2})\tau\right]p^{D}(E_{1},\dots,E_{D})
=dusinc2(uτ)q1|1D(u).\displaystyle=\int\mathrm{d}u\,\mathrm{sinc}^{2}\left(u\tau\right)q_{1|1}^{D}(u). (125)

Here pD(E1,,ED)p^{D}(E_{1},\dots,E_{D}) is the joint probability distribution of the (unordered) eigenvalues E1,EDE_{1},\dots E_{D}, and we defined q1|1D(u)q_{1|1}^{D}(u) as the probability distribution of the energy difference E1E2E_{1}-E_{2}:

q1|1D(u)=dE1dEDδ(E1E2u)pD(E1,,ED).q_{1|1}^{D}(u)=\int\int\mathrm{d}E_{1}\dots\mathrm{d}E_{D}\,\delta(E_{1}-E_{2}-u)p^{D}(E_{1},\dots,E_{D}). (126)

Hence we get:

𝔼HGUE[Tr[(ρTemp.(1)(τ)ρTemp.(1)())2]]=D1D+1dusinc2(uτ)q1|1D(u).\mathbb{E}_{H\sim\text{GUE}}\left[\mathrm{Tr}[(\rho^{(1)}_{\mathrm{Temp.}}(\tau)-\rho^{(1)}_{\mathrm{Temp.}}(\infty))^{2}]\right]=\frac{D-1}{D+1}\int\mathrm{d}u\,\mathrm{sinc}^{2}\left(u\tau\right)q_{1|1}^{D}(u). (127)

Before evaluating this integral, let us consider the case k=2k=2. By enumerating the possible sequences of indices, we get

Tr[(ρTemp.(2)(τ)ρTemp.(2)())2]\displaystyle\mathrm{Tr}[(\rho^{(2)}_{\mathrm{Temp.}}(\tau)-\rho^{(2)}_{\mathrm{Temp.}}(\infty))^{2}] =D!(D4)!𝔼HGUE[sinc2((E1+E2E3E4)τ)]𝔼HGUE[|c1|2|c2|2|c3|2|c4|2]\displaystyle=\frac{D!}{(D-4)!}\mathbb{E}_{H\sim\text{GUE}}\left[\mathrm{sinc}^{2}\left((E_{1}+E_{2}-E_{3}-E_{4})\tau\right)\right]\mathbb{E}_{H\sim\text{GUE}}\left[|c_{1}|^{2}|c_{2}|^{2}|c_{3}|^{2}|c_{4}|^{2}\right]
+2D!(D3)!𝔼HGUE[sinc2((E1+E22E3)τ)]𝔼HGUE[|c1|2|c2|2|c3|4]\displaystyle+2\frac{D!}{(D-3)!}\mathbb{E}_{H\sim\text{GUE}}\left[\mathrm{sinc}^{2}\left((E_{1}+E_{2}-2E_{3})\tau\right)\right]\mathbb{E}_{H\sim\text{GUE}}\left[|c_{1}|^{2}|c_{2}|^{2}|c_{3}|^{4}\right]
+4D!(D3)!𝔼HGUE[sinc2((E1E2)τ)]𝔼HGUE[|c1|4|c2|4]\displaystyle+4\frac{D!}{(D-3)!}\mathbb{E}_{H\sim\text{GUE}}\left[\mathrm{sinc}^{2}\left((E_{1}-E_{2})\tau\right)\right]\mathbb{E}_{H\sim\text{GUE}}\left[|c_{1}|^{4}|c_{2}|^{4}\right] (128)
+D(D1)𝔼HGUE[sinc2((2E12E2)τ)]𝔼HGUE[|c1|4|c2|4]\displaystyle+D(D-1)\mathbb{E}_{H\sim\text{GUE}}\left[\mathrm{sinc}^{2}\left((2E_{1}-2E_{2})\tau\right)\right]\mathbb{E}_{H\sim\text{GUE}}\left[|c_{1}|^{4}|c_{2}|^{4}\right]
+4D(D1)𝔼HGUE[sinc2((E1E2)τ)]𝔼HGUE[|c1|2|c2|6].\displaystyle+4D(D-1)\mathbb{E}_{H\sim\text{GUE}}\left[\mathrm{sinc}^{2}\left((E_{1}-E_{2})\tau\right)\right]\mathbb{E}_{H\sim\text{GUE}}\left[|c_{1}|^{2}|c_{2}|^{6}\right].

Similar expressions can be obtained for generic kk-th moments: in general, to study the time dependence we need to evaluate the probability distribution of linear combinations of energy eigenstates, i.e., quantities of the form

qm1ms|n1ntD(u)dE1dEDδ[(i=1smiEi)(j=1tnjEs+j)u]pD(E1,,ED).q_{m_{1}\dots m_{s}|n_{1}\dots n_{t}}^{D}(u)\equiv\int\mathrm{d}E_{1}\dots\mathrm{d}E_{D}\,\delta\left[\left(\sum_{i=1}^{s}m_{i}E_{i}\right)-\left(\sum_{j=1}^{t}n_{j}E_{s+j}\right)-u\right]p^{D}(E_{1},\dots,E_{D}). (129)

The probability distribution qm1ms|n1ntD(u)q_{m_{1}\dots m_{s}|n_{1}\dots n_{t}}^{D}(u) is then used to compute the following integral

Fm1ms|n1ntD(τ)\displaystyle F^{D}_{m_{1}\dots m_{s}|n_{1}\dots n_{t}}(\tau) =dE1dEDsinc2[(i=1smiEi)τ(j=1tnjEs+j)τ]pD(E1,,ED)\displaystyle=\int\mathrm{d}E_{1}\dots\mathrm{d}E_{D}\,\mathrm{sinc}^{2}\left[\left(\sum_{i=1}^{s}m_{i}E_{i}\right)\tau-\left(\sum_{j=1}^{t}n_{j}E_{s+j}\right)\tau\right]p^{D}(E_{1},\dots,E_{D}) (130)
=dusinc2(uτ)qm1ms|n1ntD(u).\displaystyle=\int\mathrm{d}u\,\mathrm{sinc}^{2}\left(u\tau\right)q_{m_{1}\dots m_{s}|n_{1}\dots n_{t}}^{D}(u). (131)

F.1.1 Evaluation of the integral

To evaluate the integral in Eq. (131) we use the following property: the probability distribution qm1ms|n1ntD(u)q_{m_{1}\dots m_{s}|n_{1}\dots n_{t}}^{D}(u) has, in general, the expression

qm1ms|n1ntD(u)=eγu2Pm1ms|n1ntD(γu2),q_{m_{1}\dots m_{s}|n_{1}\dots n_{t}}^{D}(u)=e^{-\gamma u^{2}}P_{m_{1}\dots m_{s}|n_{1}\dots n_{t}}^{D}(\gamma u^{2}), (132)

where

γ=D2[i=1smi2+j=1tnj2]1,\gamma=\frac{D}{2}\left[\sum_{i=1}^{s}m_{i}^{2}+\sum_{j=1}^{t}n_{j}^{2}\right]^{-1}, (133)

and Pm1ms|n1ntD(x)P_{m_{1}\dots m_{s}|n_{1}\dots n_{t}}^{D}(x) is a polynomial in xx. This expression can be obtained from the joint probability distribution of NN levels, which can be computed with standard random matrix theory techniques [119]:

pND(E1,EN)dEN+1dEDpD(E1,ED)=(DN)!D!det(KD(Ei,Ej))|i,j=1N.p^{D}_{N}(E_{1},\dots E_{N})\equiv\int\mathrm{d}E_{N+1}\dots\mathrm{d}E_{D}\,p^{D}(E_{1},\dots E_{D})=\frac{(D-N)!}{D!}\mathrm{det}\left(K_{D}(E_{i},E_{j})\right)|_{i,j=1}^{N}. (134)

Here the kernel KD(x,y)K_{D}(x,y) is defined as

KD(x,y)=eD4(x2+y2)j=0N1Hj(xD/2)Hj(yD/2)2πD2jj!,K_{D}(x,y)=e^{-\frac{D}{4}(x^{2}+y^{2})}\sum_{j=0}^{N-1}\frac{H_{j}(x\sqrt{D/2})H_{j}(y\sqrt{D/2})}{\sqrt{2\pi D}2^{j}j!}, (135)

where Hj(x)H_{j}(x) is the jj-th Hermite polynomial. Using this expression, through a change of variables we can derive in general an expression of the form in Eq. (132).

We can then evaluate the integral for each term of the polynomial Pm1ms|n1ntD(x)P_{m_{1}\dots m_{s}|n_{1}\dots n_{t}}^{D}(x) as follows:

dusinc2(uτ)eγu2(γu2)n={γ1/2τ2(π22n1(2n3)!!+eτ2/γO(γ1τ2(n1))), for n1πτ[πErf(γ1/2τ)γ1/2τ(1eτ2/γ)]. for n=0\int\mathrm{d}u\,\mathrm{sinc}^{2}\left(u\tau\right)e^{-\gamma u^{2}}(\gamma u^{2})^{n}=\begin{dcases}\frac{\gamma^{1/2}}{\tau^{2}}\left(\frac{\sqrt{\pi}}{2^{2n-1}}(2n-3)!!+e^{-\tau^{2}/\gamma}O(\gamma^{-1}\tau^{2}(n-1))\right),&\text{\;for }n\geq 1\\ \frac{\sqrt{\pi}}{\tau}\left[\sqrt{\pi}\mathrm{Erf}(\gamma^{-1/2}\tau)-\frac{\gamma^{1/2}}{\tau}(1-e^{-\tau^{2}/\gamma})\right].&\text{\;for }n=0\end{dcases} (136)

The leading term at long time hence has a power-law dependence 1/τ21/\tau^{2} if Pm1ms|n1ntD(0)=0P_{m_{1}\dots m_{s}|n_{1}\dots n_{t}}^{D}(0)=0 and 1/τ1/\tau if Pm1ms|n1ntD(0)0P_{m_{1}\dots m_{s}|n_{1}\dots n_{t}}^{D}(0)\neq 0. For the case k=1k=1, it is possible to prove (e.g., from an explicit calculation) that P1|1D(0)=0P_{1|1}^{D}(0)=0: this is the effect of level repulsion, as observed in Fig. 8(d). For generic Pm1ms|n1ntD(x)P_{m_{1}\dots m_{s}|n_{1}\dots n_{t}}^{D}(x), we do not expect similar vanishing for x=0x=0 [Fig. 8(e)]. Therefore we get the following scaling at long times τ\tau:

𝔼HGUE[ρH(k)(τ)ρH(k)F2]{1/τ2for k=1,1/τfor k>1.\mathbb{E}_{H\sim\text{GUE}}\left[\|\rho^{(k)}_{H}(\tau)-\rho^{(k)}_{H}\|_{F}^{2}\right]\propto\begin{dcases}1/\tau^{2}\qquad\text{for }k=1,\\ 1/\tau\qquad\text{for }k>1.\end{dcases} (137)

Appendix G Finite-time temporal mutual information

In Sec. IV.6.1 we derived an expression for the mutual information of temporal ensembles of finite intervals τ\tau: I(Z;T)1γπ2σHτI(Z;T)\approx 1-\gamma-\frac{\sqrt{\pi}}{2\sigma_{H}\tau}, which we provide a full derivation of here.

G.1 Bitstring correlation time

We will first introdyce the “bitstring correlation time” tct_{c}, which characterizes the time scale over which the bitstring distribution p(z,t)p(z,t) becomes independent (after rescaling their time-averaged values pavg(z)p_{\text{avg}}(z)). That is, how long does it take to form a new, independent speckle pattern? We will show that tc=σH1t_{c}=\sigma_{H}^{-1}, where σH2=tr(H2)tr(H)2\sigma_{H}^{2}=\tr(H^{2})-\tr(H)^{2}. To do this, we compute the correlator:

Az(t)\displaystyle A_{z}(t) =𝑑tp(z,t)p(z,t+t)[𝑑tp(z,t)]21=|E|z|E|2|E|Ψ0|2exp(iEt)|2|E|z|E|2|E|Ψ0|2|2+,\displaystyle=\frac{\int dt^{\prime}~p(z,t^{\prime})p(z,t^{\prime}+t)}{\left[\int dt^{\prime}p(z,t^{\prime})\right]^{2}}-1=\frac{\absolutevalue{\sum_{E}\absolutevalue{\innerproduct{z}{E}}^{2}\absolutevalue{\innerproduct{E}{\Psi_{0}}}^{2}\exp(-iEt)}^{2}}{\absolutevalue{\sum_{E}\absolutevalue{\innerproduct{z}{E}}^{2}\absolutevalue{\innerproduct{E}{\Psi_{0}}}^{2}}^{2}}+\cdots~, (138)

where we use “\cdots” to denote subleading terms. In order to make this calculation tractable, we make the following approximations:

E12πσH2𝑑Eexp((EE0)22σH2),\displaystyle\sum_{E}\rightarrow\frac{1}{\sqrt{2\pi\sigma_{H}^{2}}}\int_{-\infty}^{\infty}dE\exp(-\frac{(E-E_{0})^{2}}{2\sigma_{H}^{2}})~, (139)
|z|E|2exp(βzE)/Zz,\displaystyle\absolutevalue{\innerproduct{z}{E}}^{2}\rightarrow\exp(-\beta_{z}E)/Z_{z}~, (140)
|E|Ψ0|2exp(βΨE)/ZΨ,\displaystyle\absolutevalue{\innerproduct{E}{\Psi_{0}}}^{2}\rightarrow\exp(-\beta_{\Psi}E)/Z_{\Psi}~, (141)

where E0=tr(H)E_{0}=\tr(H), βz,βE\beta_{z},\beta_{E} are appropriate inverse temperatures of the state |z\ket{z} and |Ψ0\ket{\Psi_{0}}, and Zz,ZΨZ_{z},Z_{\Psi} are their corresponding partition functions. The first approximation takes the density of states of the spectrum to be a Gaussian [120], while the second and third approximations treat the energy populations of the states |z\ket{z} and |Ψ0\ket{\Psi_{0}} as Boltzmann distributions (Section VII). We obtain

E|z|E|2|E|Ψ0|2exp(iEt)\displaystyle\sum_{E}\absolutevalue{\innerproduct{z}{E}}^{2}\absolutevalue{\innerproduct{E}{\Psi_{0}}}^{2}\exp(-iEt) 12πσH2𝑑Eexp((EE0)22σH2)exp((βz+βE+it)E)\displaystyle\rightarrow\frac{1}{\sqrt{2\pi\sigma_{H}^{2}}}\int_{-\infty}^{\infty}dE\exp(-\frac{(E-E_{0})^{2}}{2\sigma_{H}^{2}})\exp(-(\beta_{z}+\beta_{E}+it)E) (142)
=exp((βz+βE+it)2σH2/2)\displaystyle=\exp((\beta_{z}+\beta_{E}+it)^{2}\sigma_{H}^{2}/2) (143)

Taking the ratio of the absolute values of Eq. (143) at nonzero tt and t=0t=0 gives the desired result.

G.2 Temporal mutual information

We can then derive the observed scaling of the mutual information

I(Z;T)=H(Z)H(Z|T)\displaystyle I(Z;T)=H(Z)-H(Z|T) =zp¯[t,t+τ](z)log2(p¯[t,t+τ](z))+tt+τ𝑑t1τzpt(z)log2(pt(z)),\displaystyle=-\sum_{z}\bar{p}_{[t,t+\tau]}(z)\log_{2}(\bar{p}_{[t,t+\tau]}(z))+\int_{t}^{t+\tau}dt^{\prime}\frac{1}{\tau}\sum_{z}p_{t^{\prime}}(z)\log_{2}(p_{t^{\prime}}(z))~, (144)

where p¯[t,t+τ](z)=τ1tt+τ𝑑tpt(z)\bar{p}_{[t,t+\tau]}(z)=\tau^{-1}\int_{t}^{t+\tau}dt^{\prime}p_{t^{\prime}}(z) is the outcome probability averaged over an interval. Using the facts that: (i) pt(z)pavg(z)p~p_{t^{\prime}}(z)\approx p_{\text{avg}}(z)\tilde{p}, where p~PT(1)\tilde{p}\sim\text{PT}(1), and (ii) the bitstring correlation time is tc=σH1t_{c}=\sigma_{H}^{-1}, we approximate p¯/pavg\bar{p}/p_{\text{avg}} to be the average of n=τ/tcn=\tau/t_{c} independent PT variables. The average of nn Porter-Thomas (exponential) variables follows an Erlang distribution with mean 1 [41]. Following convention, we denote this distribution as p¯/pavgErlang(1/n,n)\bar{p}/p_{\text{avg}}\sim\text{Erlang}(1/n,n), with PDF

PErlang(1/n,n)(x)dx=(nx)n1exp(nx)(n1)!ndx.P_{\text{Erlang}(1/n,n)}(x)dx=\frac{(nx)^{n-1}\exp(-nx)}{(n-1)!}ndx~. (145)

This approximation allows us to calculate the entropy of p¯\bar{p}:

zp¯[t,t+τ](z)log2(p¯[t,t+τ](z))𝔼[p¯]zpavg(z)log2(pavg(z))𝔼[p¯log2(p¯)]zpavg(z)\displaystyle-\sum_{z}\bar{p}_{[t,t+\tau]}(z)\log_{2}(\bar{p}_{[t,t+\tau]}(z))\approx-\mathbb{E}[\bar{p}]\sum_{z}p_{\text{avg}}(z)\log_{2}(p_{\text{avg}}(z))-\mathbb{E}[\bar{p}\log_{2}(\bar{p})]\sum_{z}p_{\text{avg}}(z) (146)
=Hpavg+log2(n)Ψ(n+1)n(Hpavg12n)/ln2=(Hpavg12σHτ)/ln2\displaystyle=H_{p_{\text{avg}}}+\log_{2}(n)-\Psi(n+1)\overset{n\rightarrow\infty}{\longrightarrow}(H_{p_{\text{avg}}}-\frac{1}{2n})/\ln 2=\left(H_{p_{\text{avg}}}-\frac{1}{2\sigma_{H}\tau}\right)/\ln 2 (147)

where Ψ(x)\Psi(x) is the digamma function, related to the Gamma function Γ(x)\Gamma(x) by Ψ(x)=Γ(x)/Γ(x)\Psi(x)=\Gamma^{\prime}(x)/\Gamma(x). A more careful analysis modifies the effective number nn of the Erlang distribution by considering the second moment of p¯\bar{p}:

𝔼t[p¯(z,t)2]\displaystyle\mathbb{E}_{t}[\bar{p}(z,t)^{2}] =1τ20τ𝑑t10τ𝑑t2p(z,t1)p(z,t2)\displaystyle=\frac{1}{\tau^{2}}\int_{0}^{\tau}dt_{1}\int_{0}^{\tau}dt_{2}p(z,t_{1})p(z,t_{2}) (148)
1+1τ𝑑texp(σH2t2)=1+πσHτ1+1neff,\displaystyle\approx 1+\frac{1}{\tau}\int_{-\infty}^{\infty}dt^{\prime}\exp(-\sigma_{H}^{2}t^{\prime 2})=1+\frac{\sqrt{\pi}}{\sigma_{H}\tau}\equiv 1+\frac{1}{n_{\text{eff}}}~, (149)

i.e. the effective parameter neffn_{\text{eff}} of the Erlang distribution is neff=σHτ/πn_{\text{eff}}=\sigma_{H}\tau/\sqrt{\pi}. This gives the overall temporal mutual information for a finite time interval:

I(Z;T)(1γπ2σHτ)/ln2,\displaystyle I(Z;T)\approx\left(1-\gamma-\frac{\sqrt{\pi}}{2\sigma_{H}\tau}\right)/\ln 2~, (150)

which agrees with numerical results [inset of Fig. 3(d)].

Appendix H kk-th moments of the Scrooge ensemble

Here, we derive our expression for the kk-th moments of the Scrooge ensemble [Eq. (50)] Scrooge[ρ]\text{Scrooge}[\rho]:

ρScr.,𝒎(k)=(iλi1)kΛkμm1μmk|μi=λi1,\rho^{(k)}_{\text{Scr.},\boldsymbol{m}}=\left(\prod_{i}\lambda^{-1}_{i}\right)\frac{\partial^{k}\Lambda_{k}}{\partial\mu_{m_{1}}\dots\partial\mu_{m_{k}}}{\Bigg{|}}_{\mu_{i}=\lambda_{i}^{-1}}~, (151)

where

Λk\displaystyle\Lambda_{k} j=1Dμjk2lnμjkj(μjμk),\displaystyle\equiv\sum_{j=1}^{D}\frac{\mu^{k-2}_{j}\ln\mu_{j}}{\prod_{k\neq j}(\mu_{j}-\mu_{k})}, (152)

and {λi}\{\lambda_{i}\} are the eigenvalues of the given state ρ\rho.

As an explicit example, we consider the simplest non-trivial parameters, k=2k=2 and D=2D=2. Let the eigenstates of ρ\rho be ||\uparrow\rangle and ||\downarrow\rangle, and the eigenvalues be λ,1λ\lambda,1-\lambda. Here, the second moment ρ(2)\rho^{(2)} is a 4×44\times 4 matrix. However, in the (doubled) eigenbasis of ρ\rho, only 6 entries are non-zero: |ρ(2)|ρScr.,11(2)\langle\uparrow\uparrow|\rho^{(2)}|\uparrow\uparrow\rangle\equiv\rho^{(2)}_{\text{Scr.},11}, |ρ(2)|=|ρ(2)|=|ρ(2)|=|ρ(2)|ρScr.,12(2)\langle\uparrow\downarrow|\rho^{(2)}|\uparrow\downarrow\rangle=\langle\uparrow\downarrow|\rho^{(2)}|\downarrow\uparrow\rangle=\langle\downarrow\uparrow|\rho^{(2)}|\uparrow\downarrow\rangle=\langle\downarrow\uparrow|\rho^{(2)}|\downarrow\uparrow\rangle\equiv\rho^{(2)}_{\text{Scr.},12}, and |ρ(2)|ρScr.,22(2)\langle\downarrow\downarrow|\rho^{(2)}|\downarrow\downarrow\rangle\equiv\rho^{(2)}_{\text{Scr.},22}, where

ρScr.,11(2)=λ2(2λ(4λ5)2(1λ)2log(1λλ)+3)(2λ1)3\displaystyle\rho^{(2)}_{\text{Scr.},11}=\frac{\lambda^{2}\left(2\lambda(4\lambda-5)-2(1-\lambda)^{2}\log\left(\frac{1-\lambda}{\lambda}\right)+3\right)}{(2\lambda-1)^{3}} (153)
ρScr.,12(2)=(1λ)λ(2λ+2(1λ)λlog(1λλ)1)(2λ1)3\displaystyle\rho^{(2)}_{\text{Scr.},12}=\frac{(1-\lambda)\lambda\left(2\lambda+2(1-\lambda)\lambda\log\left(\frac{1-\lambda}{\lambda}\right)-1\right)}{(2\lambda-1)^{3}} (154)
ρScr.,22(2)=(1λ)2(8λ2+2λ2log(1λλ)6λ+1)(2λ1)3\displaystyle\rho^{(2)}_{\text{Scr.},22}=-\frac{(1-\lambda)^{2}\left(8\lambda^{2}+2\lambda^{2}\log\left(\frac{1-\lambda}{\lambda}\right)-6\lambda+1\right)}{(2\lambda-1)^{3}} (155)

H.1 Second moment of the Scrooge ensemble

First, we explicitly compute the second moment of the Scrooge ensemble for a DD-dimensional density matrix ρ\rho. We write ρ\rho using its eigenvalues and eigenvectors ρ=m=1Dλm|smsm|\rho=\sum_{m=1}^{D}\lambda_{m}\ket{s_{m}}\bra{s_{m}}. We next use the definition of the Scrooge ensemble as the ρ\rho-distortion of the Haar ensemble to write the second moment of the Scrooge ensemble as

ρ(2)\displaystyle\rho^{(2)} =(Dρ|ΨΨ|Dρ)2Dρ|Ψ2PHaar(Ψ)𝑑Ψ\displaystyle=\int\frac{\Big{(}\sqrt{D\rho}\ket{\Psi}\bra{\Psi}\sqrt{D\rho}\Big{)}^{\otimes 2}}{\lVert\sqrt{D\rho}\ket{\Psi}\rVert^{2}}P_{\mathrm{Haar}}(\Psi)d\Psi (156)
=m,n,m,n=1DDλmλnλmλnsn|Ψsm|ΨΨ|snΨ|smk=1Dλk2|sk|Ψ|2PHaar(Ψ)𝑑Ψ|sn|smsn|sm|.\displaystyle=\sum_{m,n,m^{\prime},n^{\prime}=1}^{D}\int\frac{D\sqrt{\lambda_{m}\lambda_{n}\lambda_{m^{\prime}}\lambda_{n^{\prime}}}\innerproduct{s_{n}}{\Psi}\innerproduct{s_{m}}{\Psi}\innerproduct{\Psi}{s_{n^{\prime}}}\innerproduct{\Psi}{s_{m^{\prime}}}}{\sum_{k=1}^{D}\lambda_{k}^{2}|\innerproduct{s_{k}}{\Psi}|^{2}}P_{\mathrm{Haar}}(\Psi)d\Psi\ket{s_{n}}\ket{s_{m}}\bra{s_{n^{\prime}}}\bra{s_{m^{\prime}}}. (157)

From this equation, we clearly see that the only non-zero elements in the basis of the eigenvectors of ρ\rho are the ones with n=nn=n^{\prime}, m=mm=m^{\prime} and/or n=mn=m^{\prime}, m=nm=n^{\prime}. All the other terms are zero because of the average over the complex phases. We can then focus, without loss of generality, on the following matrix element of ρ(2)\rho^{(2)}

ρScr.,nm(2)\displaystyle\rho^{(2)}_{\text{Scr.},nm} sn|sm|ρ(2)|sn|sm=sn|sm|ρ(2)|sm|sn\displaystyle\equiv\bra{s_{n}}\bra{s_{m}}\rho^{(2)}\ket{s_{n}}\ket{s_{m}}=\bra{s_{n}}\bra{s_{m}}\rho^{(2)}\ket{s_{m}}\ket{s_{n}} (158)
=Dλmλn|Ψn|2|Ψm|2k=1Dλk2|Ψk|2(D1)!πDδ(1j=1D|Ψj|2)d2Ψ1d2ΨD.\displaystyle=\int\frac{D\lambda_{m}\lambda_{n}|\Psi_{n}|^{2}|\Psi_{m}|^{2}}{\sum_{k=1}^{D}\lambda_{k}^{2}|\Psi_{k}|^{2}}\frac{(D-1)!}{\pi^{D}}\delta(1-\sum_{j=1}^{D}|\Psi_{j}|^{2})d^{2}\Psi_{1}\dots d^{2}\Psi_{D}. (159)

where we used the short notation Ψj=sj|Ψ\Psi_{j}=\innerproduct{s_{j}}{\Psi} and the explicit expression of PHaarP_{\mathrm{Haar}}. Since the integrand only depends on the absolute values of the complex variables Ψ1ΨD\Psi_{1}\dots\Psi_{D} we can perform the integral over the phases and then change variable to xj=λj|Ψj|2x_{j}=\lambda_{j}|\Psi_{j}|^{2}, obtaining

ρScr.,nm(2)=(j=1Dλj1)D!0𝑑x10𝑑xDxnxmk=1Dxkδ(1j=1Dλj1xj).\rho^{(2)}_{\text{Scr.},nm}=\left(\prod_{j=1}^{D}\lambda_{j}^{-1}\right)D!\int_{0}^{\infty}dx_{1}\dots\int_{0}^{\infty}dx_{D}\frac{x_{n}x_{m}}{\sum_{k=1}^{D}x_{k}}\delta(1-\sum_{j=1}^{D}\lambda_{j}^{-1}x_{j}). (160)

To remove the delta function, it is useful to define the following quantity

fnm(2)(t)=(j=1Dλj1)D!0𝑑x10𝑑xDxnxmk=1Dxkδ(tj=1Dλj1xj)f^{(2)}_{nm}(t)=\left(\prod_{j=1}^{D}\lambda_{j}^{-1}\right)D!\int_{0}^{\infty}dx_{1}\dots\int_{0}^{\infty}dx_{D}\frac{x_{n}x_{m}}{\sum_{k=1}^{D}x_{k}}\delta(t-\sum_{j=1}^{D}\lambda_{j}^{-1}x_{j}) (161)

such that ρScr.,nm(2)=fnm(2)(1)\rho^{(2)}_{\text{Scr.},nm}=f^{(2)}_{nm}(1), and the Laplace transform

f~nm(2)(s)=0fnm(2)(t)est𝑑t=(j=1Dλj1)D!0𝑑x10𝑑xDxnxmk=1Dxkexp(sj=1Dλj1xj).\tilde{f}^{(2)}_{nm}(s)=\int_{0}^{\infty}f^{(2)}_{nm}(t)e^{-st}dt=\left(\prod_{j=1}^{D}\lambda_{j}^{-1}\right)D!\int_{0}^{\infty}dx_{1}\dots\int_{0}^{\infty}dx_{D}\frac{x_{n}x_{m}}{\sum_{k=1}^{D}x_{k}}\exp\left({-s\sum_{j=1}^{D}\lambda_{j}^{-1}x_{j}}\right). (162)

The change of variables xjxj/sx_{j}\rightarrow x_{j}/s yields

f~nm(2)(s)=(j=1Dλj1)D!sD10𝑑x10𝑑xDxnxmk=1Dxkexp(j=1Dλj1xj).\tilde{f}^{(2)}_{nm}(s)=\left(\prod_{j=1}^{D}\lambda_{j}^{-1}\right)D!s^{-D-1}\int_{0}^{\infty}dx_{1}\dots\int_{0}^{\infty}dx_{D}\frac{x_{n}x_{m}}{\sum_{k=1}^{D}x_{k}}\exp\left({-\sum_{j=1}^{D}\lambda_{j}^{-1}x_{j}}\right). (163)

Since the inverse Laplace transform of D!sD1D!s^{-D-1} is tDt^{D}, we find

ρScr.,nm(2)=fnm(2)(1)=(j=1Dλj1)0𝑑x10𝑑xDxnxmk=1Dxkexp(j=1Dλj1xj).\rho^{(2)}_{\text{Scr.},nm}=f^{(2)}_{nm}(1)=\left(\prod_{j=1}^{D}\lambda_{j}^{-1}\right)\int_{0}^{\infty}dx_{1}\dots\int_{0}^{\infty}dx_{D}\frac{x_{n}x_{m}}{\sum_{k=1}^{D}x_{k}}\exp\left({-\sum_{j=1}^{D}\lambda_{j}^{-1}x_{j}}\right). (164)

Note that these steps allowed us to get from the integral over the Haar ensemble of normalized states in Eq. (160) to an integral over the unnormalized Gaussian ensemble.

We now define

Λ2=0𝑑x10𝑑xD1k=1Dxkexp(j=1Dλj1xj)\Lambda_{2}=\int_{0}^{\infty}dx_{1}\dots\int_{0}^{\infty}dx_{D}\frac{1}{\sum_{k=1}^{D}x_{k}}\exp\left({-\sum_{j=1}^{D}\lambda_{j}^{-1}x_{j}}\right) (165)

such that we can now write all the ρScr.,nm(2)\rho^{(2)}_{\text{Scr.},nm} from this single “partition function”:

ρScr.,nm(2)=(j=1Dλj1)λn1λn1Λ2.\rho^{(2)}_{\text{Scr.},nm}=\left(\prod_{j=1}^{D}\lambda_{j}^{-1}\right)\frac{\partial}{\partial\lambda_{n}^{-1}}\frac{\partial}{\partial\lambda_{n}^{-1}}\Lambda_{2}. (166)

To find the expression for Λ2\Lambda_{2} in a closed form, we introduce an additional integration variable zz

Λ2=0𝑑x10𝑑xD0𝑑zexp[j=1D(λj1+z)xj].\Lambda_{2}=\int_{0}^{\infty}dx_{1}\dots\int_{0}^{\infty}dx_{D}\int_{0}^{\infty}dz\,\exp\left[{-\sum_{j=1}^{D}(\lambda_{j}^{-1}+z)x_{j}}\right]. (167)

The equivalence with Eq. (165) is evident upon performing the integral in zz. Since the integral is now factorized, we can easily compute the integral in the variables x1,xDx_{1},\dots x_{D}, obtaining

Λ2=0𝑑zj=1D(λj1+z)1=0𝑑zj=1DAjλj1+z\Lambda_{2}=\int_{0}^{\infty}dz\prod_{j=1}^{D}(\lambda_{j}^{-1}+z)^{-1}=\int_{0}^{\infty}dz\sum_{j=1}^{D}\frac{A_{j}}{\lambda_{j}^{-1}+z} (168)

where

Aj=kj(λk1λj1)1.A_{j}=\prod_{k\neq j}(\lambda_{k}^{-1}-\lambda_{j}^{-1})^{-1}. (169)

We can now perform the integral in zz and we get

Λ2=limMj=1DAj[ln(λj1+M)ln(λj1)].\Lambda_{2}=\lim_{M\rightarrow\infty}\sum_{j=1}^{D}A_{j}[\ln(\lambda_{j}^{-1}+M)-\ln(\lambda_{j}^{-1})]. (170)

Using the property jAj=0\sum_{j}A_{j}=0, we find

limMjAjln(λj1+M)=limM[jAjln(M)+jAjln(1+λj1M)]=0.\lim_{M\rightarrow\infty}\sum_{j}A_{j}\ln(\lambda_{j}^{-1}+M)=\lim_{M\rightarrow\infty}\left[\sum_{j}A_{j}\ln(M)+\sum_{j}A_{j}\ln\left(1+\frac{\lambda_{j}^{-1}}{M}\right)\right]=0. (171)

We then get the final expression for Λ2\Lambda_{2}

Λ2=j=1DAjln(λj)=j=1D[kj(λk1λj1)1]ln(λj).\Lambda_{2}=\sum_{j=1}^{D}A_{j}\ln(\lambda_{j})=\sum_{j=1}^{D}\left[\prod_{k\neq j}(\lambda_{k}^{-1}-\lambda_{j}^{-1})^{-1}\right]\ln(\lambda_{j}). (172)

This expression can be straightforwardly used to evaluate the matrix elements of the second moment of the Scrooge ensemble using Eq. (166).

H.2 Second moment in the limit of small purity of ρ\rho

Here, show that in the limit that ρ\rho has small purity, the second moment of the Scrooge ensemble has a simple form ρ(2)(1+𝕊)ρ2\rho^{(2)}\simeq(1+\mathbb{S})\rho^{\otimes 2}, equivalent to the product form for the temporal ensemble (Theorem 2). To do so, we compute an approximate formula for Λ2\Lambda_{2} in the case of a state ρ\rho with small purity Tr[ρ2]=kλk21\mathrm{Tr}[\rho^{2}]=\sum_{k}\lambda_{k}^{2}\ll 1. Using that kλknkλk21\sum_{k}\lambda_{k}^{n}\leq\sum_{k}\lambda_{k}^{2}\ll 1 for every n2n\geq 2, from Eq. (168) we get:

Λ2\displaystyle\Lambda_{2} =(j=1Dλj)0dzk=1D(1+λkz)1\displaystyle=\left(\prod_{j=1}^{D}\lambda_{j}\right)\int_{0}^{\infty}\mathrm{d}z\prod_{k=1}^{D}\left(1+\lambda_{k}z\right)^{-1} (173)
=(j=1Dλj)0dz{1+(k=1Dλk)z+12[(k=1Dλk)2k=1Dλk2]z2+}1\displaystyle=\left(\prod_{j=1}^{D}\lambda_{j}\right)\int_{0}^{\infty}\mathrm{d}z\left\{1+\left(\sum_{k=1}^{D}\lambda_{k}\right)z+\frac{1}{2}\left[\left(\sum_{k=1}^{D}\lambda_{k}\right)^{2}-\sum_{k=1}^{D}\lambda_{k}^{2}\right]z^{2}+\dots\right\}^{-1} (174)
(j=1Dλj)0dz[n=01n!(k=1Dλk)nzn]1\displaystyle\approx\left(\prod_{j=1}^{D}\lambda_{j}\right)\int_{0}^{\infty}\mathrm{d}z\left[\sum_{n=0}^{\infty}\frac{1}{n!}\left(\sum_{k=1}^{D}\lambda_{k}\right)^{n}z^{n}\right]^{-1} (175)
=(j=1Dλj)0dzexp(zk=1Dλk)=(j=1Dλj)(k=1Dλk)1\displaystyle=\left(\prod_{j=1}^{D}\lambda_{j}\right)\int_{0}^{\infty}\mathrm{d}z\,\exp\left(-z\sum_{k=1}^{D}\lambda_{k}\right)=\left(\prod_{j=1}^{D}\lambda_{j}\right)\left(\sum_{k=1}^{D}\lambda_{k}\right)^{-1} (176)

From this equation we obtain

ρnm(2)λnλm[(1+δnm)(1λnλm)+2λmλn]\rho_{nm}^{(2)}\approx\lambda_{n}\lambda_{m}[(1+\delta_{nm})(1-\lambda_{n}-\lambda_{m})+2\lambda_{m}\lambda_{n}] (177)

and, assuming that the Schmidt coefficients are small we have ρScr.,nm(2)λnλm(1+δnm){\rho_{\text{Scr.},nm}^{(2)}\simeq\lambda_{n}\lambda_{m}(1+\delta_{nm})}, which implies ρ(2)(1+𝕊)ρ2\rho^{(2)}\simeq(1+\mathbb{S})\rho^{\otimes 2}.

H.3 Higher moments

We now generalised the above derivation to compute the matrix elements of the kk-th moment (k<Dk<D). The non-zero matrix elements in the eigenbasis of ρ\rho are of the form

ρm1,mk(k)sm1|smk|ρ(2)|sm1|smk\rho_{m_{1},\dots m_{k}}^{(k)}\equiv\bra{s_{m_{1}}}\dots\bra{s_{m_{k}}}\rho^{(2)}\ket{s_{m_{1}^{\prime}}}\dots\ket*{s_{m_{k}^{\prime}}} (178)

where m1,,mkm_{1}^{\prime},\dots,m_{k}^{\prime} is a permutation of the indices m1,mkm_{1},\dots m_{k}. With the same procedure used above for the second moment, we get

ρm1,mk(k)=(j=1Dλj1)0𝑑x10𝑑xDxm1xmk(j=1Dxj)k1exp(j=1Dλj1xj).\rho_{m_{1},\dots m_{k}}^{(k)}=\left(\prod_{j=1}^{D}\lambda_{j}^{-1}\right)\int_{0}^{\infty}dx_{1}\dots\int_{0}^{\infty}dx_{D}\frac{x_{m_{1}}\dots x_{m_{k}}}{(\sum_{j=1}^{D}x_{j})^{k-1}}\exp\left({-\sum_{j=1}^{D}\lambda_{j}^{-1}x_{j}}\right). (179)

We can compute the this matrix element for arbitrary indices using the generating function Λk\Lambda_{k} here defined:

ρm1,mk(k)=(jλj1)kΛkλm11λmk1,\rho_{m_{1},\dots m_{k}}^{(k)}=\left(\prod_{j}\lambda_{j}^{-1}\right)\frac{\partial^{k}\Lambda_{k}}{\partial\lambda^{-1}_{m_{1}}\dots\partial\lambda^{-1}_{m_{k}}}, (180)
Λk=0𝑑x10𝑑xD(1)k(j=1Dxj)k1exp(j=1Dλj1xj)\Lambda_{k}=\int_{0}^{\infty}dx_{1}\dots\int_{0}^{\infty}dx_{D}\frac{(-1)^{k}}{(\sum_{j=1}^{D}x_{j})^{k-1}}\exp\left({-\sum_{j=1}^{D}\lambda_{j}^{-1}x_{j}}\right) (181)

Simlarly to the case k=2k=2, the integral can be computed by introducing an additional integration variable:

Λk\displaystyle\Lambda_{k} =0𝑑x10𝑑xDexp(j=1Dλj1xj)0dz(1)kzk2(k2)!exp(zjxj)\displaystyle=\int_{0}^{\infty}dx_{1}\dots\int_{0}^{\infty}dx_{D}\,\exp\left({-\sum_{j=1}^{D}\lambda^{-1}_{j}x_{j}}\right)\int_{0}^{\infty}\mathrm{d}z\frac{(-1)^{k}z^{k-2}}{(k-2)!}\,\exp\left({-z\sum_{j}x_{j}}\right) (182)
=0dz(1)kzk2(k2)!j=1D(λj1+z)1\displaystyle=\int_{0}^{\infty}\mathrm{d}z\,\frac{(-1)^{k}z^{k-2}}{(k-2)!}\prod_{j=1}^{D}(\lambda_{j}^{-1}+z)^{-1} (183)
=0dz(1)kzk2(k2)!j=1DAjλj1+z\displaystyle=\int_{0}^{\infty}\mathrm{d}z\,\frac{(-1)^{k}z^{k-2}}{(k-2)!}\sum_{j=1}^{D}\frac{A_{j}}{\lambda_{j}^{-1}+z} (184)
=j=1Dλjk+2Ajlnλj.\displaystyle=\sum_{j=1}^{D}\lambda_{j}^{-k+2}A_{j}\ln\lambda_{j}. (185)

This derivation holds for k>1k>1, but we can define in general Λk\Lambda_{k} using Eq. (185). Then the subentropy derived in Ref. [79] can be written as

Q(ρ)=(j=1Dλj1)j=1DAjλj2lnλj=(j=1Dλj1)Λ0Q(\rho)=-\left(\prod_{j=1}^{D}\lambda_{j}^{-1}\right)\sum_{j=1}^{D}A_{j}\lambda_{j}^{2}\ln\lambda_{j}=-\left(\prod_{j=1}^{D}\lambda_{j}^{-1}\right)\Lambda_{0} (186)

Appendix I Proof of Lemma 1: kk-th moments of temporal ensemble of projected states

Here, we provide the proof for Lemma 1.

Lemma 1 (Temporal kk-th moment of the projected state).

Consider the temporal ensemble comprised of the (unnormalized) projected states |Ψ~(xB,t)(𝕀AxB|)|Ψ(t)\ket*{\tilde{\Psi}(x_{B},t)}\equiv(\mathbb{I}_{A}\otimes\bra{x_{B}})\ket{\Psi(t)}. The kk-th moment of this temporal ensemble is equal to the kk-th moment of the corresponding unnormalized Scrooge ensemble, up to small corrections:

ρTemp. Proj.(k)(xB)𝔼t[|Ψ~(xB,t)Ψ~(xB,t)|k]=pd(xB)kρ~S~cr.(k)[ρ¯(xB)]+O(Δβ(xB)pd(xB)k),\displaystyle\rho^{(k)}_{\text{Temp. Proj.}}(x_{B})\equiv\mathbb{E}_{t}\left[\outerproduct*{\tilde{\Psi}(x_{B},t)}{\tilde{\Psi}(x_{B},t)}^{\otimes k}\right]=p_{d}(x_{B})^{k}\tilde{\rho}^{(k)}_{\tilde{\text{S}}\text{cr.}}[\bar{\rho}(x_{B})]+O(\Delta_{\beta}(x_{B})p_{d}(x_{B})^{k})~,

where Δβ(xB)tr[(𝕀A|xBxB|)2ρd(2)]/pd(xB)2\Delta_{\beta}(x_{B})\equiv\tr[(\mathbb{I}_{A}\otimes\outerproduct{x_{B}}{x_{B}})^{\otimes 2}\rho_{d}^{(2)}]/p_{d}(x_{B})^{2} is a correction term, that is typically exponentially small in system size. To see this, note that Δβ(xB)\Delta_{\beta}(x_{B}) can be interpreted as the purity of a probability distribution over DD elements. Specifically, we define the quantity qxB(E)|cE|2E|(𝕀A|xBxB|)|E/pd(xB)q_{x_{B}}(E)\equiv\absolutevalue{c_{E}}^{2}\bra{E}(\mathbb{I}_{A}\otimes\outerproduct{x_{B}}{x_{B}})\ket{E}/p_{d}(x_{B}) which is a probability distribution because it is non-negative and satisfies EqxB(E)=1\sum_{E}q_{x_{B}}(E)=1. Δβ(xB)=EqxB(E)2\Delta_{\beta}(x_{B})=\sum_{E}q_{x_{B}}(E)^{2} is the purity of qxB(E)q_{x_{B}}(E), which we expect to be typically O(D1)O(D^{-1})

Proof.

This is a simple application of an intermediate step from Appendix D, which states that

ρ¯(xB)kσSkPerm(σ)ρTemp. Proj.(k)\displaystyle\left\|\bar{\rho}(x_{B})^{\otimes k}\sum_{\sigma\in S_{k}}\text{Perm}(\sigma)-\rho^{(k)}_{\text{Temp. Proj.}}\right\|_{*} =sMk(D)ρ¯(xB)k(1|π(s)||Sk|)PsσSkPerm(σ)\displaystyle=\left\|\sum_{s\subset M_{k}(D)}\bar{\rho}(x_{B})^{\otimes k}\left(1-\frac{\absolutevalue{\pi(s)}}{\absolutevalue{S_{k}}}\right)P_{s}\sum_{\sigma\in S_{k}}\text{Perm}(\sigma)\right\|_{*}
Cks()sMk(D)tr[ρ¯(xB)(k2)ρ¯(2)(xB)σSkPerm(σ)]\displaystyle\leq C_{k}\sum_{\overset{s\subset M_{k}(D)}{s\neq()}}\tr\left[\bar{\rho}(x_{B})^{\otimes(k-2)}\otimes\bar{\rho}^{(2)}(x_{B})\sum_{\sigma\in S_{k}}\text{Perm}(\sigma)\right] (Ckk!)tr[ρ¯(2)(xB)]pd(xB)k2CkΔβ(xB)pd(xB)k.\displaystyle\leq(C_{k}k!)\text{tr}[\bar{\rho}^{(2)}(x_{B})]p_{d}(x_{B})^{k-2}\equiv C_{k}^{\prime}\Delta_{\beta}(x_{B})p_{d}(x_{B})^{k}~. (187)

Appendix J Proof of Theorem 4: weighted kk-th moments of projected ensemble

Here, we prove Theorem 4,

Theorem 4.

At typical (late) times tt, the weighted kk-th moment ρ~Proj.(k)(t)\tilde{\rho}^{(k)}_{\text{Proj.}}(t) of the unnormalized projected ensemble is equal to the kk-th moment of the generalized Scrooge ensemble:

ρ~Proj.(k)(t)\displaystyle\tilde{\rho}^{(k)}_{\text{Proj.}}(t) xB|Ψ~(xB,t)Ψ~(xB,t)|kpd(xB)k1\displaystyle\equiv\sum_{x_{B}}\frac{\outerproduct*{\tilde{\Psi}(x_{B},t)}{\tilde{\Psi}(x_{B},t)}^{\otimes k}}{p_{d}(x_{B})^{k-1}} (188)
=xBpd(xB)ρ~S~cr.(k)[ρ¯(xB)]+O(Δβ1/2),\displaystyle=\sum_{x_{B}}p_{d}(x_{B})\tilde{\rho}^{(k)}_{\tilde{\text{S}}\text{cr.}}[\bar{\rho}(x_{B})]+O(\Delta^{1/2}_{\beta})~, (189)

where ΔβxBpd(xB)Δβ(xB)\Delta_{\beta}\equiv\sum_{x_{B}}p_{d}(x_{B})\Delta_{\beta}(x_{B}) the weighted mean of Δβ(xB)\Delta_{\beta}(x_{B}).

Proof.

First, we compute the time average:

𝔼t[ρ~Proj.(k)(t)]xBpd(xB)ρ~Scr.(k)(ρ¯(xB))+O(Δβ).\mathbb{E}_{t}[\tilde{\rho}^{(k)}_{\text{Proj.}}(t)]\equiv\sum_{x_{B}}p_{d}(x_{B})\tilde{\rho}^{(k)}_{\text{S}\text{cr.}}(\bar{\rho}(x_{B}))+O(\Delta_{\beta})~. (190)

This follows from Appendix I. Next, we compute the variance vart[ρ~Proj.(k)(t)]\text{var}_{t}[\tilde{\rho}^{(k)}_{\text{Proj.}}(t)] by computing the trace norm

𝔼t[ρ~Proj.(k)(t)2]𝔼t[ρ~Proj.(k)(t)]2=O(tr(ρd2))=O(Δβ).\displaystyle\|\mathbb{E}_{t}[\tilde{\rho}^{(k)}_{\text{Proj.}}(t)^{\otimes 2}]-\mathbb{E}_{t}[\tilde{\rho}^{(k)}_{\text{Proj.}}(t)]^{\otimes 2}\|_{*}=O(\tr(\rho_{d}^{2}))=O(\Delta_{\beta})~. (191)

The temporal averages can be in terms of: permutations of 2k2k and kk elements respectively, and correction terms E.C. in Lemma 3. The dominant term will be from the former: permutations of 2k2k elements that cannot be written as disjoint permutations of two groups of kk elements. The largest possible contribution comes from the permutation where only two elements from each group are swapped, and all others are not permuted. This has trace

xB,xBtr[(𝕀A|xBxB|)ρd(𝕀A|xBxB|)ρd]=tr(ρd2),\sum_{x_{B},x_{B}^{\prime}}\tr[(\mathbb{I}_{A}\otimes\outerproduct{x_{B}}{x_{B}})\rho_{d}(\mathbb{I}_{A}\otimes\outerproduct{x^{\prime}_{B}}{x^{\prime}_{B}})\rho_{d}]=\tr(\rho_{d}^{2})~, (192)

and the contributions of all other relevant permutations and correction terms can be bounded by Eq. (192). Finally, we can bound tr(ρd2)Δβ\tr(\rho_{d}^{2})\leq\Delta_{\beta} as follows: we use Sedrakyan’s inequality, a specialization of the Cauchy-Schwarz inequality. It states that for positive uiu_{i} and viv_{i}, iui2/iviiui2/vi\sum_{i}u_{i}^{2}/\sum_{i}v_{i}\leq\sum_{i}u_{i}^{2}/v_{i}. Applied to our context, we have

tr(ρd2)=E|cE|4=E|cE|4(xBE|(𝕀A|xBxB|)|E)2xBpd(xB)\displaystyle\tr(\rho_{d}^{2})=\sum_{E}\absolutevalue{c_{E}}^{4}=\sum_{E}\absolutevalue{c_{E}}^{4}\frac{\left(\sum_{x_{B}}\bra{E}(\mathbb{I}_{A}\otimes\outerproduct{x_{B}}{x_{B}})\ket{E}\right)^{2}}{\sum_{x_{B}}p_{d}(x_{B})} E|cE|4xBE|(𝕀A|xBxB|)|E2pd(xB)\displaystyle\leq\sum_{E}\absolutevalue{c_{E}}^{4}\sum_{x_{B}}\frac{\bra{E}(\mathbb{I}_{A}\otimes\outerproduct{x_{B}}{x_{B}})\ket{E}^{2}}{p_{d}(x_{B})}
xBpd(xB)Δβ(xB)Δβ.\displaystyle\leq\sum_{x_{B}}p_{d}(x_{B})\Delta_{\beta}(x_{B})\equiv\Delta_{\beta}~. (193)

While the tightest general bound we can achieve is Δβ1\Delta_{\beta}\leq 1 (saturated when the initial state is an energy eigenstate |Ψ0=|E\ket{\Psi_{0}}=\ket{E}), we expect Δβ\Delta_{\beta} to be exponentially small in total system size. As discussed above, Δβ(xB)=EqxB(E)2\Delta_{\beta}(x_{B})=\sum_{E}q_{x_{B}}(E)^{2} is the purity of a distribution qxB(E)q_{x_{B}}(E) and hence is typically O(D1)O(D^{-1}). Δβ\Delta_{\beta} is the weighted average of such purities. ∎

Appendix K Additional numerical evidence

K.1 Chaotic XXZ model

In this Appendix, we support our results on the projected ensemble with another chaotic model. Specifically, we study a variant of the XXZ model [121]:

H=14j=1N1[J(XjXj+1+YjYj+1)+ΔZjZj+1]+Δ4j=1N2ZjZj+2,H=\frac{1}{4}\sum_{j=1}^{N-1}\left[J(X_{j}X_{j+1}+Y_{j}Y_{j+1})+\Delta Z_{j}Z_{j+1}\right]+\frac{\Delta^{\prime}}{4}\sum_{j=1}^{N-2}Z_{j}Z_{j+2}~, (194)

where Xj,Yj,X_{j},Y_{j}, and ZjZ_{j} are the Pauli matrices on site jj. Following Ref. [121], we have set J=2,Δ=(5+1)/4J=\sqrt{2},\Delta=(\sqrt{5}+1)/4, and Δ=1\Delta^{\prime}=1, and performed simulations on a one-dimensional chain with open boundary conditions.

This model serves as a further check of our theoretical conclusions, but also serves to illustrate the effect of conserved quantities on the projected ensemble. This model has a U(1)U(1) symmetry, leading to a conserved global magnetization MjZjM\equiv\sum_{j}Z_{j}.

The U(1)U(1) symmetry leads to significant correlations between projected state and outcome. For simplicity, we consider ZZ-basis measurements on the complementary system. If the measured bitstring zBz_{B} has magnetization mm, the projected state |Ψ(zB)\ket{\Psi(z_{B})} is guaranteed to have magnetization MmM-m. As a result, even in the case where the outcomes zBz_{B} are not “energy-revealing,” they reveal charge, another conserved quantity. We no longer expect the entire projected ensemble to be described by the Scrooge ensemble. Rather, following Ref. [29], we form the projected ensemble only out of measurement outcomes zBz_{B} with fixed magnetization mm, and conjecture that this is described by a Scrooge ensemble of the “sector reduced density matrix,” ρmzB,|zB|=m|Ψ(zB)Ψ(zB)|\rho_{m}\propto\sum_{z_{B},|z_{B}|=m}\outerproduct{\Psi(z_{B})}{\Psi(z_{B})}.

In the more general case where the outcomes zBz_{B} are energy-revealing, Theorem 4 applies without modification: each measurement outcomes zBz_{B} has its own average state ρ¯(zB)\bar{\rho}(z_{B}); the projected state |Ψ(zB)\ket{\Psi(z_{B})} is sampled from the corresponding Scrooge ensemble Scrooge[ρ¯(zB)]\text{Scrooge}[\bar{\rho}(z_{B})]. We illustrate the validity of this result in Fig. 9(a). We plot the trace distances between the third moments of the projected and generalized Scrooge ensemble, for a time-evolution of the initial state with one domain wall |Ψ0=|0(N/21)|1(N/2+1)\ket{\Psi_{0}}=|0\rangle^{\otimes(N/2-1)}|1\rangle^{\otimes(N/2+1)}. The trace distances decay exponentially with increasing system size, consistent with Theorem 4. Meanwhile, the Scrooge ensemble Scrooge[ρm]\text{Scrooge}[\rho_{m}] is a good description of the projected ensemble for smaller system sizes, but deviations become noticeable at larger system sizes, illustrated by the trace distance between the third moments of the projected and Scrooge ensemble, which saturates to a non-zero value (light green). As with the MFIM, the Porter-Thomas distribution is visible in the outcome probabilities p(zA,zB)p(z_{A},z_{B}), serving as a measurable signature of the generalized Scrooge ensemble.

K.2 Transverse field Ising model

We also demonstrate an example where the projected ensemble is not described by the generalized Scrooge ensemble, using the time-evolved states of the transverse field Ising model (TFIM), which is the mixed field Ising model [Eq. (36)] with hx=0h_{x}=0. We time-evolve the initial state |0N\ket{0}^{\otimes N}, and construct the projected ensemble on two sites near the middle of the chain, with measurements in the XBX_{B} basis. Unlike the MFIM, the projected ensemble from the TFIM is not close to the generalized Scrooge ensemble [Fig. 9(c)]. This is because the TFIM is integrable and has solution by free-fermions, and hence does not satisfy the kk-th no-resonance conditions. We have also taken care to ensure that our projected ensemble should not be described by the real Scrooge ensemble (Appendix L below): the time-reversal symmetry is broken by time evolution, and the particle-hole symmetry is broken by the choice of initial state.

Refer to caption
Figure 9: Projected ensembles in a chaotic and integrable model. (a) We time-evolve an initial state and plot the trace distances between the third moments of the Scrooge (light green) and generalized Scrooge (dark green) ensembles and the projected ensemble generated by a evolving a state under a variant of the XXZ model [Eq. (194)] for a subsystem of four qubits and measurements in the ZBZ_{B} basis. Like in the MFIM, the projected ensemble is close to the Scrooge ensemble but deviates from it at large system sizes. In contrast, the generalized Scrooge ensemble provides good agreement, even for the largest system sizes studied in this work. (b) The PT distribution is visible in the statistics of the normalized outcome probabilities p~pzA(zB)/𝔼t[pzA(zB)]\tilde{p}\equiv p_{z_{A}}(z_{B})/\mathbb{E}_{t}[p_{z_{A}}(z_{B})] (while the raw probabilities DpzA(zB)Dp_{z_{A}}(z_{B}) show small deviations) (histogram plotted over all zAz_{A} and zBz_{B}). (c) In contrast, the trace distance Δ(2)\Delta^{(2)} between the Scrooge ensemble and the projected ensemble generated from the state exp(iHTFIMt)|0N\exp(-iH_{\text{TFIM}}t)\ket{0}^{\otimes N} remains high, where HTFIMH_{\text{TFIM}} is the integrable transverse field Ising model [Eq. (36) with hx=0h_{x}=0], the subsystem AA consists of two qubits, and the complement BB is measured in the XX basis.

Appendix L Time-reversal symmetry and real Scrooge ensembles

Here, we discuss a notable deviation from the Scrooge ensemble: in the presence of relevant symmetries, states in the ensemble can be constrained to be purely real. In this case, one has to construct a maximally-entropic ensemble of real states, which we dub the real Scrooge ensemble.

A simple way that this can emerge is the presence of time-reversal symmetry: the existence of an unitary operator UU that maps the Hamiltonian HH to its complex conjugate HH^{*}: UHU=HUHU^{\dagger}=H^{*}. The eigenstates of such Hamiltonians are guaranteed to be real in an appropriate basis. Under appropriate measurement bases zBz_{B}, the projected states are guaranteed to be real. The real Scrooge ensemble can be defined as an analogous ρ\rho-distortion of the maximally entropic ensemble of real states, which has been studied in e.g. Ref. [48] and we dub the real Haar ensemble.

As an example, in analogy to Eq. (166), the second moment of the real Scrooge ensemble has non-zero entries:

ρnm(2)=(k=1dAλk1)1/2μmμnΛ2real|μi=λi1,\rho_{nm}^{(2)}=\left(\prod_{k=1}^{d_{A}}\lambda_{k}^{-1}\right)^{1/2}\frac{\partial}{\partial\mu_{m}}\frac{\partial}{\partial\mu_{n}}\Lambda_{2}^{\text{real}}\big{|}_{\mu_{i}=\lambda_{i}^{-1}}, (195)

where λi\lambda_{i} are the eigenvalues of the first moment ρ\rho, and

Λ2real=20dzk=1dA(μk+z)1/2.\Lambda_{2}^{\text{real}}=2\int_{0}^{\infty}\mathrm{d}z\prod_{k=1}^{d_{A}}(\mu_{k}+z)^{-1/2}. (196)

While there is no closed-form expression for Λ2real\Lambda_{2}^{\text{real}} analogous to Eq. (172), for a given first moment ρ\rho, we can numerically evaluate Eq. (195) to obtain the higher moments of the real Scrooge ensemble.

In Fig. 10(b), we show that the real Scrooge ensemble describes the projected ensemble of eigenstates of the MFIM. As discussed in Sec. VI.5, the MFIM is time-reversal symmetric. By measuring in the xBx_{B} basis, the projected states can be written as real states. We compute the trace distance between the second moments of the real Scrooge ensemble (blue) and the real Haar ensemble (red). The real Haar ensemble only describes the projected ensemble at infinite temperature E0E\approx 0, while the real Scrooge ensemble describes the projected ensemble at all energies/temperatures, with exponentially increasing accuracy with increasing system size [Fig. 10(b)].

In addition to trace distances, we take the real Porter-Thomas distribution as smoking-gun evidence of the real Scrooge ensemble. The real Porter-Thomas distribution [83] is an analogous universal distribution, given by

PPT,(x)dx=exp(x/2)2πxdx.P_{\text{PT},\mathbb{R}}(x)dx=\frac{\exp(-x/2)}{\sqrt{2\pi x}}dx~. (197)

In Fig. 10(d), we show the real PT distribution in relevant measurement outcomes pE(oA,xB)p_{E}(o_{A},x_{B}), following the same procedure as Fig. 6. By using the same normalization procedure in Fig. 6(e), we see even better agreement with the real PT distribution (red), indicating that the generalized real Scrooge ensemble is likely a better description than the real Scrooge ensemble, as expected.

Finally, we remark that time-evolved states can also be made to be all real. Unlike with energy eigenstates, time-reversal symmetry is not sufficient, because time-evolution breaks this symmetry. Rather, we need both time-reversal symmetry and particle-hole symmetry, which ensures the energy spectrum is symmetric about 0. With appropriate initial states, the time-evolved state will always be real, and as such the projected ensemble will be described by the real Scrooge and generalized real Scrooge ensembles. See SM of Ref. [94] for more details.

Refer to caption
Figure 10: Real Scrooge ensemble. (a) Under appropriate conditions, the states in the temporal and projected ensembles can be restricted to be always real, indicated here with single-qubit states lying in the XZX-Z plane of the Bloch sphere (orange circle). (b) We demonstrate this with the eigenstates of the MFIM, a time-reversal-symmetric Hamiltonian. We compare the second moments of the projected ensemble with those of: the real Haar ensemble (red) and the real Scrooge ensemble (blue). The real Scrooge ensemble consistently has the smaller trace distance Δ(2)\Delta^{(2)}. (c) With increasing system size, the projected ensembles appear to converge to the the real Scrooge ensemble. (d) We verify the generalized real Scrooge ensemble with a similar procedure as Fig. 6(e): the quantity p~zA(zB)=pzA(E)(zB)/𝔼E[pzA(E)(zB)]\tilde{p}_{z_{A}}(z_{B})=p^{(E)}_{z_{A}}(z_{B})/\mathbb{E}_{E}[p^{(E)}_{z_{A}}(z_{B})] (blue) obeys a real PT distribution [Eq. (197), dashed line, note cusp at p~=0\tilde{p}=0], for zA=0100z_{A}=0100, while the raw probabilities DpzA(E)(zB)Dp^{(E)}_{z_{A}}(z_{B}) do not (red).

References

  • Landau and Lifshitz [2013] L. D. Landau and E. M. Lifshitz, Statistical Physics: Volume 5, Vol. 5 (Elsevier, 2013).
  • Bekenstein [1974] J. D. Bekenstein, Generalized second law of thermodynamics in black-hole physics, Phys. Rev. D 9, 3292 (1974).
  • Cover [1999] T. M. Cover, Elements of information theory (John Wiley & Sons, 1999).
  • Mezard and Montanari [2009] M. Mezard and A. Montanari, Information, physics, and computation (Oxford University Press, 2009).
  • Bennett and Landauer [1985] C. H. Bennett and R. Landauer, The fundamental physical limits of computation, Sci. Am. 253, 48 (1985).
  • Jaynes [1957] E. T. Jaynes, Information theory and statistical mechanics, Phys. Rev. 106, 620 (1957).
  • Grandy Jr [1980] W. Grandy Jr, Principle of maximum entropy and irreversible processes, Phys. Rep. 62, 175 (1980).
  • Martyushev and Seleznev [2006] L. Martyushev and V. Seleznev, Maximum entropy production principle in physics, chemistry and biology, Phys. Rep. 426, 1 (2006).
  • Banavar et al. [2010] J. R. Banavar, A. Maritan, and I. Volkov, Applications of the principle of maximum entropy: from physics to ecology, J. Phys. Cond. Mat. 22, 063101 (2010).
  • Pressé et al. [2013] S. Pressé, K. Ghosh, J. Lee, and K. A. Dill, Principles of maximum entropy and maximum caliber in statistical physics, Rev. Mod. Phys. 85, 1115 (2013).
  • Gogolin et al. [2011] C. Gogolin, M. P. Müller, and J. Eisert, Absence of thermalization in nonintegrable systems, Phys. Rev. Lett. 106, 040401 (2011).
  • Gogolin and Eisert [2016] C. Gogolin and J. Eisert, Equilibration, thermalisation, and the emergence of statistical mechanics in closed quantum systems, Rep. Prog. Phys. 79, 056001 (2016).
  • Short [2011] A. J. Short, Equilibration of quantum systems and subsystems, N. J. Phys. 13, 053009 (2011).
  • Riera et al. [2012] A. Riera, C. Gogolin, and J. Eisert, Thermalization in nature and on a quantum computer, Phys. Rev. Lett. 108 (2012).
  • Deutsch [2018] J. M. Deutsch, Eigenstate thermalization hypothesis, Rep. Prog. Phys. 81, 082001 (2018).
  • Ueda [2020] M. Ueda, Quantum equilibration, thermalization and prethermalization in ultracold atoms, Nature Rev. Phys. 2, 669 (2020).
  • Reimann [2008a] P. Reimann, Foundation of statistical mechanics under experimentally realistic conditions, Phys. Rev. Lett. 101, 190403 (2008a).
  • Bernien et al. [2017] H. Bernien, S. Schwartz, A. Keesling, H. Levine, A. Omran, H. Pichler, S. Choi, A. S. Zibrov, M. Endres, M. Greiner, et al., Probing many-body dynamics on a 51-atom quantum simulator, Nature (London) 551, 579 (2017).
  • Zhang et al. [2017] J. Zhang, G. Pagano, P. W. Hess, A. Kyprianidis, P. Becker, H. Kaplan, A. V. Gorshkov, Z.-X. Gong, and C. Monroe, Observation of a many-body dynamical phase transition with a 53-qubit quantum simulator, Nature (London) 551, 601 (2017).
  • Neyenhuis et al. [2017] B. Neyenhuis, J. Zhang, P. W. Hess, J. Smith, A. C. Lee, P. Richerme, Z.-X. Gong, A. V. Gorshkov, and C. Monroe, Observation of prethermalization in long-range interacting spin chains, Science Adv. 3, e1700672 (2017).
  • Choi et al. [2019] J. Choi, H. Zhou, S. Choi, R. Landig, W. W. Ho, J. Isoya, F. Jelezko, S. Onoda, H. Sumiya, D. A. Abanin, et al., Probing quantum thermalization of a disordered dipolar spin ensemble with discrete time-crystalline order, Phys. Rev. Lett. 122, 043603 (2019).
  • Mi et al. [2021a] X. Mi, M. Ippoliti, C. Quintana, A. Greene, Z. Chen, J. Gross, F. Arute, K. Arya, J. Atalaya, R. Babbush, et al., Time-crystalline eigenstate order on a quantum processor, Nature (London) 601, 531 (2021a).
  • Gross and Bloch [2017] C. Gross and I. Bloch, Quantum simulations with ultracold atoms in optical lattices, Science 357, 995 (2017).
  • Chiu et al. [2019] C. S. Chiu, G. Ji, A. Bohrdt, M. Xu, M. Knap, E. Demler, F. Grusdt, M. Greiner, and D. Greif, String patterns in the doped Hubbard model, Science 365, 251 (2019).
  • Brydges et al. [2019] T. Brydges, A. Elben, P. Jurcevic, B. Vermersch, C. Maier, B. P. Lanyon, P. Zoller, R. Blatt, and C. F. Roos, Probing rényi entanglement entropy via randomized measurements, Science 364, 260 (2019).
  • Scholl et al. [2021] P. Scholl, M. Schuler, H. J. Williams, A. A. Eberharter, D. Barredo, K.-N. Schymik, V. Lienhard, L.-P. Henry, T. C. Lang, T. Lahaye, et al., Quantum simulation of 2D antiferromagnets with hundreds of Rydberg atoms, Nature (London) 595, 233 (2021).
  • Ebadi et al. [2021] S. Ebadi, T. T. Wang, H. Levine, A. Keesling, G. Semeghini, A. Omran, D. Bluvstein, R. Samajdar, H. Pichler, W. W. Ho, et al., Quantum phases of matter on a 256-atom programmable quantum simulator, Nature (London) 595, 227 (2021).
  • Choi et al. [2023] J. Choi, A. L. Shaw, I. S. Madjarov, X. Xie, R. Finkelstein, J. P. Covey, J. S. Cotler, D. K. Mark, H.-Y. Huang, A. Kale, et al., Preparing random states and benchmarking with many-body quantum chaos, Nature 613, 468 (2023).
  • Cotler et al. [2023] J. S. Cotler, D. K. Mark, H.-Y. Huang, F. Hernandez, J. Choi, A. L. Shaw, M. Endres, and S. Choi, Emergent quantum state designs from individual many-body wave functions, PRX Quantum 4, 010311 (2023).
  • Ippoliti and Ho [2023] M. Ippoliti and W. W. Ho, Dynamical purification and the emergence of quantum state designs from the projected ensemble, PRX Quantum 4 (2023).
  • Ippoliti and Ho [2022] M. Ippoliti and W. W. Ho, Solvable model of deep thermalization with distinct design times, Quantum 6, 886 (2022).
  • Claeys and Lamacraft [2022] P. W. Claeys and A. Lamacraft, Quantum 6, 738 (2022).
  • Lucas et al. [2023] M. Lucas, L. Piroli, J. De Nardis, and A. De Luca, Generalized deep thermalization for free fermions, Phys. Rev. A 107, 032215 (2023).
  • Wilming and Roth [2022] H. Wilming and I. Roth, High-temperature thermalization implies the emergence of quantum state designs, arXiv:2202.01669  (2022).
  • Bhore et al. [2023] T. Bhore, J.-Y. Desaules, and Z. Papić, Deep thermalization in constrained quantum systems, Phys. Rev. B 108 (2023).
  • Chan and Luca [2024] A. Chan and A. D. Luca, Projected state ensemble of a generic model of many-body quantum chaos, arXiv:2402.16939  (2024).
  • Varikuti and Bandyopadhyay [2024] N. D. Varikuti and S. Bandyopadhyay, Unraveling the emergence of quantum state designs in systems with symmetry, arXiv:2402.08949  (2024).
  • Vairogs and Yan [2024] C. Vairogs and B. Yan, Extracting randomness from quantum ’magic’, arXiv:2402.10181  (2024).
  • Pilatowsky-Cameo et al. [2023] S. Pilatowsky-Cameo, C. B. Dag, W. W. Ho, and S. Choi, Complete Hilbert-space ergodicity in quantum dynamics of generalized Fibonacci drives, Phys. Rev. Lett. 131 (2023).
  • Pilatowsky-Cameo et al. [2024] S. Pilatowsky-Cameo, I. Marvian, S. Choi, and W. W. Ho, Hilbert-space ergodicity in driven quantum systems: Obstructions and designs, arXiv:2402.06720  (2024).
  • Shaw et al. [2024] A. L. Shaw, D. K. Mark, J. Choi, R. Finkelstein, P. Scholl, S. Choi, and M. Endres, Universal fluctuations and noise learning from Hilbert space ergodicity, arXiv, same day.  (2024).
  • Mori and Shiraishi [2017] T. Mori and N. Shiraishi, Thermalization without eigenstate thermalization hypothesis after a quantum quench, Phys. Rev. E 96, 022153 (2017).
  • Mori et al. [2018] T. Mori, T. N. Ikeda, E. Kaminishi, and M. Ueda, Thermalization and prethermalization in isolated quantum systems: a theoretical overview, J. Phys. B 51, 112001 (2018).
  • De Palma et al. [2015] G. De Palma, A. Serafini, V. Giovannetti, and M. Cramer, Necessity of eigenstate thermalization, Phys. Rev. Lett. 115, 220401 (2015).
  • Harrow and Huang [2022] A. W. Harrow and Y. Huang, Thermalization without eigenstate thermalization, arXiv:2209.09826  (2022).
  • Nakata et al. [2012] Y. Nakata, P. S. Turner, and M. Murao, Phase-random states: Ensembles of states with fixed amplitudes and uniformly distributed phases in a fixed basis, Phys. Rev. A 86, 012301 (2012).
  • Ambainis and Emerson [2007] A. Ambainis and J. Emerson, Quantum t-designs: t-wise independence in the quantum world, arXiv:quant-ph/0701126  (2007).
  • Harrow [2013] A. W. Harrow, The church of the symmetric subspace, arXiv:1308.6595  (2013).
  • Rosenberg et al. [2023] E. Rosenberg, T. Andersen, R. Samajdar, A. Petukhov, J. Hoke, D. Abanin, A. Bengtsson, I. Drozdov, C. Erickson, P. Klimov, X. Mi, A. Morvan, M. Neeley, C. Neill, R. Acharya, I. Aleiner, R. Allen, K. Anderson, M. Ansmann, F. Arute, K. Arya, A. Asfaw, J. Atalaya, J. Bardin, A. Bilmes, G. Bortoli, A. Bourassa, J. Bovaird, L. Brill, M. Broughton, B. B. Buckley, D. Buell, T. Burger, B. Burkett, N. Bushnell, J. Campero, H.-S. Chang, Z. Chen, B. Chiaro, D. Chik, J. Cogan, R. Collins, P. Conner, W. Courtney, A. Crook, B. Curtin, D. Debroy, A. D. T. Barba, S. Demura, A. D. Paolo, A. Dunsworth, C. Earle, E. Farhi, R. Fatemi, V. Ferreira, L. Flores, E. Forati, A. Fowler, B. Foxen, G. Garcia, Élie Genois, W. Giang, C. Gidney, D. Gilboa, M. Giustina, R. Gosula, A. G. Dau, J. Gross, S. Habegger, M. Hamilton, M. Hansen, M. Harrigan, S. Harrington, P. Heu, G. Hill, M. Hoffmann, S. Hong, T. Huang, A. Huff, W. Huggins, L. Ioffe, S. Isakov, J. Iveland, E. Jeffrey, Z. Jiang, C. Jones, P. Juhas, D. Kafri, T. Khattar, M. Khezri, M. Kieferová, S. Kim, A. Kitaev, A. Klots, A. Korotkov, F. Kostritsa, J. M. Kreikebaum, D. Landhuis, P. Laptev, K. M. Lau, L. Laws, J. Lee, K. Lee, Y. Lensky, B. Lester, A. Lill, W. Liu, W. P. Livingston, A. Locharla, S. Mandrà, O. Martin, S. Martin, J. McClean, M. McEwen, S. Meeks, K. Miao, A. Mieszala, S. Montazeri, R. Movassagh, W. Mruczkiewicz, A. Nersisyan, M. Newman, J. H. Ng, A. Nguyen, M. Nguyen, M. Niu, T. O’Brien, S. Omonije, A. Opremcak, R. Potter, L. Pryadko, C. Quintana, D. Rhodes, C. Rocque, N. Rubin, N. Saei, D. Sank, K. Sankaragomathi, K. Satzinger, H. Schurkus, C. Schuster, M. Shearn, A. Shorter, N. Shutty, V. Shvarts, V. Sivak, J. Skruzny, C. Smith, R. Somma, G. Sterling, D. Strain, M. Szalay, D. Thor, A. Torres, G. Vidal, B. Villalonga, C. V. Heidweiller, T. White, B. Woo, C. Xing, J. Yao, P. Yeh, J. Yoo, G. Young, A. Zalcman, Y. Zhang, N. Zhu, N. Zobrist, H. Neven, R. Babbush, D. Bacon, S. Boixo, J. Hilton, E. Lucero, A. Megrant, J. Kelly, Y. Chen, V. Smelyanskiy, V. Khemani, S. Gopalakrishnan, T. Prosen, and P. Roushan, Dynamics of magnetization at infinite temperature in a Heisenberg spin chain, arXiv:2306.09333  (2023).
  • Wienand et al. [2023] J. F. Wienand, S. Karch, A. Impertro, C. Schweizer, E. McCulloch, R. Vasseur, S. Gopalakrishnan, M. Aidelsburger, and I. Bloch, Emergence of fluctuating hydrodynamics in chaotic quantum systems, arXiv:2306.11457  (2023).
  • Rigol et al. [2008] M. Rigol, V. Dunjko, and M. Olshanii, Thermalization and its mechanism for generic isolated quantum systems, Nature 452, 854 (2008).
  • Polkovnikov et al. [2011] A. Polkovnikov, K. Sengupta, A. Silva, and M. Vengalattore, Colloquium: Nonequilibrium dynamics of closed interacting quantum systems, Rev. Mod. Phys. 83, 863 (2011).
  • Brody and Hughston [2000] D. C. Brody and L. P. Hughston, Information content for quantum states, J. Math. Phys. 41, 2586–2592 (2000).
  • Anza and Crutchfield [2022] F. Anza and J. P. Crutchfield, Quantum information dimension and geometric entropy, PRX Quantum 3, 020355 (2022).
  • Hahn et al. [2023] D. Hahn, D. J. Luitz, and J. T. Chalker, The statistical properties of eigenstates in chaotic many-body quantum systems, arXiv:2309.12982  (2023).
  • Holevo [1998] A. S. Holevo, The capacity of the quantum channel with general signal states, IEEE Trans. Inf. Theory 44, 269 (1998).
  • Kukuljan et al. [2017] I. Kukuljan, S. Grozdanov, and T. Prosen, Weak quantum chaos, Phys. Rev. B 96, 060301 (2017).
  • Nahum et al. [2018] A. Nahum, S. Vijay, and J. Haah, Operator spreading in random unitary circuits, Phys. Rev. X 8, 021014 (2018).
  • Mi et al. [2021b] X. Mi, P. Roushan, C. Quintana, S. Mandrà, J. Marshall, C. Neill, F. Arute, K. Arya, J. Atalaya, R. Babbush, et al., Information scrambling in quantum circuits, Science 374, 1479 (2021b).
  • Barratt et al. [2022] F. Barratt, U. Agrawal, A. C. Potter, S. Gopalakrishnan, and R. Vasseur, Transitions in the learnability of global charges from local measurements, Phys. Rev. Lett. 129, 200602 (2022).
  • Ippoliti and Khemani [2023] M. Ippoliti and V. Khemani, Learnability transitions in monitored quantum dynamics via eavesdropper’s classical shadows, arXiv:2307.15011  (2023).
  • Altland et al. [2022] A. Altland, D. A. Huse, and T. Micklitz, Maximum entropy quantum state distributions, arXiv:2203.12580  (2022).
  • Nielsen and Chuang [2010] M. A. Nielsen and I. L. Chuang, Quantum Computation and Quantum Information: 10th Anniversary Edition (Cambridge University Press, 2010).
  • Fisher et al. [2023] M. P. Fisher, V. Khemani, A. Nahum, and S. Vijay, Random quantum circuits, Annu. Rev. Cond. Mat. 14, 335 (2023).
  • Potter and Vasseur [2022] A. C. Potter and R. Vasseur, Entanglement dynamics in hybrid quantum circuits, in Entanglement in Spin Chains: From Theory to Quantum Technology Applications (Springer, 2022) pp. 211–249.
  • Li et al. [2018] Y. Li, X. Chen, and M. P. A. Fisher, Quantum zeno effect and the many-body entanglement transition, Phys. Rev. B 98, 205136 (2018).
  • Skinner et al. [2019] B. Skinner, J. Ruhman, and A. Nahum, Measurement-induced phase transitions in the dynamics of entanglement, Phys. Rev. X 9, 031009 (2019).
  • Stechel and Heller [1984] E. Stechel and E. Heller, Quantum ergodicity and spectral chaos, Annu. Rev. Phys. Chem. 35, 563 (1984).
  • Srednicki [1999] M. Srednicki, The approach to thermal equilibrium in quantized chaotic systems, Journal of Physics A: Mathematical and General 32, 1163 (1999).
  • Goldstein et al. [2010] S. Goldstein, J. L. Lebowitz, R. Tumulka, and N. Zanghì, Long-time behavior of macroscopic quantum systems, Euro. Phys. J. H 35, 173 (2010).
  • Lebowitz and Penrose [1973] J. L. Lebowitz and O. Penrose, Modern ergodic theory, Phys. Today 26, 23 (1973).
  • Linden et al. [2009] N. Linden, S. Popescu, A. J. Short, and A. Winter, Quantum mechanical evolution towards thermal equilibrium, Phys. Rev. E 79, 061103 (2009).
  • Nechita and Singh [2021] I. Nechita and S. Singh, A graphical calculus for integration over random diagonal unitary matrices, Linear Algebra and its Applications 613, 46–86 (2021).
  • Arute et al. [2019] F. Arute, K. Arya, R. Babbush, D. Bacon, J. C. Bardin, R. Barends, R. Biswas, S. Boixo, F. G. Brandão, D. A. Buell, et al., Quantum supremacy using a programmable superconducting processor, Nature (London) 574, 505 (2019).
  • Brin and Stuck [2002] M. Brin and G. Stuck, Introduction to dynamical systems (Cambridge university press, 2002).
  • Goldstein et al. [2016] S. Goldstein, J. L. Lebowitz, C. Mastrodonato, R. Tumulka, and N. Zanghì, Universal probability distribution for the wave function of a quantum system entangled with its environment, Commun. Math. Phys. 342, 965 (2016).
  • Ho and Choi [2022] W. W. Ho and S. Choi, Exact emergent quantum state designs from quantum chaotic dynamics, Phys. Rev. Lett. 128, 060601 (2022).
  • Shrotriya and Ho [2023] H. Shrotriya and W. W. Ho, Nonlocality of deep thermalization, arXiv:2305.08437  (2023).
  • Jozsa et al. [1994] R. Jozsa, D. Robb, and W. K. Wootters, Lower bound for accessible information in quantum mechanics, Phys. Rev. A 49, 668 (1994).
  • Goldstein et al. [2006] S. Goldstein, J. L. Lebowitz, R. Tumulka, and N. Zanghì, On the distribution of the wave function for systems in thermal equilibrium, J. Stat. Phys. 125, 1193 (2006).
  • Reimann [2008b] P. Reimann, Typicality of pure states randomly sampled according to the Gaussian adjusted projected measure, J. Stat. Phys. 132, 921 (2008b).
  • Schumacher [1995] B. Schumacher, Quantum coding, Phys. Rev. A 51, 2738 (1995).
  • Porter and Thomas [1956] C. E. Porter and R. G. Thomas, Fluctuations of nuclear reaction widths, Phys. Rev. 104, 483 (1956).
  • Boixo et al. [2018] S. Boixo, S. V. Isakov, V. N. Smelyanskiy, R. Babbush, N. Ding, Z. Jiang, M. J. Bremner, J. M. Martinis, and H. Neven, Characterizing quantum supremacy in near-term devices, Nature Phys. 14, 595 (2018).
  • Shaw et al. [2023] A. L. Shaw, Z. Chen, J. Choi, D. K. Mark, P. Scholl, R. Finkelstein, A. Elben, S. Choi, and M. Endres, Benchmarking highly entangled states on a 60-atom analog quantum simulator, arXiv:2308.07914  (2023).
  • Christandl [2006] M. Christandl, The structure of bipartite quantum states - insights from group theory and cryptography, arXiv:quant-ph/0604183  (2006).
  • Huang et al. [2020] H.-Y. Huang, R. Kueng, and J. Preskill, Predicting many properties of a quantum system from very few measurements, Nature Phys. 16, 1050 (2020).
  • Elben et al. [2023] A. Elben, S. T. Flammia, H.-Y. Huang, R. Kueng, J. Preskill, B. Vermersch, and P. Zoller, The randomized measurement toolbox, Nature Rev. Phys. 5, 9 (2023).
  • Datta et al. [2014] N. Datta, T. Dorlas, R. Jozsa, and F. Benatti, Properties of subentropy, J. Math. Phys. 55, 062203 (2014).
  • Valiant and Valiant [2011] G. Valiant and P. Valiant, The power of linear estimators, in 2011 IEEE 52nd Annual Symposium on Foundations of Computer Science (IEEE, 2011) pp. 403–412.
  • Valiant and Valiant [2017] G. Valiant and P. Valiant, Estimating the unseen: improved estimators for entropy and other properties, J. ACM 64, 1 (2017).
  • Nation and Porras [2018] C. Nation and D. Porras, Off-diagonal observable elements from random matrix theory: distributions, fluctuations, and eigenstate thermalization, N. J. Phys. 20, 103003 (2018), publisher: IOP Publishing.
  • Nation and Porras [2019] C. Nation and D. Porras, Ergodicity probes: using time-fluctuations to measure the Hilbert space dimension, Quantum 3, 207 (2019).
  • Mark et al. [2023] D. K. Mark, J. Choi, A. L. Shaw, M. Endres, and S. Choi, Benchmarking quantum simulators using ergodic quantum dynamics, Phys. Rev. Lett. 131, 110601 (2023).
  • Liu et al. [2023] Z. Liu, Z. Hao, and H.-Y. Hu, Predicting arbitrary state properties from single hamiltonian quench dynamics, arXiv:2311.00695  (2023).
  • Kaneko et al. [2020] K. Kaneko, E. Iyoda, and T. Sagawa, Characterizing complexity of many-body quantum dynamics by higher-order eigenstate thermalization, Phys. Rev. A 101, 042126 (2020).
  • Huang [2021] Y. Huang, Extensive entropy from unitary evolution, arXiv:2104.02053  (2021).
  • Riddell and Pagliaroli [2023] J. Riddell and N. Pagliaroli, No-resonance conditions, random matrices, and quantum chaotic models, arXiv:2307.05417  (2023).
  • Schmüdgen [2017] K. Schmüdgen, The Moment Problem (Springer International Publishing, Cham, 2017) pp. 381–398.
  • Kim et al. [2014] H. Kim, T. N. Ikeda, and D. A. Huse, Testing whether all eigenstates obey the eigenstate thermalization hypothesis, Phys. Rev. E 90, 052105 (2014).
  • Parfionov and Zapatrin [2006] G. Parfionov and R. Zapatrin, ‘Lazy’ quantum ensembles, J. Phys. A 39, 10891 (2006).
  • Murthy and Srednicki [2019] C. Murthy and M. Srednicki, Structure of chaotic eigenstates and their entanglement entropy, Phys. Rev. E 100, 022131 (2019).
  • McGill [1954] W. McGill, Multivariate information transmission, Transactions of the IRE Professional Group on Information Theory 4, 93 (1954).
  • Rangamani and Rota [2015] M. Rangamani and M. Rota, Entanglement structures in qubit systems, J. Phys. A 48, 385301 (2015).
  • Rota [2016] M. Rota, Tripartite information of highly entangled states, J. High Energy Phys. 2016 (4), 1.
  • Hartmann et al. [2004] M. Hartmann, G. Mahler, and O. Hess, Gaussian quantum fluctuations in interacting many particle systems, Lett. Math. Phys. 68, 103 (2004).
  • Ryu et al. [2010] S. Ryu, A. P. Schnyder, A. Furusaki, and A. W. Ludwig, Topological insulators and superconductors: tenfold way and dimensional hierarchy, N. J. Phys. 12, 065010 (2010).
  • Schiulaz et al. [2019] M. Schiulaz, E. J. Torres-Herrera, and L. F. Santos, Thouless and relaxation time scales in many-body quantum systems, Phys. Rev. B 99, 174313 (2019).
  • Tran et al. [2023] M. C. Tran, D. K. Mark, W. W. Ho, and S. Choi, Measuring arbitrary physical properties in analog quantum simulation, Phys. Rev. X 13, 011049 (2023).
  • Coopmans et al. [2023] L. Coopmans, Y. Kikuchi, and M. Benedetti, Predicting gibbs-state expectation values with pure thermal shadows, PRX Quantum 4, 010305 (2023).
  • Bassirian et al. [2021] R. Bassirian, A. Bouland, B. Fefferman, S. Gunn, and A. Tal, On certified randomness from quantum advantage experiments, arXiv:2111.14846  (2021).
  • Aaronson and Hung [2023] S. Aaronson and S.-H. Hung, Certified randomness from quantum supremacy, arXiv:2303.01625  (2023).
  • Foini and Kurchan [2019] L. Foini and J. Kurchan, Eigenstate thermalization hypothesis and out of time order correlators, Phys. Rev. E 99, 042139 (2019).
  • Wang et al. [2022] J. Wang, M. H. Lamann, J. Richter, R. Steinigeweg, A. Dymarsky, and J. Gemmer, Eigenstate thermalization hypothesis and its deviations from random-matrix theory beyond the thermalization time, Phys. Rev. Lett. 128, 180601 (2022).
  • Pappalardi et al. [2022] S. Pappalardi, L. Foini, and J. Kurchan, Eigenstate thermalization hypothesis and free probability, Phys. Rev. Lett. 129, 170603 (2022).
  • Erdos [1942] P. Erdos, On an elementary proof of some asymptotic formulas in the theory of partitions, Ann. Math. , 437 (1942).
  • Liu [2018] J. Liu, Spectral form factors and late time quantum chaos, Phys. Rev. D 98, 086026 (2018).
  • Abramowitz and Stegun [1964] M. Abramowitz and I. A. Stegun, Handbook of mathematical functions with formulas, graphs, and mathematical tables (Dover, New York, 1964) p. 364, tenth printing ed.
  • Mehta [2004] M. L. Mehta, Random matrices (Elsevier, 2004).
  • Keating et al. [2015] J. P. Keating, N. Linden, and H. J. Wells, Spectra and eigenstates of spin chain hamiltonians, Comm. Math. Phys. 338, 81 (2015).
  • LeBlond et al. [2021] T. LeBlond, D. Sels, A. Polkovnikov, and M. Rigol, Universality in the onset of quantum chaos in many-body systems, Phys. Rev. B 104, L201117 (2021).