Eigenstate thermalization hypothesis

The eigenstate thermalization hypothesis (or ETH) is a set of ideas which purports to explain when and why an isolated quantum mechanical system can be accurately described using equilibrium statistical mechanics. In particular, it is devoted to understanding how systems which are initially prepared in far-from-equilibrium states can evolve in time to a state which appears to be in thermal equilibrium. The phrase "eigenstate thermalization" was first coined by Mark Srednicki in 1994, after similar ideas had been introduced by Josh Deutsch in 1991. The principal philosophy underlying the eigenstate thermalization hypothesis is that instead of explaining the ergodicity of a thermodynamic system through the mechanism of dynamical chaos, as is done in classical mechanics, one should instead examine the properties of matrix elements of observable quantities in individual energy eigenstates of the system.

Motivation
In statistical mechanics, the microcanonical ensemble is a particular statistical ensemble which is used to make predictions about the outcomes of experiments performed on isolated systems that are believed to be in equilibrium with an exactly known energy. The microcanonical ensemble is based upon the assumption that, when such an equilibrated system is probed, the probability for it to be found in any of the microscopic states with the same total energy have equal probability. With this assumption, the ensemble average of an observable quantity is found by averaging the value of that observable $$A_i$$ over all microstates $$i$$ with the correct total energy:
 * $$\bar{A}_{\mathrm{classical}} = \frac{1}{N}\sum_{i=1}^N A_i$$

Importantly, this quantity is independent of everything about the initial state except for its energy.

The assumptions of ergodicity are well-motivated in classical mechanics as a result of dynamical chaos, since a chaotic system will in general spend equal time in equal areas of its phase space. If we prepare an isolated, chaotic, classical system in some region of its phase space, then as the system is allowed to evolve in time, it will sample its entire phase space, subject only to a small number of conservation laws (such as conservation of total energy). If one can justify the claim that a given physical system is ergodic, then this mechanism will provide an explanation for why statistical mechanics is successful in making accurate predictions. For example, the hard sphere gas has been rigorously proven to be ergodic.

This argument cannot be straightforwardly extended to quantum systems, even ones that are analogous to chaotic classical systems, because time evolution of a quantum system does not uniformly sample all vectors in Hilbert space with a given energy. Given the state at time zero in a basis of energy eigenstates

$$ the expectation value of any observable $$\hat{A}$$ is
 * \Psi(0) \rangle = \sum_{\alpha} c_{\alpha} | E_{\alpha} \rangle,

\langle \hat{A} \rangle_t \equiv \langle \Psi(t) | \hat{A} |\Psi(t) \rangle = \sum_{\alpha, \beta} c_{\alpha}^{*} c_{\beta} A_{\alpha \beta} e^{-i \left ( E_{\beta} - E_{\alpha} \right )t / \hbar }. $$ Even if the $$E_{\alpha}$$ are incommensurate, so that this expectation value is given for long times by

\langle \hat{A} \rangle_t \overset{t\to\infty}{\approx} \sum_{\alpha} \vert c_{\alpha}\vert^2 A_{\alpha \alpha}, $$ the expectation value permanently retains knowledge of the initial state in the form of the coefficients $$c_{\alpha}$$.

In principle it is thus an open question as to whether an isolated quantum mechanical system, prepared in an arbitrary initial state, will approach a state which resembles thermal equilibrium, in which a handful of observables are adequate to make successful predictions about the system. However, a variety of experiments in cold atomic gases have indeed observed thermal relaxation in systems which are, to a very good approximation, completely isolated from their environment, and for a wide class of initial states. The task of explaining this experimentally observed applicability of equilibrium statistical mechanics to isolated quantum systems is the primary goal of the eigenstate thermalization hypothesis.

Statement
Suppose that we are studying an isolated, quantum mechanical many-body system. In this context, "isolated" refers to the fact that the system has no (or at least negligible) interactions with the environment external to it. If the Hamiltonian of the system is denoted $$\hat{H}$$, then a complete set of basis states for the system is given in terms of the eigenstates of the Hamiltonian,

\hat{H} |E_{\alpha} \rangle = E_{\alpha}|E_{\alpha} \rangle , $$ where $$|E_{\alpha} \rangle$$ is the eigenstate of the Hamiltonian with eigenvalue $$E_{\alpha} $$. We will refer to these states simply as "energy eigenstates." For simplicity, we will assume that the system has no degeneracy in its energy eigenvalues, and that it is finite in extent, so that the energy eigenvalues form a discrete, non-degenerate spectrum (this is not an unreasonable assumption, since any "real" laboratory system will tend to have sufficient disorder and strong enough interactions as to eliminate almost all degeneracy from the system, and of course will be finite in size ). This allows us to label the energy eigenstates in order of increasing energy eigenvalue. Additionally, consider some other quantum-mechanical observable $$\hat{A}$$, which we wish to make thermal predictions about. The matrix elements of this operator, as expressed in a basis of energy eigenstates, will be denoted by

A_{\alpha \beta} \equiv \langle E_{\alpha} | \hat{A} | E_{\beta} \rangle. $$

We now imagine that we prepare our system in an initial state for which the expectation value of $$\hat{A}$$ is far from its value predicted in a microcanonical ensemble appropriate to the energy scale in question (we assume that our initial state is some superposition of energy eigenstates which are all sufficiently "close" in energy). The eigenstate thermalization hypothesis says that for an arbitrary initial state, the expectation value of $$\hat{A}$$ will ultimately evolve in time to its value predicted by a microcanonical ensemble, and thereafter will exhibit only small fluctuations around that value, provided that the following two conditions are met: These conditions can be written as
 * 1) The diagonal matrix elements $$A_{\alpha \alpha}$$ vary smoothly as a function of energy, with the difference between neighboring values, $$A_{\alpha + 1,\alpha + 1} - A_{\alpha,\alpha}$$, becoming exponentially small in the system size.
 * 2) The off-diagonal matrix elements $$A_{\alpha \beta}$$, with $$\alpha \neq \beta$$, are much smaller than the diagonal matrix elements, and in particular are themselves exponentially small in the system size.

A_{\alpha\beta} \simeq \overline{A}\delta_{\alpha\beta}+\sqrt{\frac{\overline{A^2}}{\mathcal{D}}}R_{\alpha\beta}, $$ where $$\overline{A}=\overline{A}(E_\alpha)$$ and $$\overline{A^2}=\overline{A^2}(E_\alpha,E_\beta)$$ are smooth functions of energy, $$\mathcal{D}=e^{sV}$$ is the many-body Hilbert space dimension, and $$R_{\alpha\beta}$$ is a random variable with zero mean and unit variance. Conversely if a quantum many-body system satisfies the ETH, the matrix representation of any local operator in the energy eigen basis is expected to follow the above ansatz.

Equivalence of the diagonal and microcanonical ensembles
We can define a long-time average of the expectation value of the operator $$\hat A$$ according to the expression

\overline{A} \equiv \lim_{\tau \to \infty} \frac{1}{\tau}\int_{0}^{\tau}\langle \Psi(t) | \hat A |\Psi(t)\rangle~ dt. $$ If we use the explicit expression for the time evolution of this expectation value, we can write

\overline{A} = \lim_{\tau \to \infty} \frac{1}{\tau}\int_{0}^{\tau}\left [ \sum_{\alpha, \beta=1}^{D}c_{\alpha}^{*}c_{\beta} A_{\alpha \beta} e^{-i \left (E_{\beta} - E_{\alpha} \right )t/\hbar} \right ]~ dt. $$ The integration in this expression can be performed explicitly, and the result is

\overline{A} = \sum_{\alpha=1}^{D}|c_{\alpha}|^{2}A_{\alpha \alpha} + i \hbar \lim_{\tau \to \infty} \left [ \sum_{\alpha \neq \beta}^{D} \frac{c_{\alpha}^{*}c_{\beta}A_{\alpha \beta}}{E_{\beta}-E_{\alpha}}\left ( \frac{e^{-i \left (E_{\beta} - E_{\alpha} \right )\tau/\hbar}-1}{\tau} \right ) \right ]. $$ Each of the terms in the second sum will become smaller as the limit is taken to infinity. Assuming that the phase coherence between the different exponential terms in the second sum does not ever become large enough to rival this decay, the second sum will go to zero, and we find that the long-time average of the expectation value is given by

\overline{A} = \sum_{\alpha=1}^{D}|c_{\alpha}|^{2}A_{\alpha \alpha}. $$

This prediction for the time-average of the observable $$\hat A$$ is referred to as its predicted value in the diagonal ensemble, The most important aspect of the diagonal ensemble is that it depends explicitly on the initial state of the system, and so would appear to retain all of the information regarding the preparation of the system. In contrast, the predicted value in the microcanonical ensemble is given by the equally-weighted average over all energy eigenstates within some energy window centered around the mean energy of the system

\langle A \rangle_{\text{mc}} = \frac{1}{\mathcal{N}} \sum_{\alpha'=1}^{\mathcal{N}}A_{\alpha' \alpha'}, $$ where $$\mathcal{N}$$ is the number of states in the appropriate energy window, and the prime on the sum indices indicates that the summation is restricted to this appropriate microcanonical window. This prediction makes absolutely no reference to the initial state of the system, unlike the diagonal ensemble. Because of this, it is not clear why the microcanonical ensemble should provide such an accurate description of the long-time averages of observables in such a wide variety of physical systems.

However, suppose that the matrix elements $$A_{\alpha \alpha}$$ are effectively constant over the relevant energy window, with fluctuations that are sufficiently small. If this is true, this one constant value A can be effectively pulled out of the sum, and the prediction of the diagonal ensemble is simply equal to this value,

\overline{A} = \sum_{\alpha=1}^{D}|c_{\alpha}|^{2}A_{\alpha \alpha} \approx A\sum_{\alpha=1}^{D}|c_{\alpha}|^{2} = A, $$ where we have assumed that the initial state is normalized appropriately. Likewise, the prediction of the microcanonical ensemble becomes

\langle A \rangle_{\text{mc}} = \frac{1}{\mathcal{N}} \sum_{\alpha'=1}^{\mathcal{N}}A_{\alpha' \alpha'} \approx \frac{1}{\mathcal{N}} \sum_{\alpha'=1}^{\mathcal{N}}A = A. $$ The two ensembles are therefore in agreement.

This constancy of the values of $$A_{\alpha \alpha}$$ over small energy windows is the primary idea underlying the eigenstate thermalization hypothesis. Notice that in particular, it states that the expectation value of $$\hat A$$ in a single energy eigenstate is equal to the value predicted by a microcanonical ensemble constructed at that energy scale. This constitutes a foundation for quantum statistical mechanics which is radically different from the one built upon the notions of dynamical ergodicity.

Tests
Several numerical studies of small lattice systems appear to tentatively confirm the predictions of the eigenstate thermalization hypothesis in interacting systems which would be expected to thermalize. Likewise, systems which are integrable tend not to obey the eigenstate thermalization hypothesis.

Some analytical results can also be obtained if one makes certain assumptions about the nature of highly excited energy eigenstates. The original 1994 paper on the ETH by Mark Srednicki studied, in particular, the example of a quantum hard sphere gas in an insulated box. This is a system which is known to exhibit chaos classically. For states of sufficiently high energy, Berry's conjecture states that energy eigenfunctions in this many-body system of hard sphere particles will appear to behave as superpositions of plane waves, with the plane waves entering the superposition with random phases and Gaussian-distributed amplitudes (the precise notion of this random superposition is clarified in the paper). Under this assumption, one can show that, up to corrections which are negligibly small in the thermodynamic limit, the momentum distribution function for each individual, distinguishable particle is equal to the Maxwell–Boltzmann distribution

f_{\rm MB} \left ( \mathbf{p}, T_{\alpha} \right ) = \left ( 2 \pi m k T \right )^{-3/2}e^{-\mathbf{p}^{2}/2mkT_{\alpha}}, $$ where $$\mathbf{p}$$ is the particle's momentum, m is the mass of the particles, k is the Boltzmann constant, and the "temperature" $$T_{\alpha}$$ is related to the energy of the eigenstate according to the usual equation of state for an ideal gas,

E_{\alpha} = \frac{3}{2}NkT_{\alpha}, $$ where N is the number of particles in the gas. This result is a specific manifestation of the ETH, in that it results in a prediction for the value of an observable in one energy eigenstate which is in agreement with the prediction derived from a microcanonical (or canonical) ensemble. Note that no averaging over initial states whatsoever has been performed, nor has anything resembling the H-theorem been invoked. Additionally, one can also derive the appropriate Bose–Einstein or Fermi–Dirac distributions, if one imposes the appropriate commutation relations for the particles comprising the gas.

Currently, it is not well understood how high the energy of an eigenstate of the hard sphere gas must be in order for it to obey the ETH. A rough criterion is that the average thermal wavelength of each particle be sufficiently smaller than the radius of the hard sphere particles, so that the system can probe the features which result in chaos classically (namely, the fact that the particles have a finite size ). However, it is conceivable that this condition may be able to be relaxed, and perhaps in the thermodynamic limit, energy eigenstates of arbitrarily low energies will satisfy the ETH (aside from the ground state itself, which is required to have certain special properties, for example, the lack of any nodes ).

Alternatives
Three alternative explanations for the thermalization of isolated quantum systems are often proposed:


 * 1) For initial states of physical interest, the coefficients $$c_{\alpha}$$ exhibit large fluctuations from eigenstate to eigenstate, in a fashion which is completely uncorrelated with the fluctuations of $$A_{\alpha \alpha}$$ from eigenstate to eigenstate. Because the coefficients and matrix elements are uncorrelated, the summation in the diagonal ensemble is effectively performing an unbiased sampling of the values of $$A_{\alpha \alpha}$$ over the appropriate energy window. For a sufficiently large system, this unbiased sampling should result in a value which is close to the true mean of the values of $$A_{\alpha \alpha}$$ over this window, and will effectively reproduce the prediction of the microcanonical ensemble. However, this mechanism may be disfavored for the following heuristic reason. Typically, one is interested in physical situations in which the initial expectation value of $$\hat A$$ is far from its equilibrium value. For this to be true, the initial state must contain some sort of specific information about $$\hat A$$, and so it becomes suspect whether or not the initial state truly represents an unbiased sampling of the values of $$A_{\alpha \alpha}$$ over the appropriate energy window. Furthermore, whether or not this were to be true, it still does not provide an answer to the question of when arbitrary initial states will come to equilibrium, if they ever do.
 * 2) For initial states of physical interest, the coefficients $$c_{\alpha}$$ are effectively constant, and do not fluctuate at all. In this case, the diagonal ensemble is precisely the same as the microcanonical ensemble, and there is no mystery as to why their predictions are identical. However, this explanation is disfavored for much the same reasons as the first.
 * 3) Integrable quantum systems are proved to thermalize under condition of simple regular time-dependence of parameters, suggesting that cosmological expansion of the Universe and integrability of the most fundamental equations of motion are ultimately responsible for thermalization.

Temporal fluctuations of expectation values
The condition that the ETH imposes on the diagonal elements of an observable is responsible for the equality of the predictions of the diagonal and microcanonical ensembles. However, the equality of these long-time averages does not guarantee that the fluctuations in time around this average will be small. That is, the equality of the long-time averages does not ensure that the expectation value of $$\hat A$$ will settle down to this long-time average value, and then stay there for most times.

In order to deduce the conditions necessary for the observable's expectation value to exhibit small temporal fluctuations around its time-average, we study the mean squared amplitude of the temporal fluctuations, defined as

\overline{\left (A_{t} - \overline{A} \right )^{2}} \equiv \lim_{\tau \to \infty}\frac{1}{\tau} \int_{0}^{\tau } \left (A_{t} - \overline{A} \right )^{2}dt, $$ where $$A_{t}$$ is a shorthand notation for the expectation value of $$\hat A$$ at time t. This expression can be computed explicitly, and one finds that

\overline{\left (A_{t} - \overline{A} \right )^{2}} = \sum_{\alpha \neq \beta}|c_{\alpha}|^{2}|c_{\beta}|^{2}|A_{\alpha \beta}|^{2}. $$ Temporal fluctuations about the long-time average will be small so long as the off-diagonal elements satisfy the conditions imposed on them by the ETH, namely that they become exponentially small in the system size. Notice that this condition allows for the possibility of isolated resurgence times, in which the phases align coherently in order to produce large fluctuations away from the long-time average. The amount of time the system spends far away from the long-time average is guaranteed to be small so long as the above mean squared amplitude is sufficiently small. If a system poses a dynamical symmetry, however, it will periodically oscillate around the long-time average.

Quantum fluctuations and thermal fluctuations
The expectation value of a quantum mechanical observable represents the average value which would be measured after performing repeated measurements on an ensemble of identically prepared quantum states. Therefore, while we have been examining this expectation value as the principal object of interest, it is not clear to what extent this represents physically relevant quantities. As a result of quantum fluctuations, the expectation value of an observable is not typically what will be measured during one experiment on an isolated system. However, it has been shown that for an observable satisfying the ETH, quantum fluctuations in its expectation value will typically be of the same order of magnitude as the thermal fluctuations which would be predicted in a traditional microcanonical ensemble. This lends further credence to the idea that the ETH is the underlying mechanism responsible for the thermalization of isolated quantum systems.

General validity
Currently, there is no known analytical derivation of the eigenstate thermalization hypothesis for general interacting systems. However, it has been verified to be true for a wide variety of interacting systems using numerical exact diagonalization techniques, to within the uncertainty of these methods. It has also been proven to be true in certain special cases in the semi-classical limit, where the validity of the ETH rests on the validity of Shnirelman's theorem, which states that in a system which is classically chaotic, the expectation value of an operator $$\hat A$$ in an energy eigenstate is equal to its classical, microcanonical average at the appropriate energy. Whether or not it can be shown to be true more generally in interacting quantum systems remains an open question. It is also known to explicitly fail in certain integrable systems, in which the presence of a large number of constants of motion prevent thermalization.

It is also important to note that the ETH makes statements about specific observables on a case-by-case basis - it does not make any claims about whether every observable in a system will obey ETH. In fact, this certainly cannot be true. Given a basis of energy eigenstates, one can always explicitly construct an operator which violates the ETH, simply by writing down the operator as a matrix in this basis whose elements explicitly do not obey the conditions imposed by the ETH. Conversely, it is always trivially possible to find operators which do satisfy ETH, by writing down a matrix whose elements are specifically chosen to obey ETH. In light of this, one may be led to believe that the ETH is somewhat trivial in its usefulness. However, the important consideration to bear in mind is that these operators thus constructed may not have any physical relevance. While one can construct these matrices, it is not clear that they correspond to observables which could be realistically measured in an experiment, or bear any resemblance to physically interesting quantities. An arbitrary Hermitian operator on the Hilbert space of the system need not correspond to something which is a physically measurable observable.

Typically, the ETH is postulated to hold for "few-body operators," observables which involve only a small number of particles. Examples of this would include the occupation of a given momentum in a gas of particles, or the occupation of a particular site in a lattice system of particles. Notice that while the ETH is typically applied to "simple" few-body operators such as these, these observables need not be local in space - the momentum number operator in the above example does not represent a local quantity.

There has also been considerable interest in the case where isolated, non-integrable quantum systems fail to thermalize, despite the predictions of conventional statistical mechanics. Disordered systems which exhibit many-body localization are candidates for this type of behavior, with the possibility of excited energy eigenstates whose thermodynamic properties more closely resemble those of ground states. It remains an open question as to whether a completely isolated, non-integrable system without static disorder can ever fail to thermalize. One intriguing possibility is the realization of "Quantum Disentangled Liquids." It also an open question whether all eigenstates must obey the ETH in a thermalizing system.

The eigenstate thermalization hypothesis is closely connected to the quantum nature of chaos (see quantum chaos). Furthermore, since a classically chaotic system is also ergodic, almost all of its trajectories eventually explore uniformly the entire accessible phase space, which would imply the eigenstates of the quantum chaotic system fill the quantum phase space evenly (up to random fluctuations) in the semiclassical limit $$\hbar \rightarrow 0$$. In particular, there is a quantum ergodicity theorem showing that the expectation value of an operator converges to the corresponding microcanonical classical average as $$\hbar \rightarrow 0$$. However, the quantum ergodicity theorem leaves open the possibility of non-ergodic states such as quantum scars. In addition to the conventional scarring,   there are two other types of quantum scarring, which further illustrate the weak-ergodicity breaking in quantum chaotic systems: perturbation-induced   and many-body quantum scars. Since the former arise a combined effect of special nearly-degenerate unperturbed states and the localized nature of the perturbation (potential bums), the scarring can slow down the thermalization process in disordered quantum dots and wells, which is further illustrated by the fact that these quantum scars can be used to propagate quantum wave packets in a disordered nanostructure with high fidelity. On the other hand, the latter form of scarring has been speculated to be the culprit behind the unexpectedly slow thermalization of cold atoms observed experimentally.