Spectral theory is a branch of functional analysis in mathematics that generalizes the eigenvalue and eigenvector theory from finite-dimensional vector spaces to linear operators on infinite-dimensional spaces, such as Banach and Hilbert spaces.[1] It focuses on the spectrum of an operator, defined as the set of complex numbers \lambda for which the operator A - \lambda I is not invertible, providing a framework to decompose operators and solve associated equations.[2] This theory enables the representation of operators in forms amenable to analysis, such as diagonal or block-diagonal structures, reducing complex problems into manageable invariant subspaces.[3]Central to spectral theory are the classifications of the spectrum: the point spectrum consists of eigenvalues \lambda where there exists a non-zero eigenvector u satisfying Au = \lambda u; the continuous spectrum involves approximate eigenvalues without exact eigenvectors; and the residual spectrum covers values where A - \lambda I is injective but not surjective.[2] For self-adjoint operators on Hilbert spaces, the spectral theorem guarantees a unique spectral measure that integrates to reconstruct the operator, allowing it to act as multiplication by a function on a suitable space.[2] This theorem, along with extensions to normal and compact operators, forms the cornerstone for understanding operator behavior and stability.[1]Spectral theory originated in the 19th century with Fourier's work on trigonometric series for solving heat equations and Sturm-Liouville boundary value problems, which implicitly involved spectral decompositions.[4] It advanced through Fredholm's 1900 spectral theorem for integral operators and Hilbert's formulation of infinite-dimensional spaces around 1904–1912, culminating in von Neumann's 1927–1929 extensions to unbounded operators.[4] Applications span quantum mechanics, where operator spectra determine energy eigenvalues and stationary states; differential geometry, as in the analysis of Laplace-Beltrami operators for manifold shapes; and numerical methods for partial differential equations via finite element approximations.[2][3]
Introduction
Overview and scope
Spectral theory is a branch of functional analysis that studies the eigenvalues, eigenvectors, and generalized eigenspaces of linear operators acting on infinite-dimensional spaces, particularly extending the concept of diagonalization from finite-dimensional linear algebra to more complex settings.[5] In this framework, the theory seeks to classify operators by analyzing their spectral properties, allowing for a decomposition of the operator into simpler, often multiplicative, components that facilitate understanding and computation.[6]Unlike finite-dimensional spectral theory, which relies on tools like characteristic polynomials to determine eigenvalues for matrices, spectral theory in infinite dimensions addresses the absence of such finite methods and emphasizes operators—bounded or unbounded—defined on Hilbert spaces.[5] This extension is crucial because many natural operators arising in analysis and physics, such as those in partial differential equations, are unbounded and require careful treatment within the structure of Hilbert spaces to ensure well-defined spectra.[6]A central role of spectral theory lies in its use of spectral measures to decompose operators, enabling the representation of an operator as an integral with respect to these measures, which simplifies the analysis of operator actions on vectors.[5] This decomposition provides a powerful tool for solving problems where direct computation is infeasible. Key motivations include the simplification of differential equations, where spectral decompositions reveal solution behaviors through eigenvalues, and the modeling of quantum observables, where self-adjoint operators correspond to physical quantities like energy.[5] Pioneered by mathematicians such as David Hilbert and Marshall Stone, the theory has become foundational in modern mathematics and physics.[4]
Historical development
The origins of spectral theory lie in the 19th-century study of eigenvalues and eigenvectors for finite matrices. Augustin-Louis Cauchy laid early foundations in 1826 by analyzing quadratic forms in n variables, where he introduced the term "tableau" for the coefficient matrix, proved that real symmetric matrices have real eigenvalues, and demonstrated their diagonalizability over the reals.[7] These results established key algebraic properties that would underpin later generalizations. Leopold Kronecker further advanced the field in the 1850s and 1860s through his work on linear transformations and determinants, providing tools for handling matrix equations that influenced later developments.[7]David Hilbert's contributions from 1906 to 1910 marked a pivotal shift toward infinite-dimensional operators. Motivated by Fredholm's work on integral equations, Hilbert developed spectral decompositions for bounded symmetric integral operators acting on spaces of square-integrable functions, which he formalized as Hilbert spaces. In a series of six papers culminating in his 1910 work, he introduced the concept of the continuous spectrum and the resolvent operator, freeing spectral analysis from strict ties to integral equations and establishing an abstract framework.[8][4]In the 1930s, Marshall Stone generalized Hilbert's results to unbounded self-adjoint operators. Stone's 1932 book, Linear Transformations in Hilbert Space, provided a rigorous spectral theorem integrating measure theory and functional analysis, resolving open questions about unbounded operators and solidifying the role of Hilbert spaces in operator theory.[4] Concurrently, John von Neumann extended spectral theory to normal operators, including the unbounded case, in works from 1929 onward, applying it to quantum mechanics through abstract Hilbert space formulations. Post-World War II, von Neumann's advancements emphasized applications in quantum theory and ergodic theory within functional analysis, influencing computational and physical modeling.[4]Modern extensions in the mid-20th century included Béla Sz.-Nagy's dilation theorem in 1953, which showed that every contractionoperator on a Hilbert space dilates to a unitary operator, broadening spectral methods to non-self-adjoint cases and inspiring further developments in operator theory.[9]
Background Concepts
Mathematical foundations
A Hilbert space is a complete inner product space over the real or complex numbers, providing the natural setting for spectral theory due to its geometric structure and convergence properties. Specifically, it is a vector space H equipped with an inner product \langle \cdot, \cdot \rangle that satisfies linearity in the first argument, conjugate symmetry \langle x, y \rangle = \overline{\langle y, x \rangle}, and positive definiteness \langle x, x \rangle \geq 0 with equality if and only if x = 0. The induced norm is \|x\| = \sqrt{\langle x, x \rangle}, turning H into a metric space. Completeness means every Cauchy sequence \{x_n\} (where \|x_m - x_n\| \to 0 as m, n \to \infty) converges to some x \in H. Orthogonality holds when \langle x, y \rangle = 0, enabling decompositions like the orthogonal direct sum H = M \oplus M^\perp for closed subspaces M, where M^\perp = \{ y \in H : \langle x, y \rangle = 0 \ \forall x \in M \}.[10]Linear operators on a Hilbert space H map vectors to vectors, defined as T: D(T) \subseteq H \to H, with D(T) as the domain. Bounded operators satisfy D(T) = H and have finite operator norm \|T\| = \sup_{\|x\| \leq 1} \|Tx\| < \infty, implying continuity and uniform boundedness. Unbounded operators, in contrast, are defined on proper subspaces D(T) \subsetneq H and lack a uniform bound, often arising in differential equations. The adjoint T^* of a densely defined operator T is the unique operator satisfying \langle Tx, y \rangle = \langle x, T^* y \rangle for all x \in D(T), y \in D(T^*), where D(T^*) = \{ y \in H : \exists z \in H \ \langle Tx, y \rangle = \langle x, z \rangle \ \forall x \in D(T) \}. For bounded T, T^* is also bounded with \|T^*\| = \|T\|.[11][12]Self-adjoint operators on H are those with T = T^* on a common domain, equivalently \langle Tx, y \rangle = \langle x, Ty \rangle for all x, y \in D(T), ensuring real eigenvalues and orthogonal eigenspaces in finite dimensions. Normal operators satisfy TT^* = T^*T, encompassing self-adjoint and unitary operators (where T^* T = I); they preserve the inner product structure, with \|T x\| = \|T^* x\| for all x. These operators are crucial for maintaining symmetry and unitarity in infinite-dimensional settings.[11]Basic functional analysis concepts underpin the rigor of operator theory. A subspace D \subseteq H is dense if its closure \overline{D} = H, allowing operators defined on D to approximate actions on all of H via limits. Closed operators have closed graphs G(T) = \{ (x, Tx) : x \in D(T) \} \subseteq H \oplus H, meaning if x_n \to x and Tx_n \to y, then x \in D(T) and Tx = y; this ensures the domain is complete under the graph norm \|x\|_{D(T)} = \sqrt{\|x\|^2 + \|Tx\|^2}. The complex plane \mathbb{C} plays a foundational role in resolvents, where points \lambda \in \mathbb{C} outside certain regions allow invertible shifts \lambda I - T, facilitating analytic tools for operator behavior.[13]
Physical motivations
Spectral theory emerged as a crucial framework in physics to address problems involving discrete energy levels and oscillatory behaviors, driven by early 20th-century discoveries in radiation and mechanics. Max Planck's introduction of energy quantization in 1900 to resolve the blackbody radiation spectrum implied that physical systems exhibit discrete spectral lines rather than continuous distributions, laying the groundwork for analyzing atomic spectra through eigenvalue decompositions.[14] This quantization concept necessitated mathematical tools to handle the spectra of physical operators, influencing the development of spectral theory in infinite-dimensional spaces. Later, Paul Dirac's formulation in the 1930s further emphasized spectral decompositions by representing quantum observables as self-adjoint operators whose eigenvalues correspond to measurable outcomes, unifying earlier approaches in quantum mechanics.In classical mechanics, spectral theory finds motivation in the study of vibrations and waves, where normal modes represent independent oscillatory patterns that diagonalize the system's dynamics. For instance, the vibration of a string or membrane leads to eigenvalue problems for differential operators, as seen in the separation of variables for wave equations, where eigenvalues determine the frequencies of normal modes.[4] This approach, rooted in 19th-century work on Sturm-Liouville boundary value problems motivated by wave propagation and heat conduction, transforms coupled oscillations into uncoupled ones via spectral analysis, providing a physical interpretation of eigenvalues as resonant frequencies.[15]Quantum mechanics provided a profound physical impetus for spectral theory, particularly through the time-independent Schrödinger equation, which poses a spectral problem for the Hamiltonian operator: H \psi = E \psi, where E are the energy eigenvalues representing possible measurement outcomes for the system's energy.[16] Here, operators such as position and momentum, representing physical observables, have spectra that dictate the discrete results of quantum measurements, as formalized in the 1920s by Heisenberg and Schrödinger's matrix and wave mechanics.[4] John von Neumann's extensions in 1929 to unbounded operators in Hilbert space were directly inspired by these quantum needs, enabling rigorous spectral decompositions for Hamiltonians in realistic physical models.[4]
Core Definitions
Spectrum of an operator
In functional analysis, the spectrum of a bounded linear operator T on a complex Banach space X is defined as the set \sigma(T) = \{ \lambda \in \mathbb{C} : T - \lambda I \text{ does not have a bounded inverse in } \mathcal{B}(X) \}, where \mathcal{B}(X) is the algebra of bounded linear operators on X.[6] This set generalizes the notion of eigenvalues from finite-dimensional linear algebra to infinite-dimensional settings, capturing values of \lambda for which T - \lambda I fails to be bijective.[6] The spectrum is always a nonempty compact subset of \mathbb{C}.[6]The spectrum \sigma(T) decomposes into three disjoint subsets: the point spectrum \sigma_p(T), the continuous spectrum \sigma_c(T), and the residual spectrum \sigma_r(T). The point spectrum consists of eigenvalues, i.e., \sigma_p(T) = \{ \lambda \in \mathbb{C} : T - \lambda I \text{ is not injective} \}, where the kernel of T - \lambda I contains nonzero vectors.[6] The continuous spectrum includes points where T - \lambda I is injective with dense but not closed range (hence not surjective).[6] The residual spectrum comprises points where T - \lambda I is injective but the range is not dense.[6] These distinctions highlight how invertibility fails in different ways, with the point spectrum reflecting discrete spectral behavior and the others indicating more continuous or defective structures.[6]A classic example illustrating these components is the unilateral right shift operator R on the Hilbert space \ell^2(\mathbb{N}), defined by R(x_1, x_2, x_3, \dots) = (0, x_1, x_2, \dots). Here, \sigma(R) is the closed unit disk \{ \lambda \in \mathbb{C} : |\lambda| \leq 1 \}, with empty point spectrum \sigma_p(R) = \emptyset, residual spectrum \sigma_r(R) = \{ \lambda \in \mathbb{C} : |\lambda| < 1 \}, and continuous spectrum \sigma_c(R) = \{ \lambda \in \mathbb{C} : |\lambda| = 1 \}.[17] In contrast, the adjoint left shift L = R^*, given by L(x_1, x_2, x_3, \dots) = (x_2, x_3, \dots), has point spectrum \sigma_p(L) = \{ \lambda \in \mathbb{C} : |\lambda| < 1 \}, empty residual spectrum, and continuous spectrum on the unit circle, showing how adjoint operators can swap residual and point spectral components.[17]For self-adjoint operators A on a Hilbert space, the spectrum \sigma(A) lies on the real line, i.e., \sigma(A) \subseteq \mathbb{R}, reflecting the operator's symmetry.[18] Moreover, the spectral radius equals the operator norm, so \sup \{ |\lambda| : \lambda \in \sigma(A) \} = \|A\|, bounding the spectrum within [- \|A\|, \|A\| ].[18]In infinite-dimensional spaces, the geometric multiplicity of an eigenvalue \lambda \in \sigma_p(T) is the dimension of the eigenspace \ker(T - \lambda I), which can be finite or infinite and measures the "degeneracy" of the eigenvector space.[19] The algebraic multiplicity, generalizing the finite-dimensional case, is the dimension of the generalized eigenspace \bigcup_{n=1}^\infty \ker(T - \lambda I)^n, or equivalently the ascent of T - \lambda I (the smallest m such that \ker(T - \lambda I)^m = \ker(T - \lambda I)^{m+k} for all k \geq 0), which may exceed the geometric multiplicity for non-normal operators.[20] For self-adjoint operators, geometric and algebraic multiplicities coincide for each eigenvalue.[19] The resolvent set \rho(T) = \mathbb{C} \setminus \sigma(T) consists of points where T - \lambda I is invertible.[6]
Resolvent operator
The resolvent operator serves as a fundamental analytic tool in spectral theory for investigating the spectrum of a linear operator T acting on a Banach space. For a complex number \lambda belonging to the resolvent set \rho(T), which consists of those \lambda for which T - \lambda I is bijective with a bounded inverse, the resolvent is defined byR(\lambda, T) = (T - \lambda I)^{-1}.The spectrum \sigma(T) is precisely the complement of \rho(T) in the complex plane.[6] This operator-valued function encodes essential information about the operator's behavior, particularly through its singularities at points in \sigma(T).The resolvent R(\lambda, T) exhibits strong analytic properties: it is holomorphic as a function from \rho(T) to the space of bounded operators, meaning it admits a power series expansion around any point in \rho(T). Specifically, near a point \mu \in \rho(T),R(\lambda, T) = \sum_{n=0}^{\infty} (-1)^n (\lambda - \mu)^n R(\mu, T)^{n+1},valid for |\lambda - \mu| < \|R(\mu, T)\|^{-1}. The singularities of R(\lambda, T) occur at the points of \sigma(T), where isolated eigenvalues typically manifest as poles, while continuous spectrum leads to branch points or essential singularities.[21][6]A key relation is the resolvent identity, which connects the resolvents at distinct points in \rho(T): for \lambda, \mu \in \rho(T) with \lambda \neq \mu,R(\lambda, T) - R(\mu, T) = (\mu - \lambda) R(\lambda, T) R(\mu, T).This identity, also known as the first Hilbert identity, facilitates computations and proofs involving the resolvent's behavior across the complex plane.[6][22]The resolvent plays a crucial role in establishing the spectral radius formula for T. The spectral radius r(T) is defined asr(T) = \lim_{n \to \infty} \|T^n\|^{1/n} = \sup \{ |\lambda| : \lambda \in \sigma(T) \},and the resolvent set contains the exterior of the disk \{ \lambda : |\lambda| > r(T) \}, where Neumann series expansions converge:R(\lambda, T) = -\frac{1}{\lambda} \sum_{k=0}^{\infty} \left( \frac{T}{\lambda} \right)^k, \quad |\lambda| > r(T).This connection allows bounds on r(T) via resolvent norms, as \|R(\lambda, T)\| \leq 1 / (|\lambda| - r(T)) for large |\lambda|.[21][6]
Key Theoretical Frameworks
Resolution of the identity
In spectral theory, the resolution of the identity provides a measure-theoretic framework for decomposing self-adjoint operators on a Hilbert space, facilitating the construction of functional calculus through integration over the spectrum. For a self-adjoint operator A acting on a Hilbert space H, the resolution of the identity is a projection-valued measure E defined on the Borel \sigma-algebra \mathcal{B}(\mathbb{R}) of the real line, satisfying A = \int_{\mathbb{R}} \lambda \, dE(\lambda). This integral is understood in the strong operator topology, and the support of E is contained within the spectrum \sigma(A).[23]Key properties of the resolution E ensure its utility as a spectral measure. For disjoint Borel sets \Delta_1, \Delta_2 \subseteq \mathbb{R}, the projections satisfy orthogonality: E(\Delta_1) E(\Delta_2) = 0. Additionally, completeness holds via \int_{\mathbb{R}} dE(\lambda) = I, where I is the identity operator on H, reflecting the decomposition of the entire space. These properties extend the classical resolution for finite-dimensional diagonalizable matrices to infinite dimensions, enabling the representation of bounded Borel functions f as operators via f(A) = \int_{\mathbb{R}} f(\lambda) \, dE(\lambda).[23]The construction of the resolution relies on the Riesz representation theorem applied to the commutative C*-algebra generated by A and the identity. Specifically, the Gelfand transform maps this algebra to continuous functions on its spectrum \Delta = \sigma(A), and the sesquilinear forms \langle x, \varphi(A) y \rangle = \int_{\Delta} \hat{\varphi}(t) \, d\mu_{x,y}(t) induce complex measures \mu_{x,y} via Riesz-Markov-Kakutani representation. The projections are then obtained as E(\omega) = \Phi(\chi_{\omega}) for Borel sets \omega \subseteq \Delta, where \Phi is the operator-valued extension and \chi_{\omega} is the indicator function, yielding a unique resolution satisfying the integral representation.[24]A canonical example arises with multiplication operators on L^2 spaces, which illustrate the resolution explicitly. Consider the self-adjoint multiplication operator M_m on L^2(\mathbb{R}, \mu), where m: \mathbb{R} \to \mathbb{R} is a bounded measurable function and \mu is a \sigma-finite measure. The resolution of the identity is given by the projection-valued measure E(\Delta) f = \chi_{m^{-1}(\Delta)} f for Borel sets \Delta \subseteq \mathbb{R}, where \chi_{m^{-1}(\Delta)} is the indicator function of the preimage set. Here, E(\Delta) acts as multiplication by the indicator, satisfying orthogonality for disjoint \Delta and completeness over \mathbb{R}, with M_m = \int_{\mathbb{R}} \lambda \, dE(\lambda). This setup demonstrates how the spectral measure corresponds directly to level sets of the multiplier function.[24]
Spectral theorem
The spectral theorem provides a canonical decomposition for self-adjoint operators on Hilbert spaces, revealing their underlying spectral structure analogous to diagonalization for matrices. For a bounded self-adjoint operator A acting on a separable Hilbert space H, the theorem guarantees the existence of a unitary operator U: H \to L^2(\sigma(A), \mu), where \sigma(A) \subset \mathbb{R} is the spectrum of A and \mu is a positive Borel measure on \sigma(A), together with the multiplication operator M defined by (M \psi)(\lambda) = \lambda \psi(\lambda) for \psi \in L^2(\sigma(A), \mu), such that A = U M U^{-1}.[25] This representation shows that A is unitarily equivalent to multiplication by the coordinate function on a suitable L^2 space, with the measure \mu uniquely determined by the spectral projections associated to A.The theorem extends to unbounded self-adjoint operators defined on a dense domain, where the decomposition holds with the integral taken over the spectrum in the strong operator topology, ensuring the operator is well-defined on its domain.[25] For normal operators, which commute with their adjoints, the result generalizes via polar decomposition: any bounded normal operator N admits a polar form N = V |N|, where V is a partial isometry with initial space the closure of the range of |N| and |N| = \sqrt{N^* N} is self-adjoint positive; applying the spectral theorem to |N| yields a unitary equivalence |N| = U M_\phi U^{-1} for some positive function \phi, and thus N = (U V^*) M_{\phi^{1/2}} \cdot M_{\phi^{1/2}} (U V^*)^{-1}, or more directly, N is unitarily equivalent to multiplication by a complex-valued bounded measurable function on L^2(\Omega, \nu) for some measure space (\Omega, \nu). This form captures the full spectral multiplicity, with the spectrum of N as the essential range of the multiplying function.Proofs of the spectral theorem typically rely on constructing the spectral measure from analytic properties of the operator. In Stone's approach, the resolvent R(\zeta) = (A - \zeta I)^{-1} for \zeta \notin \sigma(A) is used to define projections via contour integrals, such as E(\Delta) = \frac{1}{2\pi i} \int_{\partial \Delta} (z - A)^{-1} dz for Borel sets \Delta \subset \mathbb{R}, where the family \{E(\Delta)\} forms a resolution of the identity satisfying A = \int \lambda \, dE(\lambda); the unitary equivalence then follows by mapping to the multiplication representation induced by this measure. An alternative proof leverages the Gelfand-Naimark representation theorem for C*-algebras: the C*-algebra generated by a normal operator A and the identity is commutative and isomorphic to C(\sigma(A)), the continuous functions on its spectrum; the theorem embeds this into bounded operators on L^2(\sigma(A), \mu) via multiplication, yielding the desired unitary equivalence.A key consequence of the spectral theorem is the Borel functional calculus, which defines f(A) = \int_{\sigma(A)} f(\lambda) \, dE(\lambda) for any Borel measurable function f: \sigma(A) \to \mathbb{C}, or equivalently f(A) = U M_f U^{-1} where (M_f \psi)(\lambda) = f(\lambda) \psi(\lambda); this extends the continuous functional calculus and preserves the operator norm for bounded f, enabling the construction of functions of operators like exponentials or powers essential in analysis and quantum mechanics. This calculus integrates with the resolution of the identity, where the projections E(\Delta) serve as the "integrating mechanism" for the spectral measure.
Applications and Methods
Solving operator equations
Spectral theory provides powerful tools for solving operator equations by leveraging the decomposition of operators into their spectral components, transforming complex problems into simpler multiplications or integrals over the spectrum. For self-adjointoperators on Hilbert spaces, the spectral theorem enables the representation of an operator A as A = \int_{\sigma(A)} \lambda \, dE(\lambda), where E is the spectral resolution of the identity, allowing equations involving A to be addressed through this integral form.[26]In eigenvalue problems of the form A x = \lambda x, spectral decomposition reduces the equation to a multiplication problem in the spectral basis. Specifically, applying the projection dE(\mu) to both sides yields \int (\mu - \lambda) dE(\mu) x = 0, implying that x lies in the eigenspace corresponding to eigenvalue \lambda, where the operator acts as multiplication by \lambda on that subspace. This approach diagonalizes the problem, facilitating the identification of eigenvalues as points in the spectrum where the resolvent fails to be invertible and eigenvectors as elements in the corresponding spectral subspaces.[27][28]For the time-dependent Schrödinger equation i \partial_t \psi = H \psi, where H is the self-adjointHamiltonian operator, the spectral theorem yields the explicit solution \psi(t) = \int_{\sigma(H)} e^{-i \lambda t} \, dE(\lambda) \psi(0). This integral form evolves the initial state \psi(0) by multiplying each spectral component by the phase factor e^{-i \lambda t}, capturing the unitary time evolution governed by the spectrum of H. The resolution ensures that the solution preserves the norm and orthogonality of the initial condition's projection onto eigenspaces.[28][29]Inverse problems, such as solving (A - \lambda) x = y for x when \lambda is not in the spectrum of A, rely on the resolvent operator R(\lambda) = (A - \lambda I)^{-1}, which applies the spectral decomposition to express x = \int_{\sigma(A)} (\mu - \lambda)^{-1} \, dE(\mu) y. This integral inverts the shifted operator by scaling each spectral component by the reciprocal distance from \lambda, providing a solution valid in the resolvent set. The method extends to perturbed operators, where stability of the solution depends on the spectral gap around \lambda.[6][29]A representative example is the quantum harmonic oscillator, governed by the operator H = -\frac{d^2}{dx^2} + x^2 on L^2(\mathbb{R}), whose spectrum consists of discrete eigenvalues \lambda_n = 2n + 1 for n = 0, 1, 2, \dots. The spectral decomposition H = \sum_{n=0}^\infty (2n + 1) P_n, with projections P_n onto Hermite function eigenfunctions, solves the eigenvalue equation H \phi_n = \lambda_n \phi_n directly as multiplication in this basis, and extends to time evolution via \psi(t) = \sum_{n=0}^\infty e^{-i (2n+1) t} \langle \phi_n | \psi(0) \rangle \phi_n.For the heat equation \partial_t u = \Delta u on a bounded domain with appropriate boundary conditions, spectral theory solves it through the decomposition of the Laplacian \Delta = \int_{\sigma(\Delta)} \lambda \, dE(\lambda), yielding u(t) = \int_{\sigma(\Delta)} e^{\lambda t} \, dE(\lambda) u(0), where the negative eigenvalues \lambda < 0 ensure decay. Equivalently, applying the Laplace transform in time converts the equation to (\lambda I - \Delta) \hat{u}(\lambda) = u(0), solved via the resolvent as \hat{u}(\lambda) = -R(\lambda) u(0) = -\int_{\sigma(\Delta)} (\mu - \lambda)^{-1} \, dE(\mu) u(0), with inversion recovering u(t).[28][6]
Rayleigh quotient and variational principles
The Rayleigh quotient for a self-adjoint operator A on a Hilbert space is defined asR(x) = \frac{\langle A x, x \rangle}{\langle x, x \rangle}for nonzero x in the domain of A.[30] The critical points of this quotient, where its gradient vanishes, correspond to the eigenvalues of A, with the associated eigenvectors achieving stationary values of R(x).[30]A fundamental result linking the Rayleigh quotient to the spectrum is the min-max theorem, which characterizes the k-th largest eigenvalue \lambda_k of a self-adjoint operator (or symmetric matrix) as\lambda_k = \min_{\dim V = n-k+1} \max_{\substack{x \in V \\ \|x\|=1}} R(x),where the minimum is over all subspaces V of dimension n-k+1.[31] Equivalently, it can be expressed in max-min form as\lambda_k = \max_{\dim W = k} \min_{\substack{x \in W \\ \|x\|=1}} R(x),with the maximum over subspaces W of dimension k.[31] This variational principle provides bounds on eigenvalues without explicit computation of the spectrum.The Courant-Fischer characterization, a refinement of the min-max theorem, further specifies that the eigenvalues satisfy\lambda_k = \max_{\dim S = k} \min_{\substack{x \in S \\ x \neq 0}} \frac{\langle A x, x \rangle}{\langle x, x \rangle} = \min_{\dim T = n-k+1} \max_{\substack{x \in T \\ x \neq 0}} \frac{\langle A x, x \rangle}{\langle x, x \rangle},enabling rigorous error bounds in approximations.[31] In numerical methods, this characterization underpins the Ritz-Galerkin approach, where the eigenvalue problem is projected onto a low-dimensional subspace (e.g., a Krylov subspace) via the Galerkin condition, yielding Ritz values that minimize or maximize the Rayleigh quotient over that subspace for optimal approximation.[32] Specifically, the Ritz pairs satisfy variational properties derived from Courant-Fischer, ensuring that the approximate eigenvalues interlace the true ones and providing bounds like $0 \leq \lambda_i - \mu_i \leq (\lambda_i - \lambda_n) [\zeta_i T_{k-i}(1 + 2\delta_i) \tan\theta(u_i, v)]^2, where \mu_i are Ritz values.[32]In Sturm-Liouville problems, the Rayleigh quotient facilitates bounding eigenvalues by evaluating it on trial functions satisfying the boundary conditions. For the standard form -\frac{d}{dx} \left( p(x) \frac{dy}{dx} \right) + q(x) y = \lambda \sigma(x) y, the quotient isR(y) = \frac{ -p y \frac{dy}{dx} \big|_a^b + \int_a^b \left[ p \left( \frac{dy}{dx} \right)^2 - q y^2 \right] dx }{ \int_a^b y^2 \sigma \, dx },and the smallest eigenvalue is the minimum of R(y) over admissible y, with upper bounds obtained by substituting specific trial functions.[33] For instance, in the problem \phi'' + \lambda \phi = 0 with \phi(0) = \phi(1) = 0, using the trial y(x) = x - x^2 yields R(y) = 10, bounding the fundamental eigenvalue \lambda_1 \approx 9.87 from above.[33] This variational technique proves positivity of eigenvalues when R(y) > 0 for all trial functions and extends to higher eigenvalues via orthogonal constraints.[33]