- Perturbation theory
:"This article describes perturbation theory as a general mathematical method. For perturbation theory as applied to

quantum mechanics , seeperturbation theory (quantum mechanics) ."**Perturbation theory**comprises mathematical methods that are used to find an approximate solution to a problem which cannot be solved exactly, by starting from the exact solution of a related problem. Perturbation theory is applicable if the problem at hand can be formulated by adding a "small" term to the mathematical description of the exactly solvable problem.Perturbation theory leads to an expression for the desired solution in terms of a

power series in some "small" parameter that quantifies the deviation from the exactly solvable problem. The leading term in this power series is the solution of the exactly solvable problem, while further terms describe the deviation in the solution, due to the deviation from the initial problem. Formally, we have for the approximation to the full solution A, a series in the small parameter (here called $epsilon$), like the following::$A=epsilon^0\; A\_0\; +\; epsilon^1\; A\_1\; +\; epsilon^2\; A\_2\; +\; cdots$

In this example, $A\_0$ would be the known solution to the exactly solvable initial problem and $A\_1,\; A\_2,$... represent the "higher orders" which are found iteratively by some systematic procedure. For small $epsilon$ these higher orders are presumed to become successively less important.

**Examples**Examples for the "mathematical description" are: an

algebraic equation ,adifferential equation (e.g., theequations of motion incelestial mechanics or awave equation ),a free energy (instatistical mechanics ),a Hamiltonian operator (inquantum mechanics ).Examples for the kind of solution to be found perturbatively:the solution of the equation (e.g., the

trajectory of a particle), the statistical average of somephysical quantity (e.g., average magnetization), theground state energy of a quantum mechanicalproblem.Examples for the exactly solvable problems to start with:

linear equation s, including linear equations of motion(harmonic oscillator ,linear wave equation ), statistical or quantum-mechanical systems ofnon-interacting particles (or in general, Hamiltonians or freeenergies containing only terms quadratic in all degrees of freedom).Examples of "perturbations" to deal with:Nonlinear contributions to the equations of motion,

interaction sbetween particles, terms of higher powers in the Hamiltonian/Free Energy.For physical problems involving interactions between particles,the terms of the perturbation series may be displayed (and manipulated) using

Feynman diagram s.**History**Perturbation theory has its roots in 17th century

celestial mechanics , where the theory ofepicycles was used to make small corrections to the predicted paths of planets.Fact|date=June 2007 Curiously, it was the need for more and more epicycles that eventually led to the 16th centuryCopernican revolution in the understanding of planetary orbits. The development of basic perturbation theory fordifferential equation s was fairly complete by the middle of the 19th century. It was at that time thatCharles-Eugène Delaunay was studying the perturbative expansion for theEarth-Moon-Sun system , and discovered the so-called "problem of small denominators". Here, the denominator appearing in the "n"'th term of the perturbative expansion could become arbitrarily small, causing the "n"'th correction to be as large or larger than the first-order correction. At the turn of the 20th century, this problem ledHenri Poincaré to make one of the first deductions of the existence of chaos, or what is prosaically called the "butterfly effect ": that even a very small perturbation can have a very large effect on a system.Perturbation theory saw a particularly dramatic expansion and evolution with the arrival of

quantum mechanics . Although perturbation theory was used in thesemi-classical theory of theBohr atom , the calculations were monstrously complicated, and subject to somewhat ambiguous interpretation. The discovery of Heisenberg'smatrix mechanics allowed a vast simplification of the application of perturbation theory. Notable examples are theStark effect and theZeeman effect , which have a simple enough theory to be included in standard undergraduate textbooks in quantum mechanics. Other early applications include thefine structure and thehyperfine structure in thehydrogen atom .In modern times, perturbation theory underlies much of

quantum chemistry andquantum field theory . In chemistry, perturbation theory was used to obtain the first solutions for thehelium atom .In the middle of the 20th century,

Richard Feynman realized that the perturbative expansion could be given a dramatic and beautiful graphical representation in terms of what are now calledFeynman diagrams . Although originally applied only inquantum field theory , such diagrams now find increasing use in any area where perturbative expansions are studied.A partial resolution of the small-divisor problem was given by the statement of the

KAM theorem in 1954. Developed byAndrey Kolmogorov ,Vladimir Arnold andJürgen Moser , this theorem stated the conditions under which a system of partial differential equations will have only mildly chaotic behaviour under small perturbations.In the late 20th century, broad dissatisfaction with perturbation theory in the quantum physics community, including not only the difficulty of going beyond second order in the expansion, but also questions about whether the perturbative expansion is even convergent, has led to a strong interest in the area of

non-perturbative analysis , that is, the study ofexactly solvable model s. The prototypical model is theKdV equation , a highly non-linear equation for which the interesting solutions, thesoliton s, cannot be reached by perturbation theory, even if the perturbations were carried out to infinite order. Much of the theoretical work in non-perturbative analysis goes under the name ofquantum group s andnon-commutative geometry .**Perturbation orders**The standard exposition of perturbation theory is given in terms the order to which the perturbation is carried out:

**first order perturbation theory**or**second order perturbation theory**, and whether the perturbed states are degenerate (that is, singular), in which case extra care must be taken, and the theory is slightly more difficult.:"This section needs to be expanded to include the standard textbook examples of each of the three expansions."

**First-order non-singular perturbation theory**This section develops, in simplified terms, the general theory for the perturbative solution to a

differential equation to the first order. In order to keep the exposition simple, a crucial assumption is made: that the solutions to the unperturbed system are not**degenerate**, so that the perturbation series can be inverted. There are ways of dealing with the degenerate (or "singular") case; these require extra care.Suppose one wants to solve a differential equation of the form

:$Dg(x)=lambda\; g(x)$

where "D" is some specific

differential operator , and $lambda$ is aneigenvalue . Many problems involving ordinary or partial differential equations can be cast in this form. It is presumed that the differential operator can be written in the form:$D=D^\{(0)\}+epsilon\; D^\{(1)\}$

where $epsilon$ is presumed to be small, and that furthermore, the complete set of solutions for $D^\{(0)\}$ are known. That is, one has a set of solutions $f^\{(0)\}\_n(x)$, labelled by some arbitrary index "n", such that

:$D^\{(0)\}\; f^\{(0)\}\_n\; (x)=lambda^\{(0)\}\_n\; f^\{(0)\}\_n\; (x)$.

Furthermore, one assumes that the set of solutions $\{f^\{(0)\}\_n\; (x)\}$ form an

orthonormal set::$int\; f^\{(0)\}\_m\; (x)\; f^\{(0)\}\_n\; (x)\; ,dx\; =\; delta\_\{mn\}$

with $delta\_\{mn\}$ the

Kronecker delta function.To zeroth order, one expects that the solutions $g(x)$ are then somehow "close" to one of the unperturbed solutions $f^\{(0)\}\_n\; (x)$. That is,

:$g(x)=f^\{(0)\}\_n\; (x)\; +\; mathcal\{O\}(epsilon)$

and

:$lambda=lambda^\{(0)\}\_n\; +\; mathcal\{O\}(epsilon)$.

where $mathcal\{O\}$ denotes the relative size, in

big-O notation . To solve this problem, one assumes that the solution $g(x)$ can be written as a linear combination of the $f^\{(0)\}\_n\; (x)$::$g(x)=sum\_m\; c\_m\; f^\{(0)\}\_m\; (x)$

with all of the constants $c\_m\; =mathcal\{O\}(epsilon)$ except for "n", where $c\_n\; =mathcal\{O\}(1)$. Substituting this last expansion into the differential equation, taking the inner product of the result with $f^\{(0)\}\_n\; (x)$, and making use of orthogonality, one obtains

:$c\_nlambda^\{(0)\}\_n\; +\; epsilon\; sum\_m\; c\_m\; int\; f^\{(0)\}\_n(x)\; D^\{(1)\}\; f^\{(0)\}\_m(x),dx\; =\; lambda\; c\_n$

This can be trivially re-written as a simple

linear algebra problem of finding theeigenvalue of a matrix, where:$sum\_m\; A\_\{nm\}c\_m\; =\; lambda\; c\_n!$

where the matrix elements $A\_\{nm\}$ are given by

:$A\_\{nm\}\; =\; delta\_\{nm\}lambda^\{(0)\}\_n\; +\; epsilon\; int\; f^\{(0)\}\_n(x)\; D^\{(1)\}\; f^\{(0)\}\_m(x),dx$

Rather than solving this full matrix equation, one notes that, of all the $c\_m$ in the linear equation, only one, namely $c\_n$, is not small. Thus, to the first order in $epsilon$, the linear equation may be solved trivially as

:$lambda\; =\; lambda^\{(0)\}\_n\; +\; epsilon\; int\; f^\{(0)\}\_n(x)\; D^\{(1)\}\; f^\{(0)\}\_n(x),dx$

since all of the other terms in the linear equation are of order $mathcal\{O\}(epsilon^2)$. The above gives the solution of the eigenvalue

**to first order in perturbation theory**.The function $g(x)$ to first order is obtained through similar reasoning. Substituting

:$g(x)=f^\{(0)\}\_n(x)\; +\; epsilon\; f^\{(1)\}\_n(x)$

so that

:$left(D^\{(0)\}\; +epsilon\; D^\{(1)\}\; ight)left(\; f^\{(0)\}\_n(x)\; +\; epsilon\; f^\{(1)\}\_n(x)\; ight)\; =\; left(\; lambda^\{(0)\}\_n(x)\; +\; epsilon\; lambda^\{(1)\}\_n(x)\; ight)left(\; f^\{(0)\}\_n(x)\; +\; epsilon\; f^\{(1)\}\_n(x)\; ight)$

gives an equation for $f^\{(1)\}\_n(x)$. It may be solved integrating with the

partition of unity :$delta(x-y)=sum\_n\; f^\{(0)\}\_n(x)\; f^\{(0)\}\_n(y)$

to give

:$f^\{(1)\}\_n(x)\; =\; sum\_\{m,(\; e\; n)\}\; frac\; \{f^\{(0)\}\_m\; (x)\}\{lambda^\{(0)\}\_n-\; lambda^\{(0)\}\_m\}\; int\; f^\{(0)\}\_m(y)\; D^\{(1)\}\; f^\{(0)\}\_n(y)\; ,dy$

which gives the exact solution to the perturbed differential equation

**to the first order in the perturbation $epsilon$**.Several important observations can be made about the form of this solution. First, the sum over functions with differences of eigenvalues in the denominator resembles the

resolvent inFredholm theory . This is no accident; the resolvent acts essentially as a kind ofGreen's function orpropagator , passing the perturbation along. Higher order perturbations resemble this form, with an additional sum over a resolvent appearing at each order.The form of this solution is sufficient to illustrate the idea behind the

**small-divisor problem**. If, for whatever reason, two eigenvalues are close so that difference $lambda^\{(0)\}\_n-\; lambda^\{(0)\}\_m$ become small, the corresponding term in the sum will become disproportionately large. In particular, if this happens in higher-order terms, the high order perturbation may become as large or larger in magnitude than the first-order perturbation. Such a situation calls into question the validity of doing a perturbation to begin with. This can be understood to be a fairly catastrophic situation; it is frequently encountered in chaotic dynamical systems, and requires the development of techniques other than perturbation theory to solve the problem.Curiously, the situation is not at all bad if two or more eigenvalues are exactly equal. This case is referred to as

**singular**or**degenerate perturbation theory**. The degeneracy of eigenvalues indicates that the unperturbed system has some sort ofsymmetry , and that the generators of the symmetry commute with the unperturbed differential operator. Typically, the perturbing term does not possess the symmetry; one says the perturbation "lifts" or "breaks" the degeneracy. In this case, the perturbation can still be performed; however, one must be careful to work in a basis for the unperturbed states so that these map one-to-one to the perturbed states, rather than being a mixture.**Example of second-order singular perturbation theory**Consider the following equation for the unknown variable $x$:

:$x=1+epsilon\; x^5.$

For the initial problem with $epsilon=0$, the solution is $x\_0=1$. For small $epsilon$ the lowest order approximation may be found by inserting the

ansatz :$x=x\_0+epsilon\; x\_1\; (+ldots)$

into the equation and demanding the equation to be fulfilled up to terms that involve powers of $epsilon$ higher than the first. This yields $x\_1=1$. In the same way, the higher orders may be found. However, even in this simple example it may be observed that for (arbitrarily) small $epsilon>0$ there are four other solutions to the equation (with very large magnitude). The reason we don't find these solutions in the above perturbation method is because these solutions diverge when $epsilon\; ightarrow\; 0$ while the ansatz assumes regular behavior in this limit.

The four additional solutions can be found using the methods of

. In this case this works as follows. Since the four solutions diverge at $epsilon=0$, it makes sense to rescale $x$. We putsingular perturbation theory:$x\; =\; yepsilon^\{-\; u\}$

such that in terms of $y$ the solutions stay finite. This means that we need to choose the exponent $u$ to match the rate at which the solutions diverge. In terms of $y$ the equation reads:

:$epsilon^\{-\; u\}y=1+epsilon^\{1-5\; u\}\; y^5$

The 'right' value for $u$ is obtained when the exponent of $epsilon$ in the prefactor of the term proportional to $y$ is equal to the exponent of $epsilon$ in the prefactor of the term proportional to $y^5$, i.e. when $u=1/4$. This is called 'significant degeneration'. If we choose $u$ larger, then the four solutions will collapse to zero in terms of $y$ and they will become degenerate with the solution we found above. If we choose $u$ smaller, then the four solutions will still diverge to infinity.

Putting $u=1/4$ in the above equation yields:

:$y=epsilon^\{1/4\}+y^5$

This equation can be solved using ordinary perturbation theory in the same way as regular expansion for $x$ was obtained. Since the expansion parameter is now $epsilon^\{1/4\}$ we put:

:$y=y\_0\; +\; epsilon^\{1/4\}y\_1\; +\; epsilon^\{1/2\}y\_2\; ldots$

There are 5 solutions for $y\_0$: 0, 1, -1, i and -i. We must disregard the solution $y=0$. The case $y=0$ corresponds to the original regular solution which appears to be at zero for $epsilon=0$, because in the limit $epsilon\; ightarrow\; 0$ we are rescaling by an infinite amount. The next term is $y\_1=\; -1/4$. In terms of $x$ the four solutions are thus given as:

$x\; =\; epsilon^\{-1/4\}left\; [y\_0\; -\; 1/4epsilon^\{1/4\}\; +ldots\; ight]$

**Commentary**Both regular and singular perturbation theory are frequently used in physics and engineering. Regular perturbation theory may only be used to find those solutions of a problem that evolve smoothly out of the initial solution when changing the parameter (that are "

adiabatic ally connected" to the initial solution). A well known example from physics where regular perturbation theory fails is in fluid dynamics when one treats the viscosity as a small parameter. Close to a boundary, the fluid velocity goes to zero, even for very small viscosity (theno-slip condition ). For zero viscosity, it is not possible to impose this boundary condition and a regular perturbative expansion amounts to an expansion about an unrealistic physical solution. Singular perturbation theory can, however, be applied here and this amounts to 'zooming in' at the boundaries (using themethod of matched asymptotic expansions ).Perturbation theory can fail when the system can go to a different "phase" of matter, with a qualitatively different behaviour that cannot be understood by perturbation theory (e.g., a solid crystal melting into a liquid). In some cases this failure manifests itself by divergent behavior of the perturbation series. Such divergent series can sometimes be resummed using techniques such as

Borel resummation .Perturbation techniques can be also used to find approximate solutions to non-linear differential equations. Examples of techniques used to find approximate solutions to these types of problems are the

Lindstedt-Poincaré technique , themethod of harmonic balancing , and themethod of multiple time scales .There is absolutely no guarantee perturbative methods would result in a

convergent solution. In fact,asymptotic series are the norm.**Perturbation theory in chemistry**Many of the

ab initio quantum chemistry methods use perturbation theory directly or are closely related methods.Møller-Plesset perturbation theory uses the difference between theHartree-Fock Hamiltonian and the exact non-relativistic Hamiltonian as the perturbation. The zero order energy is the sum of orbital energies. The first-order energy is the Hartree-Fock energy and electron correlation is included at second-order or higher. Calculations to second, third or forth order are very common and the code is included in most ab initio quantum chemistry programs. A related but more accurate method is thecoupled cluster method.**ee also***

Structural stability

*Eigenvalue perturbation

*Cosmological perturbation theory **External links*** [

*http://www.sm.luth.se/~johanb/applmath/chap2en/index.html Chapter II: Introduction to perturbation methods*] by Johan Byström, Lars-Erik Persson, and Fredrik Strömberg

* [*http://www.cims.nyu.edu/~eve2/reg_pert.pdf Introduction to regular perturbation theory*] by Eric Vanden-Eijnden (PDF)

* [*http://tosio.math.toronto.edu/wiki/index.php/Perturbation_theory Duality in Perturbation Theory*]

* [*http://www.scholarpedia.org/article/Multiple_Scale_Analysis Perturbation Method of Multiple Scales*]

*Wikimedia Foundation.
2010.*