- Power law
A

**power law**is anypolynomial relationship that exhibits the property of. The most common power laws relate two variables and have the formscale invariance :$f(x)\; =\; ax^k!\; +o(x^k),$

where $a$ and $k$ are constants, and $o(x^k)$ is of $x$. Here, $k$ is typically called the "scaling

exponent ", the word "scaling" denoting the fact that a power-law function satisfies $f(c\; x)\; propto\; f(x)$ where $c$ is a constant. That is, a rescaling of the function's argument changes the constant of proportionality but preserves the shape of the function itself. This point becomes clearer if we take thelogarithm of both sides::$logleft(f(x)\; ight)\; =\; k\; log\; x\; +\; log\; a.$

Notice that this expression has the form of a linear relationship with slope $k$. Rescaling the argument produces a linear shift of the function up or down but leaves both the basic form and the slope $k$ unchanged.

Power-law relations characterize a staggering number of naturally occurring phenomena, and this is one of the principal reasons why they have attracted interest. For instance,

inverse-square law s, such asgravitation and theCoulomb force , are power laws, as are many common mathematical formulae such as the quadratic law of area of the circle. However it is mainly in the study ofprobability distributions that power laws have attracted recent interest. A wide variety of observed probability distributions appear, at least approximately, to have tailsasymptotic ally following power-law forms, an observation connected closely with the study of theory of large deviations (also calledextreme value theory ), which considers the frequency of extremely rare events likestock market crash es and largenatural disaster s. It is primarily in the study of statistical distributions that the name "power law" is used; in other areas the power-law functional form is more often referred to simply as a polynomial form or polynomial function.Scientific interest in power law relations also derives from the ease with which certain general classes of mechanisms can generate them, so that the observation of a power-law relation in data often points to specific kinds of mechanisms that might underly the natural phenomenon in question, and can indicate a deep connection with other, seemingly unrelated systems (see the reference by Simon and the subsection on universality below). The ubiquity of power-law relations in physics is partly due to dimensional constraints, while in

complex systems , power laws are often thought to be signatures of hierarchy or of specificstochastic processes . A few notable examples of power laws are theGutenberg-Richter law for earthquake sizes, Pareto's law of income distribution, structural self-similarity offractals , and scaling laws in biological systems. Research on the origins of power-law relations, and efforts to observe and validate them in the real world, is an active topic of research in many fields of science, includingphysics ,computer science ,linguistics ,geophysics ,sociology ,economics and more.**Properties of power laws****cale invariance**The main property of power laws that makes them interesting is their scale invariance. Given a relation $f(x)\; =\; ax^k$, or, indeed any

homogeneous polynomial , scaling the argument $x$ by a constant factor causes only a proportionate scaling of the function itself. That is,:$f(c\; x)\; =\; a(c\; x)^k\; =\; c^\{k\}f(x)\; propto\; f(x).!$

That is, scaling by a constant simply multiplies the original power-law relation by the constant $c^k$. Thus, it follows that all power laws with a particular scaling exponent are equivalent up to constant factors, since each is simply a scaled version of the others. This behavior is what produces the linear relationship when both logarithms are taken of both $f(x)$ and $x$, and the straight-line on the log-log plot is often called the "signature" of a power law. Notably, however, with real data, such straightness is necessary, but not a sufficient condition for the data following a power-law relation. In fact, there are many ways to generate finite amounts of data that mimic this signature behavior, but, in their asymptotic limit, are not true power laws. Thus, accurately fitting and validating power-law models is an active area of research in

statistics .**Universality**The equivalence of power laws with a particular scaling exponent can have a deeper origin in the dynamical processes that generate the power-law relation. In physics, for example,

phase transition s in thermodynamic systems are associated with the emergence of power-law distributions of certain quantities, whose exponents are referred to as thecritical exponent s of the system. Diverse systems with the same critical exponents — that is, which display identical scaling behaviour as they approach criticality — can be shown, viarenormalization group theory, to share the same fundamental dynamics. For instance, the behavior of water and CO2 at their boiling points fall in the same universality class because they have identical critical exponents. In fact, almost all material phase transitions are described by a small set of universality classes. Similar observations have been made, though not as comprehensively, for various self-organized critical systems, where the critical point of the system is anattractor . Formally, this sharing of dynamics is referred to as universality, and systems with precisely the same critical exponents are said to belong to the same universality class.**Power-law functions**The general power-law function follows the polynomial form given above, and is a ubiquitous form throughout mathematics and science. Notably, however, not all polynomial functions are power laws because not all polynomials exhibit the property of scale invariance. Typically, power-law functions are polynomials in a single variable, and are explicitly used to model the scaling behavior of natural processes. For instance, allometric scaling laws for the relation of biological variables are some of the best known power-law functions in nature. In this context, the $o(x^k)$ term is most typically replaced by a deviation term $epsilon$, which can represent uncertainty in the observed values (perhaps measurement or sampling errors) or provide a simple way for observations to deviate from the no power-law function (perhaps for stochastic reasons):

:$y\; =\; ax^k\; +\; varepsilon.!$

**Estimating the exponent from empirical data**There are many methods for fitting power-law functions to data, and the best option typically depends strongly on the kind of question being asked. For instance, prediction-type questions should rely on

nonlinear regression , while descriptive-type summary questions, such as those found inallometry , should use a method that allows for uncertainty in both the $x$ and $y$ measurements. If the residuals are log normally distributed, e.g. if the spread in $y$ is multiplicative (increasing proportionally with $x$), a simple least-squareslinear regression on log-transformed data can be performed, since the log transformed residues are normally distributed after transformation. Otherwise, the logarithmic transformation produces residuals that are log-normally distributed, while the least squares method requires normally distributed errors. In this latter context, the method of standardizedmajor axis (SMA) regression (sometimes called "reduced major axis", but this term should be avoided) is preferred.The major axis is the linear equation that minimizes the sum of squares of the shortest (perpendicular) distance between data points and the equation. This axis is equivalent to the first

principal component axis of thecovariance matrix . From this observation, theestimator for the slope can be derived:$hat\{k\}\; =\; frac\{\; sigma\_y\; \}\{\; sigma\_x\; \}\; =\; sqrt\{\; frac\{\; sum\_\{i=1\}^\{N\}\; (y\_i\; -\; mu\_y)^2\; \}\{\; sum\_\{i=1\}^N\; (x\_i\; -\; mu\_\{x\})^2\; \}\; \}$

where $mu\_x$ and $mu\_y$ are the sample means of the $x$ and $y$ data, respectively.

More about this method, and the conditions under which it can be used, can be found in the Warton reference below. Further, Warton's comprehensive review article also provides [

*http://web.maths.unsw.edu.au/~dwarton/programs.html usable code*] (C++, R, and Matlab) for estimation and testing routines for power-law functions.**Examples of power law functions***The

Stefan-Boltzmann law

*The Gompertz Law of Mortality

*The Ramberg-Osgood stress-strain relationship

*TheInverse-square law ofNewtonian gravity

*TheInitial mass function

*Gamma correction relating light intensity with voltage

*Kleiber's law relating animal metabolism to size, andallometric law s in general

*Behaviour near second-order phase transitions involvingcritical exponent s

*Proposed form of experience curve effects

*The differential energy spectrum ofcosmic-ray nuclei

*Inverse-square law

*Square-cube law

*Constructal law

*Fractal s**Power-law distributions**A power-law distribution is any that, in the most general sense, has the form

:$p(x)\; propto\; L(x)\; x^\{-alpha\}$

where $alpha\; >\; 1$, and $L(x)$ is a

**slowly varying function**, which is any function that satisfies $lim\_\{x\; ightarrowinfty\}\; L(t,x)\; /\; L(x)\; =\; 1$ with $t$ constant. This property of $L(x)$ follows directly from the requirement that $p(x)$ be asymptotically scale invariant; thus, the form of $L(x)$ only controls the shape and finite extent of the lower tail. For instance, if $L(x)$ is the constant function, then we have a power-law that holds for all values of $x$. In many cases, it is convenient to assume a lower bound $x\_\{mathrm\{min$ from which the law holds. Combining these two cases, and where $x$ is a continuous variable, the power law has the form:$p(x)\; =\; frac\{alpha-1\}\{x\_min\}\; left(frac\{x\}\{x\_min\}\; ight)^\{-alpha\},$

where the constant is necessary to guarantee that the distribution is properly normalized. Briefly, we can consider several properties of this distribution.

In general, the moments of this distribution are given by

:$langle\; x^\{m\}\; angle\; =\; int\_\{x\_min\}^\{infty\}\; x^\{m\}\; p(x)\; ,mathrm\{d\}x\; =\; frac\{alpha-1\}\{alpha-1-m\}x\_min^m$

which is only well defined for $m\; <\; alpha\; -1$. That is, all moments $m\; geq\; alpha\; -\; 1$ diverge: when $alpha<2$, the average and all higher-order moments are infinite; when $23\; math>,\; the\; mean\; exists,\; but\; the\; variance\; and\; higher-order\; moments\; are\; infinite,\; etc.\; For\; finite-size\; samples\; drawn\; from\; such\; distribution,\; this\; behavior\; implies\; that\; the\; central\; moment\; estimators\; (like\; the\; mean\; and\; the\; variance)\; for\; diverging\; moments\; will\; never\; converge\; -\; as\; more\; data\; is\; accumulated,\; they\; continue\; to\; grow.$

Another kind of power-law distribution, which does not satisfy the general form above, is the power law with an exponential cutoff

:$p(x)\; propto\; L(x)\; x^\{-alpha\}\; mathrm\{e\}^\{-lambda\; x\}$

where we introduce an exponential decay term $mathrm\{e\}^\{-lambda\; x\}$ that overwhelms the power-law behavior at large values of $x$. This distribution does not scale and is thus not asymptotically a power law; however, it does approximately scale over a finite region before the cutoff. (Note that the pure form above is a subset of this family, with $lambda=0$.) This distribution is a common alternative to the asymptotic power-law distribution because it naturally captures finite-size effects. For instance, although the

Gutenberg-Richter Law is commonly cited as an example of a power-law distribution, the distribution of earthquake magnitudes cannot scale as a power law in the limit $x\; ightarrowinfty$ because there is a finite amount of energy in the Earth's crust. Thus, there must be some maximum size earthquake, and the scaling behavior must taper off as it approaches this size.**Plotting power-law distributions**In general, power-law distributions are plotted on doubly logarithmic axes, which emphasizes the upper tail region. The most convenient way to do this is via the (complementary) cumulative distribution (cdf), $P(x)\; =\; mathrm\{Pr\}(X\; >\; x)$,

:$P(x)\; =\; Pr(X\; >\; x)\; =\; C\; int\_x^infty\; p(X),mathrm\{d\}X\; =\; frac\{alpha-1\}\{x\_min^\{-alpha+1\; int\_x^infty\; X^\{-alpha\},mathrm\{d\}X\; =\; left(frac\{x\}\{x\_min\}\; ight)^\{(-alpha+1)\}.$

Note that the cdf is also a power-law function, but with a smaller scaling exponent. For data, an equivalent form of the cdf is the rank-frequency approach, in which we first sort the $n$ observed values in ascending order, and plot them against the vector $left\; [1,frac\{n-1\}\{n\},frac\{n-2\}\{n\},dots,frac\{1\}\{n\}\; ight]$.

Although it can be convenient to log-bin the data, or otherwise smooth the probability density (mass) function directly, these methods introduce an implicit bias in the representation of the data, and thus should be avoided. The cdf, on the other hand, introduces no bias in the data and preserves the linear signature on doubly logarithmic axes.

**Estimating the exponent from empirical data**There are many ways of estimating the value of the scaling exponent for a power-law tail, however not all of them yield unbiased and consistent answers. The most reliable techniques are often based on the method of maximum likelihood. Alternative methods are often based on making a linear regression on either the log-log probability, the log-log cumulative distribution function, or on log-binned data, but these approaches should be avoided as they can all lead to highly biased estimates of the scaling exponent (see the Clauset et al. reference below).

For real-valued data, we fit a power-law distribution of the form

:$p(x)\; =\; frac\{alpha-1\}\{x\_min\}\; left(frac\{x\}\{x\_min\}\; ight)^\{-alpha\}$

to the data $xgeq\; x\_min$. Given a choice for $x\_min$, a simple derivation by this method yields the estimator equation

:$hat\{alpha\}\; =\; 1\; +\; n\; left\; [\; sum\_\{i=1\}^\{n\}\; ln\; frac\{x\_i\}\{x\_min\}\; ight]\; ^\{-1\}$

where $\{x\_i\}$ are the $n$ data points $x\_\{i\}geq\; x\_min$. (For a more detailed derivation, see Hall or Newman below.) This estimator exhibits a small finite sample-size bias of order $O(n^\{-1\})$, which is small when "n" > 100. Further, the uncertainty in the estimation can be derived from the maximum likelihood argument, and has the form $sigma\; =\; frac\{alpha-1\}\{sqrt\{n$. This estimator is equivalent to the popular Hill estimator from

quantitative finance andextreme value theory .For a set of "n" integer-valued data points $\{x\_i\}$, again where each $x\_igeq\; x\_min$, the maximum likelihood exponent is the solution to the transcendental equation

:$frac\{zeta\text{'}(hatalpha,x\_min)\}\{zeta(hat\{alpha\},x\_min)\}\; =\; -frac\{1\}\{n\}\; sum\_\{i=1\}^n\; ln\; frac\{x\_i\}\{x\_min\}$

where $zeta(alpha,x\_\{mathrm\{min)$ is the incomplete zeta function. The uncertainty in this estimate follows the same formula as for the continuous equation. However, the two equations for $hat\{alpha\}$ are not equivalent, and the continuous version should not be applied to discrete data, nor vice versa.

Further, both of these estimators require the choice of $x\_min$. For functions with a non-trivial $L(x)$ function, choosing $x\_min$ too small produces a significant bias in $hatalpha$, while choosing it too small increases the uncertainty in $hat\{alpha\}$, and reduces the

statistical power of our model. In general, the optimum choice of $x\_min$ depends strongly on the particular form of the lower tail, represented by $L(x)$ above.More about these methods, and the conditions under which they can be used, can be found in the Clauset et al. reference below. Further, this comprehensive review article provides [

*http://www.santafe.edu/~aaronc/powerlaws/ usable code*] (Matlab and R) for estimation and testing routines for power-law distributions.**Examples of power-law distributions***

Pareto distribution (continuous)

*Zeta distribution (discrete)

*Yule–Simon distribution (discrete)

*Student's t-distribution (continuous), of which theCauchy distribution is a special case

*Zipf's law and its generalization, theZipf-Mandelbrot law (discrete)

**Lotka's law

*Thescale-free network model

*Bibliogram s

*Gutenberg-Richter law ofearthquake magnitudes

*Horton's laws describing river systems

*Richardson's Law for the severity of violent conflicts (wars and terrorism)

*population of cities

*numbers of religious adherents

*net worth of individuals

*frequency of words in a textA great many power-law distributions have been conjectured in recent years. For instance, power laws are thought to characterize the behavior of the upper tails for the popularity of websites, number of species per genus, the popularity of given names, the size of financial returns, and many others. However, much debate remains as to which of these tails are actually power-law distributed and which are not. For instance, it is commonly accepted now that the famous

Gutenberg-Richter Law decays more rapidly than a pure power-law tail because of a finite exponential cutoff in the upper tail.**Validating power laws**Although power-law relations are attractive for many theoretical reasons, demonstrating that data do indeed follow a power-law relation requires more than simply fitting such a model to the data. In general, many alternative functional forms can appear to follow a power-law form for some extent. Thus, the preferred method for validation of power-law relations is by testing many orthogonal predictions of a particular generative mechanism against data, and not simply fitting a power-law relation to a particular kind of data. As such, the validation of power-law claims remains a very active field of research in many areas of modern science.

**ee also**

*Allometric law

*Extreme value theory

*Lognormal distribution

*Fat tail

*Heavy-tailed distribution s

*80-20 rule

*The Long Tail

*Wealth condensation

*Keston process

*Levy skew alpha-stable distribution

*Lévy flight

*Kleiber's law

*Power law fluid

*Simon Model

*Stevens' power law

*Zipf's law **Bibliography*** cite journal

author = Simon, H. A.

year = 1955

title = On a Class of Skew Distribution Functions

journal =Biometrika

volume = 42

pages = 425–440

url = http://links.jstor.org/sici?sici=0006-3444%28195512%2942%3A3%2F4%3C425%3AOACOSD%3E2.0.CO%3B2-M

doi = 10.2307/2333389

* cite journal

author = Hall, P.

year = 1982

title = On Some Simple Estimates of an Exponent of Regular Variation

journal = Journal of theRoyal Statistical Society , Series B (Methodological)

volume = 44

pages = 37–42

url = http://links.jstor.org/sici?sici=0035-9246(1982)44%3A1%3C37%3AOSSEOA%3E2.0.CO%3B2-4

issue = 1

* cite journal

author = Mitzenmacher, M.

year = 2003

title = A brief history of generative models for power law and lognormal distributions

journal = Internet Mathematics

volume = 1

pages = 226–251

url = http://www.internetmathematics.org/volumes/1/2/pp226_251.pdf

* cite journal

author = Newman, M. E. J.

year = 2005

title = Power laws, Pareto distributions and Zipf's law

journal = Contemporary Physics

volume = 46

pages = 323–351

url = http://www.journalsonline.tandf.co.uk/openurl.asp?genre=article&doi=10.1080/00107510500052444

doi = 10.1080/00107510500052444

* cite journal

author = Warton, D. I., Wright, I. J., Falster, D. S., and Westoby, M.

year = 2006

title = Bivariate line-fitting methods for allometry

journal = Biological Reviews

volume = 81

pages = 259–291

url = http://www.maths.unsw.edu.au/statistics/files/preprint-2005-02.pdf

doi = 10.1017/S1464793106007007

* cite journal

author = Clauset, A., Shalizi, C. R. and Newman, M. E. J.

year = 2007

title = Power-law distributions in empirical data

url = http://arxiv.org/abs/0706.1062**External links*** [

*http://www.nslij-genetics.org/wli/zipf/ Zipf's law*]

* [*http://aps.arxiv.org/abs/cond-mat/0412004/ Power laws, Pareto distributions and Zipf's law*]

* [*http://www.hpl.hp.com/research/idl/papers/ranking/ranking.html Zipf, Power-laws, and Pareto - a ranking tutorial*]

* [*http://simscience.org/crackling/Advanced/Earthquakes/GutenbergRichter.html Gutenberg-Richter Law*]

* [*http://www.physicalgeography.net/fundamentals/10ab.html Stream Morphometry and Horton's Laws*]

* [*http://www.fooledbyrandomness.com/fortune.pdf "How the Finance Gurus Get Risk All Wrong"*] by Benoit Mandelbrot & Nassim Nicholas Taleb. "Fortune", July 11, 2005.

* [*http://www.newyorker.com/fact/content/articles/060213fa_fact "Million-dollar Murray":*] power-law distributions in homelessness and other social problems; by Malcolm Gladwell. "The New Yorker", February 13, 2006.

*Benoit Mandelbrot & Richard Hudson: The Misbehaviour of Markets (2004)

*Philip Ball: [*http://www.agrfoto.com/philipball/criticalmass.php Critical Mass: How one thing leads to another*] (2005)

* [*http://econophysics.blogspot.com/2006/07/tyranny-of-power-law-and-why-we-should.html "Tyranny of the Power Law"*] from [*http://econophysics.blogspot.com The Econophysics Blog*]

*Wikimedia Foundation.
2010.*