- Sum of normally distributed random variables
In
probability theory , if "X" and "Y" are independentrandom variable s that are normally distributed, then "X" + "Y" is also normally distributed; i.e. if:
and
:
and "X" and "Y" are independent, then
:
Proofs
This proposition may be proved by any of several methods.
Proof using convolutions
By the total probability theorem, we have the probability density function of "z"
:
and since "X" and "Y" are independent, we get
:
But ƒ"Z"("z"|"x,y") is trivially equal to "δ"("z" − ("x" + "y")), so
:
where "δ" is
Dirac's delta function . We substitute ("z" − "x") for "y"::
which we recognize as a
convolution of ƒ"X" with ƒ"Y".Therefore the
probability density function of the sum of two independent random variables "X" and "Y" with probability density functions ƒ and "g" is theconvolution :
No generality is lost by assuming the two
expected value s μ and ν are zero. Thus the two densities are:
The convolution is
::
:
In simplifying this expression it saves some effort to recall this obvious fact that the context might later make easy to forget: The integral
:
actually does not depend on "A". This is seen be a simple substitution: "w" = "u" − "A", "dw" = "du", and the bounds of integration remain −∞ and +∞.
Now we have
:
where "constant" in this context means not depending on "x". The last integral does not depend on "x" because of the "obvious fact" mentioned above.
A probability density function that is a constant multiple of
:
is the density of a normal distribution with variance σ2 + τ2. Although we did not explicitly develop the constant in this derivation, this is indeed the case.
Proof using characteristic functions
:
of the sum of two independent random variables "X" and "Y" is just the product of the two separate characteristic functions:
:
and
:
of "X" and "Y".
The characteristic function of the normal distribution with expected value μ and variance σ2 is
:
So
::
:
This is the characteristic function of the normal distribution with expected value μ + ν and variance σ2 + τ2.
Finally, recall that no two distinct distributions can both have the same characteristic function, so the distribution of "X" + "Y" must be just this normal distribution.
Correlated random variables
In the event that the variables "X" and "Y" are jointly normally distributed random variables, then "X" + "Y" is still normally distributed (see
Multivariate normal distribution ). In this case, one needs to consider:
As above, one makes the substitution
This integral is more complicated to simplify analytically, but can be done easily using a symbolic mathematics program. The probability distribution "ƒ""Z"("z") is given in this case by
:
If one considers instead "Z" = "X" − "Y", then one obtains :
The standard deviations of each distribution are obvious by comparison with the standard normal distribution.
ee also
:*
List of convolutions of probability distributions
Wikimedia Foundation. 2010.