# Interval arithmetic

Interval arithmetic

Interval arithmetic, also called "interval mathematics", "interval analysis", and "interval computation", is a method in mathematics. It has been developed by mathematicians since the 1950s and 1960s as an approach to putting bounds on rounding errors in mathematical computation and thus developing numerical methods that yield very reliable results.

Where classical arithmetic defines operations on individual numbers, interval arithmetic defines a set of operations on intervals:

:T · S = { "x" | there is some "y" in "T", and some "z" in "S", such that "x" = "y" · "z" }.

The basic operations of interval arithmetic are, for two intervals ["a", "b"] and ["c", "d"] that are subsets of the real line (-∞,∞) ,

* ["a","b"] + ["c","d"] = ["a" + "c", "b" + "d"]
* ["a","b"] − ["c", "d"] = ["a" − "d", "b" −"c"]
* ["a","b"] &times; ["c","d"] = [min ("ac", "ad", "bc", "bd"), max ("ac", "ad", "bc", "bd")]
* ["a","b"] / ["c","d"] = [min ("a/c", "a/d", "b/c", "b/d"), max ("a/c", "a/d", "b/c", "b/d")]

Division by an interval containing zero is not defined under the basic interval arithmetic. The addition and multiplication operations are commutative, associative and sub-distributive: the set "X" ( "Y" + "Z" ) is a subset of "XY" + "XZ".

Instead of working with an uncertain real $x$ we work with the two ends of the interval $\left[a,b\right]$ which contains $x$: $x$ lies between $a$ and $b$, or could be one of them. Similarly a function $f$ when applied to $x$ is also uncertain. Instead, in interval arithmetic $f$ produces an interval $\left[c,d\right]$ which is all the possible values for $f\left(x\right)$ for all $x in \left[a,b\right]$.

This concept is suitable, inter alia, for the treatment of rounding errors directly during the calculation and of uncertainties in the knowledge of the exact values of physical and technical parameters. The latter often arise from measurement errors and tolerances for components. Interval arithmetic also helps find reliable and guaranteed solutions to equations and optimization problems.

Take as an example the calculation of body mass index (BMI). The BMI is the body weight in kilograms divided by the square of height in metres. Measuring the mass with bathroom scales may have an accuracy of one kilogram. We will not know intermediate values - about 79.6 kg or 80.3 kg - but information rounded to the nearest whole number. It is unlikely that you really weigh 80.0 kg exactly when it appears. In normal rounding to the nearest value, the scales showing 80 kg indicates a weight between 79.5 kg and 80.5 kg. The relevant range is that of all real numbers that are greater than or equal to 79.5, while less than or equal to 80.5, or in other words the interval [79.5,80.5] .

For a man who weighs 80 kg and is 1.80 m tall, the BMI is about 24.7. With a weight of 79.5 kg and the same height the value is 24.5, while 80.5 kilograms gives almost 24.9. So the actual BMI is in the range [24.5,24.9] . The error in this case does not affect the conclusion (normal weight), but this is not always the position. For example, weight fluctuates in the course of a day so that the BMI can vary between 24 (normal weight) and 25 (overweight). Without detailed analysis it is no possible to always exclude questions as to whether an error ultimately is large enough to have significant influence.

Interval arithmetic states the range of possible outcomes explicitly. Simply put, results are no longer stated as numbers, but as intervals which represent imprecise values. The size of the intervals are similar to error bars to a metric in expressing the extent of uncertainty. Simple arithmetic operations, such as basic arithmetic and trigonometric functions, enable the calculation of outer limits of intervals.

Introduction

The main focus in the interval arithmetic is on the simplest way to calculate upper and lower endpoints for the the range of values of a function in one or more variables. These barriers need be not necessarily the supremum or infimum, since the precise calculation of those values are often too difficult; it can be shown that that task is in general NP-hard.

Treatment is typically limited to real intervals, so quantities of form:$\left[a,b\right] = \left\{x in mathbb\left\{R\right\} ,|, a le x le b\right\}$,where $a = \left\{-infty\right\}$ and $b = \left\{infty\right\}$ are allowed; with one of then infinite we would have an unbounded interval, while with both infinite we would have the whole real number line.

As with traditional calculations with real numbers, simple arithmetic operations and functions on elementary intervals must first be defined (Lit.: Kulisch, 1989). More complicated functions can be calculated from these basic elements (Lit.: Kulish, 1989).

imple arithmetic

Returning to the earlier BMI example, in determining the body mass index, height and body weight both affect the result. For height, measurements are usually in round centimetres: a recorded measurement of 1.80 metres actually means a height somewhere between 1.795 m and 1.805 m. This uncertainty must be combined with the fluctuation range in weight between 79.5 kg and 80.5 kg. The BMI is defined as the weight in kilograms divided by the square of height in metre. Using either 79.5 kg and 1.795 m or 80.5 kg and 1.805 m gives approximately 24.7. But the person in question may only be 1.795 m tall, with a weight of 80.5 kilograms - or 1.805 m and 79.5 kilograms: all combinations of all possible intermediate values must be considered. Using the interval arithmetic methods described below, the BMI lies in the interval:$\left[79\left\{.\right\}5; 80\left\{.\right\}5\right] /\left( \left[1\left\{.\right\}795; 1\left\{.\right\}805\right] \right)^2 = \left[24\left\{.\right\}4; 25\left\{.\right\}0\right] .$

An operation $\left\{langle!mathrm\left\{op\right\}! angle\right\}$ on two intervals , with $\left\{langle!mathrm\left\{op\right\}! angle\right\}$ for example being addition or multiplication, is defined by

:$\left[x_1, x_2\right] \left\{,langle!mathrm\left\{op\right\}! angle,\right\} \left[y_1, y_2\right] = \left\{ x \left\{,langle!mathrm\left\{op\right\}! angle,\right\} y , | , x in \left[x_1, x_2\right] ,mbox\left\{and\right\}, y in \left[y_1, y_2\right] \right\}$.For the four basic arithmetic operations this can become:

provided that $x \left\{,langle!mathrm\left\{op\right\}! angle,\right\} y$ is allowed for all $xin \left[x_1, x_2\right]$ and $y in \left[y_1, y_2\right]$.

For practical applications this can be simplified further:

* Addition: $\left[x_1, x_2\right] + \left[y_1, y_2\right] = \left[x_1+y_1, x_2+y_2\right]$
* Subtraction: $\left[x_1, x_2\right] - \left[y_1, y_2\right] = \left[x_1-y_2, x_2-y_1\right]$
* Multiplication: $\left[x_1, x_2\right] cdot \left[y_1, y_2\right] = \left[min\left(x_1 y_1,x_1 y_2,x_2 y_1,x_2 y_2\right), max\left(x_1 y_1,x_1 y_2,x_2 y_1,x_2 y_2\right)\right]$
* Division: $\left[x_1, x_2\right] / \left[y_1, y_2\right] = \left[x_1, x_2\right] cdot \left(1/ \left[y_1, y_2\right] \right)$, where $1/ \left[y_1, y_2\right] = \left[1/y_2, 1/y_1\right]$ if $0 otin \left[y_1, y_2\right]$. For division by an interval including zero, first define: $1/ \left[y_1, 0\right] = \left[-infty, 1/y_1\right]$ and $1/ \left[0, y_2\right] = \left[1/y_2, infty\right]$. For $y_1 < 0 < y2$, we get $1/ \left[y_1, y_2\right] = \left[-infty, 1/y_1\right] cup \left[1/y_2, infty\right]$ which as a single interval gives $1/ \left[y_1, y_2\right] = \left[-infty, infty\right]$; this loses useful information about $\left(1/y_1, 1/y_2\right)$. So typically it is common to work with $\left[-infty, 1/y_1\right]$ and $\left[1/y_2, infty\right]$ as separate intervals.

Because several such divisions may occur in an interval arithmetic calculation, it is sometimes useful to do the calculation with so-called "multi-intervals" of the form . The corresponding "multi-interval arithmetic" maintains a disjoint set of intervals and also provides for overlapping intervals to unite (Lit.: Dreyer, 2005).

Since a number $rin mathbb\left\{R\right\}$ can be interpreted as the interval $\left[r,r\right]$, you can combine intervals and real numbers.

With the help of these definitions, it is already possible to calculate the range of simple functions, such as $f\left(a,b,x\right) = a cdot x + b$. If, for example$a = \left[1,2\right]$, $b = \left[5,7\right]$ and $x = \left[2,3\right]$, it is clear

:$f\left(a,b,x\right) = \left( \left[1,2\right] cdot \left[2,3\right] \right) + \left[5,7\right] = \left[1cdot 2, 2cdot 3\right] + \left[5,7\right] = \left[7,13\right]$.

Interpreting this as a function $f\left(a,b,x\right)$ of the variable $x$ with interval parameters $a$ and $b$, them it is possible to find the roots of this function. It is then

:$f\left( \left[1,2\right] , \left[5,7\right] ,x\right) = \left( \left[1,2\right] cdot x\right) + \left[5,7\right] = 0Leftrightarrow \left[1,2\right] cdot x = \left[-7, -5\right] Leftrightarrow x = \left[-7, -5\right] / \left[1,2\right]$,the possible zeros are in the interval $\left[-7, \left\{-2.5\right\}\right]$.

As in the above example, the multiplication of intervals often only requires two multiplications. It is in fact

:$\left[x_1, x_2\right] cdot \left[y_1, y_2\right] = \left[x_1 cdot y_1, x_2 cdot y_2\right]$, if $x_1, y_1 geq 0$.

The multiplication can be see as a destination area of a rectangle with varying edges. The result interval covers all levels from the smallest to the largest.

The same applies when one of the two intervals is non-positive and the other non-negative. Generally, multiplication can produce results as wide as $\left[\left\{-infty\right\}, \left\{infty\right\}\right]$, for example if $0 cdot infty$ is squared. This also occurs, for example, in a division, if the numerator and denominator both contain zero.

Notation

To make the notation of intervals smaller in formulae, brackets can be used.

So we can use $\left[x\right] equiv \left[x_1, x_2\right]$ to represent an interval. For the set of all finite intervals, we can use :as an abbreviation. For a vector of intervals we can also used a bold font: $\left[mathbf\left\{x\right\}\right]$.

In such a compact notation, you should note that $\left[x\right]$ should not be confused between a so-called improper or single point interval $\left[x_1, x_1\right]$ and the lower and upper limit.

Elementary functions

Interval methods can also apply to functions which do not just use simple arithmetic, and we must also use other basic functions for redefining intervals, using already known monotonicity properties.

For monotonic functions in one variable, the range of values is also easy. If $f: mathbb\left\{R\right\} ightarrow mathbb\left\{R\right\}$ is monotonically rising or falling in the interval $\left[x_1, x_2\right]$, then for all values in the interval $y_1, y_2 in \left[x_1, x_2\right]$ such that $y_1 leq y_2$, one of the following inequalities applies::$f\left(y_1\right) leq f\left(y_2\right)$, or $f\left(y_1\right) geq f\left(y_2\right)$.

The range corresponding to the interval $\left[y_1, y_2\right] subseteq \left[x_1, x_2\right]$ can be calculated by applying the function to the endpoints $y_1$ and $y_2$::.

From this the following basic features for interval functions can easily be defined:
* Exponential function: $a^\left\{ \left[x_1, x_2\right] \right\} = \left[a^\left\{x_1\right\},a^\left\{x_2\right\}\right]$, for $a > 1$,
* Logarithm: , for positive intervals $\left[x_1, x_2\right]$ and $a>1$
* Odd powers: $\left\{ \left[x_1, x_2\right] \right\}^n = \left[\left\{x_1\right\}^n,\left\{x_2\right\}^n\right]$, for odd $nin mathbb\left\{N\right\}$.

For even powers, the range of values being considered is important, and needs to be dealt with before doing any multiplication.For example $x^n$ for $x in \left[-1,1\right]$ should produce the interval $\left[0,1\right]$ when $n = 2, 4, 6, ldots$. But if you take $\left[-1,1\right] ^n$ by applying interval multiplication of form $\left[-1,1\right] cdot ldots cdot \left[-1,1\right]$ then the result will appear to be $\left[-1,1\right]$, wider than necessary.

Instead consider the function $x^n$ as a monotonically decreasing function for $x < 0$ and a monotonically increasing function for $x > 0$. So for even $nin mathbb\left\{N\right\}$:

* $\left\{ \left[x_1, x_2\right] \right\}^n = \left[x_1^n, x_2^n\right]$, if $x_1 geq 0$,
* $\left\{ \left[x_1, x_2\right] \right\}^n = \left[x_2^n, x_1^n\right]$, if $x_2 < 0$,
* $\left\{ \left[x_1, x_2\right] \right\}^n = \left[0, max \left\{x_1^n, x_2^n \right\} \right]$, otherwise.

More generally, one can say that for piecewise monotonic functions it is sufficient to consider the endpoints $x_1, x_2$ of the interval $\left[x_1, x_2\right]$, together with the so-called "critical points" within the interval being those points where the monotonicity of the function changes direction.

For the sine and cosine functions, the critical points are at $left\left( \left\{\right\}^1!!/!\left\{\right\}_2 + \left\{n\right\} ight\right) cdot pi$ or $\left\{n\right\} cdot pi$ for all $n in mathbb\left\{Z\right\}$ respectively. Only up to five points matter as the resulting interval will be $\left[-1,1\right]$ if at least half a period is in the input interval. For sine and cosine, only the endpoints need full evaluation as the critical points lead to easily pre-calculated values – namely -1, 0 , +1.

Interval extensions of general functions

In general, it may not be easy to find such a simple description of the output interval for many functions. But it may still be possible to extend functions to interval arithmetic. If $f:mathbb\left\{R\right\}^n ightarrow mathbb\left\{R\right\}$ is a function from a real vector to a real number, then $\left[f\right] : \left[mathbb\left\{R\right\}\right] ^n ightarrow \left[mathbb\left\{R\right\}\right]$ is called an "interval extension" of $f$ if:$\left[f\right] \left( \left[mathbf\left\{x\right\}\right] \right) supseteq \left\{f\left(mathbf\left\{y\right\}\right) | mathbf\left\{y\right\} in \left[mathbf\left\{x\right\}\right] \right\}$.

This definition of the interval extension does not give a precise result. For example, both $\left[f\right] \left( \left[x_1,x_2\right] \right) = \left[e^\left\{x_1\right\}, e^\left\{x_2\right\}\right]$ and $\left[g\right] \left( \left[x_1,x_2\right] \right) = \left[\left\{-infty\right\}, \left\{infty\right\}\right]$ are allowable extensions of the exponential function. Extensions as tight as possible are desirable, taking into the relative costs of calculation and imprecision; in this case $\left[f\right]$ should be chosen as it give the tightest possible result.

The "natural interval extension" is achieved by combining the function rule $f\left(x_1, cdots, x_n\right)$ with the equivalents of the basic arithmetic and elementary functions.

The "Taylor interval extension" (of degree $k$ ) is a $k+1$ times differentiable function $f$ defined by

:$\left[f\right] \left( \left[mathbf\left\{x\right\}\right] \right) := f\left(mathbf\left\{y\right\}\right) + sum_\left\{i=1\right\}^kfrac\left\{1\right\}\left\{i!\right\}mathrm\left\{D\right\}^i f\left(mathbf\left\{y\right\}\right) cdot \left( \left[mathbf\left\{x\right\}\right] - mathbf\left\{y\right\}\right)^i + \left[r\right] \left( \left[mathbf\left\{x\right\}\right] , \left[mathbf\left\{x\right\}\right] , mathbf\left\{y\right\}\right)$,for some $mathbf\left\{y\right\} in \left[mathbf\left\{x\right\}\right]$, where $mathrm\left\{D\right\}^i f\left(mathbf\left\{y\right\}\right)$ is the $i$th order differential of $f$ at the point $mathbf\left\{y\right\}$ and $\left[r\right]$ is an interval extension of the "Taylor remainder"

:$r\left(mathbf\left\{x\right\}, xi, mathbf\left\{y\right\}\right) = frac\left\{1\right\}\left\{\left(k+1\right)!\right\}mathrm\left\{D\right\}^\left\{k+1\right\} f\left(xi\right) cdot \left(mathbf\left\{x\right\}-mathbf\left\{y\right\}\right)^\left\{k+1\right\}.$ The vector $xi$ lies between $mathbf\left\{x\right\}$ and $mathbf\left\{y\right\}$ with $mathbf\left\{x\right\}, mathbf\left\{y\right\} in \left[mathbf\left\{x\right\}\right]$, $xi$ is protected by $\left[mathbf\left\{x\right\}\right]$.Usually you choose $mathbf\left\{y\right\}$ to be the midpoint of the interval and use the natural interval extension to assess the remainder.

The special case of the Taylor interval extension of degree $k = 0$ is also referred to as the "average interval extension".For an interval extension of the Jacobian $\left[J_f\right] \left(mathbf\left\{ \left[x\right] \right\}\right)$you get

:$\left[f\right] \left( \left[mathbf\left\{x\right\}\right] \right) := f\left(mathbf\left\{y\right\}\right) + \left[J_f\right] \left(mathbf\left\{ \left[x\right] \right\}\right) cdot \left( \left[mathbf\left\{x\right\}\right] - mathbf\left\{y\right\}\right)$.

A nonlinear function can be defined by linear features.

Interval methods

The methods of classical numerical analysis can not be transferred one-to-one into interval-valued algorithms, as dependencies between numerical values are usually not taken into account.

Rounded interval arithmetic

To working effectively in a real-life implementation, intervals must be compatible to floating point computing. The earlier operations were based on exact arithmetic, but in general fast numerical solution methods may not be available. The range of values of the function $f\left(x, y\right) = x + y$for $x in \left[0.1, 0.8\right]$ and $y in \left[0.06, 0.08\right]$ are for example $\left[0.16, 0.88\right]$. Where the same calculation is done with single digit precision, the result would normally be $\left[0.2, 0.9\right]$. But $\left[0.2, 0.9\right] otsupseteq \left[0.16, 0.88\right]$,so this approach would contradict the basic principles of interval arithmetic contradict, as a part of the domain of $f\left( \left[0.1, 0.8\right] , \left[0.06, 0.08\right] \right)$ would be lost.Instead, it is the outward rounded solution $\left[0.1, 0.9\right]$ which is used.

The standard IEEE 754 for binary floating-point arithmetic also sets out procedures for the implementation of rounding. An IEEE 754 compliant system allows programmers to round to the nearest floating point number; alternatives are rounding towards 0 (truncating), rounding toward positive infinity (i.e. up), or rounding towards negative infinity (i.e. down).

The required "external rounding" for interval arithmetic can thus be achieved by changing the rounding settings of the processor in the calculation of the upper limit (up) and lower limit (down). Alternatively, an appropriate small interval $\left[varepsilon_1, varepsilon_2\right]$ can be added.

Dependency problem

The so-called "dependency problem" is a major obstacle to the application of interval arithmetic. Although interval methods can determine the range of elementary arithmetic operations and functions very accurately, this is not always true with more complicated functions. If an interval occurs several times in a calculation using parameters, and each occurrence is taken independently then this can lead to an unwanted expansion of the resulting intervals.

As an illustration, take the function $f$ defined by $f\left(x\right) = x^2 + x$. The values of this function over the interval $\left[-1, 1\right]$ are really $\left[-1/4 , 2\right]$. As the natural interval extension, it is calculated as $\left[-1, 1\right] ^2 + \left[-1, 1\right] = \left[0,1\right] + \left[-1,1\right] = \left[-1,2\right]$, which is slightly larger; we have instead calculated the infimum and supremum of the function $h\left(x, y\right)= x^2+y$ over $x,y in \left[-1,1\right]$. There is a better expression of $f$ in which the variable $x$ only appears once, namely by rewriting $f\left(x\right) = x^2 + x$ as addition and squaring in the quadratic $f\left(x\right) = left\left(x + frac\left\{1\right\}\left\{2\right\} ight\right)^2 -frac\left\{1\right\}\left\{4\right\}$. So the suitable interval calculation is :$left\left( \left[-1,1\right] + frac\left\{1\right\}\left\{2\right\} ight\right)^2 -frac\left\{1\right\}\left\{4\right\} = left \left[-frac\left\{1\right\}\left\{2\right\}, frac\left\{3\right\}\left\{2\right\} ight\right] ^2 -frac\left\{1\right\}\left\{4\right\} = left \left[0, frac\left\{9\right\}\left\{4\right\} ight\right] -frac\left\{1\right\}\left\{4\right\} = left \left[-frac\left\{1\right\}\left\{4\right\},2 ight\right]$and gives the correct values.

In general, it can be shown that the exact range of values can be achieved, if each variable appears only once. However, not every function can be rewritten this way.

The dependency of the problem causing over-estimation of the value range can go as far as covering a large range, preventing more meaningful conclusions.

An additional increase in the range stems from the solution of areas that do not take the form of an interval vector. The solution set of the linear system: for $pin \left[-1,1\right]$is precisely the line between the points $\left(-1,-1\right)$ and $\left(1,1\right)$. Interval methods deliver the best case, but in the square $\left[-1,1\right] imes \left[-1,1\right]$, The real solution is contained in this square (this is known as the "wrapping effect").

Linear interval systems

A linear interval system consists of a matrix interval extension $\left[mathbf\left\{A\right\}\right] in \left[mathbb\left\{R\right\}\right] ^\left\{n imes m\right\}$ and an interval vector $\left[mathbf\left\{b\right\}\right] in \left[mathbb\left\{R\right\}\right] ^\left\{n\right\}$. We want the smallest cuboid $\left[mathbf\left\{x\right\}\right] in \left[mathbb\left\{R\right\}\right] ^\left\{m\right\}$, for all vectors $mathbf\left\{x\right\} in mathbb\left\{R\right\}^\left\{m\right\}$ which there is a pair $\left(mathbf\left\{A\right\}, mathbf\left\{b\right\}\right)$ with $mathbf\left\{A\right\} in \left[mathbf\left\{A\right\}\right]$ and $mathbf\left\{b\right\} in \left[mathbf\left\{b\right\}\right]$ satisfying:$mathbf\left\{A\right\} cdot mathbf\left\{x\right\} = mathbf\left\{b\right\}$.

For quadratic systems - in other words, for $n = m$ - there can be such an interval vector $\left[mathbf\left\{x\right\}\right]$, which covers all possible solutions, found simply with the interval Gauss method. This replaces the numerical operations, in that the linear algebra method known as Gaussian elimination becomes its interval version. However, since this method uses the interval entities$\left[mathbf\left\{A\right\}\right]$ and $\left[mathbf\left\{b\right\}\right]$ repeatedly in the calculation, it can produce poor results for some problems Hence using the result of the interval-valued Gauss only provides first rough estimates, since although it contains the entire solution set, it also has a large area outside it.

A rough solution $\left[mathbf\left\{x\right\}\right]$ can often be improved by an interval version of the Gauss–Seidel method. The motivation for this is that the $i$-th row of the interval extension of the linear equation:can be determined by the variable $x_i$ if the division $1/ \left[a_\left\{ii\right\}\right]$ is allowed. It is therefore simultaneously :$x_j in \left[x_j\right]$ and $x_j in frac\left\{ \left[b_i\right] - sumlimits_\left\{k ot= j\right\} \left[a_\left\{ik\right\}\right] cdot \left[x_k\right] \right\}\left\{ \left[a_\left\{ij\right\}\right] \right\}$.So you can now replace $\left[x_j\right]$ by :$\left[x_j\right] cap frac\left\{ \left[b_i\right] - sumlimits_\left\{k ot= j\right\} \left[a_\left\{ik\right\}\right] cdot \left[x_k\right] \right\}\left\{ \left[a_\left\{ij\right\}\right] \right\}$, and so the vector $\left[mathbf\left\{x\right\}\right]$ by each element.Since the procedure is more efficient for a diagonally dominant matrix, instead of the system $\left[mathbf\left\{A\right\}\right] cdot mathbf\left\{x\right\} = \left[mathbf\left\{b\right\}\right] mbox\left\{,\right\}$ you can often try multiplying it by an appropriate rational matrix $mathbf\left\{M\right\}$ with the resulting matrix equation:$\left(mathbf\left\{M\right\}cdot \left[mathbf\left\{A\right\}\right] \right)cdot mathbf\left\{x\right\} = mathbf\left\{M\right\}cdot \left[mathbf\left\{b\right\}\right]$left to solve. If you choose, for example, $mathbf\left\{M\right\} = mathbf\left\{A\right\}^\left\{-1\right\}$ for the central matrix $mathbf\left\{A\right\} in \left[mathbf\left\{A\right\}\right]$, then $mathbf\left\{M\right\} cdot \left[mathbf\left\{A\right\}\right]$ is outer extension of the identity matrix.

These methods only work well if the widths of the intervals occurring are sufficiently small. For wider intervals it can be useful to use an interval-linear system on finite (albeit large) real number equivalent linear systems. If all the matrices $mathbf\left\{A\right\} in \left[mathbf\left\{A\right\}\right]$ are invertible, it is sufficient to consider all possible combinations (upper and lower) of the endpoints occurring in the intervals. The resulting problems can be resolved using conventional numerical methods. Interval arithmetic is still used to determine rounding errors.

This is only suitable for systems of smaller dimension, since with a fully occupied $n imes n$ matrix, $2^\left\{n^2\right\}$ real matrices need to be inverted, with $2^n$ vectors for the right hand side. This approach was developed by Jiri Rohn and is still being developed. [ [http://www.cs.cas.cz/rohn/publist/000home.htm Jiri Rohn, List of publications] ]

Interval Newton method

An interval variant of Newton's method for finding the zeros in an interval vector $\left[mathbf\left\{x\right\}\right]$ can be derived from the average value extension (Lit.: Hansen, 1992). For an unknown vector $mathbf\left\{z\right\}in \left[mathbf\left\{x\right\}\right]$ applied to $mathbf\left\{y\right\}in \left[mathbf\left\{x\right\}\right]$, gives:$f\left(mathbf\left\{z\right\}\right) in f\left(mathbf\left\{y\right\}\right) + \left[J_f\right] \left(mathbf\left\{ \left[x\right] \right\}\right) cdot \left(mathbf\left\{z\right\} - mathbf\left\{y\right\}\right)$.For a zero $mathbf\left\{z\right\}$, that is $f\left(z\right)=0$, and thus must satisfy:$f\left(mathbf\left\{y\right\}\right) + \left[J_f\right] \left(mathbf\left\{ \left[x\right] \right\}\right) cdot \left(mathbf\left\{z\right\} - mathbf\left\{y\right\}\right)=0$.This is equivalent to $mathbf\left\{z\right\} in mathbf\left\{y\right\} - \left[J_f\right] \left(mathbf\left\{ \left[x\right] \right\}\right)^\left\{-1\right\}cdot f\left(mathbf\left\{y\right\}\right)$.An outer estimate of $\left[J_f\right] \left(mathbf\left\{ \left[x\right] \right\}\right)^\left\{-1\right\}cdot f\left(mathbf\left\{y\right\}\right)\right)$ can be determined using linear methods.

In each step of the interval Newton method, an approximate starting value $\left[mathbf\left\{x\right\}\right] in \left[mathbb\left\{R\right\}\right] ^n$ is replaced by $\left[mathbf\left\{x\right\}\right] cap left\left(mathbf\left\{y\right\} - \left[J_f\right] \left(mathbf\left\{ \left[x\right] \right\}\right)^\left\{-1\right\}cdot f\left(mathbf\left\{y\right\}\right) ight\right)$ and so the result can be improved iteratively. In contrast to traditional methods, the interval method approaches the result by containing the zeros. This guarantees that the result will produce all the zeros in the initial range. Conversely, it will prove that no zeros of $f$ were in the initial range $\left[mathbf\left\{x\right\}\right]$ if a Newton step produces the empty set.

The method converges on all zeros in the starting region. Division by zero can lead to separation of distinct zeros, though the separation may not be complete; it can be complemented by the bisection method.

As an example, consider the function $f\left(x\right)= x^2-2$, the starting range $\left[x\right] = \left[-2,2\right]$, and the point $y= 0$. You then have $J_f\left(x\right) = 2, x$ and the first Newton step gives :$\left[-2,2\right] cap left\left(0 - frac\left\{1\right\}\left\{2cdot \left[-2,2\right] \right\} \left(0-2\right) ight\right) = \left[-2,2\right] cap Big\left( \left[\left\{-infty\right\}, \left\{-0.5\right\}\right] cup \left[\left\{0.5\right\}, \left\{infty\right\}\right] Big\right)$.There is therefore a zero in .More Newton steps are used separately on $xin \left[\left\{-2\right\}, \left\{-0.5\right\}\right]$ and $\left[\left\{0.5\right\}, \left\{2\right\}\right]$. These converge to arbitrarily small intervals around $-sqrt\left\{2\right\}$ and $+sqrt\left\{2\right\}$.

The Interval Newton method can also be used with "thick functions" such as $g\left(x\right)= x^2- \left[2,3\right]$, which would in any case have interval results. The result then produces intervals containing $left \left[-sqrt\left\{3\right\},-sqrt\left\{2\right\} ight\right] cup left \left[sqrt\left\{2\right\},sqrt\left\{3\right\} ight\right]$.

Bisection and covers

The various interval methods deliver conservative results as dependencies between the sizes of different intervals extensions are not taken into account. However the dependency problem becomes less significant for narrower intervals.

Covering an interval vector $\left[mathbf\left\{x\right\}\right]$ by smaller boxes $\left[mathbf\left\{x\right\}_1\right] , dots , \left[mathbf\left\{x\right\}_k\right] mbox\left\{,\right\}$ so that is then valid for the range of values So for the interval extensions described above, is valid.Since $\left[f\right] \left( \left[mathbf\left\{x\right\}\right] \right)$ is often a genuine superset of the right-hand side, this usually leads to an improved estimate.

Such a cover can be generated by the bisection method such as thick elements $\left[x_\left\{i1\right\}, x_\left\{i2\right\}\right]$ of the interval vector $\left[mathbf\left\{x\right\}\right] = \left( \left[x_\left\{11\right\}, x_\left\{12\right\}\right] , dots, \left[x_\left\{n1\right\}, x_\left\{n2\right\}\right] \right)$ by splitting in the centre into the two intervals $\left[x_\left\{i1\right\}, \left(x_\left\{i1\right\}+x_\left\{i2\right\}\right)/2\right]$ and $\left[\left(x_\left\{i1\right\}+x_\left\{i2\right\}\right)/2, x_\left\{i2\right\}\right]$. It the result is still not suitable then further gradual subdivision is possible. Note that a cover of $2^r$ intervals results from $r$ divisions of vector elements, substantially increasing the computation costs.

With very wide intervals, it can be helpful to split all intervals into several subintervals with a constant (and smaller) width, a method known as "mincing". This then avoids the calculations for intermediate bisection steps. Both methods are only suitable for problems of low dimension.

Application

Interval arithmetic can be use in various areas, in order to be treated estimates for which no exact numerical values can stated (Lit.: Jaulin et al., 2001).

Rounding error analysis

Interval arithmetic is used with error analysis, to control rounding errors arising from each calculation. The advantage of interval arithmetic is that after each operation there is an interval which reliably includes the true result. The distance between the interval boundaries gives the current calculation of rounding errors directly:: Error = $mathrm\left\{abs\right\}\left(a-b\right)$ for a given interval $\left[a,b\right]$.Interval analysis adds to rather than substituting for traditional methods for error reduction, such as pivoting.

Tolerance analysis

Parameters for which no exact figures can be allocated often arise during the simulation of technical and physical processes. The production process of technical components allows certain tolerances, so some parameters fluctuate within intervals.In addition, many fundamental constants not are not known precisely (Lit.: Dreyer, 2005).

If the behavior of such a system affected by tolerances satisfies, for example, $f\left(mathbf\left\{x\right\}, mathbf\left\{p\right\}\right) = 0$, for $mathbf\left\{p\right\} in \left[mathbf\left\{p\right\}\right]$ and unknown $mathbf\left\{x\right\}$ then the set of possible solutions :$\left\{mathbf\left\{x\right\},|, exists mathbf\left\{p\right\} in \left[mathbf\left\{p\right\}\right] , f\left(mathbf\left\{x\right\}, mathbf\left\{p\right\}\right)= 0\right\}$,can be found by interval methods. This provides an alternative to traditional propagation of error analysis. Unlike point methods, such as Monte Carlo simulation, interval arithmetic methodology ensures that no part of the solution area can be overlooked.However, the result is always a worst case analysis for the distribution of error, as other probability-based distributions are not considered.

Fuzzy interval arithmetic

Interval arithmetic can also be used with affiliation functions for fuzzy quantities as they are used in fuzzy logic. Apart from the strict statements $xin \left[x\right]$ and $x otin \left[x\right]$, intermediate values are also possible, to which real numbers $mu in \left[0,1\right]$ are assigned. $mu = 1$ corresponds to definite membership while $mu = 0$ is non-membership. A distribution function assigns uncertainty which can be understood as a further interval.

For "fuzzy arithmetic" [ [http://www.iam.uni-stuttgart.de/Mitarbeiter/Hanss/hanss_en.htm Application of Fuzzy Arithmetic to Quantifying the Effects of Uncertain Model Parameters, Michael Hanss] , University of Stuttgart] only a finite number of discrete affiliation stages $mu_i in \left[0,1\right]$ are considered. The form of such a distribution for an indistinct value can then represented by a sequence of intervals:$left \left[x^\left\{\left(1\right)\right\} ight\right] supset left \left[x^\left\{\left(2\right)\right\} ight\right] supset cdots supset left \left[x^\left\{\left(k\right)\right\} ight\right]$. The interval $\left[x^\left\{\left(i\right)\right\}\right]$ corresponds exactly to the fluctuation range for the stage $mu_i$.

The appropriate distribution for a function $f\left(x_1, cdots, x_n\right)$ concerning indistinct values$x_1, cdots, x_n$ and the corresponding sequences $left \left[x_1^\left\{\left(1\right)\right\} ight\right] supset cdots supset left \left[x_1^\left\{\left(k\right)\right\} ight\right] , cdots ,left \left[x_n^\left\{\left(1\right)\right\} ight\right] supset cdots supset left \left[x_n^\left\{\left(k\right)\right\} ight\right]$ can be approximated by the sequence $left \left[y^\left\{\left(1\right)\right\} ight\right] supset cdots supset left \left[y^\left\{\left(k\right)\right\} ight\right]$.The values $left \left[y^\left\{\left(i\right)\right\} ight\right]$ are given by $left \left[y^\left\{\left(i\right)\right\} ight\right] = f left\left( left \left[x_\left\{1\right\}^\left\{\left(i\right)\right\} ight\right] , cdots left \left[x_\left\{n\right\}^\left\{\left(i\right)\right\} ight\right] ight\right)$ and can be calculated by interval methods. The value $left \left[y^\left\{\left(1\right)\right\} ight\right]$ corresponds to the result of an interval calculation.

History

Interval arithmetic is not a completely new phenomenon in mathematics; it has appeared several times under different names in the course of history. For example Archimedes calculated lower and upper bounds 223/71 < π < 22/7 in the 3rd century BC.Actual calculations with intervals has neither been as popular as other numerical techniques, nor been completely forgotten.

Rules for calculating with intervals and other subsets of the real numbers were published in a 1931 work by Rosalind Cicely Young, a doctoral candidate at the University of Cambridge. Arithmetic work on range numbers to improve reliability of digital systems were then published in a 1951 textbook on linear algebra by Paul Dwyer (University of Michigan); intervals were used to measure rounding errors associated with floating-point numbers.

The birth of modern interval arithmetic was marked by the appearance of the book "Interval Analysis" by Ramon E. Moore in 1966 (Lit.: Moore). He had the idea in Spring 1958, and a year later he published an article about computer interval arithmetic. [ [http://interval.louisiana.edu/Moores_early_papers/bibliography.html Publications Related to Early Interval Work of R. E. Moore] ] . Its merit was that stating with a simple principle it provided is a general method for automated error analysis, not just errors resulting from rounding.

Independently in 1956, Mieczyslaw Warmus suggested formulae for calculations with intervals [ [http://www.ippt.gov.pl/~zkulpa/quaphys/warmus.html Precursory papers on interval analysis by M. Warmus] ] , though Moore found the first non-trivial applications.

In the following twenty years German groups of researchers carried out pioneering work around Götz Alefeld (Lit.: Alefeld and Herzberger) and Ulrich Kulisch (Lit.: Kulisch) at the University of Karlsruhe and later also at the Bergische University of Wuppertal. For example, Karl Nickel explored more effective implementations, while improved containment procedures for the solution set of systems of equations were due to Arnold Neumaier among others. [ [http://www.mat.univie.ac.at/~neum/publist.html Publications by Arnold Neumaier] ] . In the 1960s Eldon R. Hansen dealt with interval extensions for linear equations and then provided crucial contributions to global optimisation (Lit.: Hansen). Classical methods in this often are have the problem of determining the largest (or smallest) global value, but could only find a local optimum and could not find better values; Helmut Ratschek and Jon George Rokne developed branch and bound methods, which till then had only applies to integer values, by using intervals to provide applications for continuous values [ [http://pages.cpsc.ucalgary.ca/~rokne/#SEC3 Some publications of Jon Rokne] ] .

In 1988 Rudolf Lohner developed Fortran-based software for reliable solutions for initial value problems using ordinary differential equations. [ [http://fam-pape.de/raw/ralph/studium/dgl/dglsem.html Bounds for ordinary differential equations of Rudolf Lohner] (in German)]

The journal "Reliable Computing" (originally "Interval Computations") has been published since the 1990s , dedicated to the reliability of computer-aided computations. As lead editor, R. Baker Kearfott, in addition to his work on global optimisation, has contributed significantly to the unification of notation and terminology used in interval arithmetic (Web: Kearfott).

In recent years work has concentrated in particular on the estimation of preimages of parameterised functions and to robust control theory by the COPRIN working group of INRIA in Sophia Antipolis in France (Web: INRIA).

Patents

One of the main sponsors of the interval arithmetic, G. William Walster of Sun Microsystems, has - in part with Ramon E. Moore and Eldon R. Hansen - lodged several patents in the field of interval arithmetic atthe U.S. Patent and Trademark Office in the years 2002-04 [ [http://www.mat.univie.ac.at/coconut-environment/#patents Patent Issues in Interval Arithmetic] ] . The validity of these patent applications have been disputed in the interval arithmetic research community, since they may possibly only show the past state of the art.

Implementations

There are many software packages which permit the development of numerical applications using interval arithmetic [ [http://www.cs.utep.edu/interval-comp/main.html Software for Interval Computations collected by Vladik Kreinovich] , University of Texas at El Paso] .These are usually provided in the form of program libraries. [ [http://docs.sun.com/source/816-2465/iapgCusing.html#26326 C++ Interval Arithmetic Programming Reference] from Sun Microsystems] There are also C++ and Fortran compilers handle interval data types and suitable operations as a language extension, [ [http://developers.sun.com/sunstudio/overview/topics/numerics_index.html C++ and Fortran compilers with Interval data types] from Sun Microsystems] so interval arithmetic is supported directly.

Since 1967 "Extensions for Scientific Computation" (XSC) have been developed in the University of Karlsruhe for various programming languages, such as C++, Fortran and Pascal. [ [http://www.math.uni-wuppertal.de/org/WRST/xsc/history.html History of XSC-Languages] ] The first platform was a Zuse Z 23, for which a new interval data type with appropriate elementary operators was made available. There followed in 1976 Pascal-SC, a Pascal variant on a Zilog Z80 which it made possible to create fast complicated routines for automated result verification. Then came the Fortran 77-based ACRITH XSC for the System/370 architecture, which was later delivered by IBM. Starting from 1991 one could produce code for C compilers with Pascal XSC; a year later the C++ class library supported C-XSC on many different computer systems. In 1997 all XSC variants were made available under the General Public License. At the beginning of 2000 C-XSC 2.0 was released under the leadership of the working group for scientific computation at the Bergische University of Wuppertal, in order to correspond to the improved C++ standard.

Another C++- class library was created in 1993 at the Hamburg University of Technology called "Profil/BIAS" (Programmer's Runtime Optimized Fast Interval Library, Basic Interval Arithmetic), which made the usual interval operations more user friendly. It emphasised the efficient use of hardware, portability and independence of a particular presentation of intervals.

The Boost collection of C++ libraries contains a template class for intervals. Its authors are aiming to have interval arithmetic in the standard C++ language. [ [http://www-sop.inria.fr/geometrica/team/Sylvain.Pion/cxx/ A Proposal to add Interval Arithmetic to the C++ Standard Library] ]

In addition computer algebra systems, such as Mathematica, Maple and MuPAD, can handle intervals. There is a Matlab extension "Intlab" which builds on BLAS routines, as well as the Toolbox b4m which makes a Profil/BIAS interface. [ [http://www.ti3.tu-harburg.de/~rump/intlab/ INTerval LABoratory] and [http://www.ti3.tu-harburg.de/zemke/b4m/ b4m] ] .

* Automatic differentiation
* Multigrid Method
* Monte-Carlo simulation

Further information

Literature

* Götz Alefeld und Jürgen Herzberger: "Einführung in die Intervallrechnung". Bibliographisches Institut, Reihe Informatik, Band 12, B.I.-Wissenschaftsverlag, Mannheim - Wien - Zürich, ISBN 3-411-01466-0
* Alexander Dreyer: "Interval Analysis of Analog Circuits with Component Tolerances". Doctoral thesis, Shaker Verlag, Aachen, 2003, ISBN 3-8322-4555-3.
* Eldon Hansen and G. William Walster: "Global Optimization using Interval Analysis, Second Edition, Revised and Expanded", Marcel Dekker, New York, 2004, ISBN 0-8247-4059-9.
* L. Jaulin, M. Kieffer, O. Didrit, and É.Walter: "Applied Interval Analysis: With examples in parameter estimation robust control and robotics". Springer, London, 2001, ISBN 1-85233-219-0.
* Ulrich Kulisch: "Wissenschaftliches Rechnen mit Ergebnisverifikation. Eine Einführung", Vieweg-Verlag, Wiesbaden 1989, ISBN 3-528-08943-1.
* R. E. Moore: "Interval Analysis". Prentice-Hall, Englewood Cliff, NJ, 1966, ISBN 0-13-476853-1.

* [http://www.cs.utep.edu/interval-comp/hayes.pdf Brian Hayes, 'A Lucid Interval', a good introduction (pdf)]
* [http://www-sop.inria.fr/coprin/logiciels/ALIAS/Movie/movie_undergraduate.mpg Introductory Film (mpeg)] of the [http://www-sop.inria.fr/coprin/index_english.html COPRIN] teams of INRIA, Sophia Antipolis
* [http://interval.louisiana.edu/kearfott.html Bibliography of R. Baker Kearfott] , University of Louisiana at Lafayette
* [http://www.mat.univie.ac.at/~neum/interval.html Interval Methods from Arnold Neumaier] , University of Vienna

References

Wikimedia Foundation. 2010.

### Look at other dictionaries:

• Interval (mathematics) — This article is about intervals of real numbers. For intervals in general mathematics, see Partially ordered set. For other uses, see Interval. In mathematics, a (real) interval is a set of real numbers with the property that any number that lies …   Wikipedia

• Arithmetic coding — is a method for lossless data compression. Normally, a string of characters such as the words hello there is represented using a fixed number of bits per character, as in the ASCII code. Like Huffman coding, arithmetic coding is a form of… …   Wikipedia

• Arithmetic mean — In mathematics and statistics, the arithmetic mean, often referred to as simply the mean or average when the context is clear, is a method to derive the central tendency of a sample space. The term arithmetic mean is preferred in mathematics and… …   Wikipedia

• Interval class — In musical set theory, an interval class (usual abbreviation: ic) is the shortest distance in pitch class space between two unordered pitch classes. For example, the interval class between pitch classes 4 and 9 is 5 because 9 4 = 5 is less than 4 …   Wikipedia

• Affine arithmetic — (AA) is a model for self validated numerical analysis. In AA, the quantities of interest are represented as affine combinations (affine forms) of certain primitive variables, which stand for sources of uncertainty in the data or approximations… …   Wikipedia

• Significance arithmetic — is a set of rules (sometimes called significant figure rules) for approximating the propagation of uncertainty in scientific or statistical calculations. These rules can be used to find the appropriate number of significant figures to use to… …   Wikipedia

• Greek arithmetic, geometry and harmonics: Thales to Plato — Ian Mueller INTRODUCTION: PROCLUS’ HISTORY OF GEOMETRY In a famous passage in Book VII of the Republic starting at Socrates proposes to inquire about the studies (mathēmata) needed to train the young people who will become leaders of the ideal… …   History of philosophy

• Arbitrary-precision arithmetic — In computer science, arbitrary precision arithmetic indicates that calculations are performed on numbers whose digits of precision are limited only by the available memory of the host system. This contrasts with the faster fixed precision… …   Wikipedia

• Confidence interval — This article is about the confidence interval. For Confidence distribution, see Confidence Distribution. In statistics, a confidence interval (CI) is a particular kind of interval estimate of a population parameter and is used to indicate the… …   Wikipedia

• Credible interval — Bayesian statistics Theory Bayesian probability Probability interpretations Bayes theorem Bayes rule · Bayes factor Bayesian inference Bayesian network Prior · Posterior · Likelihood …   Wikipedia