Without justification or historical context, traditional linear algebra texts will often define the determinant as the first step of an elaborate sequence of definitions and theorems leading up to the solution of linear systems, Cramer's rule and matrix inversion.
An alternative treatment is to axiomatically introduce the wedge product, and then demonstrate that this can be used directly to solve linear systems. This is shown below, and does not require sophisticated math skills to understand.
It is then possible to define determinants as nothing more than the coefficients of the wedge product in terms of "unit k-vectors" ( terms) expansions as above.
:A one by one determinant is the coefficient of for an 1-vector.:A two-by-two determinant is the coefficient of for an bivector:A three-by-three determinant is the coefficient of for an trivector:...
When linear system solution is introduced via the wedge product, Cramer's rule follows as a side effect, and there is no need to lead up to the end results with definitions of minors, matrices, matrix invertibility, adjoints, cofactors, Laplace expansions, theorems on determinant multiplication and row column exchanges, and so forth.
Equation of a plane
For the plane of all points through the plane passing through three independent points , , and , the normal form of the equation is
:
The equivalent wedge product equation is:
Projective and rejective components of a vector
For three dimensions the projective and rejective components of a vector with respect to an arbitrary non-zero unit vector, can be expressed in terms of the dot and cross product
:
For the general case the same result can be written in terms of the dot and wedge product and the geometric product of that and the unit vector
:
It's also worthwhile to point out that this result can also be expressed using right or left vector division as defined by the geometric product
::
Area of the parallelogram defined by u and v
If A is the area of the parallelogram defined by u and v, then :
and:
Note that this squared bivector is a geometric product.
Angle between two vectors
::
Volume of the parallelopiped formed by three vectors
:
Derivative of a unit vector
It can be shown that a unit vector derivative can be expressed using the cross product
:
The equivalent geometric product generalization is
:
Thus this derivative is the component of in the direction perpendicular to . In other words this is minus the projection of that vector onto .
This intuitively make sense (but a picture would help) since a unit vector is constrained to circular motion, and any change to a unit vector due to a change in its generating vector has to be in the direction of the rejection of from . That rejection has to be scaled by 1/|r| to get the final result.
When the objective isn't comparing to the cross product, it's also notable that this unit vector derivative can be written
:
Some properties and examples
Some fundamental geometric algebra manipulations will be provided below, showing how this vector product can be used in calculation of projections, area, and rotations. How some of these tie together and correlate concepts from other branches of mathematics, such as complex numbers, will also be shown.
In some cases these examples provide details used above in the cross product and geometric product comparisons.
Inversion of a vector
One of the powerful properties of the Geometric product is that it provides the capability to express the inverse of a non-zero vector. This is expressed by:
:-1
Dot and wedge products defined in terms of the geometric product
Given a definition of the geometric product in terms of the dot and wedge products, adding and subtracting and demonstrates that the dot and wedge product of two vectors can also be defined in terms of the geometric product
The dot product
:
This is the symmetric component of the geometric product. When two vectors are colinear the geometric and dot products of those vectors are equal.
As a motivation for the dot product it is normal to show that this quantity occurs in the solution of the length of a general triangle where the third side is the vector sum of the first and second sides .
:
The last sum is then given the name the dot product and other properties of this quantity are then shown (projection, angle between vectors, ...).
This can also be expressed using the geometric product
:
By comparison, the following equality exists
:.
Without requiring expansion by components one can define the dot product exclusively in terms of the geometric product due to its properties of contraction, distribution and associativity. This is arguably a more natural way to define the geometric product, especially since the wedge product is not familiar to many people with traditional vector algebra background, and there is no immediate requirement to add two dissimilar terms (ie: scalar and bivector).
The wedge product
:
This is the antisymmetric component of the geometric product. When two vectors are orthogonal the geometric and wedge products of those vectors are equal.
Switching the order of the vectors negates this antisymmetric geometric product component, and contraction property shows that this is zero if the vectors are equal. These are the defining properties of the wedge product.
Note on symmetric and antisymmetric dot and wedge product formulas
A generalization of the dot product that allows computation of the component of a vector "in the direction" of a plane (bivector), or other k-vectors can be found below. Since the signs change depending on the grades of the terms being multiplied, care is required with the formulas above to ensure that they are only used for a pair of vectors.
Dot and wedge products compared to the real and imaginary parts of a complex number
Reversing the order of multiplication of two vectors has the effect of the inverting the sign of just the wedge product term of the geometric product.
It is not a coincidence that this is a similar operation to the conjugate operation of complex numbers.
The reverse of a product is written in the following fashion
::
Thus, the dot product is
:
This is the symmetric component of the geometric product. When two vectors are colinear the geometric and dot products of those vectors are equal. The antisymmetric component is represented by the wedge product:
:
These symmetric and antisymmetric components extract the scalar and bivector components of a geometric product in the same fashion as the real and imaginary components of a complex number are extracted by its symmetric and antisymmetric components
::
This extraction of components also applies to higher order geometric product terms. For example
:
Orthogonal decomposition of a vector
Using the Gram-Schmidt process a single vector can be decomposed into two components with respect to a reference vector, namely the projection onto a unit vector in a reference direction, and the difference between the vector and that projection.
With, , the projection of onto is
:
Orthogonal to that vector is the difference, designated the rejection,
:
The rejection can be expressed as a single geometric algebraic product in a few different ways
:
The similarity in form between the projection and the rejection is notable. The sum of these recovers the original vector
:
Here the projection is in its customary vector form. An alternate formulation is possible that puts the projection in a form that differs from the usual vector formulation
:
A quicker way to the end result
Working backwards from the end result, it can be observed that this orthogonal decomposition result can in fact follow more directly from the definition of the geometric product itself.
:
With this approach, the original geometrical consideration is not necessarily obvious, but it is a much quicker way to get at the same algebraic result.
However, the hint that one can work backwards, coupled with the knowledge that the wedge product can be used to solve sets of linear equations (see: [http://www.grassmannalgebra.info/grassmannalgebra/book/bookpdf/TheExteriorProduct.pdf] ), the problem of orthogonal decomposition can be posed directly,
Let , where . To discard the portions of that are colinear with , take the wedge product
:
Here the geometric product can be employed
:
Because the geometric product is invertible, this can be solved for x
:
The same techniques can be applied to similar problems, such as calculation of the component of a vector in a plane and perpendicular to the plane.
Area of parallelogram spanned by two vectors
The area of a parallelogram spanned between one vector and another equals the length of one of those vectors multiplied by the length of the rejection of that vector from the second.
:
The length of this vector is the area of the spanned parallelogram, and in the square is
:
There are a couple things of note here. One is that the area can easily be expressed in terms of the square of a bivector. The other is that the square of a bivector has the same property as a purely imaginary number, a negative square.
Expansion of a bivector and a vector rejection in terms of the standard basis
If a vector is factored directly into projective and rejective terms using the geometric product , then it is not necessarily obvious that the rejection term, a product of vector and bivector is even a vector. Expansion of the vector bivector product in terms of the standard basis vectors has the following form
:Let
It can be shown that: of the imaginary unit complex number.
This allows the point to be specified as a complex exponential
:= mathbf hat{u} r ( cos heta + mathbf{I}_{mathbf{u},mathbf{v sin heta )= mathbf hat{u} r exp( mathbf{I}_{mathbf{u},mathbf{v heta )
Complex numbers could be expressed in terms of the mathbb R^2unit bivector mathbf {e_1} wedge mathbf {e_2}. However this isomorphism really only requires a pair of linearly independent vectors in a plane (of arbitrary dimension).
Quaternions
Like complex numbers, quaternions may be written as a multivector with scalar and bivector components (a 0,2-multivector).
:q = alpha + mathbf{B}
Where the complex number has one bivector component, and the quaternions have three.
One can describe quaternions as 0,2-multivectors where the basis for the bivector part is left-handed. There isn't really anything special about quaternion multiplication, or complex number multiplication, for that matter. Both are just a specific examples of a 0,2-multivector multiplication. Other quaternion operations can also be found to have natural multivector equivalents. The most important of which is likely the quaternion conjugate, since it implies the norm and the inverse. As a multivector, like complex numbers, the conjugate operation is reversal:
:overline{q} = q^dagger = alpha - mathbf{B}
Thus {vert{q}vert}^2 = qoverline{q} = alpha^2 - mathbf{B}^2. Note that this norm is a positive definite as expected since a bivector square is negative.
To be more specific about the left-handed basis property of quaternions one can note that the quaternion bivector basis is usually defined in terms of the following properties
:mathbf{i}^2 = mathbf{j}^2 = mathbf{k}^2 = -1:mathbf{i}mathbf{j} = -mathbf{j}mathbf{i}, mathbf{i}mathbf{k} = -mathbf{k}mathbf{i}, mathbf{j}mathbf{k} = -mathbf{k}mathbf{j}:mathbf{i}mathbf{j} = mathbf{k}
The first two properties are satisfied by any set of orthogonal unit bivectors for the space. The last property, which could also be written mathbf{i}mathbf{j}mathbf{k} = -1, amounts to a choice for the orientation of this bivector basis of the 2-vector part of the quaternion.
As an example suppose one picks
:mathbf{i} = mathbf{e}_2mathbf{e}_3:mathbf{j} = mathbf{e}_3mathbf{e}_1
Then the third bivector required to complete the basis set subject to the properties above is
:mathbf{i}mathbf{j} = mathbf{e}_2mathbf{e}_1 = mathbf{k}.
Suppose that, instead of the above, one picked a slightly more natural bivector basis, the duals of the unit vectors obtained by multiplication with the pseudoscalar (mathbf{e}_1mathbf{e}_2mathbf{e}_3mathbf{e}_i). These bivectors are
:mathbf{i}=mathbf{e}_2mathbf{e}_3, mathbf{j}=mathbf{e}_3mathbf{e}_1, mathbf{k}=mathbf{e}_1mathbf{e}_2.
A 0,2-multivector with this as the basis for the bivector part would have properties similar to the standard quaternions (anti-commutative unit quaternions, negation for unit quaternion square, same congugate, norm and inversion operations, ...), however the triple product would have the value mathbf{i}mathbf{j}mathbf{k} = 1, instead of -1.
Cross product as outer product
The cross product of traditional vector algebra (on mathbb{R}^3) find its place in geometric algebra mathcal{G}_3 as a scaled outer product
:mathbf{a} imesmathbf{b} = -i(mathbf{a}wedgemathbf{b})
(this is antisymmetric). Relevant is the distinction between axial and polar vectors in vector algebra, which is natural in geometric algebra as the mere distinction between vectors and bivectors (elements of grade two).
The i here is a unit pseudoscalar of Euclidean 3-space, which establishes a duality between the vectors and the bivectors, and is named so because of the expected property
:i^2 = (mathbf {e_1}mathbf {e_2}mathbf {e_3})^2 = mathbf {e_1}mathbf {e_2}mathbf {e_3}mathbf {e_1}mathbf {e_2}mathbf {e_3}= -mathbf {e_1}mathbf {e_2}mathbf {e_1}mathbf {e_3}mathbf {e_2}mathbf {e_3}= mathbf {e_1}mathbf {e_1}mathbf {e_2}mathbf {e_3}mathbf {e_2}mathbf {e_3}= -mathbf {e_3}mathbf {e_2}mathbf {e_2}mathbf {e_3}= -1
The equivalence of the mathbb{R}^3 cross product and the wedge product expression above can be confirmed by direct multiplication of -i = -mathbf {e_1}mathbf {e_2}mathbf {e_3} with a determinant expansion of the wedge product
:mathbf u wedge mathbf v = sum_{1<=iSee also Cross product#Cross product as an exterior product. Essentially, the geometric product of a bivector and the pseudoscalar of Euclidean 3-space provides a method of calculation of the hodge dual.
Intersection of a line and a plane
As a meaningful result one can consider a fixed non-zero vector mathbf v , from a point chosen as the origin, in the usual 3-D Euclidean space, mathbb{R}^3. The set of all vectors mathbf x such that mathbf x wedge mathbf v = mathbf B , mathbf B denoting a given bivector containing mathbf v , determines a line l parallel to mathbf v . Since mathbf B is a "directed" area, l is uniquely determined with respect to the chosen origin. The set of all vectors mathbf x such that mathbf x cdot mathbf v = s , s denoting a given (real) scalar, determines a plane P orthogonal to mathbf v . Again, P is uniquely determined with respect to the chosen origin. The two information pieces, mathbf B and s , can be set independently of one another. Now, what is (if any) the vector mathbf x that satisfies the system {mathbf x wedge mathbf v = mathbf B , mathbf x cdot mathbf v = s}? Geometrically, the answer is plain: it is the vector that departs from the origin and arrives at the intersection of l and P. By geometric algebra, even the algebraic answer is simple: : mathbf x mathbf v = s + mathbf B implies mathbf x = (s + mathbf B )/ mathbf v = (s + mathbf B ) mathbf v -1.
Note that the division by a vector transforms the multivector s + mathbf B into the sum of two vectors. Namely, s mathbf v -1 is the projection of mathbf x on mathbf v , and mathbf B mathbf v -1 is the rejection of mathbf x from mathbf v (i.e. the component of mathbf x orthogonal to mathbf v ). Note also that the structure of the solution does not depend on the chosen origin.
Torque
Torque is generally defined as the magnitude of the perpendicular force component times distance, or work per unit angle.
Suppose a circular path in an arbitrary plane containing orthonormal vectors hat{mathbf u} and hat{mathbf v} is parameterized by angle.
:mathbf r = r(hat{mathbf u} cos heta + hat{mathbf v} sin heta) = r hat{mathbf u}(cos heta + hat{mathbf u} hat{mathbf v} sin heta)
By designating the unit bivector of this plane as the imaginary number
:mathbf{i} = hat{mathbf u} hat{mathbf v} = hat{mathbf u} wedge hat{mathbf v}:mathbf{i}^2 = -1
this path vector can be conveniently written in complex exponential form
:mathbf r = r hat{mathbf u} e^{mathbf{i} heta}
and the derivative with respect to angle is
:frac{d mathbf r}{d heta} = r hat{mathbf u} mathbf{i} e^{mathbf{i} heta} = mathbf{r} mathbf{i}
So the torque, the rate of change of work "W", due to a force "F", is
: au = frac{dW}{d heta} = mathbf F cdot frac{d mathbf r}{d heta} = mathbf F cdot (mathbf{r} mathbf{i})
Unlike the cross product description of torque, mathbf au = mathbf r imes mathbf F no vector in a normal direction had to be introduced, a normal that doesn't exist in two dimensions or in greater than three dimensions. The unit bivector describes the plane and the orientation of the rotation, and the sense of the rotation is relative to the angle between the vectors mathbf{hat{u and mathbf{hat{v.
Expanding the result in terms of components
At a glance this doesn't appear much like the familiar torque as a determinant or cross product, but this can be expanded to demonstrate its equivalence (the cross product is hiding there in the bivector mathbf i = hat{mathbf u} wedge hat{mathbf v}). Expanding the position vector in terms of the planar unit vectors
:mathbf r mathbf i =left(r_u hat{mathbf u} + r_v hat{mathbf v}
ight)hat{mathbf u} hat{mathbf v}= r_u hat{mathbf v} - r_v hat{mathbf u}
and expanding the force by components in the same direction plus the possible perpendicular remainder term
:mathbf F = F_u hat{mathbf u} + F_v hat{mathbf v} + mathbf{F}_{perp hat{mathbf u},hat{mathbf v
and then taking dot products yields is the torque
: au = mathbf F cdot (mathbf{r} mathbf{i}) = r_u F_v - r_v F_u.
This determinant may be familiar from derivations with mathbf{hat{u = mathbf{e}_1, and mathbf{hat{v = mathbf{e}_2 (See the Feynman lectures Volume I for example).
Geometrical description
When the magnitude of the "rotational arm" is factored out, the torque can be written as
: au = mathbf F cdot (mathbf{r} mathbf{i}) = |mathbf{r}| (mathbf F cdot (mathbf{hat{r mathbf{i}))
The vector mathbf{hat{r mathbf{i} is the unit vector perpendicular to the mathbf{r}. Thus the torque can also be described as the product of the magnitude of the rotational arm times the component of the force that is in the direction of the rotation (ie: the work done rotating something depends on length of the lever, and the size of the useful part of the force pushing on it).
Application of the force to a lever not in the rotation plane
If the rotational arm that the force is applied to is not in the plane of rotation then only the components of the lever arm direction and the component of the force that are in the plane will contribute to the work done. The calculation above allowed for a force applied in an arbitrary direction, so to generalize this, a calculation that discards the component of the level arm direction not in the plane.
When mathbf{r} is allowed to lie outside of the plane of rotation the component in the plane (bivector) mathbf{i} can be described with the geometric product nicely
mathbf{r}_{mathbf{i = (mathbf{r} cdot mathbf{i}) frac{1}{mathbf{i = -(mathbf{r} cdot mathbf{i}) mathbf{i}
Thus, the vector with this magnitude that is perpendicular to this in the plane of the rotation is
mathbf{r}_{mathbf{i mathbf{i} = -(mathbf{r} cdot mathbf{i}) mathbf{i}^2= (mathbf{r} cdot mathbf{i})
and the total torque is thus
au= mathbf{F} cdot (mathbf{r} cdot mathbf{i})
This makes sense when once considers that only the dot product part of mathbf{r} mathbf{i} = mathbf{r} cdot mathbf{i} + mathbf{r} wedge mathbf{i} contributes to the component of mathbf{r} in the plane, and when the lever is in the rotational plane this wedge product component of mathbf{r} mathbf{i} is zero.
Matrix inversion and determinants
Matrix inversion (Cramer's rule) and determinants can be naturally expressed in terms of the wedge product.
The use of the wedge product in the solution of linear equations can be quite useful for various geometric product calculations.
Traditionally, instead of using the wedge product, Cramer's rule is usually presented as a generic algorithm that can be used to solve linear equations of the form mathbf A mathbf x = mathbf b (or equivalently to invert a matrix). Namely
:mathbf x = frac{1}operatorname{adj}(mathbf A)mathbf b.
This is a useful theoretic result. For numerical problems row reduction with pivots and other methods are more stable and efficient.
When the wedge product is coupled with the Clifford product and put into a natural geometric context, the fact that the determinants are used in the expression of {mathbb R}^N parallelogram area and parallelepiped volumes (and higher dimensional generalizations of these) also comes as a nice side effect.
As is also shown below, results such as Cramer's rule also follow directly from the property of the wedge product that it selects non identical elements. The end result is then simple enough that it could be derived easily if required instead of having to remember or look up a rule.
Two variables example
:egin{bmatrix}mathbf a & mathbf bend{bmatrix}egin{bmatrix}x \ yend{bmatrix}= mathbf a x + mathbf b y = mathbf c
Pre and post multiplying by mathbf a and mathbf b
: ( mathbf a x + mathbf b y ) wedge mathbf b = (mathbf a wedge mathbf b) x = mathbf c wedge mathbf b:mathbf a wedge ( mathbf a x + mathbf b y ) = (mathbf a wedge mathbf b) y = mathbf a wedge mathbf c
Provided mathbf a wedge mathbf b
eq 0 the solution is
:egin{bmatrix}x \ yend{bmatrix}= frac{1}{mathbf a wedge mathbf b}egin{bmatrix}mathbf c wedge mathbf b \ mathbf a wedge mathbf cend{bmatrix}
For mathbf a, mathbf b in {mathbb R}^2, this is Cramer's rule since the mathbf{e}_1 wedge mathbf{e}_2 factors of the wedge products
:mathbf u wedge mathbf v = egin{vmatrix}u_1 & u_2 \ v_1 & v_2 end{vmatrix} mathbf{e}_1 wedge mathbf{e}_2
divide out.
Similarly, for three, or N variables, the same ideas hold
:egin{bmatrix}mathbf a & mathbf b & mathbf cend{bmatrix}egin{bmatrix}x \ y \ zend{bmatrix} = mathbf d
:egin{bmatrix}x \ y \ zend{bmatrix} = frac{1}{mathbf a wedge mathbf b wedge mathbf c}egin{bmatrix}mathbf d wedge mathbf b wedge mathbf c \mathbf a wedge mathbf d wedge mathbf c \mathbf a wedge mathbf b wedge mathbf dend{bmatrix}
Again, for the three variable three equation case this is Cramer's rule since the mathbf{e}_1 wedge mathbf{e}_2 wedge mathbf{e}_3 factors of all the wedge products divide out, leaving the familiar determinants.
A numeric example with three equations and two unknowns
When there are more equations than variables case, if the equations have a solution, each of the k-vector quotients will be scalars
To illustrate here is the solution of a simple example with three equations andtwo unknowns.
:egin{bmatrix} 1 \ 1 \ 0 end{bmatrix}x + egin{bmatrix} 1 \ 1 \ 1 end{bmatrix}y = egin{bmatrix} 1 \ 1 \ 2 end{bmatrix}
The right wedge product with (1, 1, 1) solves for x
:egin{bmatrix} 1 \ 1 \ 0 end{bmatrix}wedgeegin{bmatrix} 1 \ 1 \ 1 end{bmatrix}x = egin{bmatrix} 1 \ 1 \ 2 end{bmatrix}wedgeegin{bmatrix} 1 \ 1 \ 1 end{bmatrix}
and a left wedge product with (1, 1, 0) solves for y
:egin{bmatrix} 1 \ 1 \ 0 end{bmatrix}wedgeegin{bmatrix} 1 \ 1 \ 1 end{bmatrix}y = egin{bmatrix} 1 \ 1 \ 0 end{bmatrix}wedgeegin{bmatrix} 1 \ 1 \ 2 end{bmatrix}
Observe that both of these equations have the same factor, soone can compute this only once (if this was zero it wouldindicate the system of equations has no solution).
Collection of results forx and y yields a Cramers rule like form:
:egin{bmatrix} x \ y end{bmatrix}=frac{1}{(1, 1, 0) wedge (1, 1, 1)}egin{bmatrix}(1, 1, 2) wedge (1, 1, 1) \(1, 1, 0) wedge (1, 1, 2)end{bmatrix}
Writing mathbf{e} _i wedge mathbf{e} _j = mathbf{e} _{ij}, we have the end result:
:egin{bmatrix} x \ y end{bmatrix}=frac{1}{mathbf{e}_{13} + mathbf{e}_{23egin{bmatrix}{-mathbf{e}_{13} - mathbf{e}_{23 \{2mathbf{e}_{13} +2mathbf{e}_{23 \end{bmatrix}=egin{bmatrix} -1 \ 2 end{bmatrix}
The contraction rule
The connection between Clifford algebras and quadratic forms come from the contraction property. This rule also gives the space a metric defined by the naturally derived inner product. It is to be noted that in geometric algebra in all its generality there is no restriction whatsoever on the value of the scalar, it can very well be negative, even zero (in that case, the possibility of an inner product is ruled out if you require langle x, x
angle ge 0).
The contraction rule can be put in the form::Q(mathbf a) = mathbf a^2 = epsilon_a {Vert mathbf a Vert}^2where Vert mathbf a Vert is the modulus of vector a, and epsilon_a=0, , pm1 is called the "signature" of vector a. This is especially useful in the construction of a Minkowski space (the spacetime of special relativity) through mathbb{R}_{1,3}. In that context, null-vectors are called "lightlike vectors", vectors with negative signature are called "spacelike vectors" and vectors with positive signature are called "timelike vectors" (these last two denominations are exchanged when using mathbb{R}_{3,1} instead).
Applications of geometric algebra
A useful example is mathbb{R}_{3, 1}, and to generate mathcal{G}_{3, 1}, an instance of geometric algebra sometimes called spacetime algebra. [Cf. Hestenes (1966).] The electromagnetic field tensor, in this context, becomes just a bivector mathbf{E} + imathbf{B} where the imaginary unit is the volume element, giving an example of the geometric reinterpretation of the traditional "tricks".
Boosts in this Lorenzian metric space have the same expression e^{mathbf{eta as rotation in Euclidean space, where mathbf{eta} is of course the bivector generated by the time and the space directions involved, whereas in the Euclidean case it is the bivector generated by the two space directions, strengthening the "analogy" to almost identity.
History
The concinnity of geometry and algebra dates as far back at least to Euclid's "Elements" in the 3rd century B.C. [Euclid, Books II and VI.] It was not, however, until 1844 that algebra would be used in a "systematic way" to describe the geometrical properties and transformations of a space. In that year, Hermann Grassmann introduced the idea of a geometrical algebra in full generality as a certain calculus (analogous to the propositional calculus) which encoded all of the geometrical information of a space. Grassmann's algebraic system could be applied to a number of different kinds of spaces: the chief among them being Euclidean space, affine space, and projective space. Following Grassmann, in 1878 William Kingdon Clifford examined Grassmann's algebraic system alongside the quaternions of William Rowan Hamilton. From his point of view, the quaternions described certain "transformations" (which he called "rotors"), whereas Grassmann's algebra described certain "properties" (or "Strecken" such as length, area, and volume). His contribution was to define a new product — the geometric product — on an existing Grassmann algebra, which realized the quaternions as living within that algebra. Subsequently Rudolf Lipschitz in 1886 generalized Clifford's interpretation of the quaternions and applied them to the geometry of rotations in "n" dimensions. Later these developments would lead other 20th-century mathematicians to formalize and explore the properties of the Clifford algebra.
Nevertheless, another revolutionary development of the 19th-century would completely overshadow the geometric algebras: that of vector analysis, developed independently by Josiah Willard Gibbs and Oliver Heaviside. Vector analysis was motivated by James Clerk Maxwell's studies of electromagnetism, and specifically the need to express and manipulate conveniently certain differential equations. Vector analysis had a certain intuitive appeal compared to the rigors of the new algebras. Physicists and mathematicians alike readily adopted it as their geometrical toolkit of choice. Progress on the study of Clifford algebras quietly advanced through the twentieth century, although largely due to the work of abstract algebraists such as Hermann Weyl and Claude Chevalley.
The "geometrical" approach to geometric algebras has seen a number of 20th-century revivals. In mathematics, Emil Artin's "Geometric Algebra" discusses the algebra associated with each of a number of geometries, including affine geometry, projective geometry, symplectic geometry, and orthogonal geometry. In physics, geometric algebras have been revived as a "new" way to do classical mechanics and electromagnetism. [Hestenes, "et al" (1984).] David Hestenes reinterpreted the Pauli and Dirac matrices as vectors in ordinary space and spacetime, respectively. In computer graphics, geometric algebras have been revived in order to represent efficiently rotations (and other transformations) on computer hardware [Dorst, "et al" (2007).] .
ee also
* Clifford algebra
* Algebra of physical space
* Spinor
* Quaternion
* Algebraic geometry
Notes
References
*
*
* Leo Dorst, Daniel Fontijne, Stephen Mann, " [http://www.geometricalgebra.net/ Geometric Algebra for Computer Science: An Object-Oriented Approach to Geometry] " (The Morgan Kaufmann Series in Computer Graphics), Morgan Kaufmann (April 19, 2007), ISBN-10: 0123694655, ISBN-13: 978-0123694652.
* (The Linear Extension Theory - A new Branch of Mathematics)
*
* David Hestenes and Sobczyk, G., 1984. "Clifford Algebra to Geometric Calculus", Springer Verlag ISBN 90-277-1673-0
Further reading
* Baylis, W. E., ed., 1996. "Clifford (Geometric) Algebra with Applications to Physics, Mathematics, and Engineering". Boston: Birkhäuser.
* Baylis, W. E., 2002. "Electrodynamics: A Modern Geometric Approach", 2nd ed. Birkhäuser. ISBN 0-8176-4025-8
* Nicolas Bourbaki, 1980. "Eléments de Mathématique. Algèbre". Chpt. 9, "Algèbres de Clifford". Paris: Hermann.
* Hestenes, D., 1999. "New Foundations for Classical Mechanics", 2nd ed. Springer Verlag ISBN 0-7923-5302-1
* Lasenby, J., Lasenby, A. N., and Doran, C. J. L., 2000, " [http://www.mrao.cam.ac.uk/%7Eclifford/publications/ps/dll_millen.pdf A Unified Mathematical Language for Physics and Engineering in the 21st Century] ," "Philosophical Transactions of the Royal Society of London A 358": 1-18.
*cite book |author=Chris Doran & Anthony Lasenby |title=Geometric algebra for physicists |year=2003 |publisher=Cambridge University Press |isbn=978-0-521-71595-9
External links
Research groups
* [http://sinai.mech.fukui-u.ac.jp/gcj/gc_int.html Geometric Calculus International] . Links to Research groups, Software, and Conferences, worldwide.
* [http://www.mrao.cam.ac.uk/~clifford/ Cambridge Geometric Algebra group] . Full-text online publications, and other material.
* [http://www.science.uva.nl/ga/ University of Amsterdam group]
* [http://modelingnts.la.asu.edu/GC_R&D.html Geometric Calculus research & development] (Arizona State University).
* [http://gaupdate.wordpress.com/ GA-Net blog] and [http://sinai.mech.fukui-u.ac.jp/GA-Net/archive/index.html newsletter archive] . Geometric Algebra/Clifford Algebra development news.
Online reading
* [http://www.mrao.cam.ac.uk/~clifford/introduction/intro/intro.html Imaginary Numbers are not Real - the Geometric Algebra of Spacetime] . Introduction (Cambridge GA group).
* [http://www.mrao.cam.ac.uk/~clifford/ptIIIcourse/ Physical Applications of Geometric Algebra] . Final-year undergraduate course (Cambridge GA group; see also [http://www.mrao.cam.ac.uk/~clifford/ptIIIcourse/course99/ 1999 version] ).
* [http://www.iancgbell.clara.net/maths/ Maths for (Games) Programmers: 5 - Multivector methods] . Comprehensive introduction and reference for programmers, from Ian Bell.
* [http://web.archive.org/web/20040610223908/home.student.utwente.nl/j.suter/ga_primer.pdf A Geometric Algebra Primer] , especially for computer scientists.
*
* [http://www.bayarea.net/~kins/thomas_briggs/ Exploring Hyperspace with the Geometric Product] - Highlights applications to higher dimensions and cosmology that includes wormholes.