text stringlengths 13 991 |
|---|
Tensors were conceived in 1900 by Tullio Levi-Civita and Gregorio Ricci-Curbastro, who continued the earlier work of Bernhard Riemann and Elwin Bruno Christoffel and others, as part of the "absolute differential calculus". The concept enabled an alternative formulation of the intrinsic differential geometry of a manifold in the form of the Riemann curvature tensor. |
Although seemingly different, the various approaches to defining tensors describe the same geometric concept using different language and at different levels of abstraction. For example, tensors are defined and discussed for statistical and machine learning applications. |
Just as the components of a vector change when we change the basis of the vector space, the components of a tensor also change under such a transformation. Each type of tensor comes equipped with a "transformation law" that details how the components of the tensor respond to a change of basis. The components of a vector can respond in two distinct ways to a change of basis (see covariance and contravariance of vectors), where the new basis vectors formula_1 are expressed in terms of the old basis vectors formula_2 as, |
Here "R"" j""i" are the entries of the change of basis matrix, and in the rightmost expression the summation sign was suppressed: this is the Einstein summation convention, which will be used throughout this article. The components "v""i" of a column vector v transform with the inverse of the matrix "R", |
where the hat denotes the components in the new basis. This is called a "contravariant" transformation law, because the vector components transform by the "inverse" of the change of basis. In contrast, the components, "w""i", of a covector (or row vector), w transform with the matrix "R" itself, |
This is called a "covariant" transformation law, because the covector components transforms by the "same matrix" as the change of basis matrix. The components of a more general tensor transform by some combination of covariant and contravariant transformations, with one transformation law for each index. If the transformation matrix of an index is the inverse matrix of the basis transformation, then the index is called "contravariant" and is conventionally denoted with an upper index (superscript). If the transformation matrix of an index is the basis transformation itself, then the index is called "covariant" and is denoted with a lower index (subscript). |
As a simple example, the matrix of a linear operator with respect to a basis is a rectangular array formula_6 that transforms under a change of basis matrix formula_7 by formula_8. For the individual matrix entries, this transformation law has the form formula_9 so the tensor corresponding to the matrix of a linear operator has one covariant and one contravariant index: it is of type (1,1). |
Combinations of covariant and contravariant components with the same index allow us to express geometric invariants. For example, the fact that a vector is the same object in different coordinate systems can be captured by the following equations, using the formulas defined above: |
where formula_11 is the Kronecker delta, which functions similarly to the identity matrix, and has the effect of renaming indices ("j" into "k" in this example). This shows several features of the component notation: the ability to re-arrange terms at will (commutativity), the need to use different indices when working with multiple objects in the same expression, the ability to rename indices, and the manner in which contravariant and covariant tensors combine so that all instances of the transformation matrix and its inverse cancel, so that expressions like formula_12 can immediately be seen to be geometrically identical in all coordinate systems. |
Similarly, a linear operator, viewed as a geometric object, does not actually depend on a basis: it is just a linear map that accepts a vector as an argument and produces another vector. The transformation law for how the matrix of components of a linear operator changes with the basis is consistent with the transformation law for a contravariant vector, so that the action of a linear operator on a contravariant vector is represented in coordinates as the matrix product of their respective coordinate representations. That is, the components formula_13 are given by formula_14. These components transform contravariantly, since |
The transformation law for an order tensor with "p" contravariant indices and "q" covariant indices is thus given as, |
Here the primed indices denote components in the new coordinates, and the unprimed indices denote the components in the old coordinates. Such a tensor is said to be of order or "type" . The terms "order", "type", "rank", "valence", and "degree" are all sometimes used for the same concept. Here, the term "order" or "total order" will be used for the total dimension of the array (or its generalization in other definitions), in the preceding example, and the term "type" for the pair giving the number of contravariant and covariant indices. A tensor of type is also called a -tensor for short. |
This discussion motivates the following formal definition: |
to each basis of an "n"-dimensional vector space such that, if we apply the change of basis |
then the multidimensional array obeys the transformation law |
The definition of a tensor as a multidimensional array satisfying a transformation law traces back to the work of Ricci. |
An equivalent definition of a tensor uses the representations of the general linear group. There is an action of the general linear group on the set of all ordered bases of an "n"-dimensional vector space. If formula_23 is an ordered basis, and formula_24 is an invertible formula_25 matrix, then the action is given by |
Let "F" be the set of all ordered bases. Then "F" is a principal homogeneous space for GL("n"). Let "W" be a vector space and let formula_27 be a representation of GL("n") on "W" (that is, a group homomorphism formula_28). Then a tensor of type formula_27 is an equivariant map formula_30. Equivariance here means that |
When formula_27 is a tensor representation of the general linear group, this gives the usual definition of tensors as multidimensional arrays. This definition is often used to describe tensors on manifolds, and readily generalizes to other groups. |
A downside to the definition of a tensor using the multidimensional array approach is that it is not apparent from the definition that the defined object is indeed basis independent, as is expected from an intrinsically geometric object. Although it is possible to show that transformation laws indeed ensure independence from the basis, sometimes a more intrinsic definition is preferred. One approach that is common in differential geometry is to define tensors relative to a fixed (finite-dimensional) vector space "V", which is usually taken to be a particular vector space of some geometrical significance like the tangent space to a manifold. In this approach, a type tensor "T" is defined as a multilinear map, |
where "V"∗ is the corresponding dual space of covectors, which is linear in each of its arguments. The above assumes "V" is a vector space over the real numbers, ℝ. More generally, "V" can be taken over an arbitrary field of numbers, "F" (e.g. the complex numbers) with a one-dimensional vector space over "F" replacing ℝ as the codomain of the multilinear maps. |
By applying a multilinear map "T" of type to a basis {e"j"} for "V" and a canonical cobasis {ε"i"} for "V"∗, |
a -dimensional array of components can be obtained. A different choice of basis will yield different components. But, because "T" is linear in all of its arguments, the components satisfy the tensor transformation law used in the multilinear array definition. The multidimensional array of components of "T" thus form a tensor according to that definition. Moreover, such an array can be realized as the components of some multilinear map "T". This motivates viewing multilinear maps as the intrinsic objects underlying tensors. |
In viewing a tensor as a multilinear map, it is conventional to identify the double dual "V"∗∗ of the vector space "V", i.e., the space of linear functionals on the dual vector space "V"∗, with the vector space "V". There is always a natural linear map from "V" to its double dual, given by evaluating a linear form in "V"∗ against a vector in "V". This linear mapping is an isomorphism in finite dimensions, and it is often then expedient to identify "V" with its double dual. |
For some mathematical applications, a more abstract approach is sometimes useful. This can be achieved by defining tensors in terms of elements of tensor products of vector spaces, which in turn are defined through a universal property. A type tensor is defined in this context as an element of the tensor product of vector spaces, |
A basis of and basis of naturally induce a basis of the tensor product . The components of a tensor are the coefficients of the tensor with respect to the basis obtained from a basis for and its dual basis , i.e. |
Using the properties of the tensor product, it can be shown that these components satisfy the transformation law for a type tensor. Moreover, the universal property of the tensor product gives a -to- correspondence between tensors defined in this way and tensors defined as multilinear maps. |
Tensor products can be defined in great generality – for example, involving arbitrary modules over a ring. In principle, one could define a "tensor" simply to be an element of any tensor product. However, the mathematics literature usually reserves the term "tensor" for an element of a tensor product of any number of copies of a single vector space and its dual, as above. |
In many applications, especially in differential geometry and physics, it is natural to consider a tensor with components that are functions of the point in a space. This was the setting of Ricci's original work. In modern mathematical terminology such an object is called a tensor field, often referred to simply as a tensor. |
In this context, a coordinate basis is often chosen for the tangent vector space. The transformation law may then be expressed in terms of partial derivatives of the coordinate functions, |
This table shows important examples of tensors on vector spaces and tensor fields on manifolds. The tensors are classified according to their type , where "n" is the number of contravariant indices, "m" is the number of covariant indices, and gives the total order of the tensor. For example, a bilinear form is the same thing as a -tensor; an inner product is an example of a -tensor, but not all -tensors are inner products. In the -entry of the table, "M" denotes the dimensionality of the underlying vector space or manifold because for each dimension of the space, a separate index is needed to select that dimension to get a maximally covariant antisymmetric tensor. |
Raising an index on an -tensor produces an -tensor; this corresponds to moving diagonally down and to the left on the table. Symmetrically, lowering an index corresponds to moving diagonally up and to the right on the table. Contraction of an upper with a lower index of an -tensor produces an -tensor; this corresponds to moving diagonally up and to the left on the table. |
Assuming a basis of a real vector space, e.g., a coordinate frame in the ambient space, a tensor can be represented as an organized multidimensional array of numerical values with respect to this specific basis. Changing the basis transforms the values in the array in a characteristic way that allows to "define" tensors as objects adhering to this transformational behavior. For example, there are invariants of tensors that must be preserved under any change of the basis, thereby making only certain multidimensional arrays of numbers a tensor. Compare this to the array representing formula_40 not being a tensor, for the sign change under transformations changing the orientation. |
Because the components of vectors and their duals transform differently under the change of their dual bases, there is a covariant and/or contravariant transformation law that relates the arrays, which represent the tensor with respect to one basis and that with respect to the other one. The numbers of, respectively, (contravariant indices) and dual (covariant indices) in the input and output of a tensor determine the "type" (or "valence") of the tensor, a pair of natural numbers , which determine the precise form of the transformation law. The "" of a tensor is the sum of these two numbers. |
The collection of tensors on a vector space and its dual forms a tensor algebra, which allows products of arbitrary tensors. Simple applications of tensors of order , which can be represented as a square matrix, can be solved by clever arrangement of transposed vectors and by applying the rules of matrix multiplication, but the tensor product should not be confused with this. |
There are several notational systems that are used to describe tensors and perform calculations involving them. |
Ricci calculus is the modern formalism and notation for tensor indices: indicating inner and outer products, covariance and contravariance, summations of tensor components, symmetry and antisymmetry, and partial and covariant derivatives. |
The Einstein summation convention dispenses with writing summation signs, leaving the summation implicit. Any repeated index symbol is summed over: if the index is used twice in a given term of a tensor expression, it means that the term is to be summed for all . Several distinct pairs of indices may be summed this way. |
Penrose graphical notation is a diagrammatic notation which replaces the symbols for tensors with shapes, and their indices by lines and curves. It is independent of basis elements, and requires no symbols for the indices. |
The abstract index notation is a way to write tensors such that the indices are no longer thought of as numerical, but rather are indeterminates. This notation captures the expressiveness of indices and the basis-independence of index-free notation. |
A component-free treatment of tensors uses notation that emphasises that tensors do not rely on any basis, and is defined in terms of the tensor product of vector spaces. |
There are several operations on tensors that again produce a tensor. The linear nature of tensor implies that two tensors of the same type may be added together, and that tensors may be multiplied by a scalar with results analogous to the scaling of a vector. On components, these operations are simply performed component-wise. These operations do not change the type of the tensor; but there are also operations that produce a tensor of different type. |
The tensor product takes two tensors, "S" and "T", and produces a new tensor, , whose order is the sum of the orders of the original tensors. When described as multilinear maps, the tensor product simply multiplies the two tensors, i.e. |
which again produces a map that is linear in all its arguments. On components, the effect is to multiply the components of the two input tensors pairwise, i.e. |
If is of type and is of type , then the tensor product has type . |
Tensor contraction is an operation that reduces a type tensor to a type tensor, of which the trace is a special case. It thereby reduces the total order of a tensor by two. The operation is achieved by summing components for which one specified contravariant index is the same as one specified covariant index to produce a new component. Components for which those two indices are different are discarded. For example, a -tensor formula_43 can be contracted to a scalar through formula_44. Where the summation is again implied. When the -tensor is interpreted as a linear map, this operation is known as the trace. |
The contraction is often used in conjunction with the tensor product to contract an index from each tensor. |
The contraction can also be understood using the definition of a tensor as an element of a tensor product of copies of the space "V" with the space "V"∗ by first decomposing the tensor into a linear combination of simple tensors, and then applying a factor from "V"∗ to a factor from "V". For example, a tensor formula_45 can be written as a linear combination |
The contraction of "T" on the first and last slots is then the vector |
In a vector space with an inner product (also known as a metric) "g", the term contraction is used for removing two contravariant or two covariant indices by forming a trace with the metric tensor or its inverse. For example, a -tensor formula_48 can be contracted to a scalar through formula_49 (yet again assuming the summation convention). |
When a vector space is equipped with a nondegenerate bilinear form (or "metric tensor" as it is often called in this context), operations can be defined that convert a contravariant (upper) index into a covariant (lower) index and vice versa. A metric tensor is a (symmetric) (-tensor; it is thus possible to contract an upper index of a tensor with one of the lower indices of the metric tensor in the product. This produces a new tensor with the same index structure as the previous tensor, but with lower index generally shown in the same position of the contracted upper index. This operation is quite graphically known as "lowering an index". |
Conversely, the inverse operation can be defined, and is called "raising an index". This is equivalent to a similar contraction on the product with a -tensor. This "inverse metric tensor" has components that are the matrix inverse of those of the metric tensor. |
If a particular surface element inside the material is singled out, the material on one side of the surface will apply a force on the other side. In general, this force will not be orthogonal to the surface, but it will depend on the orientation of the surface in a linear manner. This is described by a tensor of type, in linear elasticity, or more precisely by a tensor field of type , since the stresses may vary from point to point. |
Applications of tensors of order > 2. |
The concept of a tensor of order two is often conflated with that of a matrix. Tensors of higher order do however capture ideas important in science and engineering, as has been shown successively in numerous areas as they develop. This happens, for instance, in the field of computer vision, with the trifocal tensor generalizing the fundamental matrix. |
The field of nonlinear optics studies the changes to material polarization density under extreme electric fields. The polarization waves generated are related to the generating electric fields through the nonlinear susceptibility tensor. If the polarization P is not linearly proportional to the electric field E, the medium is termed "nonlinear". To a good approximation (for sufficiently weak fields, assuming no permanent dipole moments are present), P is given by a Taylor series in E whose coefficients are the nonlinear susceptibilities: |
Here formula_51 is the linear susceptibility, formula_52 gives the Pockels effect and second harmonic generation, and formula_53 gives the Kerr effect. This expansion shows the way higher-order tensors arise naturally in the subject matter. |
The vector spaces of a tensor product need not be the same, and sometimes the elements of such a more general tensor product are called "tensors". For example, an element of the tensor product space is a second-order "tensor" in this more general sense, and an order- tensor may likewise be defined as an element of a tensor product of different vector spaces. A type tensor, in the sense defined previously, is also a tensor of order in this more general sense. The concept of tensor product can be extended to arbitrary modules over a ring. |
The notion of a tensor can be generalized in a variety of ways to infinite dimensions. One, for instance, is via the tensor product of Hilbert spaces. Another way of generalizing the idea of tensor, common in nonlinear analysis, is via the multilinear maps definition where instead of using finite-dimensional vector spaces and their algebraic duals, one uses infinite-dimensional Banach spaces and their continuous dual. Tensors thus live naturally on Banach manifolds and Fréchet manifolds. |
Suppose that a homogeneous medium fills , so that the density of the medium is described by a single scalar value in . The mass, in kg, of a region is obtained by multiplying by the volume of the region , or equivalently integrating the constant over the region: |
where the Cartesian coordinates are measured in m. If the units of length are changed into cm, then the numerical values of the coordinate functions must be rescaled by a factor of 100: |
The numerical value of the density must then also transform by formula_56 to compensate, so that the numerical value of the mass in kg is still given by integral of formula_57. Thus formula_58 (in units of ). |
More generally, if the Cartesian coordinates undergo a linear transformation, then the numerical value of the density must change by a factor of the reciprocal of the absolute value of the determinant of the coordinate transformation, so that the integral remains invariant, by the change of variables formula for integration. Such a quantity that scales by the reciprocal of the absolute value of the determinant of the coordinate transition map is called a scalar density. To model a non-constant density, is a function of the variables (a scalar field), and under a curvilinear change of coordinates, it transforms by the reciprocal of the Jacobian of the coordinate change. For more on the intrinsic meaning, see Density on a manifold. |
A tensor density transforms like a tensor under a coordinate change, except that it in addition picks up a factor of the absolute value of the determinant of the coordinate transition: |
Here "w" is called the weight. In general, any tensor multiplied by a power of this function or its absolute value is called a tensor density, or a weighted tensor. An example of a tensor density is the current density of electromagnetism. |
Under an affine transformation of the coordinates, a tensor transforms by the linear part of the transformation itself (or its inverse) on each index. These come from the rational representations of the general linear group. But this is not quite the most general linear transformation law that such an object may have: tensor densities are non-rational, but are still semisimple representations. A further class of transformations come from the logarithmic representation of the general linear group, a reducible but not semisimple representation, consisting of an with the transformation law |
The transformation law for a tensor behaves as a functor on the category of admissible coordinate systems, under general linear transformations (or, other transformations within some class, such as local diffeomorphisms.) This makes a tensor a special case of a geometrical object, in the technical sense that it is a function of the coordinate system transforming functorially under coordinate changes. Examples of objects obeying more general kinds of transformation laws are jets and, more generally still, natural bundles. |
Succinctly, spinors are elements of the spin representation of the rotation group, while tensors are elements of its tensor representations. Other classical groups have tensor representations, and so also tensors that are compatible with the group, but all non-compact classical groups have infinite-dimensional unitary representations as well. |
The concepts of later tensor analysis arose from the work of Carl Friedrich Gauss in differential geometry, and the formulation was much influenced by the theory of algebraic forms and invariants developed during the middle of the nineteenth century. The word "tensor" itself was introduced in 1846 by William Rowan Hamilton to describe something different from what is now meant by a tensor. The contemporary usage was introduced by Woldemar Voigt in 1898. |
Tensor calculus was developed around 1890 by Gregorio Ricci-Curbastro under the title "absolute differential calculus", and originally presented by Ricci-Curbastro in 1892. It was made accessible to many mathematicians by the publication of Ricci-Curbastro and Tullio Levi-Civita's 1900 classic text "Méthodes de calcul différentiel absolu et leurs applications" (Methods of absolute differential calculus and their applications). |
In the 20th century, the subject came to be known as "tensor analysis", and achieved broader acceptance with the introduction of Einstein's theory of general relativity, around 1915. General relativity is formulated completely in the language of tensors. Einstein had learned about them, with great difficulty, from the geometer Marcel Grossmann. Levi-Civita then initiated a correspondence with Einstein to correct mistakes Einstein had made in his use of tensor analysis. The correspondence lasted 1915–17, and was characterized by mutual respect: |
Tensors were also found to be useful in other fields such as continuum mechanics. Some well-known examples of tensors in differential geometry are quadratic forms such as metric tensors, and the Riemann curvature tensor. The exterior algebra of Hermann Grassmann, from the middle of the nineteenth century, is itself a tensor theory, and highly geometric, but it was some time before it was seen, with the theory of differential forms, as naturally unified with tensor calculus. The work of Élie Cartan made differential forms one of the basic kinds of tensors used in mathematics. |
From about the 1920s onwards, it was realised that tensors play a basic role in algebraic topology (for example in the Künneth theorem). Correspondingly there are types of tensors at work in many branches of abstract algebra, particularly in homological algebra and representation theory. Multilinear algebra can be developed in greater generality than for scalars coming from a field. For example, scalars can come from a ring. But the theory is then less geometric and computations more technical and less algorithmic. Tensors are generalized within category theory by means of the concept of monoidal category, from the 1960s. |
Frame-dragging is an effect on spacetime, predicted by Albert Einstein's general theory of relativity, that is due to non-static stationary distributions of mass–energy. A stationary field is one that is in a steady state, but the masses causing that field may be non-static — rotating, for instance. More generally, the subject that deals with the effects caused by mass–energy currents is known as gravitoelectromagnetism, which is analogous to the magnetism of classical electromagnetism. |
The first frame-dragging effect was derived in 1918, in the framework of general relativity, by the Austrian physicists Josef Lense and Hans Thirring, and is also known as the Lense–Thirring effect. They predicted that the rotation of a massive object would distort the spacetime metric, making the orbit of a nearby test particle precess. This does not happen in Newtonian mechanics for which the gravitational field of a body depends only on its mass, not on its rotation. The Lense–Thirring effect is very small – about one part in a few trillion. To detect it, it is necessary to examine a very massive object, or build an instrument that is very sensitive. |
In 2015, new general-relativistic extensions of Newtonian rotation laws were formulated to describe geometric dragging of frames which incorporates a newly discovered antidragging effect. |
Rotational frame-dragging (the Lense–Thirring effect) appears in the general principle of relativity and similar theories in the vicinity of rotating massive objects. Under the Lense–Thirring effect, the frame of reference in which a clock ticks the fastest is one which is revolving around the object as viewed by a distant observer. This also means that light traveling in the direction of rotation of the object will move past the massive object faster than light moving against the rotation, as seen by a distant observer. It is now the best known frame-dragging effect, partly thanks to the Gravity Probe B experiment. Qualitatively, frame-dragging can be viewed as the gravitational analog of electromagnetic induction. |
Another interesting consequence is that, for an object constrained in an equatorial orbit, but not in freefall, it weighs more if orbiting anti-spinward, and less if orbiting spinward. For example, in a suspended equatorial bowling alley, a bowling ball rolled anti-spinward would weigh more than the same ball rolled in a spinward direction. Note, frame dragging will neither accelerate nor slow down the bowling ball in either direction. It is not a "viscosity". Similarly, a stationary plumb-bob suspended over the rotating object will not list. It will hang vertically. If it starts to fall, induction will push it in the spinward direction. |
Linear frame dragging is the similarly inevitable result of the general principle of relativity, applied to linear momentum. Although it arguably has equal theoretical legitimacy to the "rotational" effect, the difficulty of obtaining an experimental verification of the effect means that it receives much less discussion and is often omitted from articles on frame-dragging (but see Einstein, 1921). |
Static mass increase is a third effect noted by Einstein in the same paper. The effect is an increase in inertia of a body when other masses are placed nearby. While not strictly a frame dragging effect (the term frame dragging is not used by Einstein), it is demonstrated by Einstein that it derives from the same equation of general relativity. It is also a tiny effect that is difficult to confirm experimentally. |
In 1976 Van Patten and Everitt proposed to implement a dedicated mission aimed to measure the Lense–Thirring node precession of a pair of counter-orbiting spacecraft to be placed in terrestrial polar orbits with drag-free apparatus. A somewhat equivalent, cheaper version of such an idea was put forth in 1986 by Ciufolini who proposed to launch a passive, geodetic satellite in an orbit identical to that of the LAGEOS satellite, launched in 1976, apart from the orbital planes which should have been displaced by 180 deg apart: the so-called butterfly configuration. The measurable quantity was, in this case, the sum of the nodes of LAGEOS and of the new spacecraft, later named LAGEOS III, LARES, WEBER-SAT. |
The Gravity Probe B experiment was a satellite-based mission by a Stanford group and NASA, used to experimentally measure another gravitomagnetic effect, the Schiff precession of a gyroscope, to an expected 1% accuracy or better. Unfortunately such accuracy was not achieved. The first preliminary results released in April 2007 pointed towards an accuracy of 256–128%, with the hope of reaching about 13% in December 2007. |
In 2008 the Senior Review Report of the NASA Astrophysics Division Operating Missions stated that it was unlikely that Gravity Probe B team will be able to reduce the errors to the level necessary to produce a convincing test of currently untested aspects of General Relativity (including frame-dragging). |
On May 4, 2011, the Stanford-based analysis group and NASA announced the final report, and in it the data from GP-B demonstrated the frame-dragging effect with an error of about 19 percent, and Einstein's predicted value was at the center of the confidence interval. |
NASA published claims of success in verification of frame dragging for the GRACE twin satellites and Gravity Probe B, both of which claims are still in public view. A research group in Italy, USA, and UK also claimed success in verification of frame dragging with the Grace gravity model, published in a peer reviewed journal. All the claims include recommendations for further research at greater accuracy and other gravity models. |
In the case of stars orbiting close to a spinning, supermassive black hole, frame dragging should cause the star's orbital plane to precess about the black hole spin axis. This effect should be detectable within the next few years via astrometric monitoring of stars at the center of the Milky Way galaxy. |
By comparing the rate of orbital precession of two stars on different orbits, it is possible in principle to test the no-hair theorems of general relativity, in addition to measuring the spin of the black hole. |
Relativistic jets may provide evidence for the reality of frame-dragging. Gravitomagnetic forces produced by the Lense–Thirring effect (frame dragging) within the ergosphere of rotating black holes combined with the energy extraction mechanism by Penrose have been used to explain the observed properties of relativistic jets. The gravitomagnetic model developed by Reva Kay Williams predicts the observed high energy particles (~GeV) emitted by quasars and active galactic nuclei; the extraction of X-rays, γ-rays, and relativistic e−– e+ pairs; the collimated jets about the polar axis; and the asymmetrical formation of jets (relative to the orbital plane). |
The Lense–Thirring effect has been observed in a binary system that consists of a massive white dwarf and a pulsar. |
Frame-dragging may be illustrated most readily using the Kerr metric, which describes the geometry of spacetime in the vicinity of a mass "M" rotating with angular momentum "J", and Boyer–Lindquist coordinates (see the link for the transformation): |
and where the following shorthand variables have been introduced for brevity |
In the non-relativistic limit where "M" (or, equivalently, "r""s") goes to zero, the Kerr metric becomes the orthogonal metric for the oblate spheroidal coordinates |
We may rewrite the Kerr metric in the following form |
This metric is equivalent to a co-rotating reference frame that is rotating with angular speed Ω that depends on both the radius "r" and the colatitude "θ" |
In the plane of the equator this simplifies to: |
Thus, an inertial reference frame is entrained by the rotating central mass to participate in the latter's rotation; this is frame-dragging. |
An extreme version of frame dragging occurs within the ergosphere of a rotating black hole. The Kerr metric has two surfaces on which it appears to be singular. The inner surface corresponds to a spherical event horizon similar to that observed in the Schwarzschild metric; this occurs at |
where the purely radial component "grr" of the metric goes to infinity. The outer surface can be approximated by an oblate spheroid with lower spin parameters, and resembles a pumpkin-shape with higher spin parameters. It touches the inner surface at the poles of the rotation axis, where the colatitude "θ" equals 0 or π; its radius in Boyer-Lindquist coordinates is defined by the formula |
where the purely temporal component "gtt" of the metric changes sign from positive to negative. The space between these two surfaces is called the ergosphere. A moving particle experiences a positive proper time along its worldline, its path through spacetime. However, this is impossible within the ergosphere, where "gtt" is negative, unless the particle is co-rotating with the interior mass "M" with an angular speed at least of Ω. However, as seen above, frame-dragging occurs about every rotating mass and at every radius "r" and colatitude "θ", not only within the ergosphere. |
The Lense–Thirring effect inside a rotating shell was taken by Albert Einstein as not just support for, but a vindication of Mach's principle, in a letter he wrote to Ernst Mach in 1913 (five years before Lense and Thirring's work, and two years before he had attained the final form of general relativity). A reproduction of the letter can be found in Misner, Thorne, Wheeler. The general effect scaled up to cosmological distances, is still used as a support for Mach's principle. |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.