Revision as of 13:56, 3 September 2012 editSlawekb (talk | contribs)Extended confirmed users11,467 editsmNo edit summary← Previous edit | Latest revision as of 01:04, 1 December 2024 edit undoBD2412 (talk | contribs)Autopatrolled, IP block exemptions, Administrators2,457,142 editsm clean up spacing around commas and other punctuation, replaced: ,X → , X (2), ,Y → , Y, ,j → , j, ,n → , n, ,v → , v (2), , → , (3), ; → ;Tag: AWB | ||
(323 intermediate revisions by more than 100 users not shown) | |||
Line 1: | Line 1: | ||
{{Short description|Vector behavior under coordinate changes}} | |||
:''For other uses of "covariant" or "contravariant", see ].'' | |||
{{About||use of "covariance" in the context of special relativity|Lorentz covariance|other uses of "covariant" or "contravariant"|Covariance and contravariance (disambiguation)}} | |||
[[File:Vector 1-form.svg| |
[[File:Vector 1-form.svg|upright=1.35|thumb| A {{color box|#CC0000}} vector, '''v''', represented in terms of | ||
{{glossary}} | |||
{{term|tangent basis}} | |||
{{defn|{{color box|orange}} '''e'''{{sub|1}}, '''e'''{{sub|2}}, '''e'''{{sub|3}} to the {{color box|black}} coordinate curves (''left''),}} | |||
{{term|dual basis, covector basis, or reciprocal basis}} | |||
{{defn|{{color box|blue}} '''e'''{{sup|1}}, '''e'''{{sup|2}}, '''e'''{{sup|3}} to {{color box|#4B545B}} coordinate surfaces (''right''),}} | |||
{{glossary end}} | |||
in ] general '']'' {{nowrap|(''q''{{sup|1}}, ''q''{{sup|2}}, ''q''{{sup|3}})}}, a ] of numbers to define a point in a ]. Note the basis and cobasis coincide only when the basis is ].<ref> | |||
{{cite book | |||
| title=Gravitation |title-link=Gravitation (book) | |||
| first3=J.A. |last3=Wheeler | |||
| first1=C. |last1=Misner | |||
| first2=K.S. |last2=Thorne | |||
| publisher=W.H. Freeman | |||
| year=1973 | |||
| isbn=0-7167-0344-0 | |||
}}</ref>{{specify}} | |||
]] | |||
In ], especially in ] and ], '''covariance''' and '''contravariance''' describe how the quantitative description of certain geometric or physical entities changes with a ].<ref>{{Cite book |last=Frankel |first=Theodore |url=https://www.worldcat.org/oclc/739094283 |title=The geometry of physics : an introduction |date=2012 |publisher=Cambridge University Press |isbn=978-1-107-60260-1 |page=42 |location=Cambridge |oclc=739094283}}</ref> Briefly, a contravariant vector is a list of numbers that transforms oppositely to a change of basis, and a covariant vector is a list of numbers that transforms in the same way. Contravariant vectors are often just called ''vectors'' and covariant vectors are called ''covectors'' or ''dual vectors''. The terms ''covariant'' and ''contravariant'' were introduced by ] in 1851.<ref>{{cite book |first=J.J. |last=Sylvester |chapter=On the general theory of associated algebraical forms |chapter-url=https://books.google.com/books?id=mZ9EAAAAcAAJ&pg=PA289 |editor= |title=Cambridge and Dublin Mathematical Journal |publisher= |location= |date=1851 |isbn= |volume=6 |pages=289–293 |url=}}</ref><ref>{{Cite book|title=The collected mathematical papers of James Joseph Sylvester |volume=3, 1870–1883|last=Sylvester|first=J.J. University Press|date=16 February 2012 |publisher=Cambridge University Press |isbn=978-1107661431|oclc=758983870}}</ref> | |||
<p>'''tangent basis vectors''' (left: '''e'''<sub>1</sub>, '''e'''<sub>2</sub>, '''e'''<sub>3</sub>) to the coordinate curves (yellow),</p> | |||
<p>'''dual basis, covector basis, or cobasis''' (right: '''e'''<sup>1</sup>, '''e'''<sup>2</sup>, '''e'''<sup>3</sup>), normal vectors to coordinate surfaces (green),</p> | |||
]s, such as ] or ], are often used in physical and geometric problems. Associated with any coordinate system is a natural choice of coordinate basis for vectors based at each point of the space, and covariance and contravariance are particularly important for understanding how the coordinate description of a vector changes by passing from one coordinate system to another. ]s are objects in ] that can have aspects of both covariance and contravariance. | |||
in ] general ''']''' (''q''<sup>1</sup>, ''q''<sup>2</sup>, ''q''<sup>3</sup>), a ] of numbers to define point in a ]. Note the basis and cobasis do not coincide unless the basis is ].<ref>{{cite book|title=Gravitation|author=J.A. Wheeler, C. Misner, K.S. Thorne|publisher=W.H. Freeman & Co|year=1973|isbn=0-7167-0344-0}}</ref> ]] | |||
In ] and ], '''covariance''' and '''contravariance''' describe how the quantitative description of certain geometric or physical entities changes with a ]. For ], this is determined by a change from one ] to another. When an ] is ] into another orthogonal basis, this distinction is invisible. However, when considering more general coordinate systems such as ], ], and coordinate systems on ]s, the distinction is significant. | |||
==Introduction== | |||
* For a ] (such as a ] or ] vector) to be ], the components of the vector must ''contra-vary'' with a change of basis to compensate. That is, the components must vary with the inverse transformation to that of the change of basis. The components of vectors (as opposed to of dual vectors) are said to be '''contravariant'''. Examples of vectors with ''contravariant components'' include the position of an object relative to an observer, or any ] of position with respect to time, including ], ], and ]. In ], contravariant components are denoted with ''upper indices'' as in | |||
In physics, a vector typically arises as the outcome of a measurement or series of measurements, and is represented as a list (or ]) of numbers such as | |||
::<math>\mathbf{v} = v^i \mathbf{e}_i .</math> | |||
:<math>(v_1,v_2,v_3) .</math> | |||
* For a ] (also called a ''covector'') to be basis-independent, the components of the dual vector must ''co-vary'' with a change of basis to remain representing the same covector. That is, the components must vary by the same transformation as the change of basis. The components of dual vectors (as opposed to of vectors) are said to be covariant. Examples of ''covariant'' vectors generally appear when taking a ] of a function. In ], covariant components are denoted with ''lower indices'' as in | |||
::<math>\mathbf{v} = v_i \mathbf{e}^i .</math> | |||
The numbers in the list depend on the choice of ]. For instance, if the vector represents position with respect to an observer (]), then the coordinate system may be obtained from a system of rigid rods, or reference axes, along which the components ''v''<sub>1</sub>, ''v''<sub>2</sub>, and ''v''<sub>3</sub> are measured. For a vector to represent a geometric object, it must be possible to describe how it looks in any other coordinate system. That is to say, the components of the vectors will ''transform'' in a certain way in passing from one coordinate system to another. | |||
In physics, vectors often have units of distance or distance times some other unit (such as the velocity), whereas covectors have units the inverse of distance or the inverse of distance times some other unit. The distinction between covariant and contravariant vectors is particularly important for computations with ]s, which can have mixed variance. This means that they have components that are both covariant and contravariant. The valence or ] of a tensor gives the number of covariant and contravariant component indices. | |||
A simple illustrative case is that of a ]. For a vector, once a set of basis vectors has been defined, then the components of that vector will always vary ''opposite'' to that of the basis vectors. That vector is therefore defined as a ''contravariant'' tensor. Take a standard position vector for example. By changing the scale of the reference axes from meters to centimeters (that is, ''dividing'' the scale of the reference axes by 100, so that the basis vectors now are <math>.01</math> meters long), the components of the measured ] ] are ''multiplied'' by 100. A vector's components change scale ''inversely'' to changes in scale to the reference axes, and consequently a vector is called a ''contravariant'' tensor. | |||
The terms covariant and contravariant were introduced by ] in ] in order to study algebraic ]. In this context, for instance, a system of ] is contravariant in the variables. The use of both terms in the modern context of ] is a specific example of ] in ]. | |||
A ''vector'', which is an example of a ''contravariant'' tensor, has components that transform inversely to the transformation of the reference axes, (with example transformations including ] and ]). ]; instead, the components of the vector change in a way that cancels the change in the spatial axes. In other words, if the reference axes were rotated in one direction, the component representation of the vector would rotate in exactly the opposite way. Similarly, if the reference axes were stretched in one direction, the components of the vector, would reduce in an exactly compensating way. Mathematically, if the coordinate system undergoes a transformation described by an <math>n\times n</math> ] ''M'', so that the basis vectors transform according to <math>\begin{bmatrix} \mathbf{e}_1^\prime\ \mathbf{e}_2^\prime\ ... \ \mathbf{e}_n^\prime\end{bmatrix}=\begin{bmatrix} \mathbf{e}_1\ \mathbf{e}_2\ ... \ \mathbf{e}_n\end{bmatrix}M</math>, then the components of a vector '''v''' in the original basis ( <math>v^i</math> ) must be similarly transformed via <math>\begin{bmatrix} v^1{^\prime} \\ v^2{^\prime} \\ ... \\ v^n{^\prime} \end{bmatrix}=M^{-1}\begin{bmatrix} v^1 \\ v^2 \\ ... \\ v^n \end{bmatrix}</math>. The components of a ''vector'' are often represented arranged in a column. | |||
==Introduction== | |||
In physics, a ] typically arises as the outcome of a measurement or series of measurements, and is represented as a list (or ]) of numbers such as | |||
By contrast, a ''covector'' has components that transform like the reference axes. It lives in the dual vector space, and represents a linear map from vectors to scalars. The dot product operator involving vectors is a good example of a covector. To illustrate, assume we have a covector defined as <math>\mathbf{v}\ \cdot</math> , where <math>\mathbf{v}</math> is a vector. The components of this covector in some arbitrary basis are <math>\begin{bmatrix} \mathbf{v}\cdot\mathbf{e}_1 & \mathbf{v}\cdot\mathbf{e}_2 & ... & \mathbf{v}\cdot\mathbf{e}_n \end{bmatrix}</math>, with <math>\begin{bmatrix} \mathbf{e}_1\ \mathbf{e}_2\ ... \ \mathbf{e}_n\end{bmatrix}</math> being the basis vectors in the corresponding vector space. (This can be derived by noting that we want to get the correct answer for the dot product operation when multiplying by an arbitrary vector <math>\mathbf{w}</math> , with components <math>\begin{bmatrix} w^1 \\ w^2 \\ ... \\ w^n \end{bmatrix}</math>). The covariance of these covector components is then seen by noting that if a transformation described by an <math>n\times n</math> ] ''M'' were to be applied to the basis vectors in the corresponding vector space, <math>\begin{bmatrix} \mathbf{e}_1^\prime\ \mathbf{e}_2^\prime\ ... \ \mathbf{e}_n^\prime\end{bmatrix}=\begin{bmatrix} \mathbf{e}_1\ \mathbf{e}_2\ ... \ \mathbf{e}_n\end{bmatrix}M</math>, then the components of the covector <math>\mathbf{v}\ \cdot</math> will transform with the same matrix <math>M</math>, namely, <math>\begin{bmatrix} \mathbf{v}\cdot\mathbf{e}_1^\prime & \mathbf{v}\cdot\mathbf{e}_2^\prime & ... & \mathbf{v}\cdot\mathbf{e}_n^\prime \end{bmatrix}=\begin{bmatrix} \mathbf{v}\cdot\mathbf{e}_1 & \mathbf{v}\cdot\mathbf{e}_2 & ... & \mathbf{v}\cdot\mathbf{e}_n \end{bmatrix}M</math>. The components of a ''covector'' are often represented arranged in a row. | |||
:<math>(v_1,v_2,v_3). \, </math> | |||
A third concept related to covariance and contravariance is ]. A ] (also called type-0 or rank-0 tensor) is an object that does not vary with the change in basis. An example of a physical ] that is a scalar is the ] of a particle. The single, scalar value of mass is independent to changes in basis vectors and consequently is called ''invariant''. The magnitude of a vector (such as ]) is another example of an invariant, because it remains fixed even if geometrical vector components vary. (For example, for a position vector of length <math>3</math> meters, if all ] basis vectors are changed from <math>1</math> meters in length to <math>.01</math> meters in length, the length of the position vector remains unchanged at <math>3</math> meters, although the vector components will all increase by a factor of <math>100</math>). The scalar product of a vector and a covector is invariant, because one has components that vary with the base change, and the other has components that vary oppositely, and the two effects cancel out. One thus says that covectors are ''dual'' to vectors. | |||
This list of numbers depends on the choice of ]. For instance, if the vector represents position with respect to an observer (]), then the coordinate system may be obtained from a system of rigid rods, or reference axes, along which the components ''v''<sub>1</sub>, ''v''<sub>2</sub>, and ''v''<sub>3</sub> are measured. For a vector to represent a geometric object, it must be possible to describe how it looks in any other coordinate system. That is to say, the components of the vectors will ''transform'' in a certain way in passing from one coordinate system to another. | |||
Thus, to summarize: | |||
A ''contravariant vector'' is required to have components that "transform in the same way as the coordinates" (the opposite way as the reference axes) under changes of coordinates such as ] and dilation. ]; instead, the components of the vector make a change that cancels the change in the spatial axes, in the same way that co-ordinates change. In other words, if the reference axes were rotated in one direction, the component representation of the vector would rotate in exactly the opposite way. Similarly, if the reference axes were stretched in one direction, the components of the vector, like the co-ordinates, would reduce in an exactly compensating way. Mathematically, if the coordinate system undergoes a transformation described by an ] ''M'', so that a ] '''x''' is transformed to '''x'''′ = ''M'''''x''', then a contravariant vector '''v''' must be similarly transformed via '''v'''′ = ''M'''''v'''. This important requirement is what distinguishes a contravariant vector from any other triple of physically meaningful quantities. For example, if ''v'' consists of the ''x'', ''y'', and ''z''-components of ], then ''v'' is a contravariant vector: if the coordinates of space are stretched, rotated, or twisted, then the components of the velocity transform in the same way. On the other hand, for instance, a triple consisting of the length, width, and height of a rectangular box could make up the three components of an abstract ], but this vector would not be contravariant, since rotating the box does not change the box's length, width, and height. Examples of contravariant vectors include ], ], ], ], and ]. | |||
* A vector or ], has components that ''contra-vary'' with a change of basis to compensate. That is, the matrix that transforms the vector components must be the inverse of the matrix that transforms the basis vectors. The components of vectors (as opposed to those of covectors) are said to be '''contravariant'''. In ] (implicit summation over repeated index), contravariant components are denoted with ''upper indices'' as in | |||
*:<math>\mathbf{v} = v^i \mathbf{e}_i</math> | |||
By contrast, a ''covariant vector'' has components that change oppositely to the coordinates or, equivalently, transform like the reference axes. For instance, the components of the ] vector of a function | |||
* A covector or ] has components that ''co-vary'' with a change of basis in the corresponding (initial) vector space. That is, the components must be transformed by the same matrix as the change of basis matrix in the corresponding (initial) vector space. The components of covectors (as opposed to those of vectors) are said to be '''covariant'''. In ], covariant components are denoted with ''lower indices'' as in | |||
:<math>\nabla f = \frac{\partial f}{\partial x_1}\widehat{x}_1+\frac{\partial f}{\partial x_2}\widehat{x}_2+\frac{\partial f}{\partial x_3}\widehat{x}_3</math> | |||
*:<math>\mathbf{w} = w_i \mathbf{e}^i.</math> | |||
transform like the reference axes themselves. When only ] of the spatial are considered, the components of contravariant and covariant vectors behave in the same way. It is only when other transformations are allowed that the difference becomes apparent. | |||
* The scalar product of a vector and covector is the scalar <math>v^iw_i</math>, which is invariant. It is the duality pairing of vectors and covectors. | |||
==Definition== | ==Definition== | ||
] | |||
The general formulation of covariance and contravariance refers to how the components of a ] transform under a ] (]). Thus let ''V'' be a ] of dimension ''n'' over the field of ] ''S'', and let each of '''f''' = (''X''<sub>1</sub>,...,''X''<sub>''n''</sub>) and '''f'''' = (''Y''<sub>1</sub>,...,''Y''<sub>''n''</sub>) be a ] of ''V''.<ref group="note">A basis '''f''' may here profitably be viewed as a ] from '''R'''<sup>''n''</sup> to ''V''. Regarding '''f''' as a row vector whose entries are the elements of the basis, the associated linear isomorphism is then <math>\mathbf{x}\mapsto \mathbf{f}\mathbf{x}.</math></ref> Also, let the ] from '''f''' to '''f′''' be given by | |||
The general formulation of covariance and contravariance refers to how the components of a coordinate vector transform under a ] (]). Thus let ''V'' be a ] of dimension ''n'' over a ] of ] ''S'', and let each of {{nowrap|1='''f''' = (''X''<sub>1</sub>, ..., ''X''<sub>''n''</sub>)}} and {{nowrap|1='''f'''′ = (''Y''<sub>1</sub>, ..., ''Y''<sub>''n''</sub>)}} be a ] of ''V''.<ref group="note">A basis '''f''' may here profitably be viewed as a ] from '''R'''<sup>''n''</sup> to ''V''. Regarding '''f''' as a row vector whose entries are the elements of the basis, the associated linear isomorphism is then <math>\mathbf{x}\mapsto \mathbf{f}\mathbf{x}.</math></ref> Also, let the ] from '''f''' to '''f'''′ be given by | |||
{{NumBlk|:|<math>\mathbf{f}\mapsto \mathbf{f}' = \ |
{{NumBlk|:|<math>\mathbf{f}\mapsto \mathbf{f}' = \biggl(\sum_i a^i_1X_i,\dots,\sum_i a^i_nX_i\biggr) = \mathbf{f}A</math>|{{EquationRef|1}}}} | ||
for some ] ''n'' |
for some ] ''n''×''n'' matrix ''A'' with entries <math>a^i_j</math>. | ||
Here, each vector ''Y''<sub>''j''</sub> of the '''f''' |
Here, each vector ''Y''<sub>''j''</sub> of the '''f'''′ basis is a linear combination of the vectors ''X''<sub>''i''</sub> of the '''f''' basis, so that | ||
:<math>Y_j=\sum_i a^i_jX_i.</math> | :<math>Y_j=\sum_i a^i_jX_i.</math> | ||
{{Multicol}} | |||
===Contravariant transformation=== | ===Contravariant transformation=== | ||
{{Main|Covariant_transformation#Contravariant_transformation|l1=Contravariant transformation}} | |||
A ] ''v'' in ''V'' is expressed uniquely as a ] of the elements of the '''f''' basis as | |||
A vector <math>v</math> in ''V'' is expressed uniquely as a ] of the elements <math>X_i</math> of the '''f''' basis as | |||
{{NumBlk|:|<math>v |
{{NumBlk|:|<math>v=\sum_i v^iX_i,</math>|{{EquationRef|2}}}} | ||
where ''v'' |
where ''v''{{i sup|''i''}} are elements of the field ''S'' known as the '''components''' of ''v'' in the '''f''' basis. Denote the ] of components of ''v'' by '''v''': | ||
:<math>\mathbf{v} = \begin{bmatrix}v^1\\v^2\\\vdots\\v^n\end{bmatrix}</math> | :<math>\mathbf{v} = \begin{bmatrix}v^1\\v^2\\\vdots\\v^n\end{bmatrix}</math> | ||
Line 57: | Line 72: | ||
:<math>v = \mathbf{f}\, \mathbf{v}.</math> | :<math>v = \mathbf{f}\, \mathbf{v}.</math> | ||
The vector ''v'' may also be expressed in terms of the '''f''' |
The vector ''v'' may also be expressed in terms of the '''f'''′ basis, so that | ||
:<math>v = \mathbf{f'}\, \mathbf{v}.</math> | :<math>v = \mathbf{f'}\, \mathbf{v}.</math> | ||
Line 65: | Line 80: | ||
:<math>\mathbf{f}\, \mathbf{v} = v = \mathbf{f'}\, \mathbf{v}.</math> | :<math>\mathbf{f}\, \mathbf{v} = v = \mathbf{f'}\, \mathbf{v}.</math> | ||
The invariance of ''v'' combined with the relationship ({{EquationNote|1}}) between '''f''' and '''f''' |
The invariance of ''v'' combined with the relationship ({{EquationNote|1}}) between '''f''' and '''f'''′ implies that | ||
:<math>\mathbf{f}\, \mathbf{v} = \mathbf{f}A\, \mathbf{v},</math> | :<math>\mathbf{f}\, \mathbf{v} = \mathbf{f}A\, \mathbf{v},</math> | ||
Line 71: | Line 86: | ||
giving the transformation rule | giving the transformation rule | ||
:<math>\mathbf{v} = A^{-1}\mathbf{v}.</math> | :<math>\mathbf{v} = \mathbf{v} = A^{-1}\mathbf{v}.</math> | ||
In terms of components, | In terms of components, | ||
Line 82: | Line 97: | ||
The way ''A'' relates the two pairs is depicted in the following informal diagram using an arrow. The reversal of the arrow indicates a contravariant change: | The way ''A'' relates the two pairs is depicted in the following informal diagram using an arrow. The reversal of the arrow indicates a contravariant change: | ||
:<math>\mathbf{f}\longrightarrow \mathbf{f'} |
:<math>\begin{align} | ||
\mathbf{f} &\longrightarrow \mathbf{f'} \\ | |||
v &\longleftarrow v | |||
\end{align}</math> | |||
{{Multicol-break}} | |||
===Covariant transformation=== | ===Covariant transformation=== | ||
{{main|Covariant transformation}} | |||
A ] α on ''V'' is expressed uniquely in terms of its '''components''' (scalars in ''S'') in the '''f''' basis as | |||
A ] ''α'' on ''V'' is expressed uniquely in terms of its '''components''' (elements in ''S'') in the '''f''' basis as | |||
:<math>\alpha(X_i) = \alpha_i , \quad i=1,2,\dots,n.</math> | :<math>\alpha(X_i) = \alpha_i , \quad i=1,2,\dots,n.</math> | ||
These components are the action of α on the basis vectors ''X''<sub>''i''</sub> of the '''f''' basis. | These components are the action of ''α'' on the basis vectors ''X''<sub>''i''</sub> of the '''f''' basis. | ||
Under the change of basis from '''f''' to '''f''' |
Under the change of basis from '''f''' to '''f'''′ (via {{EquationNote|1}}), the components transform so that | ||
{{NumBlk|:|<math>\begin{ |
{{NumBlk|:|<math>\begin{align} | ||
\alpha_i & |
\alpha_i &= \alpha(Y_i) \\ | ||
&= \alpha\biggl(\sum_j a^j_i X_j\biggr) \\ | |||
&= \sum_j a^j_i \alpha(X_j) \\ | |||
&= \sum_j a^j_i \alpha_j. | |||
\end{ |
\end{align}</math>|{{EquationRef|3}}}} | ||
Denote the ] of components of |
Denote the ] of components of ''α'' by ''α'': | ||
:<math>\mathbf{\alpha} = \begin{bmatrix}\alpha_1,\alpha_2,\dots,\alpha_n\end{bmatrix}</math> | :<math>\mathbf{\alpha} = \begin{bmatrix}\alpha_1,\alpha_2,\dots,\alpha_n\end{bmatrix}</math> | ||
Line 114: | Line 130: | ||
The way ''A'' relates the two pairs is depicted in the following informal diagram using an arrow. A covariant relationship is indicated since the arrows travel in the same direction: | The way ''A'' relates the two pairs is depicted in the following informal diagram using an arrow. A covariant relationship is indicated since the arrows travel in the same direction: | ||
:<math>\begin{align} | |||
:<math>\mathbf{f}\longrightarrow \mathbf{f'}</math> | |||
\mathbf{f} &\longrightarrow \mathbf{f'} \\ | |||
\alpha &\longrightarrow \alpha | |||
\end{align}</math> | |||
Had a column vector representation been used instead, the transformation law would be the ] | Had a column vector representation been used instead, the transformation law would be the ] | ||
:<math>\alpha^\mathrm{T} = A^\mathrm{T}\alpha^\mathrm{T}.</math> | :<math>\alpha^\mathrm{T} = A^\mathrm{T}\alpha^\mathrm{T}.</math> | ||
{{Multicol-end}} | |||
==Coordinates== | ==Coordinates== | ||
Line 126: | Line 143: | ||
The coordinates on ''V'' are therefore contravariant in the sense that | The coordinates on ''V'' are therefore contravariant in the sense that | ||
:<math>x^i = \sum_{k=1}^n \tilde{a}^i_kx^k.</math> | :<math>x^i = \sum_{k=1}^n \tilde{a}^i_kx^k.</math> | ||
Conversely, a system of ''n'' quantities ''v''<sup>''i''</sup> that transform like the coordinates ''x''<sup>''i''</sup> on ''V'' defines a contravariant vector. A system of ''n'' quantities that transform oppositely to the coordinates is then a covariant vector. | Conversely, a system of ''n'' quantities ''v''<sup>''i''</sup> that transform like the coordinates ''x''<sup>''i''</sup> on ''V'' defines a contravariant vector (or simply vector). A system of ''n'' quantities that transform oppositely to the coordinates is then a covariant vector (or covector). | ||
This formulation of contravariance and covariance is often more natural in applications in which there is a coordinate space (a ]) on which vectors live as ]s or ]s. Given a local coordinate system ''x''<sup>''i''</sup> on the manifold, the reference axes for the coordinate system are the ]s | This formulation of contravariance and covariance is often more natural in applications in which there is a coordinate space (a ]) on which vectors live as ]s or ]s. Given a local coordinate system ''x''<sup>''i''</sup> on the manifold, the reference axes for the coordinate system are the ]s | ||
:<math>X_1 = \frac{\partial}{\partial x^1},\dots,X_n=\frac{\partial}{\partial x^n}.</math> | :<math>X_1 = \frac{\partial}{\partial x^1},\dots,X_n=\frac{\partial}{\partial x^n}.</math> | ||
This gives rise to the frame '''f''' |
This gives rise to the frame {{nowrap|1='''f''' = (''X''<sub>1</sub>, ..., ''X''<sub>''n''</sub>)}} at every point of the coordinate patch. | ||
If ''y''<sup>''i''</sup> is a different coordinate system and | If ''y''<sup>''i''</sup> is a different coordinate system and | ||
:<math>Y_1=\frac{\partial}{\partial y^1},\dots,Y_n=\frac{\partial}{\partial y^n},</math> | :<math>Y_1=\frac{\partial}{\partial y^1},\dots,Y_n = \frac{\partial}{\partial y^n},</math> | ||
then the frame '''f'''' is related to the frame '''f''' by the inverse of the ] of the coordinate transition: | then the frame '''f'''' is related to the frame '''f''' by the inverse of the ] of the coordinate transition: | ||
:<math>\mathbf{f}' = \mathbf{f}J^{-1},\quad J=\left(\frac{\partial y^i}{\partial x^j}\right)_{i,j=1}^n.</math> | :<math>\mathbf{f}' = \mathbf{f}J^{-1},\quad J = \left(\frac{\partial y^i}{\partial x^j}\right)_{i,j=1}^n.</math> | ||
Or, in indices, | Or, in indices, | ||
:<math>\frac{\partial}{\partial y^i} = \sum_{j=1}^n\frac{\partial x^j}{\partial y^i}\frac{\partial}{\partial x^j}.</math> | :<math>\frac{\partial}{\partial y^i} = \sum_{j=1}^n\frac{\partial x^j}{\partial y^i}\frac{\partial}{\partial x^j}.</math> | ||
A |
A tangent vector is by definition a vector that is a linear combination of the coordinate partials <math>\partial/\partial x^i</math>. Thus a tangent vector is defined by | ||
:<math>v = \sum_{i=1}^n v^i X_i = \mathbf{f}\ \mathbf{v}.</math> | :<math>v = \sum_{i=1}^n v^i X_i = \mathbf{f}\ \mathbf{v}.</math> | ||
Such a vector is contravariant with respect to change of frame. Under changes in the coordinate system, one has | Such a vector is contravariant with respect to change of frame. Under changes in the coordinate system, one has | ||
:<math>\mathbf{v} = \mathbf{v} = J\, \mathbf{v}.</math> | :<math>\mathbf{v}\left = \mathbf{v}\left = J\, \mathbf{v}.</math> | ||
Therefore the components of a tangent vector transform via | |||
Therefore, the components of a tangent vector transform via | |||
:<math>v^i = \sum_{j=1}^n \frac{\partial y^i}{\partial x^j}v^j.</math> | |||
:<math>v^i\left = \sum_{j=1}^n \frac{\partial y^i}{\partial x^j}v^j.</math> | |||
Accordingly, a system of ''n'' quantities ''v''<sup>''i''</sup> depending on the coordinates that transform in this way on passing from one coordinate system to another is called a contravariant vector. | Accordingly, a system of ''n'' quantities ''v''<sup>''i''</sup> depending on the coordinates that transform in this way on passing from one coordinate system to another is called a contravariant vector. | ||
==Covariant and contravariant components of a vector with a metric== | ==Covariant and contravariant components of a vector with a metric== | ||
] onto the coordinate axes. The covariant components {{color box|blue}} are obtained by projecting onto the normal lines to the coordinate hyperplanes.]] | |||
In a ] ''V'' over a field ''K'' with a ] {{nowrap|''g'' : ''V'' × ''V'' → ''K''}} (which may be referred to as the ]), there is little distinction between covariant and contravariant vectors, because the ] allows covectors to be identified with vectors. That is, a vector ''v'' uniquely determines a covector α via | |||
In a finite-dimensional ] ''V'' over a field ''K'' with a symmetric ] {{nowrap|''g'' : ''V'' × ''V'' → ''K''}} (which may be referred to as the ]), there is little distinction between covariant and contravariant vectors, because the ] allows covectors to be identified with vectors. That is, a vector ''v'' uniquely determines a covector ''α'' via | |||
:<math>\alpha(w) = g(v, w) </math> | :<math>\alpha(w) = g(v, w) </math> | ||
for all vectors ''w''. Conversely, each covector α determines a unique vector ''v'' by this equation. Because of this identification of vectors with covectors, one may speak of the '''covariant components''' or '''contravariant components''' of a vector, that is, they are just representations of the same vector using reciprocal |
for all vectors ''w''. Conversely, each covector ''α'' determines a unique vector ''v'' by this equation. Because of this identification of vectors with covectors, one may speak of the '''covariant components''' or '''contravariant components''' of a vector, that is, they are just representations of the same vector using the ]. | ||
Given a basis '''f''' |
Given a basis {{nowrap|1='''f''' = (''X''<sub>1</sub>, ..., ''X''<sub>''n''</sub>)}} of ''V'', there is a unique reciprocal basis {{nowrap|1='''f'''<sup>#</sup> = (''Y''<sup>1</sup>, ..., ''Y''<sup>''n''</sup>)}} of ''V'' determined by requiring that | ||
:<math>Y^i |
:<math>g(Y^i,X_j) = \delta^i_j,</math> | ||
the ]. In terms of these bases, any vector ''v'' can be written in two ways: | the ]. In terms of these bases, any vector ''v'' can be written in two ways: | ||
:<math>\begin{align} | :<math>\begin{align} | ||
Line 163: | Line 185: | ||
:<math>\mathbf{v} = A^{-1}\mathbf{v},\quad \mathbf{v}^\sharp = A^T\mathbf{v}^\sharp.</math> | :<math>\mathbf{v} = A^{-1}\mathbf{v},\quad \mathbf{v}^\sharp = A^T\mathbf{v}^\sharp.</math> | ||
] onto the coordinate axes (yellow). The covariant components are obtained by projecting onto the normal lines to the coordinate hyperplanes (blue).]] | |||
{{clr}} | |||
===Euclidean plane=== | ===Euclidean plane=== | ||
In the Euclidean plane, the ] allows for vectors to be identified with covectors. If <math>\mathbf{e}_1,\mathbf{e}_2</math> is a basis, then the dual basis <math>\mathbf{e}^1,\mathbf{e}^2</math> satisfies | In the Euclidean plane, the ] allows for vectors to be identified with covectors. If <math>\mathbf{e}_1,\mathbf{e}_2</math> is a basis, then the dual basis <math>\mathbf{e}^1,\mathbf{e}^2</math> satisfies | ||
:<math>\begin{align} | :<math>\begin{align} | ||
\mathbf{e}^1\cdot\mathbf{e}_1=1, &\quad\mathbf{e}^1\cdot\mathbf{e}_2=0\\ | \mathbf{e}^1\cdot\mathbf{e}_1 = 1, &\quad \mathbf{e}^1\cdot\mathbf{e}_2 = 0 \\ | ||
\mathbf{e}^2\cdot\mathbf{e}_1=0, &\quad \mathbf{e}^2\cdot\mathbf{e}_2=1. | \mathbf{e}^2\cdot\mathbf{e}_1 = 0, &\quad \mathbf{e}^2\cdot\mathbf{e}_2 = 1. | ||
\end{align} | \end{align}</math> | ||
</math> | |||
Thus, '''e'''<sup>1</sup> and '''e'''<sub>2</sub> are perpendicular to each other, as are '''e'''<sup>2</sup> and '''e'''<sub>1</sub>, and the lengths of '''e'''<sup>1</sup> and '''e'''<sup>2</sup> normalized against '''e'''<sub>1</sub> and '''e'''<sub>2</sub>, respectively. | Thus, '''e'''<sup>1</sup> and '''e'''<sub>2</sub> are perpendicular to each other, as are '''e'''<sup>2</sup> and '''e'''<sub>1</sub>, and the lengths of '''e'''<sup>1</sup> and '''e'''<sup>2</sup> normalized against '''e'''<sub>1</sub> and '''e'''<sub>2</sub>, respectively. | ||
====Example==== | ====Example==== | ||
For example,<ref>{{cite |
For example,<ref>{{cite book |last1=Bowen |first1=Ray |first2=C.-C. |last2=Wang |chapter=§3.14 Reciprocal Basis and Change of Basis |chapter-url=https://books.google.com/books?id=Hr5bhIVWr4wC&pg=PA76 |title=Introduction to Vectors and Tensors |year=2008 |orig-year=1976 |publisher=Dover |pages=78, 79, 81 |isbn=9780486469140}}</ref> suppose that we are given a basis '''e'''<sub>1</sub>, '''e'''<sub>2</sub> consisting of a pair of vectors making a 45° angle with one another, such that '''e'''<sub>1</sub> has length 2 and '''e'''<sub>2</sub> has length 1. Then the dual basis vectors are given as follows: | ||
* '''e'''<sup>2</sup> is the result of rotating '''e'''<sub>1</sub> through an angle of 90° (where the sense is measured by assuming the pair '''e'''<sub>1</sub>, '''e'''<sub>2</sub> to be positively oriented), and then rescaling so that {{nowrap|'''e'''<sup>2</sup> |
* '''e'''<sup>2</sup> is the result of rotating '''e'''<sub>1</sub> through an angle of 90° (where the sense is measured by assuming the pair '''e'''<sub>1</sub>, '''e'''<sub>2</sub> to be positively oriented), and then rescaling so that {{nowrap|'''e'''<sup>2</sup> ⋅ '''e'''<sub>2</sub> {{=}} 1}} holds. | ||
* '''e'''<sup>1</sup> is the result of rotating '''e'''<sub>2</sub> through an angle of 90°, and then rescaling so that {{nowrap|'''e'''<sup>1</sup> |
* '''e'''<sup>1</sup> is the result of rotating '''e'''<sub>2</sub> through an angle of 90°, and then rescaling so that {{nowrap|'''e'''<sup>1</sup> ⋅ '''e'''<sub>1</sub> {{=}} 1}} holds. | ||
Applying these rules, we find | Applying these rules, we find | ||
:<math>\mathbf{e}^1 = \frac{1}{2}\mathbf{e}_1 - \frac{1}{\sqrt{2}}\mathbf{e}_2</math> | :<math>\mathbf{e}^1 = \frac{1}{2}\mathbf{e}_1 - \frac{1}{\sqrt{2}}\mathbf{e}_2</math> | ||
and | and | ||
:<math>\mathbf{e}^2 = -\frac{1}{\sqrt{2}}\mathbf{e}_1+2\mathbf{e}_2.</math> | :<math>\mathbf{e}^2 = -\frac{1}{\sqrt{2}}\mathbf{e}_1 + 2\mathbf{e}_2.</math> | ||
Thus the change of basis matrix in going from the original basis to the reciprocal basis is | Thus the change of basis matrix in going from the original basis to the reciprocal basis is | ||
:<math>R = \begin{bmatrix} |
:<math>R = \begin{bmatrix} | ||
\frac{1}{2} & -\frac{1}{\sqrt{2}} \\ | |||
-1/\sqrt{2} & 2 | |||
-\frac{1}{\sqrt{2}} & 2 | |||
\end{bmatrix},</math> | \end{bmatrix},</math> | ||
since | since | ||
:<math> = \begin{bmatrix} |
:<math> = \begin{bmatrix} | ||
\frac{1}{2} & -\frac{1}{\sqrt{2}} \\ | |||
-1/\sqrt{2} & 2 | |||
-\frac{1}{\sqrt{2}} & 2 | |||
\end{bmatrix}.</math> | \end{bmatrix}.</math> | ||
Line 198: | Line 219: | ||
is a vector with contravariant components | is a vector with contravariant components | ||
:<math>v^1 = \frac{3}{2},\quad v^2 = 2.</math> | :<math>v^1 = \frac{3}{2},\quad v^2 = 2.</math> | ||
The covariant components are obtained by equating the two expressions for the vector ''v'': | The covariant components are obtained by equating the two expressions for the vector ''v'': | ||
:<math>v = v_1\mathbf{e}^1 + v_2\mathbf{e}^2 = v^1\mathbf{e}_1 + v^2\mathbf{e}_2</math> | |||
:<math> | |||
v = v_1\mathbf{e}^1 + v_2\mathbf{e}^2 = v^1\mathbf{e}_1+v^2\mathbf{e}_2</math> | |||
so | so | ||
:<math>\begin{align} | :<math>\begin{align} | ||
\begin{bmatrix}v_1\\ v_2\end{bmatrix} &= R^{-1}\begin{bmatrix}v^1\\ v^2\end{bmatrix} \\ | \begin{bmatrix}v_1\\ v_2\end{bmatrix} &= R^{-1}\begin{bmatrix}v^1 \\ v^2\end{bmatrix} \\ | ||
&= \begin{bmatrix}4&\sqrt{2}\\ \sqrt{2}&1\end{bmatrix} |
&= \begin{bmatrix}4 & \sqrt{2} \\ \sqrt{2} & 1\end{bmatrix} | ||
\begin{bmatrix}v^1 \\ v^2\end{bmatrix} \\ | |||
&= \begin{bmatrix}6 + 2\sqrt{2} \\ 2 + \frac{3}{\sqrt{2}}\end{bmatrix} | |||
\end{align}.</math> | |||
===Three-dimensional Euclidean space=== | ===Three-dimensional Euclidean space=== | ||
In the three-dimensional ], one can also determine explicitly the dual basis to a given set of ]s '''e'''<sub>1</sub>, '''e'''<sub>2</sub>, '''e'''<sub>3</sub> of ''E''<sub>3</sub> that are not necessarily assumed to be orthogonal nor of unit norm. |
In the three-dimensional ], one can also determine explicitly the dual basis to a given set of ]s '''e'''<sub>1</sub>, '''e'''<sub>2</sub>, '''e'''<sub>3</sub> of ''E''<sub>3</sub> that are not necessarily assumed to be orthogonal nor of unit norm. The dual basis vectors are: | ||
:<math> \mathbf{e}^1 = \frac{\mathbf{e}_2 \times \mathbf{e}_3}{\mathbf{e}_1 \cdot (\mathbf{e}_2 \times \mathbf{e}_3)} ; \qquad \mathbf{e}^2 = \frac{\mathbf{e}_3 \times \mathbf{e}_1}{\mathbf{e}_2 \cdot (\mathbf{e}_3 \times \mathbf{e}_1)}; \qquad \mathbf{e}^3 = \frac{\mathbf{e}_1 \times \mathbf{e}_2}{\mathbf{e}_3 \cdot (\mathbf{e}_1 \times \mathbf{e}_2)}. | :<math> \mathbf{e}^1 = \frac{\mathbf{e}_2 \times \mathbf{e}_3}{\mathbf{e}_1 \cdot (\mathbf{e}_2 \times \mathbf{e}_3)} ; \qquad \mathbf{e}^2 = \frac{\mathbf{e}_3 \times \mathbf{e}_1}{\mathbf{e}_2 \cdot (\mathbf{e}_3 \times \mathbf{e}_1)}; \qquad \mathbf{e}^3 = \frac{\mathbf{e}_1 \times \mathbf{e}_2}{\mathbf{e}_3 \cdot (\mathbf{e}_1 \times \mathbf{e}_2)}. | ||
</math> | </math> | ||
Even when the '''e'''<sub>i</sub> and '''e'''<sup>i</sup> are not ], they are still mutually |
Even when the '''e'''<sub>i</sub> and '''e'''<sup>i</sup> are not ], they are still mutually reciprocal: | ||
:<math>\mathbf{e}^i \cdot \mathbf{e}_j = \delta^i_j,</math> | :<math>\mathbf{e}^i \cdot \mathbf{e}_j = \delta^i_j,</math> | ||
Then the contravariant |
Then the contravariant components of any vector '''v''' can be obtained by the ] of '''v''' with the dual basis vectors: | ||
:<math> q^1 = \mathbf{v} \cdot \mathbf{e}^1; \qquad q^2 = \mathbf{v} \cdot \mathbf{e}^2; \qquad q^3 = \mathbf{v} \cdot \mathbf{e}^3. |
:<math> q^1 = \mathbf{v} \cdot \mathbf{e}^1; \qquad q^2 = \mathbf{v} \cdot \mathbf{e}^2; \qquad q^3 = \mathbf{v} \cdot \mathbf{e}^3 .</math> | ||
Likewise, the covariant components of ''' |
Likewise, the covariant components of '''v''' can be obtained from the dot product of '''v''' with basis vectors, viz. | ||
:<math> q_1 = \mathbf{v} \cdot \mathbf{e}_1; \qquad q_2 = \mathbf{v} \cdot \mathbf{e}_2; \qquad q_3 = \mathbf{v} \cdot \mathbf{e}_3. |
:<math> q_1 = \mathbf{v} \cdot \mathbf{e}_1; \qquad q_2 = \mathbf{v} \cdot \mathbf{e}_2; \qquad q_3 = \mathbf{v} \cdot \mathbf{e}_3 .</math> | ||
Then ''' |
Then '''v''' can be expressed in two (reciprocal) ways, viz. | ||
:<math> \mathbf{v} = q^i \mathbf{e}_i = q^1 \mathbf{e}_1 + q^2 \mathbf{e}_2 + q^3 \mathbf{e}_3 .</math> | |||
:<math> \mathbf{v} = q_i \mathbf{e}^i = q_1 \mathbf{e}^1 + q_2 \mathbf{e}^2 + q_3 \mathbf{e}^3 \, </math> | |||
or | or | ||
:<math> \mathbf{v} = |
:<math> \mathbf{v} = q_i \mathbf{e}^i = q_1 \mathbf{e}^1 + q_2 \mathbf{e}^2 + q_3 \mathbf{e}^3 </math> | ||
Combining the above relations, we have | Combining the above relations, we have | ||
:<math> \mathbf{v} = (\mathbf{v} \cdot \mathbf{e} |
:<math> \mathbf{v} = (\mathbf{v} \cdot \mathbf{e}^i) \mathbf{e}_i = (\mathbf{v} \cdot \mathbf{e}_i) \mathbf{e}^i </math> | ||
and we can convert |
and we can convert between the basis and dual basis with | ||
:<math>q_i = \mathbf{v}\cdot \mathbf{e}_i = (q^j \mathbf{e}_j)\cdot \mathbf{e}_i = (\mathbf{e}_j\cdot\mathbf{e}_i) q^j |
:<math>q_i = \mathbf{v}\cdot \mathbf{e}_i = (q^j \mathbf{e}_j)\cdot \mathbf{e}_i = (\mathbf{e}_j\cdot\mathbf{e}_i) q^j </math> | ||
and | and | ||
:<math>q^i = \mathbf{v}\cdot \mathbf{e}^i = (q_j \mathbf{e}^j)\cdot \mathbf{e}^i = (\mathbf{e}^j\cdot\mathbf{e}^i) q_j. |
:<math>q^i = \mathbf{v}\cdot \mathbf{e}^i = (q_j \mathbf{e}^j)\cdot \mathbf{e}^i = (\mathbf{e}^j\cdot\mathbf{e}^i) q_j .</math> | ||
If the basis vectors are ], then they are the same as the dual basis vectors. | |||
===General Euclidean spaces=== | ===General Euclidean spaces=== | ||
More generally, in an ''n''-dimensional Euclidean space ''V'', if a basis is | More generally, in an ''n''-dimensional Euclidean space ''V'', if a basis is | ||
:<math>\mathbf{e}_1,\dots,\mathbf{e}_n</math> |
:<math>\mathbf{e}_1,\dots,\mathbf{e}_n,</math> | ||
the reciprocal basis is given by | the reciprocal basis is given by (double indices are summed over), | ||
:<math>\mathbf{e}^i= |
:<math>\mathbf{e}^i=g^{ij}\mathbf{e}_j</math> | ||
where the coefficients '' |
where the coefficients ''g''<sup>''ij''</sup> are the entries of the inverse matrix of | ||
:<math> |
:<math>g_{ij} = \mathbf{e}_i\cdot\mathbf{e}_j .</math> | ||
Indeed, we then have | Indeed, we then have | ||
:<math>\mathbf{e}^i\cdot\mathbf{e}_k= |
:<math>\mathbf{e}^i\cdot\mathbf{e}_k=g^{ij}\mathbf{e}_j\cdot\mathbf{e}_k=g^{ij}g_{jk} = \delta^i_k .</math> | ||
The covariant and contravariant components of any vector | The covariant and contravariant components of any vector | ||
Line 253: | Line 276: | ||
are related as above by | are related as above by | ||
:<math>q_i = \mathbf{v}\cdot \mathbf{e}_i = (q^j \mathbf{e}_j)\cdot \mathbf{e}_i = q^ |
:<math>q_i = \mathbf{v}\cdot \mathbf{e}_i = (q^j \mathbf{e}_j)\cdot \mathbf{e}_i = q^jg_{ji}</math> | ||
and | and | ||
:<math>q^i = \mathbf{v}\cdot \mathbf{e}^i = (q_j\mathbf{e}^j)\cdot \mathbf{e}^i = |
:<math>q^i = \mathbf{v}\cdot \mathbf{e}^i = (q_j\mathbf{e}^j)\cdot \mathbf{e}^i = q_jg^{ji} .</math> | ||
==Informal usage== | |||
In the field of ], the ] '''covariant''' is often used informally as a synonym for ]. For example, the ] does not keep its written form under the coordinate transformations of ]. Thus, a physicist might say that the Schrödinger equation is ''not covariant''. In contrast, the ] and the ] do keep their written form under these coordinate transformations. Thus, a physicist might say that these equations are ''covariant''. | |||
Despite this usage of "covariant", it is more accurate to say that the Klein–Gordon and Dirac equations are invariant, and that the Schrödinger equation is not invariant. Additionally, to remove ambiguity, the transformation by which the invariance is evaluated should be indicated. | |||
Because the components of vectors are contravariant and those of covectors are covariant, the vectors themselves are often referred to as being contravariant and the covectors as covariant. This usage may be misleading, however, since vectors ] – are covariant under ] – and covectors ] – are contravariant under diffeomorphism. See ] for details. | |||
==Use in tensor analysis== | ==Use in tensor analysis== | ||
The distinction between covariance and contravariance is particularly important for computations with ]s, which often have '''mixed variance'''. This means that they have both covariant and contravariant components, or both vector and |
The distinction between covariance and contravariance is particularly important for computations with ]s, which often have '''mixed variance'''. This means that they have both covariant and contravariant components, or both vector and covector components. The valence of a tensor is the number of covariant and contravariant terms, and in ], covariant components have lower indices, while contravariant components have upper indices. The duality between covariance and contravariance intervenes whenever a vector or tensor quantity is represented by its components, although modern ] uses more sophisticated ]. | ||
In ], a '''covariant''' vector varies more or less reciprocally to a corresponding contravariant vector. Expressions for lengths, areas and volumes of objects in the vector space can then be given in terms of tensors with covariant and contravariant indices. Under simple expansions and contractions of the coordinates, the reciprocity is exact; under affine transformations the components of a vector intermingle on going between covariant and contravariant expression. | In ], a '''covariant''' vector varies more or less reciprocally to a corresponding contravariant vector. Expressions for lengths, areas and volumes of objects in the vector space can then be given in terms of tensors with covariant and contravariant indices. Under simple expansions and contractions of the coordinates, the reciprocity is exact; under affine transformations the components of a vector intermingle on going between covariant and contravariant expression. | ||
On a ], a ] will typically have multiple |
On a ], a ] will typically have multiple, upper and lower indices, where Einstein notation is widely used. When the manifold is equipped with a ], covariant and contravariant indices become very closely related to one another. Contravariant indices can be turned into covariant indices by ] with the metric tensor. The reverse is possible by contracting with the (matrix) inverse of the metric tensor. Note that in general, no such relation exists in spaces not endowed with a metric tensor. Furthermore, from a more abstract standpoint, a tensor is simply "there" and its components of either kind are only calculational artifacts whose values depend on the chosen coordinates. | ||
The explanation in geometric terms is that a general tensor will have contravariant indices as well as covariant indices, because it has parts that live in the ] as well as the ]. | The explanation in geometric terms is that a general tensor will have contravariant indices as well as covariant indices, because it has parts that live in the ] as well as the ]. | ||
A contravariant vector is one which transforms like <math>\frac{dx^{\mu}}{d\tau}</math>, where <math>x^{\mu} \!</math> are the coordinates of a particle at its ] <math>\tau |
A contravariant vector is one which transforms like <math>\frac{dx^{\mu}}{d\tau}</math>, where <math>x^{\mu} \!</math> are the coordinates of a particle at its ] <math>\tau</math>. A covariant vector is one which transforms like <math>\frac{\partial \varphi}{\partial x^{\mu}}</math>, where <math>\varphi</math> is a scalar field. | ||
==Algebra and geometry== | ==Algebra and geometry== | ||
In ], there are ]s and ]s. The ] |
In ], there are ]s and ]s. The assignment of the ] to a vector space is a standard example of a contravariant functor. Contravariant (resp. covariant) vectors are contravariant (resp. covariant) functors from a <math>\text{GL}(n)</math>-] to the fundamental representation of <math>\text{GL}(n)</math>. Similarly, tensors of higher degree are functors with values in other representations of <math>\text{GL}(n)</math>. However, some constructions of ] are of "mixed" variance, which prevents them from being functors. | ||
In ], the components of a vector relative to a basis of the ] are covariant if they change with the same linear transformation as a change of basis. They are contravariant if they change by the inverse transformation. This is sometimes a source of confusion for two distinct but related reasons. The first is that vectors whose components are covariant (called covectors or ]s) actually ] under smooth functions, meaning that the operation assigning the space of covectors to a smooth manifold is actually a ''contravariant'' functor. Likewise, vectors whose components are contravariant ] under smooth mappings, so the operation assigning the space of (contravariant) vectors to a smooth manifold is a ''covariant'' functor. Secondly, in the classical approach to differential geometry, it is not bases of the tangent bundle that are the most primitive object, but rather changes in the coordinate system. Vectors with contravariant components transform in the same way as changes in the coordinates (because these actually change oppositely to the induced change of basis). Likewise, vectors with covariant components transform in the opposite way as changes in the coordinates. | |||
In ], the same map in/map out distinction is helpful in assessing the variance of constructions. A ] to a ] ''M'' is, to begin with, a ] mapping smoothly into ''M'' and passing through a given point ''P''. It is therefore covariant, with respect to smooth mappings of ''M''. A contravariant vector, or ], is in the same way constructed from a smooth mapping from ''M'' to the real line, near ''P''. It is in the ], built up from the ]s of the ]s. Its ''components with respect to'' a local basis of one-forms ''dx<sub>i</sub>'' will be covariant; but one-forms and ]s in general are contravariant, in the sense that they ] under smooth mappings. This is crucial to how they are applied; for example a differential form can be ''restricted'' to any ], while this does not make the same sense for a field of tangent vectors. | |||
Covariant and contravariant components transform in different ways under ]. By considering a coordinate transformation on a manifold as a map from the manifold to itself, the transformation of covariant indices of a tensor are given by a ], and the transformation properties of the contravariant indices is given by a ]. | |||
==See also== | ==See also== | ||
* ] | |||
* ] | |||
* ] | * ] | ||
* ], which generalizes this notion to tensors that have indices not just in the primal and dual space but in other vector spaces (such as other tangent spaces on the same manifold). | |||
* ] | * ] | ||
* ], a generalization allowing indices to reference multiple vector bases | |||
==Notes== | ==Notes== | ||
{{Reflist|group="note"}} | {{Reflist|group="note"}} | ||
== |
==Citations== | ||
{{Reflist}} | {{Reflist}} | ||
* {{citation |last1=Arfken |first1=George B. |first2=Hans J. |last2=Weber |title=Mathematical Methods for Physicists |edition=6th |publisher=Harcourt |publication-place=San Diego |year=2005) |isbn=0-12-059876-0}}. | |||
==References== | |||
* {{Citation | last1=Dodson | first1=C. T. J. | last2=Poston | first2=T. | title=Tensor geometry | publisher=] | location=Berlin, New York | edition=2nd | series=Graduate Texts in Mathematics | isbn=978-3-540-52018-4 | mr=1223091 | year=1991 | volume=130}}. | |||
{{refbegin}} | |||
* {{Citation | last1=Greub | first1=Werner Hildbert | title=Multilinear algebra | publisher=Springer-Verlag New York, Inc., New York | series=Die Grundlehren der Mathematischen Wissenschaften, Band 136 | mr=0224623 | year=1967}}. | |||
* {{ |
* {{citation |last1=Kusse |first1=Bruce R. |last2=Westwig |first2=Erik A. |title=Mathematical Physics: Applied Mathematics for Scientists and Engineers |edition=2nd |publisher=Wiley |year=2010 |isbn=978-3-527-61814-9}}. | ||
* {{citation |last1=Arfken |first1=George B.|authorlink1=George B. Arfken |first2=Hans J. |last2=Weber |title=Mathematical Methods for Physicists |edition=6th |publisher=Harcourt |year=2005 |isbn=0-12-059876-0}}. | |||
* {{citation |doi=10.1098/rstl.1853.0018 |first=J.J. |last=Sylvester |jstor=108572 |title=On a Theory of the Syzygetic Relations of Two Rational Integral Functions, Comprising an Application to the Theory of Sturm's Functions, and That of the Greatest Algebraical Common Measure |year=1853 |journal=Philosophical Transactions of the Royal Society of London |volume=143 |pages=407–548 |publisher=The Royal Society}}. | |||
* {{Citation | last1=Dodson | first1=C. T. J. | last2=Poston | first2=T. | title=Tensor geometry | publisher=Springer | edition=2nd | series=Graduate Texts in Mathematics | isbn=978-3-540-52018-4 | mr=1223091 | year=1991 | volume=130}}. | |||
* {{Citation | last1=Greub | first1=Werner Hildbert | title=Multilinear algebra | publisher=Springer | series=Die Grundlehren der Mathematischen Wissenschaften, Band 136 | mr=0224623 | year=1967 |isbn=9780387038278}}. | |||
* {{Citation | last1=Sternberg | first1=Shlomo | author1-link=Shlomo Sternberg | title=Lectures on differential geometry | publisher=Chelsea | isbn=978-0-8284-0316-0 | year=1983}}. | |||
* {{citation |doi=10.1098/rstl.1853.0018 |first=J.J. |last=Sylvester |jstor=108572 |title=On a Theory of the Syzygetic Relations of Two Rational Integral Functions, Comprising an Application to the Theory of Sturm's Functions, and That of the Greatest Algebraical Common Measure |year=1853 |journal=Philosophical Transactions of the Royal Society of London |volume=143 |pages=407–548 |url=https://zenodo.org/record/1432412 |doi-access=free }}. | |||
* {{Citation | last1=Weinreich | first1=Gabriel | title=Geometrical Vectors | publisher=The University of Chicago Press | series=Chicago Lectures in Physics | year=1998 | isbn=9780226890487 | pages=126}} | |||
{{refend}} | |||
==External links== | ==External links== | ||
* {{springer|title=Covariant tensor|id=p/c026880}} | |||
* {{springer|title=Contravariant tensor|id=p/c025970}} | |||
* {{MathWorld | urlname=CovariantTensor | title=Covariant Tensor}} | * {{MathWorld | urlname=CovariantTensor | title=Covariant Tensor}} | ||
* {{MathWorld | urlname=ContravariantTensor | title=Contravariant Tensor}} | * {{MathWorld | urlname=ContravariantTensor | title=Contravariant Tensor}} | ||
* | * | ||
* {{cite web |first1=Kees |last1=Dullemond |first2=Kasper |last2=Peeters |title=Introduction to tensor calculus |date=2010 |publisher= |url=http://www.ita.uni-heidelberg.de/~dullemond/lectures/tensor/tensor.pdf}} | |||
{{tensors}} | |||
] | ] | ||
] | ] | ||
] | ] | ||
] | ] | ||
] | |||
] | |||
] | |||
] | |||
] | |||
] | |||
] |
Latest revision as of 01:04, 1 December 2024
Vector behavior under coordinate changes For use of "covariance" in the context of special relativity, see Lorentz covariance. For other uses of "covariant" or "contravariant", see Covariance and contravariance (disambiguation).In physics, especially in multilinear algebra and tensor analysis, covariance and contravariance describe how the quantitative description of certain geometric or physical entities changes with a change of basis. Briefly, a contravariant vector is a list of numbers that transforms oppositely to a change of basis, and a covariant vector is a list of numbers that transforms in the same way. Contravariant vectors are often just called vectors and covariant vectors are called covectors or dual vectors. The terms covariant and contravariant were introduced by James Joseph Sylvester in 1851.
Curvilinear coordinate systems, such as cylindrical or spherical coordinates, are often used in physical and geometric problems. Associated with any coordinate system is a natural choice of coordinate basis for vectors based at each point of the space, and covariance and contravariance are particularly important for understanding how the coordinate description of a vector changes by passing from one coordinate system to another. Tensors are objects in multilinear algebra that can have aspects of both covariance and contravariance.
Introduction
In physics, a vector typically arises as the outcome of a measurement or series of measurements, and is represented as a list (or tuple) of numbers such as
The numbers in the list depend on the choice of coordinate system. For instance, if the vector represents position with respect to an observer (position vector), then the coordinate system may be obtained from a system of rigid rods, or reference axes, along which the components v1, v2, and v3 are measured. For a vector to represent a geometric object, it must be possible to describe how it looks in any other coordinate system. That is to say, the components of the vectors will transform in a certain way in passing from one coordinate system to another.
A simple illustrative case is that of a Euclidean vector. For a vector, once a set of basis vectors has been defined, then the components of that vector will always vary opposite to that of the basis vectors. That vector is therefore defined as a contravariant tensor. Take a standard position vector for example. By changing the scale of the reference axes from meters to centimeters (that is, dividing the scale of the reference axes by 100, so that the basis vectors now are meters long), the components of the measured position vector are multiplied by 100. A vector's components change scale inversely to changes in scale to the reference axes, and consequently a vector is called a contravariant tensor.
A vector, which is an example of a contravariant tensor, has components that transform inversely to the transformation of the reference axes, (with example transformations including rotation and dilation). The vector itself does not change under these operations; instead, the components of the vector change in a way that cancels the change in the spatial axes. In other words, if the reference axes were rotated in one direction, the component representation of the vector would rotate in exactly the opposite way. Similarly, if the reference axes were stretched in one direction, the components of the vector, would reduce in an exactly compensating way. Mathematically, if the coordinate system undergoes a transformation described by an invertible matrix M, so that the basis vectors transform according to , then the components of a vector v in the original basis ( ) must be similarly transformed via . The components of a vector are often represented arranged in a column.
By contrast, a covector has components that transform like the reference axes. It lives in the dual vector space, and represents a linear map from vectors to scalars. The dot product operator involving vectors is a good example of a covector. To illustrate, assume we have a covector defined as , where is a vector. The components of this covector in some arbitrary basis are , with being the basis vectors in the corresponding vector space. (This can be derived by noting that we want to get the correct answer for the dot product operation when multiplying by an arbitrary vector , with components ). The covariance of these covector components is then seen by noting that if a transformation described by an invertible matrix M were to be applied to the basis vectors in the corresponding vector space, , then the components of the covector will transform with the same matrix , namely, . The components of a covector are often represented arranged in a row.
A third concept related to covariance and contravariance is invariance. A scalar (also called type-0 or rank-0 tensor) is an object that does not vary with the change in basis. An example of a physical observable that is a scalar is the mass of a particle. The single, scalar value of mass is independent to changes in basis vectors and consequently is called invariant. The magnitude of a vector (such as distance) is another example of an invariant, because it remains fixed even if geometrical vector components vary. (For example, for a position vector of length meters, if all Cartesian basis vectors are changed from meters in length to meters in length, the length of the position vector remains unchanged at meters, although the vector components will all increase by a factor of ). The scalar product of a vector and a covector is invariant, because one has components that vary with the base change, and the other has components that vary oppositely, and the two effects cancel out. One thus says that covectors are dual to vectors.
Thus, to summarize:
- A vector or tangent vector, has components that contra-vary with a change of basis to compensate. That is, the matrix that transforms the vector components must be the inverse of the matrix that transforms the basis vectors. The components of vectors (as opposed to those of covectors) are said to be contravariant. In Einstein notation (implicit summation over repeated index), contravariant components are denoted with upper indices as in
- A covector or cotangent vector has components that co-vary with a change of basis in the corresponding (initial) vector space. That is, the components must be transformed by the same matrix as the change of basis matrix in the corresponding (initial) vector space. The components of covectors (as opposed to those of vectors) are said to be covariant. In Einstein notation, covariant components are denoted with lower indices as in
- The scalar product of a vector and covector is the scalar , which is invariant. It is the duality pairing of vectors and covectors.
Definition
The general formulation of covariance and contravariance refers to how the components of a coordinate vector transform under a change of basis (passive transformation). Thus let V be a vector space of dimension n over a field of scalars S, and let each of f = (X1, ..., Xn) and f′ = (Y1, ..., Yn) be a basis of V. Also, let the change of basis from f to f′ be given by
1 |
for some invertible n×n matrix A with entries . Here, each vector Yj of the f′ basis is a linear combination of the vectors Xi of the f basis, so that
Contravariant transformation
Main article: Contravariant transformationA vector in V is expressed uniquely as a linear combination of the elements of the f basis as
2 |
where v are elements of the field S known as the components of v in the f basis. Denote the column vector of components of v by v:
so that (2) can be rewritten as a matrix product
The vector v may also be expressed in terms of the f′ basis, so that
However, since the vector v itself is invariant under the choice of basis,
The invariance of v combined with the relationship (1) between f and f′ implies that
giving the transformation rule
In terms of components,
where the coefficients are the entries of the inverse matrix of A.
Because the components of the vector v transform with the inverse of the matrix A, these components are said to transform contravariantly under a change of basis.
The way A relates the two pairs is depicted in the following informal diagram using an arrow. The reversal of the arrow indicates a contravariant change:
Covariant transformation
Main article: Covariant transformationA linear functional α on V is expressed uniquely in terms of its components (elements in S) in the f basis as
These components are the action of α on the basis vectors Xi of the f basis.
Under the change of basis from f to f′ (via 1), the components transform so that
3 |
Denote the row vector of components of α by α:
so that (3) can be rewritten as the matrix product
Because the components of the linear functional α transform with the matrix A, these components are said to transform covariantly under a change of basis.
The way A relates the two pairs is depicted in the following informal diagram using an arrow. A covariant relationship is indicated since the arrows travel in the same direction:
Had a column vector representation been used instead, the transformation law would be the transpose
Coordinates
The choice of basis f on the vector space V defines uniquely a set of coordinate functions on V, by means of
The coordinates on V are therefore contravariant in the sense that
Conversely, a system of n quantities v that transform like the coordinates x on V defines a contravariant vector (or simply vector). A system of n quantities that transform oppositely to the coordinates is then a covariant vector (or covector).
This formulation of contravariance and covariance is often more natural in applications in which there is a coordinate space (a manifold) on which vectors live as tangent vectors or cotangent vectors. Given a local coordinate system x on the manifold, the reference axes for the coordinate system are the vector fields
This gives rise to the frame f = (X1, ..., Xn) at every point of the coordinate patch.
If y is a different coordinate system and
then the frame f' is related to the frame f by the inverse of the Jacobian matrix of the coordinate transition:
Or, in indices,
A tangent vector is by definition a vector that is a linear combination of the coordinate partials . Thus a tangent vector is defined by
Such a vector is contravariant with respect to change of frame. Under changes in the coordinate system, one has
Therefore, the components of a tangent vector transform via
Accordingly, a system of n quantities v depending on the coordinates that transform in this way on passing from one coordinate system to another is called a contravariant vector.
Covariant and contravariant components of a vector with a metric
In a finite-dimensional vector space V over a field K with a symmetric bilinear form g : V × V → K (which may be referred to as the metric tensor), there is little distinction between covariant and contravariant vectors, because the bilinear form allows covectors to be identified with vectors. That is, a vector v uniquely determines a covector α via
for all vectors w. Conversely, each covector α determines a unique vector v by this equation. Because of this identification of vectors with covectors, one may speak of the covariant components or contravariant components of a vector, that is, they are just representations of the same vector using the reciprocal basis.
Given a basis f = (X1, ..., Xn) of V, there is a unique reciprocal basis f = (Y, ..., Y) of V determined by requiring that
the Kronecker delta. In terms of these bases, any vector v can be written in two ways:
The components v are the contravariant components of the vector v in the basis f, and the components vi are the covariant components of v in the basis f. The terminology is justified because under a change of basis,
Euclidean plane
In the Euclidean plane, the dot product allows for vectors to be identified with covectors. If is a basis, then the dual basis satisfies
Thus, e and e2 are perpendicular to each other, as are e and e1, and the lengths of e and e normalized against e1 and e2, respectively.
Example
For example, suppose that we are given a basis e1, e2 consisting of a pair of vectors making a 45° angle with one another, such that e1 has length 2 and e2 has length 1. Then the dual basis vectors are given as follows:
- e is the result of rotating e1 through an angle of 90° (where the sense is measured by assuming the pair e1, e2 to be positively oriented), and then rescaling so that e ⋅ e2 = 1 holds.
- e is the result of rotating e2 through an angle of 90°, and then rescaling so that e ⋅ e1 = 1 holds.
Applying these rules, we find
and
Thus the change of basis matrix in going from the original basis to the reciprocal basis is
since
For instance, the vector
is a vector with contravariant components
The covariant components are obtained by equating the two expressions for the vector v:
so
Three-dimensional Euclidean space
In the three-dimensional Euclidean space, one can also determine explicitly the dual basis to a given set of basis vectors e1, e2, e3 of E3 that are not necessarily assumed to be orthogonal nor of unit norm. The dual basis vectors are:
Even when the ei and e are not orthonormal, they are still mutually reciprocal:
Then the contravariant components of any vector v can be obtained by the dot product of v with the dual basis vectors:
Likewise, the covariant components of v can be obtained from the dot product of v with basis vectors, viz.
Then v can be expressed in two (reciprocal) ways, viz.
or
Combining the above relations, we have
and we can convert between the basis and dual basis with
and
If the basis vectors are orthonormal, then they are the same as the dual basis vectors.
General Euclidean spaces
More generally, in an n-dimensional Euclidean space V, if a basis is
the reciprocal basis is given by (double indices are summed over),
where the coefficients g are the entries of the inverse matrix of
Indeed, we then have
The covariant and contravariant components of any vector
are related as above by
and
Use in tensor analysis
The distinction between covariance and contravariance is particularly important for computations with tensors, which often have mixed variance. This means that they have both covariant and contravariant components, or both vector and covector components. The valence of a tensor is the number of covariant and contravariant terms, and in Einstein notation, covariant components have lower indices, while contravariant components have upper indices. The duality between covariance and contravariance intervenes whenever a vector or tensor quantity is represented by its components, although modern differential geometry uses more sophisticated index-free methods to represent tensors.
In tensor analysis, a covariant vector varies more or less reciprocally to a corresponding contravariant vector. Expressions for lengths, areas and volumes of objects in the vector space can then be given in terms of tensors with covariant and contravariant indices. Under simple expansions and contractions of the coordinates, the reciprocity is exact; under affine transformations the components of a vector intermingle on going between covariant and contravariant expression.
On a manifold, a tensor field will typically have multiple, upper and lower indices, where Einstein notation is widely used. When the manifold is equipped with a metric, covariant and contravariant indices become very closely related to one another. Contravariant indices can be turned into covariant indices by contracting with the metric tensor. The reverse is possible by contracting with the (matrix) inverse of the metric tensor. Note that in general, no such relation exists in spaces not endowed with a metric tensor. Furthermore, from a more abstract standpoint, a tensor is simply "there" and its components of either kind are only calculational artifacts whose values depend on the chosen coordinates.
The explanation in geometric terms is that a general tensor will have contravariant indices as well as covariant indices, because it has parts that live in the tangent bundle as well as the cotangent bundle.
A contravariant vector is one which transforms like , where are the coordinates of a particle at its proper time . A covariant vector is one which transforms like , where is a scalar field.
Algebra and geometry
In category theory, there are covariant functors and contravariant functors. The assignment of the dual space to a vector space is a standard example of a contravariant functor. Contravariant (resp. covariant) vectors are contravariant (resp. covariant) functors from a -torsor to the fundamental representation of . Similarly, tensors of higher degree are functors with values in other representations of . However, some constructions of multilinear algebra are of "mixed" variance, which prevents them from being functors.
In differential geometry, the components of a vector relative to a basis of the tangent bundle are covariant if they change with the same linear transformation as a change of basis. They are contravariant if they change by the inverse transformation. This is sometimes a source of confusion for two distinct but related reasons. The first is that vectors whose components are covariant (called covectors or 1-forms) actually pull back under smooth functions, meaning that the operation assigning the space of covectors to a smooth manifold is actually a contravariant functor. Likewise, vectors whose components are contravariant push forward under smooth mappings, so the operation assigning the space of (contravariant) vectors to a smooth manifold is a covariant functor. Secondly, in the classical approach to differential geometry, it is not bases of the tangent bundle that are the most primitive object, but rather changes in the coordinate system. Vectors with contravariant components transform in the same way as changes in the coordinates (because these actually change oppositely to the induced change of basis). Likewise, vectors with covariant components transform in the opposite way as changes in the coordinates.
See also
- Active and passive transformation
- Mixed tensor
- Two-point tensor, a generalization allowing indices to reference multiple vector bases
Notes
- A basis f may here profitably be viewed as a linear isomorphism from R to V. Regarding f as a row vector whose entries are the elements of the basis, the associated linear isomorphism is then
Citations
- Misner, C.; Thorne, K.S.; Wheeler, J.A. (1973). Gravitation. W.H. Freeman. ISBN 0-7167-0344-0.
- Frankel, Theodore (2012). The geometry of physics : an introduction. Cambridge: Cambridge University Press. p. 42. ISBN 978-1-107-60260-1. OCLC 739094283.
- Sylvester, J.J. (1851). "On the general theory of associated algebraical forms". Cambridge and Dublin Mathematical Journal. Vol. 6. pp. 289–293.
- Sylvester, J.J. University Press (16 February 2012). The collected mathematical papers of James Joseph Sylvester. Vol. 3, 1870–1883. Cambridge University Press. ISBN 978-1107661431. OCLC 758983870.
- Bowen, Ray; Wang, C.-C. (2008) . "§3.14 Reciprocal Basis and Change of Basis". Introduction to Vectors and Tensors. Dover. pp. 78, 79, 81. ISBN 9780486469140.
References
- Kusse, Bruce R.; Westwig, Erik A. (2010), Mathematical Physics: Applied Mathematics for Scientists and Engineers (2nd ed.), Wiley, ISBN 978-3-527-61814-9.
- Arfken, George B.; Weber, Hans J. (2005), Mathematical Methods for Physicists (6th ed.), Harcourt, ISBN 0-12-059876-0.
- Dodson, C. T. J.; Poston, T. (1991), Tensor geometry, Graduate Texts in Mathematics, vol. 130 (2nd ed.), Springer, ISBN 978-3-540-52018-4, MR 1223091.
- Greub, Werner Hildbert (1967), Multilinear algebra, Die Grundlehren der Mathematischen Wissenschaften, Band 136, Springer, ISBN 9780387038278, MR 0224623.
- Sternberg, Shlomo (1983), Lectures on differential geometry, Chelsea, ISBN 978-0-8284-0316-0.
- Sylvester, J.J. (1853), "On a Theory of the Syzygetic Relations of Two Rational Integral Functions, Comprising an Application to the Theory of Sturm's Functions, and That of the Greatest Algebraical Common Measure", Philosophical Transactions of the Royal Society of London, 143: 407–548, doi:10.1098/rstl.1853.0018, JSTOR 108572.
- Weinreich, Gabriel (1998), Geometrical Vectors, Chicago Lectures in Physics, The University of Chicago Press, p. 126, ISBN 9780226890487
External links
- "Covariant tensor", Encyclopedia of Mathematics, EMS Press, 2001
- "Contravariant tensor", Encyclopedia of Mathematics, EMS Press, 2001
- Weisstein, Eric W. "Covariant Tensor". MathWorld.
- Weisstein, Eric W. "Contravariant Tensor". MathWorld.
- Invariance, Contravariance, and Covariance
- Dullemond, Kees; Peeters, Kasper (2010). "Introduction to tensor calculus" (PDF).
Tensors | |||||
---|---|---|---|---|---|
Glossary of tensor theory | |||||
Scope |
| ||||
Notation | |||||
Tensor definitions | |||||
Operations | |||||
Related abstractions | |||||
Notable tensors |
| ||||
Mathematicians |