In mathematics, especially the usage of linear algebra in mathematical physics and differential geometry, Einstein notation (also known as the Einstein summation convention or Einstein summation notation) is a notational convention that implies summation over a set of indexed terms in a formula, thus achieving brevity. As part of mathematics it is a notational subset of Ricci calculus; however, it is often used in physics applications that do not distinguish between tangent and cotangent spaces. It was introduced to physics by Albert Einstein in 1916.<ref name=Ein1916>Template:Cite journal</ref>
IntroductionEdit
Statement of conventionEdit
According to this convention, when an index variable appears twice in a single term and is not otherwise defined (see Free and bound variables), it implies summation of that term over all the values of the index. So where the indices can range over the set Template:Math, <math display="block">y = \sum_{i = 1}^3 x^i e_i = x^1 e_1 + x^2 e_2 + x^3 e_3 </math> is simplified by the convention to: <math display="block">y = x^i e_i </math>
The upper indices are not exponents but are indices of coordinates, coefficients or basis vectors. That is, in this context Template:Math should be understood as the second component of Template:Math rather than the square of Template:Math (this can occasionally lead to ambiguity). The upper index position in Template:Math is because, typically, an index occurs once in an upper (superscript) and once in a lower (subscript) position in a term (see Template:Section link below). Typically, Template:Math would be equivalent to the traditional Template:Math.
In general relativity, a common convention is that
- the Greek alphabet is used for space and time components, where indices take on values 0, 1, 2, or 3 (frequently used letters are Template:Math),
- the Latin alphabet is used for spatial components only, where indices take on values 1, 2, or 3 (frequently used letters are Template:Math),
In general, indices can range over any indexing set, including an infinite set. This should not be confused with a typographically similar convention used to distinguish between tensor index notation and the closely related but distinct basis-independent abstract index notation.
An index that is summed over is a summation index, in this case "Template:Math". It is also called a dummy index since any symbol can replace "Template:Math" without changing the meaning of the expression (provided that it does not collide with other index symbols in the same term).
An index that is not summed over is a free index and should appear only once per term. If such an index does appear, it usually also appears in every other term in an equation. An example of a free index is the "Template:Math" in the equation <math>v_i = a_i b_j x^j</math>, which is equivalent to the equation <math display="inline">v_i = \sum_j(a_{i} b_{j} x^{j})</math>.
ApplicationEdit
Einstein notation can be applied in slightly different ways. Typically, each index occurs once in an upper (superscript) and once in a lower (subscript) position in a term; however, the convention can be applied more generally to any repeated indices within a term.<ref name="wolfram">{{#invoke:citation/CS1|citation |CitationClass=web }}</ref> When dealing with covariant and contravariant vectors, where the position of an index indicates the type of vector, the first case usually applies; a covariant vector can only be contracted with a contravariant vector, corresponding to summation of the products of coefficients. On the other hand, when there is a fixed coordinate basis (or when not considering coordinate vectors), one may choose to use only subscripts; see Template:Section link below.
Vector representationsEdit
Superscripts and subscripts versus only subscriptsEdit
In terms of covariance and contravariance of vectors,
- upper indices represent components of contravariant vectors (vectors),
- lower indices represent components of covariant vectors (covectors).
They transform contravariantly or covariantly, respectively, with respect to change of basis.
In recognition of this fact, the following notation uses the same symbol both for a vector or covector and its components, as in: <math display="block">\begin{align} v = v^i e_i = \begin{bmatrix} e_1 & e_2 & \cdots & e_n \end{bmatrix} \begin{bmatrix} v^1 \\ v^2 \\ \vdots \\ v^n \end{bmatrix} \\ w = w_i e^i = \begin{bmatrix} w_1 & w_2 & \cdots & w_n \end{bmatrix} \begin{bmatrix} e^1 \\ e^2 \\ \vdots \\ e^n \end{bmatrix} \end{align}</math>
where <math> v </math> is the vector and <math> v^i </math> are its components (not the <math> i </math>th covector <math> v </math>), <math> w </math> is the covector and <math> w_i </math> are its components. The basis vector elements <math>e_i</math> are each column vectors, and the covector basis elements <math>e^i</math> are each row covectors. (See also Template:Slink; duality, below and the examples)
In the presence of a non-degenerate form (an isomorphism Template:Math, for instance a Riemannian metric or Minkowski metric), one can raise and lower indices.
A basis gives such a form (via the dual basis), hence when working on Template:Math with a Euclidean metric and a fixed orthonormal basis, one has the option to work with only subscripts.
However, if one changes coordinates, the way that coefficients change depends on the variance of the object, and one cannot ignore the distinction; see Covariance and contravariance of vectors.
MnemonicsEdit
In the above example, vectors are represented as Template:Math matrices (column vectors), while covectors are represented as Template:Math matrices (row covectors).
When using the column vector convention:
- "Upper indices go up to down; lower indices go left to right."
- "Covariant tensors are row vectors that have indices that are below (co-row-below)."
- Covectors are row vectors: <math display="block">\begin{bmatrix} w_1 & \cdots & w_k \end{bmatrix}.</math> Hence the lower index indicates which column you are in.
- Contravariant vectors are column vectors: <math display="block">\begin{bmatrix} v^1 \\ \vdots \\ v^k \end{bmatrix}</math> Hence the upper index indicates which row you are in.
Abstract descriptionEdit
The virtue of Einstein notation is that it represents the invariant quantities with a simple notation.
In physics, a scalar is invariant under transformations of basis. In particular, a Lorentz scalar is invariant under a Lorentz transformation. The individual terms in the sum are not. When the basis is changed, the components of a vector change by a linear transformation described by a matrix. This led Einstein to propose the convention that repeated indices imply the summation is to be done.
As for covectors, they change by the inverse matrix. This is designed to guarantee that the linear function associated with the covector, the sum above, is the same no matter what the basis is.
The value of the Einstein convention is that it applies to other vector spaces built from Template:Math using the tensor product and duality. For example, Template:Math, the tensor product of Template:Math with itself, has a basis consisting of tensors of the form Template:Math. Any tensor Template:Math in Template:Math can be written as: <math display="block">\mathbf{T} = T^{ij}\mathbf{e}_{ij}.</math>
Template:Math, the dual of Template:Math, has a basis Template:Math, Template:Math, ..., Template:Math which obeys the rule <math display="block">\mathbf{e}^i (\mathbf{e}_j) = \delta^i_j.</math> where Template:Math is the Kronecker delta. As <math display="block">\operatorname{Hom}(V, W) = V^* \otimes W</math> the row/column coordinates on a matrix correspond to the upper/lower indices on the tensor product.
Common operations in this notationEdit
In Einstein notation, the usual element reference <math>A_{mn}</math> for the <math>m</math>-th row and <math>n</math>-th column of matrix <math>A</math> becomes <math>{A^m}_{n}</math>. We can then write the following operations in Einstein notation as follows.
Inner productEdit
The inner product of two vectors is the sum of the products of their corresponding components, with the indices of one vector lowered (see #Raising and lowering indices): <math display="block">\langle\mathbf u,\mathbf v\rangle = \langle\mathbf e_i, \mathbf e_j\rangle u^i v^j = u_j v^j</math> In the case of an orthonormal basis, we have <math>u^j = u_j</math>, and the expression simplifies to: <math display="block">\langle\mathbf u,\mathbf v\rangle = \sum_j u^j v^j = u_j v^j</math>
Vector cross productEdit
In three dimensions, the cross product of two vectors with respect to a positively oriented orthonormal basis, meaning that <math>\mathbf e_1\times\mathbf e_2=\mathbf e_3</math>, can be expressed as: <math display="block">\mathbf{u} \times \mathbf{v} = \varepsilon^i_{\,jk} u^j v^k \mathbf{e}_i</math>
Here, <math>\varepsilon^i_{\,jk} = \varepsilon_{ijk}</math> is the Levi-Civita symbol. Since the basis is orthonormal, raising the index <math>i</math> does not alter the value of <math>\varepsilon_{ijk}</math>, when treated as a tensor.
Matrix-vector multiplicationEdit
The product of a matrix Template:Math with a column vector Template:Math is: <math display="block">\mathbf{u}_{i} = (\mathbf{A} \mathbf{v})_{i} = \sum_{j=1}^N A_{ij} v_{j}</math> equivalent to <math display="block">u^i = {A^i}_j v^j </math>
This is a special case of matrix multiplication.
Matrix multiplicationEdit
The matrix product of two matrices Template:Math and Template:Math is: <math display="block">\mathbf{C}_{ik} = (\mathbf{A} \mathbf{B})_{ik} =\sum_{j=1}^N A_{ij} B_{jk}</math>
equivalent to <math display="block">{C^i}_k = {A^i}_j {B^j}_k</math>
TraceEdit
For a square matrix Template:Math, the trace is the sum of the diagonal elements, hence the sum over a common index Template:Math.
Outer productEdit
The outer product of the column vector Template:Math by the row vector Template:Math yields an Template:Math matrix Template:Math: <math display="block">{A^i}_j = u^i v_j = {(u v)^i}_j</math>
Since Template:Math and Template:Math represent two different indices, there is no summation and the indices are not eliminated by the multiplication.
Raising and lowering indicesEdit
Given a tensor, one can raise an index or lower an index by contracting the tensor with the metric tensor, Template:Math. For example, taking the tensor Template:Math, one can lower an index: <math display="block">g_{\mu\sigma} {T^\sigma}_\beta = T_{\mu\beta}</math>
Or one can raise an index: <math display="block">g^{\mu\sigma} {T_\sigma}^\alpha = T^{\mu\alpha}</math>
See alsoEdit
- Tensor
- Abstract index notation
- Bra–ket notation
- Penrose graphical notation
- Levi-Civita symbol
- DeWitt notation
NotesEdit
- Template:Note labelThis applies only for numerical indices. The situation is the opposite for abstract indices. Then, vectors themselves carry upper abstract indices and covectors carry lower abstract indices, as per the example in the introduction of this article. Elements of a basis of vectors may carry a lower numerical index and an upper abstract index.
ReferencesEdit
BibliographyEdit
External linksEdit
- Template:Cite news
- {{#invoke:citation/CS1|citation
|CitationClass=web }}
- {{#invoke:citation/CS1|citation
|CitationClass=web }}