Trace (linear algebra)
Template:Short description Template:More citations needed In linear algebra, the trace of a square matrix Template:Math, denoted Template:Math,<ref name=":1">{{#invoke:citation/CS1|citation |CitationClass=web }}</ref> is the sum of the elements on its main diagonal, <math>a_{11} + a_{22} + \dots + a_{nn}</math>. It is only defined for a square matrix (Template:Math).
The trace of a matrix is the sum of its eigenvalues (counted with multiplicities). Also, Template:Math for any matrices Template:Math and Template:Math of the same size. Thus, similar matrices have the same trace. As a consequence, one can define the trace of a linear operator mapping a finite-dimensional vector space into itself, since all matrices describing such an operator with respect to a basis are similar.
The trace is related to the derivative of the determinant (see Jacobi's formula).
DefinitionEdit
The trace of an Template:Math square matrix Template:Math is defined as<ref name=":1"/><ref name=":2">Template:Cite encyclopedia </ref><ref name=LipschutzLipson>Template:Cite book</ref>Template:Rp <math display="block">\operatorname{tr}(\mathbf{A}) = \sum_{i=1}^n a_{ii} = a_{11} + a_{22} + \dots + a_{nn}</math> where Template:Math denotes the entry on the Template:Nobr row and Template:Nobr column of Template:Math. The entries of Template:Math can be real numbers, complex numbers, or more generally elements of a field Template:Mvar. The trace is not defined for non-square matrices.
ExampleEdit
Let Template:Math be a matrix, with <math display="block">\mathbf{A} = \begin{pmatrix}
a_{11} & a_{12} & a_{13} \\ a_{21} & a_{22} & a_{23} \\ a_{31} & a_{32} & a_{33}
\end{pmatrix} = \begin{pmatrix}
1 & 0 & 3 \\ 11 & 5 & 2 \\ 6 & 12 & -5
\end{pmatrix} </math>
Then <math display="block">\operatorname{tr}(\mathbf{A}) = \sum_{i=1}^{3} a_{ii} = a_{11} + a_{22} + a_{33} = 1 + 5 + (-5) = 1</math>
PropertiesEdit
Basic propertiesEdit
The trace is a linear mapping. That is,<ref name=":1" /><ref name=":2" /> <math display="block">\begin{align} \operatorname{tr}(\mathbf{A} + \mathbf{B}) &= \operatorname{tr}(\mathbf{A}) + \operatorname{tr}(\mathbf{B}) \\ \operatorname{tr}(c\mathbf{A}) &= c \operatorname{tr}(\mathbf{A}) \end{align}</math> for all square matrices Template:Math and Template:Math, and all scalars Template:Mvar.<ref name="LipschutzLipson"/>Template:Rp
A matrix and its transpose have the same trace:<ref name=":1" /><ref name=":2" /><ref name="LipschutzLipson"/>Template:Rp <math display="block">\operatorname{tr}(\mathbf{A}) = \operatorname{tr}\left(\mathbf{A}^\mathsf{T}\right).</math>
This follows immediately from the fact that transposing a square matrix does not affect elements along the main diagonal.
Trace of a productEdit
The trace of a square matrix which is the product of two matrices can be rewritten as the sum of entry-wise products of their elements, i.e. as the sum of all elements of their Hadamard product. Phrased directly, if Template:Math and Template:Math are two Template:Math matrices, then: <math display="block"> \operatorname{tr}\left(\mathbf{A}^\mathsf{T}\mathbf{B}\right) = \operatorname{tr}\left(\mathbf{A}\mathbf{B}^\mathsf{T}\right) = \operatorname{tr}\left(\mathbf{B}^\mathsf{T}\mathbf{A}\right) = \operatorname{tr}\left(\mathbf{B}\mathbf{A}^\mathsf{T}\right) = \sum_{i=1}^m \sum_{j=1}^n a_{ij}b_{ij} \; . </math>
If one views any real Template:Math matrix as a vector of length Template:Mvar (an operation called vectorization) then the above operation on Template:Math and Template:Math coincides with the standard dot product. According to the above expression, Template:Math is a sum of squares and hence is nonnegative, equal to zero if and only if Template:Math is zero.<ref name="HornJohnson">Template:Cite book</ref>Template:Rp Furthermore, as noted in the above formula, Template:Math. These demonstrate the positive-definiteness and symmetry required of an inner product; it is common to call Template:Math the Frobenius inner product of Template:Math and Template:Math. This is a natural inner product on the vector space of all real matrices of fixed dimensions. The norm derived from this inner product is called the Frobenius norm, and it satisfies a submultiplicative property, as can be proven with the Cauchy–Schwarz inequality: <math display="block">0 \leq \left[\operatorname{tr}(\mathbf{A} \mathbf{B})\right]^2 \leq \operatorname{tr}\left(\mathbf{A}^\mathsf{T} \mathbf{A}\right) \operatorname{tr}\left(\mathbf{B}^\mathsf{T} \mathbf{B}\right) ,</math> if Template:Math and Template:Math are real matrices such that Template:Math is a square matrix. The Frobenius inner product and norm arise frequently in matrix calculus and statistics.
The Frobenius inner product may be extended to a hermitian inner product on the complex vector space of all complex matrices of a fixed size, by replacing Template:Math by its complex conjugate.
The symmetry of the Frobenius inner product may be phrased more directly as follows: the matrices in the trace of a product can be switched without changing the result. If Template:Math and Template:Math are Template:Math and Template:Math real or complex matrices, respectively, then<ref name=":1" /><ref name=":2" /><ref name="LipschutzLipson"/>Template:Rp<ref group="note">This is immediate from the definition of the matrix product: <math display="block">\operatorname{tr}(\mathbf{A}\mathbf{B}) = \sum_{i=1}^m \left(\mathbf{A}\mathbf{B}\right)_{ii} = \sum_{i=1}^m \sum_{j=1}^n a_{ij} b_{ji} = \sum_{j=1}^n \sum_{i=1}^m b_{ji} a_{ij} = \sum_{j=1}^n \left(\mathbf{B}\mathbf{A}\right)_{jj} = \operatorname{tr}(\mathbf{B}\mathbf{A}).</math> </ref>
This is notable both for the fact that Template:Math does not usually equal Template:Math, and also since the trace of either does not usually equal Template:Math.<ref group="note">For example, if <math display="block"> \mathbf{A} = \begin{pmatrix}
0 & 1 \\ 0 & 0
\end{pmatrix},\quad \mathbf{B} = \begin{pmatrix}
0 & 0 \\ 1 & 0
\end{pmatrix}, </math>
then the product is <math display="block">\mathbf{AB} = \begin{pmatrix}
1 & 0 \\ 0 & 0
\end{pmatrix},</math> and the traces are Template:Math.</ref> The similarity-invariance of the trace, meaning that Template:Math for any square matrix Template:Math and any invertible matrix Template:Math of the same dimensions, is a fundamental consequence. This is proved by <math display="block"> \operatorname{tr}\left(\mathbf{P}^{-1}(\mathbf{A}\mathbf{P})\right) = \operatorname{tr}\left((\mathbf{A} \mathbf{P})\mathbf{P}^{-1}\right) = \operatorname{tr}(\mathbf{A}). </math> Similarity invariance is the crucial property of the trace in order to discuss traces of linear transformations as below.
Additionally, for real column vectors <math>\mathbf{a}\in\mathbb{R}^n</math> and <math>\mathbf{b}\in\mathbb{R}^n</math>, the trace of the outer product is equivalent to the inner product: Template:Equation box 1
Cyclic propertyEdit
More generally, the trace is invariant under circular shifts, that is,
This is known as the cyclic property.
Arbitrary permutations are not allowed: in general, <math display="block">\operatorname{tr}(\mathbf{A}\mathbf{B}\mathbf{C}\mathbf{D}) \ne \operatorname{tr}(\mathbf{A}\mathbf{C}\mathbf{B}\mathbf{D}) ~.</math>
However, if products of three symmetric matrices are considered, any permutation is allowed, since: <math display="block">\operatorname{tr}(\mathbf{A}\mathbf{B}\mathbf{C}) = \operatorname{tr}\left(\left(\mathbf{A}\mathbf{B}\mathbf{C}\right)^{\mathsf T}\right) = \operatorname{tr}(\mathbf{C}\mathbf{B}\mathbf{A}) = \operatorname{tr}(\mathbf{A}\mathbf{C}\mathbf{B}),</math> where the first equality is because the traces of a matrix and its transpose are equal. Note that this is not true in general for more than three factors.
Trace of a Kronecker productEdit
The trace of the Kronecker product of two matrices is the product of their traces: <math display="block">\operatorname{tr}(\mathbf{A} \otimes \mathbf{B}) = \operatorname{tr}(\mathbf{A})\operatorname{tr}(\mathbf{B}).</math>
Characterization of the traceEdit
The following three properties: <math display="block">\begin{align} \operatorname{tr}(\mathbf{A} + \mathbf{B}) &= \operatorname{tr}(\mathbf{A}) + \operatorname{tr}(\mathbf{B}), \\ \operatorname{tr}(c\mathbf{A}) &= c \operatorname{tr}(\mathbf{A}), \\ \operatorname{tr}(\mathbf{A}\mathbf{B}) &= \operatorname{tr}(\mathbf{B}\mathbf{A}), \end{align}</math> characterize the trace up to a scalar multiple in the following sense: If <math>f</math> is a linear functional on the space of square matrices that satisfies <math>f(xy) = f(yx),</math> then <math>f</math> and <math>\operatorname{tr}</math> are proportional.<ref group="note">Proof: Let <math>e_{ij}</math> the standard basis and note that <math>f\left(e_{ij}\right) = f\left(e_{i} e_{j}^\top\right) = f\left(e_i e_1^\top e_1 e_j^\top\right) = f\left(e_1 e_j^\top e_i e_1^\top\right) = f\left(0\right) = 0</math> if <math>i \neq j</math> and <math>f\left(e_{jj}\right) = f\left(e_{11}\right)</math> <math display="block">f(\mathbf{A}) = \sum_{i, j} [\mathbf{A}]_{ij} f\left(e_{ij}\right) = \sum_i [\mathbf{A}]_{ii} f\left(e_{11}\right) = f\left(e_{11}\right) \operatorname{tr}(\mathbf{A}).</math>
More abstractly, this corresponds to the decomposition <math display="block">\mathfrak{gl}_n = \mathfrak{sl}_n \oplus k,</math> as <math>\operatorname{tr}(AB) = \operatorname{tr}(BA)</math> (equivalently, <math>\operatorname{tr}([A, B]) = 0</math>) defines the trace on <math>\mathfrak{sl}_n,</math> which has complement the scalar matrices, and leaves one degree of freedom: any such map is determined by its value on scalars, which is one scalar parameter and hence all are multiple of the trace, a nonzero such map.</ref>
For <math>n\times n</math> matrices, imposing the normalization <math>f(\mathbf{I}) = n</math> makes <math>f</math> equal to the trace.
Trace as the sum of eigenvaluesEdit
Given any Template:Math matrix Template:Math, there is
where Template:Math are the eigenvalues of Template:Math counted with multiplicity. This holds true even if Template:Math is a real matrix and some (or all) of the eigenvalues are complex numbers. This may be regarded as a consequence of the existence of the Jordan canonical form, together with the similarity-invariance of the trace discussed above.
Trace of commutatorEdit
When both Template:Math and Template:Math are Template:Math matrices, the trace of the (ring-theoretic) commutator of Template:Math and Template:Math vanishes: Template:Math, because Template:Math and Template:Math is linear. One can state this as "the trace is a map of Lie algebras Template:Math from operators to scalars", as the commutator of scalars is trivial (it is an Abelian Lie algebra). In particular, using similarity invariance, it follows that the identity matrix is never similar to the commutator of any pair of matrices.
Conversely, any square matrix with zero trace is a linear combination of the commutators of pairs of matrices.<ref group="note">Proof: <math>\mathfrak{sl}_n</math> is a semisimple Lie algebra and thus every element in it is a linear combination of commutators of some pairs of elements, otherwise the derived algebra would be a proper ideal.</ref> Moreover, any square matrix with zero trace is unitarily equivalent to a square matrix with diagonal consisting of all zeros.
Traces of special kinds of matricesEdit
Relationship to the characteristic polynomialEdit
The trace of an <math>n \times n</math> matrix <math>A</math> is the coefficient of <math>t^{n-1}</math> in the characteristic polynomial, possibly changed of sign, according to the convention in the definition of the characteristic polynomial.
Relationship to eigenvaluesEdit
If Template:Math is a linear operator represented by a square matrix with real or complex entries and if Template:Math are the eigenvalues of Template:Math (listed according to their algebraic multiplicities), then
This follows from the fact that Template:Math is always similar to its Jordan form, an upper triangular matrix having Template:Math on the main diagonal. In contrast, the determinant of Template:Math is the product of its eigenvalues; that is, <math display="block">\det(\mathbf{A}) = \prod_i \lambda_i.</math>
Everything in the present section applies as well to any square matrix with coefficients in an algebraically closed field.
Derivative relationshipsEdit
If Template:Math is a square matrix with small entries and Template:Math denotes the identity matrix, then we have approximately
<math display="block">\det(\mathbf{I}+\mathbf{\Delta A})\approx 1 + \operatorname{tr}(\mathbf{\Delta A}).</math>
Precisely this means that the trace is the derivative of the determinant function at the identity matrix. Jacobi's formula
<math display="block">d\det(\mathbf{A}) = \operatorname{tr} \big(\operatorname{adj}(\mathbf{A})\cdot d\mathbf{A}\big)</math>
is more general and describes the differential of the determinant at an arbitrary square matrix, in terms of the trace and the adjugate of the matrix.
From this (or from the connection between the trace and the eigenvalues), one can derive a relation between the trace function, the matrix exponential function, and the determinant:<math display="block">\det(\exp(\mathbf{A})) = \exp(\operatorname{tr}(\mathbf{A})).</math>
A related characterization of the trace applies to linear vector fields. Given a matrix Template:Math, define a vector field Template:Math on Template:Math by Template:Math. The components of this vector field are linear functions (given by the rows of Template:Math). Its divergence Template:Math is a constant function, whose value is equal to Template:Math.
By the divergence theorem, one can interpret this in terms of flows: if Template:Math represents the velocity of a fluid at location Template:Math and Template:Mvar is a region in Template:Math, the net flow of the fluid out of Template:Mvar is given by Template:Math, where Template:Math is the volume of Template:Mvar.
The trace is a linear operator, hence it commutes with the derivative: <math display="block">d \operatorname{tr} (\mathbf{X}) = \operatorname{tr}(d\mathbf{X}) .</math>
Trace of a linear operatorEdit
In general, given some linear map Template:Math (where Template:Mvar is a finite-dimensional vector space), we can define the trace of this map by considering the trace of a matrix representation of Template:Mvar, that is, choosing a basis for Template:Mvar and describing Template:Mvar as a matrix relative to this basis, and taking the trace of this square matrix. The result will not depend on the basis chosen, since different bases will give rise to similar matrices, allowing for the possibility of a basis-independent definition for the trace of a linear map.
Such a definition can be given using the canonical isomorphism between the space Template:Math of linear maps on Template:Mvar and Template:Math, where Template:Math is the dual space of Template:Mvar. Let Template:Mvar be in Template:Mvar and let Template:Mvar be in Template:Mvar. Then the trace of the indecomposable element Template:Math is defined to be Template:Math; the trace of a general element is defined by linearity. The trace of a linear map Template:Math can then be defined as the trace, in the above sense, of the element of Template:Math corresponding to f under the above mentioned canonical isomorphism. Using an explicit basis for Template:Mvar and the corresponding dual basis for Template:Math, one can show that this gives the same definition of the trace as given above.
Numerical algorithmsEdit
Stochastic estimatorEdit
The trace can be estimated unbiasedly by "Hutchinson's trick":<ref>Template:Cite journal</ref>
Given any matrix <math>\boldsymbol W\in \R^{n\times n}</math>, and any random <math>\boldsymbol u\in \R^n</math> with <math>\mathbb E[\boldsymbol u\boldsymbol u^\intercal] = \mathbf I</math>, we have <math>\mathbb E[\boldsymbol u^\intercal\boldsymbol W\boldsymbol u ] = \operatorname{tr}\boldsymbol W</math>.
For a proof expand the expectation directly.
Usually, the random vector is sampled from <math>\operatorname N(\mathbf 0,\mathbf I)</math> (normal distribution) or <math>\{\pm n^{-1/2}\}^n</math> (Rademacher distribution).
More sophisticated stochastic estimators of trace have been developed.<ref>Template:Cite journal</ref>
ApplicationsEdit
If a 2 x 2 real matrix has zero trace, its square is a diagonal matrix.
The trace of a 2 × 2 complex matrix is used to classify Möbius transformations. First, the matrix is normalized to make its determinant equal to one. Then, if the square of the trace is 4, the corresponding transformation is parabolic. If the square is in the interval Template:Nowrap, it is elliptic. Finally, if the square is greater than 4, the transformation is loxodromic. See classification of Möbius transformations.
The trace is used to define characters of group representations. Two representations Template:Math of a group Template:Mvar are equivalent (up to change of basis on Template:Mvar) if Template:Math for all Template:Math.
The trace also plays a central role in the distribution of quadratic forms.
Lie algebraEdit
The trace is a map of Lie algebras <math>\operatorname{tr}:\mathfrak{gl}_n\to K</math> from the Lie algebra <math>\mathfrak{gl}_n</math> of linear operators on an Template:Mvar-dimensional space (Template:Math matrices with entries in <math>K</math>) to the Lie algebra Template:Mvar of scalars; as Template:Mvar is Abelian (the Lie bracket vanishes), the fact that this is a map of Lie algebras is exactly the statement that the trace of a bracket vanishes: <math display="block">\operatorname{tr}([\mathbf{A}, \mathbf{B}]) = 0 \text{ for each }\mathbf A,\mathbf B\in\mathfrak{gl}_n.</math>
The kernel of this map, a matrix whose trace is zero, is often said to be Template:Visible anchor or Template:Visible anchor, and these matrices form the simple Lie algebra <math>\mathfrak{sl}_n</math>, which is the Lie algebra of the special linear group of matrices with determinant 1. The special linear group consists of the matrices which do not change volume, while the special linear Lie algebra is the matrices which do not alter volume of infinitesimal sets.
In fact, there is an internal direct sum decomposition <math>\mathfrak{gl}_n = \mathfrak{sl}_n \oplus K</math> of operators/matrices into traceless operators/matrices and scalars operators/matrices. The projection map onto scalar operators can be expressed in terms of the trace, concretely as: <math display="block">\mathbf{A} \mapsto \frac{1}{n}\operatorname{tr}(\mathbf{A})\mathbf{I}.</math>
Formally, one can compose the trace (the counit map) with the unit map <math>K\to\mathfrak{gl}_n</math> of "inclusion of scalars" to obtain a map <math>\mathfrak{gl}_n\to\mathfrak{gl}_n</math> mapping onto scalars, and multiplying by Template:Mvar. Dividing by Template:Mvar makes this a projection, yielding the formula above.
In terms of short exact sequences, one has <math display="block">0 \to \mathfrak{sl}_n \to \mathfrak{gl}_n \overset{\operatorname{tr}}{\to} K \to 0</math> which is analogous to <math display="block">1 \to \operatorname{SL}_n \to \operatorname{GL}_n \overset{\det}{\to} K^* \to 1</math> (where <math>K^*=K\setminus\{0\}</math>) for Lie groups. However, the trace splits naturally (via <math>1/n</math> times scalars) so <math>\mathfrak{gl}_n=\mathfrak{sl}_n\oplus K</math>, but the splitting of the determinant would be as the Template:Mvarth root times scalars, and this does not in general define a function, so the determinant does not split and the general linear group does not decompose: <math display="block">\operatorname{GL}_n \neq \operatorname{SL}_n \times K^*.</math>
Bilinear formsEdit
The bilinear form (where Template:Math, Template:Math are square matrices) <math display="block">B(\mathbf{X}, \mathbf{Y}) = \operatorname{tr}(\operatorname{ad}(\mathbf{X})\operatorname{ad}(\mathbf{Y}))</math>
- where <math>\operatorname{ad}(\mathbf{X})\mathbf{Y} = [\mathbf{X}, \mathbf{Y}] = \mathbf{X}\mathbf{Y} - \mathbf{Y}\mathbf{X}</math>
- and for orientation, if <math>\operatorname{det} \mathbf{Y} \ne 0 </math>
- then <math>\operatorname{ad}(\mathbf{X}) = \mathbf{X} - \mathbf{Y}\mathbf{X}\mathbf{Y}^{-1} ~.</math>
<math> B(\mathbf{X}, \mathbf{Y})</math> is called the Killing form; it is used to classify Lie algebras.
The trace defines a bilinear form: <math display="block">(\mathbf{X}, \mathbf{Y}) \mapsto \operatorname{tr}(\mathbf{X}\mathbf{Y}) ~.</math>
The form is symmetric, non-degenerate<ref group=note>This follows from the fact that Template:Math if and only if Template:Math.</ref> and associative in the sense that: <math display="block">\operatorname{tr}(\mathbf{X}[\mathbf{Y}, \mathbf{Z}]) = \operatorname{tr}([\mathbf{X}, \mathbf{Y}]\mathbf{Z}).</math>
For a complex simple Lie algebra (such as Template:Math), every such bilinear form is proportional to each other; in particular, to the Killing formTemplate:Citation needed.
Two matrices Template:Math and Template:Math are said to be trace orthogonal if <math display="block">\operatorname{tr}(\mathbf{X}\mathbf{Y}) = 0.</math>
There is a generalization to a general representation <math>(\rho,\mathfrak{g},V)</math> of a Lie algebra <math>\mathfrak{g}</math>, such that <math>\rho</math> is a homomorphism of Lie algebras <math>\rho: \mathfrak{g} \rightarrow \text{End}(V).</math> The trace form <math>\text{tr}_V</math> on <math>\text{End}(V)</math> is defined as above. The bilinear form <math display="block">\phi(\mathbf{X},\mathbf{Y}) = \text{tr}_V(\rho(\mathbf{X})\rho(\mathbf{Y}))</math> is symmetric and invariant due to cyclicity.
GeneralizationsEdit
The concept of trace of a matrix is generalized to the trace class of compact operators on Hilbert spaces, and the analog of the Frobenius norm is called the Hilbert–Schmidt norm.
If <math>K</math> is a trace-class operator, then for any orthonormal basis <math>\{e_n\}_{n=1}</math>, the trace is given by <math display="block">\operatorname{tr}(K) = \sum_n \left\langle e_n, Ke_n \right\rangle,</math> and is finite and independent of the orthonormal basis.<ref>Template:Cite book</ref>
The partial trace is another generalization of the trace that is operator-valued. The trace of a linear operator <math>Z</math> which lives on a product space <math>A\otimes B</math> is equal to the partial traces over <math>A</math> and <math>B</math>: <math display="block">\operatorname{tr}(Z) = \operatorname{tr}_A \left(\operatorname{tr}_B(Z)\right) = \operatorname{tr}_B \left(\operatorname{tr}_A(Z)\right).</math>
For more properties and a generalization of the partial trace, see traced monoidal categories.
If <math>A</math> is a general associative algebra over a field <math>k</math>, then a trace on <math>A</math> is often defined to be any functional <math>\operatorname{tr}:A\to k</math> which vanishes on commutators; <math>\operatorname{tr}([a,b])=0</math> for all <math>a,b\in A</math>. Such a trace is not uniquely defined; it can always at least be modified by multiplication by a nonzero scalar.
A supertrace is the generalization of a trace to the setting of superalgebras.
The operation of tensor contraction generalizes the trace to arbitrary tensors.
Gomme and Klein (2011) define a matrix trace operator <math>\operatorname{trm}</math> that operates on block matrices and use it to compute second-order perturbation solutions to dynamic economic models without the need for tensor notation.<ref>Template:Cite journal</ref>
Traces in the language of tensor productsEdit
Given a vector space Template:Mvar, there is a natural bilinear map Template:Math given by sending Template:Math to the scalar Template:Math. The universal property of the tensor product Template:Math automatically implies that this bilinear map is induced by a linear functional on Template:Math.<ref name="kassel">Template:Cite book</ref>
Similarly, there is a natural bilinear map Template:Math given by sending Template:Math to the linear map Template:Math. The universal property of the tensor product, just as used previously, says that this bilinear map is induced by a linear map Template:Math. If Template:Mvar is finite-dimensional, then this linear map is a linear isomorphism.<ref name="kassel" /> This fundamental fact is a straightforward consequence of the existence of a (finite) basis of Template:Mvar, and can also be phrased as saying that any linear map Template:Math can be written as the sum of (finitely many) rank-one linear maps. Composing the inverse of the isomorphism with the linear functional obtained above results in a linear functional on Template:Math. This linear functional is exactly the same as the trace.
Using the definition of trace as the sum of diagonal elements, the matrix formula Template:Math is straightforward to prove, and was given above. In the present perspective, one is considering linear maps Template:Mvar and Template:Mvar, and viewing them as sums of rank-one maps, so that there are linear functionals Template:Math and Template:Math and nonzero vectors Template:Math and Template:Math such that Template:Math and Template:Math for any Template:Mvar in Template:Mvar. Then
- <math>(S\circ T)(u)=\sum_i\varphi_i\left(\sum_j\psi_j(u)w_j\right)v_i=\sum_i\sum_j\psi_j(u)\varphi_i(w_j)v_i </math>
for any Template:Mvar in Template:Mvar. The rank-one linear map Template:Math has trace Template:Math and so
- <math>\operatorname{tr}(S\circ T)=\sum_i\sum_j\psi_j(v_i)\varphi_i(w_j)=\sum_j\sum_i\varphi_i(w_j)\psi_j(v_i).</math>
Following the same procedure with Template:Mvar and Template:Mvar reversed, one finds exactly the same formula, proving that Template:Math equals Template:Math.
The above proof can be regarded as being based upon tensor products, given that the fundamental identity of Template:Math with Template:Math is equivalent to the expressibility of any linear map as the sum of rank-one linear maps. As such, the proof may be written in the notation of tensor products. Then one may consider the multilinear map Template:Math given by sending Template:Math to Template:Math. Further composition with the trace map then results in Template:Math, and this is unchanged if one were to have started with Template:Math instead. One may also consider the bilinear map Template:Math given by sending Template:Math to the composition Template:Math, which is then induced by a linear map Template:Math. It can be seen that this coincides with the linear map Template:Math. The established symmetry upon composition with the trace map then establishes the equality of the two traces.<ref name="kassel" />
For any finite dimensional vector space Template:Mvar, there is a natural linear map Template:Math; in the language of linear maps, it assigns to a scalar Template:Mvar the linear map Template:Math. Sometimes this is called coevaluation map, and the trace Template:Math is called evaluation map.<ref name="kassel" /> These structures can be axiomatized to define categorical traces in the abstract setting of category theory.
See alsoEdit
- Trace of a tensor with respect to a metric tensor
- Characteristic function
- Field trace
- Golden–Thompson inequality
- Singular trace
- Specht's theorem
- Trace class
- Trace identity
- Trace inequalities
- von Neumann's trace inequality