Open main menu
Home
Random
Recent changes
Special pages
Community portal
Preferences
About Wikipedia
Disclaimers
Incubator escapee wiki
Search
User menu
Talk
Dark mode
Contributions
Create account
Log in
Editing
Polar decomposition
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
{{Short description|Representation of invertible matrices as unitary operator multiplying a Hermitian operator}} In [[mathematics]], the '''polar decomposition''' of a square [[real number|real]] or [[complex number|complex]] [[matrix (mathematics)|matrix]] <math>A</math> is a [[matrix decomposition|factorization]] of the form <math>A = U P</math>, where <math>U</math> is a [[unitary matrix]], and <math>P</math> is a [[positive semi-definite matrix|positive semi-definite]] [[Hermitian matrix]] (<math>U</math> is an [[orthogonal matrix]], and <math>P</math> is a positive semi-definite [[symmetric matrix]] in the real case), both square and of the same size.<ref>{{harvnb|Hall|2015|loc=Section 2.5}}.</ref> If a real <math>n \times n</math> matrix <math>A</math> is interpreted as a [[linear transformation]] of <math>n</math>-dimensional [[Cartesian space|space]] <math>\mathbb{R}^n</math>, the polar decomposition separates it into a [[rotation (geometry)|rotation]] or [[reflection (geometry)|reflection]] <math>U</math> of <math>\mathbb{R}^n</math> and a [[scaling (geometry)|scaling]] of the space along a set of <math>n</math> orthogonal axes. The polar decomposition of a square matrix <math>A</math> always exists. If <math>A</math> is [[invertible matrix|invertible]], the decomposition is unique, and the factor <math>P</math> will be [[positive-definite matrix|positive-definite]]. In that case, <math>A</math> can be written uniquely in the form <math>A = U e^X</math>, where <math>U</math> is unitary, and <math>X</math> is the unique self-adjoint [[logarithm of a matrix|logarithm]] of the matrix <math>P</math>.<ref>{{harvnb|Hall|2015|loc=Theorem 2.17}}.</ref> This decomposition is useful in computing the [[fundamental group]] of (matrix) [[Lie group]]s.<ref>{{harvnb|Hall|2015|loc=Section 13.3}}.</ref> The polar decomposition can also be defined as <math>A = P' U</math>, where <math>P' = U P U^{-1}</math> is a symmetric positive-definite matrix with the same eigenvalues as <math>P</math> but different eigenvectors. The polar decomposition of a matrix can be seen as the matrix analog of the [[complex number#Polar form|polar form]] of a [[complex number]] <math>z</math> as <math>z = u r</math>, where <math>r</math> is its [[absolute value#Complex numbers|absolute value]] (a non-negative [[real number]]), and <math>u</math> is a complex number with unit norm (an element of the [[circle group]]). The definition <math>A = UP</math> may be extended to rectangular matrices <math>A \in \mathbb{C}^{m \times n}</math> by requiring <math>U \in \mathbb{C}^{m \times n}</math> to be a [[Semi-orthogonal matrix|semi-unitary]] matrix, and <math>P \in \mathbb{C}^{n \times n}</math> to be a positive-semidefinite Hermitian matrix. The decomposition always exists, and <math>P</math> is always unique. The matrix <math>U</math> is unique if and only if <math>A</math> has full rank.<ref name="higham1990"/> ==Geometric interpretation== A real square <math>m\times m</math> matrix <math>A</math> can be interpreted as the [[linear transformation]] of <math>\mathbb{R}^m</math> that takes a column vector <math>x</math> to <math>A x</math>. Then, in the polar decomposition <math>A = RP</math>, the factor <math>R</math> is an <math>m\times m</math> real orthogonal matrix. The polar decomposition then can be seen as expressing the linear transformation defined by <math>A</math> into a [[scaling (geometry)|scaling]] of the space <math>\mathbb{R}^m</math> along each eigenvector <math>e_i</math> of <math>P</math> by a scale factor <math>\sigma_i</math> (the action of <math>P</math>), followed by a rotation of <math>\mathbb{R}^m</math> (the action of <math>R</math>). Alternatively, the decomposition <math>A=P R</math> expresses the transformation defined by <math>A</math> as a rotation (<math>R</math>) followed by a scaling (<math>P</math>) along certain orthogonal directions. The scale factors are the same, but the directions are different. ==Properties== The polar decomposition of the [[complex conjugate]] of <math>A</math> is given by <math>\overline{A} = \overline{U}\overline{P}.</math> Note that <math display="block"> \det A = \det U \det P = e^{i\theta} r </math> gives the corresponding polar decomposition of the [[determinant]] of ''A'', since <math>\det U = e^{i\theta},</math> and <math>\det P = r = |\det A|.</math> In particular, if <math>A</math> has determinant 1, then both <math>U</math> and <math>P</math> have determinant 1. The positive-semidefinite matrix ''P'' is always unique, even if ''A'' is [[Singular matrices|singular]], and is denoted as <math display="block"> P = (A^* A)^{1/2}, </math> where <math>A^*</math> denotes the [[conjugate transpose]] of <math>A</math>. The uniqueness of ''P'' ensures that this expression is well-defined. The uniqueness is guaranteed by the fact that <math>A^* A</math> is a positive-semidefinite Hermitian matrix and, therefore, has a unique positive-semidefinite Hermitian [[square root of a matrix|square root]].<ref>{{harvnb|Hall|2015|loc=Lemma 2.18}}.</ref> If ''A'' is invertible, then ''P'' is positive-definite, thus also invertible, and the matrix ''U'' is uniquely determined by <math display="block"> U = AP^{-1}. </math> ===Relation to the SVD=== In terms of the [[singular value decomposition]] (SVD) of <math>A</math>, <math>A = W\Sigma V^*</math>, one has <math display="block">\begin{align} P &= V\Sigma V^*, \\ U &= WV^*, \end{align}</math> where <math>U</math>, <math>V</math>, and <math>W</math> are unitary matrices ([[Orthogonal matrix|orthogonal]] if the field is the reals <math>\mathbb{R}</math>). This confirms that <math>P</math> is positive-definite, and <math>U</math> is unitary. Thus, the existence of the SVD is equivalent to the existence of polar decomposition. One can also decompose <math>A</math> in the form <math display="block"> A = P'U. </math> Here <math>U</math> is the same as before, and <math>P'</math> is given by <math display="block"> P' = UPU^{-1} = (AA^*)^{1/2} = W \Sigma W^*. </math> This is known as the left polar decomposition, whereas the previous decomposition is known as the right polar decomposition. Left polar decomposition is also known as reverse polar decomposition. The '''polar decomposition''' of a square invertible real matrix <math>A</math> is of the form <math display="block"> A = [A] R, </math> where <math>[A] \equiv \left(AA^\mathsf{T}\right)^{1/2}</math> is a [[positive-semidefinite matrix|positive-definite]] [[Hermitian matrix|matrix]], and <math>R = [A]^{-1} A</math> is an orthogonal matrix. ===Relation to normal matrices=== The matrix <math>A</math> with polar decomposition <math>A = UP</math> is [[normal matrix|normal]] if and only if <math>U</math> and <math>P</math> [[Commuting matrices|commute]] (<math>UP = PU</math>), or equivalently, they are [[Diagonalizable matrix#Simultaneous diagonalization|simultaneously diagonalizable]]. == Construction and proofs of existence == The core idea behind the construction of the polar decomposition is similar to that used to compute the [[singular-value decomposition]]. === Derivation for normal matrices === If <math>A</math> is [[Normal matrix|normal]], then it is unitarily equivalent to a diagonal matrix: <math>A = V\Lambda V^*</math> for some unitary matrix <math>V</math> and some diagonal matrix <math>\Lambda ~.</math> This makes the derivation of its polar decomposition particularly straightforward, as we can then write <math display="block">A = V\Phi_\Lambda|\Lambda|V^* = \underbrace{\left( V\Phi_\Lambda V^* \right)}_{\equiv U}\underbrace{\left(V|\Lambda|V^* \right)}_{\equiv P},</math> where <math>|\Lambda|</math> is the matrix of absolute diagonal values, and <math>\Phi_\Lambda</math> is a diagonal matrix containing the ''phases'' of the elements of <math>\Lambda,</math> that is, <math>(\Phi_\Lambda)_{ii}\equiv \Lambda_{ii}/ |\Lambda_{ii}|</math> when <math>\Lambda_{ii} \neq 0,</math>, and <math>(\Phi_\Lambda)_{ii} = 0</math> when <math>\Lambda_{ii} = 0 ~.</math> The polar decomposition is thus <math>A=UP,</math> with <math>U</math> and <math>P</math> diagonal in the eigenbasis of <math>A</math> and having eigenvalues equal to the phases and absolute values of those of <math>A,</math> respectively. === Derivation for invertible matrices === From the [[singular-value decomposition]], it can be shown that a matrix <math>A</math> is invertible if and only if <math>A^* A</math> (equivalently, <math>AA^*</math>) is. Moreover, this is true if and only if the eigenvalues of <math>A^* A</math> are all not zero.<ref>Note how this implies, by the positivity of <math>A^* A</math>, that the eigenvalues are all real and strictly positive.</ref> In this case, the polar decomposition is directly obtained by writing <math display="block">A = A\left(A^* A\right)^{-1/2}\left(A^* A\right)^{1/2},</math> and observing that <math>A\left(A^* A\right)^{-1/2}</math> is unitary. To see this, we can exploit the spectral decomposition of <math>A^* A</math> to write <math>A\left(A^* A\right)^{-1/2} = AVD^{-1/2}V^*</math>. In this expression, <math>V^*</math> is unitary because <math>V</math> is. To show that also <math>AVD^{-1/2}</math> is unitary, we can use the [[singular-value decomposition|SVD]] to write <math>A = WD^{1/2}V^*</math>, so that <math display="block">AV D^{-1/2} = WD^{1/2}V^* VD^{-1/2} = W,</math> where again <math>W</math> is unitary by construction. Yet another way to directly show the unitarity of <math>A\left(A^* A\right)^{-1/2}</math> is to note that, writing the [[singular-value decomposition|SVD]] of <math>A</math> in terms of rank-1 matrices as <math display="inline">A = \sum_k s_k v_k w_k^*</math>, where <math>s_k</math>are the singular values of <math>A</math>, we have <math display="block">A\left(A^* A\right)^{-1/2} = \left(\sum_j \lambda_j v_j w_j^*\right)\left(\sum_k |\lambda_k|^{-1} w_k w_k^*\right) = \sum_k \frac{\lambda_k}{|\lambda_k|} v_k w_k^*,</math> which directly implies the unitarity of <math>A\left(A^* A\right)^{-1/2}</math> because a matrix is unitary if and only if its singular values have unitary absolute value. Note how, from the above construction, it follows that ''the unitary matrix in the polar decomposition of an invertible matrix is uniquely defined''. === General derivation === The SVD of a square matrix <math>A</math> reads <math>A = W D^{1/2} V^*</math>, with <math>W, V</math> unitary matrices, and <math>D</math> a diagonal, positive semi-definite matrix. By simply inserting an additional pair of <math>W</math>s or <math>V</math>s, we obtain the two forms of the polar decomposition of <math>A</math>:<math display="block"> A = WD^{1/2}V^* = \underbrace{\left(W D^{1/2} W^*\right)}_P \underbrace{\left(W V^*\right)}_U = \underbrace{\left(W V^*\right)}_U \underbrace{\left(VD^{1/2} V^*\right)}_{P'}. </math>More generally, if <math> A </math> is some rectangular <math> n\times m </math> matrix, its SVD can be written as <math> A=WD^{1/2}V^* </math> where now <math> W </math> and <math> V </math> are isometries with dimensions <math> n\times r </math> and <math> m\times r </math>, respectively, where <math> r\equiv\operatorname{rank}(A) </math>, and <math> D </math> is again a diagonal positive semi-definite square matrix with dimensions <math> r\times r </math>. We can now apply the same reasoning used in the above equation to write <math> A=PU=UP'</math>, but now <math> U\equiv WV^* </math> is not in general unitary. Nonetheless, <math> U </math> has the same support and range as <math> A </math>, and it satisfies <math> U^* U=VV^* </math> and <math> UU^*=WW^* </math>. This makes <math> U </math> into an isometry when its action is restricted onto the support of <math> A </math>, that is, it means that <math> U </math> is a [[partial isometry]]. As an explicit example of this more general case, consider the SVD of the following matrix:<math display="block"> A\equiv \begin{pmatrix}1&1\\2&-2\\0&0\end{pmatrix} = \underbrace{\begin{pmatrix}1&0\\0&1\\0&0\end{pmatrix}}_{\equiv W} \underbrace{\begin{pmatrix}\sqrt2&0\\0&\sqrt8\end{pmatrix}}_{\sqrt D} \underbrace{\begin{pmatrix}\frac1{\sqrt2} & \frac1{\sqrt2} \\ \frac1{\sqrt2} & -\frac1{\sqrt2}\end{pmatrix}}_{V^\dagger}. </math>We then have<math display="block"> WV^\dagger = \frac1{\sqrt2}\begin{pmatrix}1&1 \\ 1&-1 \\ 0&0\end{pmatrix} </math>which is an isometry, but not unitary. On the other hand, if we consider the decomposition of<math display="block"> A\equiv \begin{pmatrix}1&0&0\\0&2&0\end{pmatrix} = \begin{pmatrix}1&0\\0&1\end{pmatrix} \begin{pmatrix}1&0\\0&2\end{pmatrix} \begin{pmatrix}1&0&0\\0&1&0\end{pmatrix}, </math>we find<math display="block"> WV^\dagger =\begin{pmatrix}1&0&0\\0&1&0\end{pmatrix}, </math>which is a partial isometry (but not an isometry). ==Bounded operators on Hilbert space== The '''polar decomposition''' of any [[bounded linear operator]] ''A'' between complex [[Hilbert space]]s is a canonical factorization as the product of a [[partial isometry]] and a non-negative operator. The polar decomposition for matrices generalizes as follows: if ''A'' is a bounded linear operator then there is a unique factorization of ''A'' as a product ''A'' = ''UP'' where ''U'' is a partial isometry, ''P'' is a non-negative self-adjoint operator and the initial space of ''U'' is the closure of the range of ''P''. The operator ''U'' must be weakened to a partial isometry, rather than unitary, because of the following issues. If ''A'' is the [[shift operator|one-sided shift]] on ''l''<sup>2</sup>('''N'''), then |''A''| = {''A{{sup|*}}A''}<sup>1/2</sup> = ''I''. So if ''A'' = ''U'' |''A''|, ''U'' must be ''A'', which is not unitary. The existence of a polar decomposition is a consequence of [[Douglas' lemma]]: {{math theorem|name=Lemma|math_statement= If ''A'', ''B'' are bounded operators on a Hilbert space ''H'', and ''A{{sup|*}}A'' β€ ''B{{sup|*}}B'', then there exists a contraction ''C'' such that ''A = CB''. Furthermore, ''C'' is unique if ker(''B{{sup|*}}'') ⊂ ker(''C'').}} The operator ''C'' can be defined by ''C''(''Bh'') := ''Ah'' for all ''h'' in ''H'', extended by continuity to the closure of ''Ran''(''B''), and by zero on the orthogonal complement to all of ''H''. The lemma then follows since ''A{{sup|*}}A'' β€ ''B{{sup|*}}B'' implies ker(''B'') β ker(''A''). In particular. If ''A{{sup|*}}A'' = ''B{{sup|*}}B'', then ''C'' is a partial isometry, which is unique if ker(''B{{sup|*}}'') β ker(''C''). In general, for any bounded operator ''A'', <math display="block">A^*A = \left(A^*A\right)^{1/2} \left(A^*A\right)^{1/2},</math> where (''A{{sup|*}}A'')<sup>1/2</sup> is the unique positive square root of ''A{{sup|*}}A'' given by the usual [[functional calculus]]. So by the lemma, we have <math display="block">A = U\left(A^*A\right)^{1/2}</math> for some partial isometry ''U'', which is unique if ker(''A{{sup|*}}'') β ker(''U''). Take ''P'' to be (''A{{sup|*}}A'')<sup>1/2</sup> and one obtains the polar decomposition ''A'' = ''UP''. Notice that an analogous argument can be used to show ''A = P'U{{'}}'', where ''P' '' is positive and ''U{{'}}'' a partial isometry. When ''H'' is finite-dimensional, ''U'' can be extended to a unitary operator; this is not true in general (see example above). Alternatively, the polar decomposition can be shown using the operator version of [[singular value decomposition#Bounded operators on Hilbert spaces|singular value decomposition]]. By property of the [[continuous functional calculus]], |''A''| is in the [[C*-algebra]] generated by ''A''. A similar but weaker statement holds for the partial isometry: ''U'' is in the [[von Neumann algebra]] generated by ''A''. If ''A'' is invertible, the polar part ''U'' will be in the [[C*-algebra]] as well. ==Unbounded operators== If ''A'' is a closed, densely defined [[unbounded operator]] between complex Hilbert spaces then it still has a (unique) '''polar decomposition''' <math display="block"> A = U |A|, </math> where |''A''| is a (possibly unbounded) non-negative self-adjoint operator with the same domain as ''A'', and ''U'' is a partial isometry vanishing on the orthogonal complement of the range ran(|''A''|). The proof uses the same lemma as above, which goes through for unbounded operators in general. If dom(''A''{{sup|*}}''A'') = dom(''B{{sup|*}}B''), and ''A''{{sup|*}}''Ah'' = ''B''{{sup|*}}''Bh'' for all ''h'' β dom(''A''{{sup|*}}''A''), then there exists a partial isometry ''U'' such that ''A'' = ''UB''. ''U'' is unique if ran(''B'')<sup>β₯</sup> β ker(''U''). The operator ''A'' being closed and densely defined ensures that the operator ''A''{{sup|*}}''A'' is self-adjoint (with dense domain) and therefore allows one to define (''A''{{sup|*}}''A'')<sup>1/2</sup>. Applying the lemma gives polar decomposition. If an unbounded operator ''A'' is [[affiliated operator|affiliated]] to a von Neumann algebra '''M''', and ''A'' = ''UP'' is its polar decomposition, then ''U'' is in '''M''' and so is the spectral projection of ''P'', 1<sub>''B''</sub>(''P''), for any Borel set ''B'' in {{closed-open|0, β}}. ==Quaternion polar decomposition== The polar decomposition of [[quaternion]]s <math>\mathbb{H}</math> with [[orthonormal basis]] quaternions <math>1, \hat\imath, \hat\jmath, \hat k</math> depends on the unit 2-dimensional sphere <math>\hat r \in \{x \hat \imath + y \hat \jmath + z \hat k \in \mathbb{H} \setminus \mathbb{R} : x^2 + y^2 + z^2 = 1\}</math> of [[quaternion#Square roots of β1|square roots of minus one]], known as ''[[right versor]]s''. Given any <math>\hat r</math> on this sphere and an angle {{nobr|{{math|β{{pi}} < ''a'' β€ {{pi}}}},}} the [[versor]] <math>e^{a \hat r} = \cos a + \hat r \sin a</math> is on the unit [[3-sphere]] of <math>\mathbb{H}.</math> For {{nobr|{{math|''a'' {{=}} 0}}}} and {{nobr|{{math|''a'' {{=}} {{pi}}}},}} the versor is 1 or β1, regardless of which {{mvar|r}} is selected. The [[norm (mathematics)|norm]] {{mvar|t}} of a quaternion {{mvar|q}} is the [[Euclidean distance]] from the origin to {{mvar|q}}. When a quaternion is not just a real number, then there is a ''unique'' polar decomposition: <math display="block"> q = t \exp(a \hat r). </math> Here {{mvar|r}}, {{mvar|a}}, {{mvar|t}} are all uniquely determined such that {{mvar|r}} is a right versor {{nobr|({{math|''r''<sup>2</sup> {{=}} β1}}),}} {{mvar|a}} satisfies {{nobr|{{math|0 < ''a'' < {{pi}}}},}} and {{nobr|{{math|''t'' > 0}}.}} ==Alternative planar decompositions== In the [[Cartesian plane]], alternative planar [[ring (mathematics)|ring]] decompositions arise as follows: * If {{math|''x'' β 0}}, {{math|1=''z'' = ''x''(1 + Ξ΅(''y''/''x''))}} is a polar decomposition of a [[dual number]] {{math|1=''z'' = ''x'' + ''yΞ΅''}}, where {{math|1=''Ξ΅''<sup>2</sup> = 0}}; i.e., ''Ξ΅'' is [[nilpotent]]. In this polar decomposition, the unit circle has been replaced by the line {{math|1=''x'' = 1}}, the polar angle by the [[slope]] ''y''/''x'', and the radius ''x'' is negative in the left half-plane. * If {{math|''x''<sup>2</sup> β ''y''<sup>2</sup>}}, then the [[unit hyperbola]] {{math|1=''x''<sup>2</sup> β ''y''<sup>2</sup> = 1}}, and [[conjugate hyperbola|its conjugate]] {{math|1=''x''<sup>2</sup> β ''y''<sup>2</sup> = β1}} can be used to form a polar decomposition based on the branch of the unit hyperbola through {{math|(1, 0)}}. This branch is parametrized by the [[hyperbolic angle]] ''a'' and is written <math display="block"> \cosh a + j \sinh a = \exp(aj) = e^{aj}, </math> where {{math|1=''j''<sup>2</sup> = +1}}, and the arithmetic<ref>Sobczyk, G. (1995) "Hyperbolic Number Plane", ''[[College Mathematics Journal]]'' 26:268β280.</ref> of [[split-complex number]]s is used. The branch through {{math|(β1, 0)}} is traced by β''e''<sup>''aj''</sup>. Since the operation of multiplying by ''j'' reflects a point across the line {{math|1=''y'' = ''x''}}, the conjugate hyperbola has branches traced by ''je''<sup>''aj''</sup> or β''je''<sup>''aj''</sup>. Therefore a point in one of the quadrants has a polar decomposition in one of the forms: <math display="block"> r e^{aj}, -re^{aj}, rje^{aj}, -rje^{aj}, \quad r > 0. </math> The set {{math|{1, β1, ''j'', β''j''}<nowiki/>}} has products that make it isomorphic to the [[Klein four-group]]. Evidently polar decomposition in this case involves an element from that group. Polar decomposition of an element of the [[algebra over a field|algebra]] M(2, R) of 2 Γ 2 real matrices uses these alternative planar decompositions since any planar [[subalgebra]] is isomorphic to dual numbers, split-complex numbers, or ordinary complex numbers. ==Numerical determination of the matrix polar decomposition== To compute an approximation of the polar decomposition ''A'' = ''UP'', usually the unitary factor ''U'' is approximated.<ref name="higham1986" /><ref name="byers2008" /> The iteration is based on [[Heron's method]] for the square root of ''1'' and computes, starting from <math>U_0 = A</math>, the sequence <math display="block">U_{k+1} = \frac{1}{2}\left(U_k + \left(U_k^*\right )^{-1}\right),\qquad k = 0, 1, 2, \ldots</math> The combination of inversion and Hermite conjugation is chosen so that in the singular value decomposition, the unitary factors remain the same and the iteration reduces to Heron's method on the singular values. This basic iteration may be refined to speed up the process: {{bulleted list | Every step or in regular intervals, the range of the singular values of <math>U_k</math> is estimated and then the matrix is rescaled to <math>\gamma_k U_k</math> to center the singular values around ''1''. The scaling factor <math>\gamma_k</math> is computed using matrix norms of the matrix and its inverse. Examples of such scale estimates are: <math display="block">\gamma_k = \sqrt[4]{\frac{\left\|U_k^{-1}\right\|_1 \left\|U_k^{-1}\right\|_\infty}{\left\|U_k\right\|_1 \left\|U_k\right\|_\infty} }</math> using the row-sum and column-sum [[induced norm|matrix norms]] or <math display="block">\gamma_k = \sqrt{\frac{\left\|U_k^{-1}\right\|_F}{\left\|U_k\right\|_F}}</math> using the [[Frobenius norm]]. Including the scale factor, the iteration is now <math display="block">U_{k+1} = \frac{1}{2}\left(\gamma_k U_k + \frac{1}{\gamma_k} \left(U_k^*\right)^{-1}\right), \qquad k = 0, 1, 2, \ldots</math> | The [[QR decomposition]] can be used in a preparation step to reduce a singular matrix ''A'' to a smaller regular matrix, and inside every step to speed up the computation of the inverse. | Heron's method for computing roots of <math>x^2 - 1 = 0</math> can be replaced by higher order methods, for instance based on [[Halley's method]] of third order, resulting in <math display="block">U_{k+1} = U_k\left(I + 3U_k^* U_k\right)^{-1}\left(3I + U_k^* U_k\right),\qquad k = 0, 1, 2, \ldots</math> This iteration can again be combined with rescaling. This particular formula has the benefit that it is also applicable to singular or rectangular matrices ''A''. }} ==See also== *[[Cartan decomposition]] *[[Matrix decomposition#Algebraic polar decomposition|Algebraic polar decomposition]] *[[Complex measure#Variation of a complex measure and polar decomposition|Polar decomposition of a complex measure]] *[[Lie group decomposition]] == References == {{reflist|refs= <ref name="higham1986"> {{cite journal |last = Higham |first = Nicholas J. |title = Computing the polar decomposition with applications |journal = SIAM J. Sci. Stat. Comput. |volume = 7 |number = 4 |year = 1986 |issn = 0196-5204 |pages = 1160β1174 |doi = 10.1137/0907079 |citeseerx = 10.1.1.137.7354 |publisher = Society for Industrial and Applied Mathematics |location = Philadelphia, PA, USA }} </ref> <ref name="higham1990"> {{cite journal |last1 = Higham |first1 = Nicholas J. |last2 = Schreiber |first2 = Robert S. |title = Fast polar decomposition of an arbitrary matrix |journal = SIAM J. Sci. Stat. Comput. |volume = 11 |number = 4 |year = 1990 |issn = 0196-5204 |pages = 648β655 |doi = 10.1137/0911038 |citeseerx = 10.1.1.111.9239 |publisher = Society for Industrial and Applied Mathematics |location = Philadelphia, PA, USA |s2cid = 14268409 }} </ref> <ref name="byers2008"> {{cite journal | last=Byers | first=Ralph |author2=Hongguo Xu | title=A New Scaling for Newton's Iteration for the Polar Decomposition and its Backward Stability | journal=SIAM J. Matrix Anal. Appl. | volume=30 | number = 2 | year = 2008 | issn = 0895-4798 | pages = 822β843 | doi = 10.1137/070699895 | publisher = Society for Industrial and Applied Mathematics | location = Philadelphia, PA, USA | citeseerx=10.1.1.378.6737 }} </ref> }} * {{cite book |author-link=John B. Conway |last=Conway |first=J. B. |title=A Course in Functional Analysis |series=[[Graduate Texts in Mathematics]] |place=New York |publisher=Springer |year=1990 |doi=10.1007/978-1-4757-4383-8 }} * {{cite journal |author-link=Ronald G. Douglas |last=Douglas |first=R. G. |title=On Majorization, Factorization, and Range Inclusion of Operators on Hilbert Space |journal=[[Proceedings of the American Mathematical Society|Proc. Amer. Math. Soc.]] |volume=17 |pages=413β415 |year=1966 |doi=10.1090/S0002-9939-1966-0203464-1 }} * {{cite book |first=Brian C. |last=Hall |title=Lie Groups, Lie Algebras, and Representations: An Elementary Introduction |edition=2nd |series=Graduate Texts in Mathematics |volume=222 |publisher=Springer |year=2015 |isbn=978-3319134666}} * {{cite book |first=Sigurdur |last=Helgason |authorlink=Sigurdur Helgason (mathematician) |title=Differential geometry, Lie groups, and symmetric spaces |year=1978 |publisher=Academic Press |isbn= 0-8218-2848-7 }} {{Functional Analysis}} {{SpectralTheory}} [[Category:Lie groups]] [[Category:Operator theory]] [[Category:Matrix theory]] [[Category:Matrix decompositions]]
Edit summary
(Briefly describe your changes)
By publishing changes, you agree to the
Terms of Use
, and you irrevocably agree to release your contribution under the
CC BY-SA 4.0 License
and the
GFDL
. You agree that a hyperlink or URL is sufficient attribution under the Creative Commons license.
Cancel
Editing help
(opens in new window)
Pages transcluded onto the current version of this page
(
help
)
:
Template:'
(
edit
)
Template:Bulleted list
(
edit
)
Template:Cite book
(
edit
)
Template:Cite journal
(
edit
)
Template:Closed-open
(
edit
)
Template:Functional Analysis
(
edit
)
Template:Harvnb
(
edit
)
Template:Math
(
edit
)
Template:Math theorem
(
edit
)
Template:Mvar
(
edit
)
Template:Nobr
(
edit
)
Template:Reflist
(
edit
)
Template:Short description
(
edit
)
Template:SpectralTheory
(
edit
)
Template:Sup
(
edit
)