Open main menu
Home
Random
Recent changes
Special pages
Community portal
Preferences
About Wikipedia
Disclaimers
Incubator escapee wiki
Search
User menu
Talk
Dark mode
Contributions
Create account
Log in
Editing
Lindemann–Weierstrass theorem
(section)
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
===Proof=== The proof relies on two preliminary [[Lemma_(mathematics)|lemmas]]. Notice that Lemma B itself is already sufficient to deduce the original statement of Lindemann–Weierstrass theorem. ====Preliminary lemmas==== {{math_theorem|name=Lemma A.|math_statement=Let {{math|''c''(1), ..., ''c''(''r'')}} be [[integer]]s and, for every {{mvar|k}} between {{math|1}} and {{mvar|r}}, let {{math|{''γ''(''k'')<sub>1</sub>, ..., ''γ''(''k'')<sub>''m''(''k'')</sub>} }} be the roots of a non-zero [[polynomial]] with integer coefficients <math>T_k(x)</math>. If {{math|''γ''(''k'')<sub>''i''</sub> ≠ ''γ''(''u'')<sub>''v''</sub> }} whenever {{math|(''k'', ''i'') ≠ (''u'', ''v'')}}, then :<math>c(1)\left (e^{\gamma(1)_1}+\cdots+ e^{\gamma(1)_{m(1)}} \right ) + \cdots + c(r) \left (e^{\gamma(r)_1}+\cdots+ e^{\gamma(r)_{m(r)}} \right) = 0</math> has only the trivial solution <math>c(i)=0</math> for all <math>i = 1, \dots, r.</math>}} '''Proof of Lemma A.''' To simplify the notation set: :<math> \begin{align} & n_0 =0, & & \\ & n_i =\sum\nolimits_{k=1}^i m(k), & & i=1,\ldots,r \\ & n=n_r, & & \\ & \alpha_{n_{i-1}+j} =\gamma(i)_j, & & 1\leq i\leq r,\ 1\leq j\leq m(i) \\ & \beta_{n_{i-1}+j} =c(i). \end{align} </math> Then the statement becomes :<math>\sum_{k=1}^n \beta_k e^{\alpha_k}\neq 0.</math> Let {{mvar|p}} be a [[prime number]] and define the following polynomials: : <math>f_i(x) = \frac {\ell^{np} (x-\alpha_1)^p \cdots (x-\alpha_n)^p}{(x-\alpha_i)},</math> where {{mvar|ℓ}} is a non-zero integer such that <math>\ell\alpha_1,\ldots,\ell\alpha_n</math> are all [[algebraic integers#Facts|algebraic integers]]. Define<ref>Up to a factor, this is the same integral appearing in [[Transcendental number#A proof that e is transcendental|the proof that {{mvar|e}} is a transcendental number]], where {{math|''β''<sub>1</sub> {{=}} 1, ..., ''β<sub>m</sub>'' {{=}} ''m''.}} The rest of the proof of the Lemma is analog to that proof.</ref> : <math>I_i(s) = \int^s_0 e^{s-x} f_i(x) \, dx.</math> Using [[integration by parts]] we arrive at : <math>I_i(s) = e^s \sum_{j=0}^{np-1} f_i^{(j)}(0) - \sum_{j=0}^{np-1} f_i^{(j)}(s),</math> where <math>np-1</math> is the [[Degree of a polynomial|degree]] of <math>f_i</math>, and <math>f_i^{(j)}</math> is the ''j''-th derivative of <math>f_i</math>. This also holds for ''s'' complex (in this case the integral has to be intended as a contour integral, for example along the straight segment from 0 to ''s'') because :<math>-e^{s-x} \sum_{j=0}^{np-1} f_i^{(j)}(x)</math> is a primitive of <math>e^{s-x} f_i(x)</math>. Consider the following sum: :<math>\begin{align} J_i &=\sum_{k=1}^n\beta_k I_i(\alpha_k)\\[5pt] &= \sum_{k=1}^n\beta_k \left ( e^{\alpha_k} \sum_{j=0}^{np-1} f_i^{(j)}(0) - \sum_{j=0}^{np-1} f_i^{(j)}(\alpha_k)\right ) \\[5pt] &=\left(\sum_{j=0}^{np-1}f_i^{(j)}(0)\right)\left(\sum_{k=1}^n \beta_k e^{\alpha_k}\right)-\sum_{k=1}^n\sum_{j=0}^{np-1} \beta_kf_i^{(j)}(\alpha_k)\\[5pt] &= -\sum_{k=1}^n \sum_{j=0}^{np-1} \beta_kf_i^{(j)}(\alpha_k) \end{align}</math> In the last line we assumed that the conclusion of the Lemma is false. In order to complete the proof we need to reach a contradiction. We will do so by estimating <math>|J_1\cdots J_n|</math> in two different ways. First <math>f_i^{(j)}(\alpha_k)</math> is an [[algebraic integer]] which is divisible by ''p''! for <math>j\geq p</math> and vanishes for <math>j<p</math> unless <math>j=p-1</math> and <math>k=i</math>, in which case it equals :<math>\ell^{np}(p-1)!\prod_{k\neq i}(\alpha_i-\alpha_k)^p.</math> This is not divisible by ''p'' when ''p'' is large enough because otherwise, putting :<math>\delta_i=\prod_{k\neq i}(\ell\alpha_i-\ell\alpha_k)</math> (which is a non-zero algebraic integer) and calling <math>d_i\in\mathbb Z</math> the product of its conjugates (which is still non-zero), we would get that ''p'' divides <math>\ell^p(p-1)!d_i^p</math>, which is false. So <math>J_i</math> is a non-zero algebraic integer divisible by (''p'' − 1)!. Now :<math>J_i=-\sum_{j=0}^{np-1}\sum_{t=1}^r c(t)\left(f_i^{(j)}(\alpha_{n_{t-1}+1}) + \cdots + f_i^{(j)}(\alpha_{n_t})\right).</math> Since each <math>f_i(x)</math> is obtained by dividing a fixed polynomial with integer coefficients by <math>(x-\alpha_i)</math>, it is of the form :<math>f_i(x)=\sum_{m=0}^{np-1}g_m(\alpha_i)x^m, </math> where <math>g_m</math> is a polynomial (with integer coefficients) independent of ''i''. The same holds for the derivatives <math>f_i^{(j)}(x)</math>. Hence, by the [[fundamental theorem of symmetric polynomials]], :<math>f_i^{(j)}(\alpha_{n_{t-1}+1})+\cdots+f_i^{(j)}(\alpha_{n_t})</math> is a fixed polynomial with rational coefficients evaluated in <math>\alpha_i</math> (this is seen by grouping the same powers of <math>\alpha_{n_{t-1}+1},\dots,\alpha_{n_t}</math> appearing in the expansion and using the fact that these algebraic numbers are a complete set of conjugates). So the same is true of <math>J_i</math>, i.e. it equals <math>G(\alpha_i)</math>, where ''G'' is a polynomial with rational coefficients independent of ''i''. Finally <math>J_1\cdots J_n=G(\alpha_1)\cdots G(\alpha_n)</math> is rational (again by the fundamental theorem of symmetric polynomials) and is a non-zero algebraic integer divisible by <math>(p-1)!^n</math> (since the <math>J_i</math>'s are algebraic integers divisible by <math>(p-1)!</math>). Therefore :<math>|J_1\cdots J_n|\geq (p-1)!^n.</math> However one clearly has: :<math>|I_i(\alpha_k)| \leq {|\alpha_k|} e^{|\alpha_k|} F_i({|\alpha_k|}),</math> where {{mvar|F<sub>i</sub>}} is the polynomial whose coefficients are the absolute values of those of ''f''<sub>''i''</sub> (this follows directly from the definition of <math>I_i(s)</math>). Thus :<math>|J_i|\leq \sum_{k=1}^n \left |\beta_k\alpha_k \right |e^{|\alpha_k|}F_i \left ( \left |\alpha_k \right| \right )</math> and so by the construction of the <math>f_i</math>'s we have <math>|J_1\cdots J_n|\le C^p</math> for a sufficiently large ''C'' independent of ''p'', which contradicts the previous inequality. This proves Lemma A. ∎ {{math_theorem|name=Lemma B.|math_statement=If ''b''(1), ..., ''b''(''n'') are integers and ''γ''(1), ..., ''γ''(''n''), are distinct [[algebraic number]]s, then :<math>b(1)e^{\gamma(1)}+\cdots+ b(n)e^{\gamma(n)} = 0</math> has only the trivial solution <math>b(i)=0</math> for all <math>i = 1, \dots, n.</math>}} '''Proof of Lemma B:''' Assuming :<math>b(1)e^{\gamma(1)}+\cdots+ b(n)e^{\gamma(n)}= 0,</math> we will derive a contradiction, thus proving Lemma B. Let us choose a polynomial with integer coefficients which vanishes on all the <math>\gamma(k)</math>'s and let <math>\gamma(1),\ldots,\gamma(n),\gamma(n+1),\ldots,\gamma(N)</math> be all its distinct roots. Let ''b''(''n'' + 1) = ... = ''b''(''N'') = 0. The polynomial :<math>P(x_1,\dots,x_N)=\prod_{\sigma\in S_N}(b(1) x_{\sigma(1)}+\cdots+b(N) x_{\sigma(N)})</math> vanishes at <math>(e^{\gamma(1)},\dots,e^{\gamma(N)})</math> by assumption. Since the product is symmetric, for any <math>\tau\in S_N</math> the monomials <math>x_{\tau(1)}^{h_1}\cdots x_{\tau(N)}^{h_N}</math> and <math>x_1^{h_1}\cdots x_N^{h_N}</math> have the same coefficient in the expansion of ''P''. Thus, expanding <math>P(e^{\gamma(1)},\dots,e^{\gamma(N)})</math> accordingly and grouping the terms with the same exponent, we see that the resulting exponents <math>h_1\gamma(1)+\dots+h_N\gamma(N)</math> form a complete set of conjugates and, if two terms have conjugate exponents, they are multiplied by the same coefficient. So we are in the situation of Lemma A. To reach a contradiction it suffices to see that at least one of the coefficients is non-zero. This is seen by equipping {{math|'''C'''}} with the lexicographic order and by choosing for each factor in the product the term with non-zero coefficient which has maximum exponent according to this ordering: the product of these terms has non-zero coefficient in the expansion and does not get simplified by any other term. This proves Lemma B. ∎ ====Final step==== We turn now to prove the theorem: Let ''a''(1), ..., ''a''(''n'') be non-zero [[algebraic number]]s, and ''α''(1), ..., ''α''(''n'') distinct algebraic numbers. Then let us assume that: : <math>a(1)e^{\alpha(1)}+\cdots + a(n)e^{\alpha(n)} = 0.</math> We will show that this leads to contradiction and thus prove the theorem. The proof is very similar to that of Lemma B, except that this time the choices are made over the ''a''(''i'')'s: For every ''i'' ∈ {1, ..., ''n''}, ''a''(''i'') is algebraic, so it is a root of an [[irreducible polynomial]] with integer coefficients of degree ''d''(''i''). Let us denote the distinct roots of this polynomial ''a''(''i'')<sub>1</sub>, ..., ''a''(''i'')<sub>''d''(''i'')</sub>, with ''a''(''i'')<sub>1</sub> = ''a''(''i''). Let S be the functions σ which choose one element from each of the sequences (1, ..., ''d''(1)), (1, ..., ''d''(2)), ..., (1, ..., ''d''(''n'')), so that for every 1 ≤ ''i'' ≤ ''n'', σ(''i'') is an integer between 1 and ''d''(''i''). We form the polynomial in the variables <math>x_{11},\dots,x_{1d(1)},\dots,x_{n1},\dots,x_{nd(n)},y_1,\dots,y_n</math> : <math>Q(x_{11},\dots,x_{nd(n)},y_1,\dots,y_n)=\prod\nolimits_{\sigma\in S}\left(x_{1\sigma(1)}y_1+\dots+x_{n\sigma(n)}y_n\right).</math> Since the product is over all the possible choice functions σ, ''Q'' is symmetric in <math>x_{i1},\dots,x_{id(i)}</math> for every ''i''. Therefore ''Q'' is a polynomial with integer coefficients in elementary symmetric polynomials of the above variables, for every ''i'', and in the variables ''y''<sub>''i''</sub>. Each of the latter symmetric polynomials is a rational number when evaluated in <math>a(i)_1,\dots,a(i)_{d(i)}</math>. The evaluated polynomial <math>Q(a(1)_1,\dots,a(n)_{d(n)},e^{\alpha(1)},\dots,e^{\alpha(n)})</math> vanishes because one of the choices is just σ(''i'') = 1 for all ''i'', for which the corresponding factor vanishes according to our assumption above. Thus, the evaluated polynomial is a sum of the form : <math>b(1)e^{\beta(1)}+ b(2)e^{\beta(2)}+ \cdots + b(N)e^{\beta(N)}= 0,</math> where we already grouped the terms with the same exponent. So in the left-hand side we have distinct values β(1), ..., β(''N''), each of which is still algebraic (being a sum of algebraic numbers) and coefficients <math>b(1),\dots,b(N)\in\mathbb Q</math>. The sum is nontrivial: if <math>\alpha(i)</math> is maximal in the lexicographic order, the coefficient of <math>e^{|S|\alpha(i)}</math> is just a product of ''a''(''i'')<sub>''j''</sub>'s (with possible repetitions), which is non-zero. By multiplying the equation with an appropriate integer factor, we get an identical equation except that now ''b''(1), ..., ''b''(''N'') are all integers. Therefore, according to Lemma B, the equality cannot hold, and we are led to a contradiction which completes the proof. ∎ Note that Lemma A is sufficient to prove that ''e'' is [[irrational number|irrational]], since otherwise we may write ''e'' = ''p'' / ''q'', where both ''p'' and ''q'' are non-zero integers, but by Lemma A we would have ''qe'' − ''p'' ≠ 0, which is a contradiction. Lemma A also suffices to prove that {{pi}} is irrational, since otherwise we may write {{pi}} = ''k'' / ''n'', where both ''k'' and ''n'' are integers) and then ±''i''{{pi}} are the roots of ''n''<sup>2</sup>''x''<sup>2</sup> + ''k''<sup>2</sup> = 0; thus 2 − 1 − 1 = 2''e''<sup>0</sup> + ''e''<sup>''i''{{pi}}</sup> + ''e''<sup>−''i''{{pi}}</sup> ≠ 0; but this is false. Similarly, Lemma B is sufficient to prove that ''e'' is transcendental, since Lemma B says that if ''a''<sub>0</sub>, ..., ''a''<sub>''n''</sub> are integers not all of which are zero, then : <math>a_ne^n+\cdots+a_0e^0\ne 0.</math> Lemma B also suffices to prove that {{pi}} is transcendental, since otherwise we would have 1 + ''e''<sup>''i''{{pi}}</sup> ≠ 0. ====Equivalence of the two statements==== Baker's formulation of the theorem clearly implies the first formulation. Indeed, if <math>\alpha(1),\ldots,\alpha(n)</math> are algebraic numbers that are linearly independent over <math>\Q</math>, and :<math>P(x_1, \ldots, x_n)= \sum b_{i_1,\ldots, i_n} x_1^{i_1}\cdots x_n^{i_n}</math> is a polynomial with rational coefficients, then we have :<math>P\left(e^{\alpha(1)},\dots,e^{\alpha(n)}\right)= \sum b_{i_1,\dots,i_n} e^{i_1 \alpha(1) + \cdots + i_n \alpha(n)},</math> and since <math>\alpha(1),\ldots,\alpha(n)</math> are algebraic numbers which are linearly independent over the rationals, the numbers <math>i_1 \alpha(1) + \cdots + i_n \alpha(n)</math> are algebraic and they are distinct for distinct ''n''-tuples <math>(i_1,\dots,i_n)</math>. So from Baker's formulation of the theorem we get <math> b_{i_1,\ldots,i_n}=0</math> for all ''n''-tuples <math>(i_1,\dots,i_n)</math>. Now assume that the first formulation of the theorem holds. For <math>n=1</math> Baker's formulation is trivial, so let us assume that <math>n>1</math>, and let <math>a(1),\ldots,a(n)</math> be non-zero algebraic numbers, and <math>\alpha(1),\ldots,\alpha(n)</math> distinct algebraic numbers such that: :<math>a(1)e^{\alpha(1)} + \cdots + a(n)e^{\alpha(n)} = 0.</math> As seen in the previous section, and with the same notation used there, the value of the polynomial :<math>Q(x_{11},\ldots,x_{nd(n)},y_1,\dots,y_n)=\prod\nolimits_{\sigma\in S}\left(x_{1\sigma(1)}y_1+\dots+x_{n\sigma(n)}y_n\right),</math> at :<math>\left (a(1)_1,\ldots,a(n)_{d(n)},e^{\alpha(1)},\ldots,e^{\alpha(n)} \right)</math> has an expression of the form : <math>b(1)e^{\beta(1)}+ b(2)e^{\beta(2)}+ \cdots + b(M)e^{\beta(M)}= 0,</math> where we have grouped the exponentials having the same exponent. Here, as proved above, <math>b(1),\ldots, b(M)</math> are rational numbers, not all equal to zero, and each exponent <math>\beta(m)</math> is a linear combination of <math>\alpha(i)</math> with integer coefficients. Then, since <math>n>1</math> and <math>\alpha(1),\ldots,\alpha(n)</math> are pairwise distinct, the <math>\Q</math>-vector subspace <math>V</math> of <math>\C</math> generated by <math>\alpha(1),\ldots,\alpha(n)</math> is not trivial and we can pick <math>\alpha(i_1),\ldots,\alpha(i_k)</math> to form a basis for <math>V.</math> For each <math>m=1,\dots,M</math>, we have :<math>\begin{align} \beta(m) = q_{m,1} \alpha(i_1) + \cdots + q_{m,k} \alpha(i_k), && q_{m,j} = \frac{c_{m,j}}{d_{m,j}}; \qquad c_{m,j}, d_{m,j} \in \Z. \end{align}</math> For each <math>j=1,\ldots, k,</math> let <math>d_j</math> be the least common multiple of all the <math>d_{m,j}</math> for <math>m=1,\ldots, M</math>, and put <math>v_j = \tfrac{1}{d_j} \alpha(i_j)</math>. Then <math>v_1,\ldots,v_k </math> are algebraic numbers, they form a basis of <math>V</math>, and each <math>\beta(m)</math> is a linear combination of the <math>v_j</math> with integer coefficients. By multiplying the relation :<math>b(1)e^{\beta(1)}+ b(2)e^{\beta(2)}+ \cdots + b(M)e^{\beta(M)}= 0,</math> by <math>e^{N(v_1+ \cdots + v_k)}</math>, where <math>N</math> is a large enough positive integer, we get a non-trivial algebraic relation with rational coefficients connecting <math>e^{v_1},\cdots,e^{v_k}</math>, against the first formulation of the theorem.
Edit summary
(Briefly describe your changes)
By publishing changes, you agree to the
Terms of Use
, and you irrevocably agree to release your contribution under the
CC BY-SA 4.0 License
and the
GFDL
. You agree that a hyperlink or URL is sufficient attribution under the Creative Commons license.
Cancel
Editing help
(opens in new window)