Open main menu
Home
Random
Recent changes
Special pages
Community portal
Preferences
About Wikipedia
Disclaimers
Incubator escapee wiki
Search
User menu
Talk
Dark mode
Contributions
Create account
Log in
Editing
Linear combination
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
{{Short description|Sum of terms, each multiplied with a scalar}}{{Redirects|Superposition|3=superposition (disambiguation)}} [[File:Linjcomb.png|thumb|300x300px|<math>\mathbf v </math> is the linear combination of vectors <math>\mathbf u_1 </math> and <math>\mathbf u_2 </math> such that <math>\mathbf v = 2 \cdot \mathbf u_1 + 1.5 \cdot \mathbf u_2 </math>]] In [[mathematics]], a '''linear combination''' or '''superposition''' is an [[Expression (mathematics)|expression]] constructed from a [[Set (mathematics)|set]] of terms by multiplying each term by a constant and adding the results (e.g. a linear combination of ''x'' and ''y'' would be any expression of the form ''ax'' + ''by'', where ''a'' and ''b'' are constants).<ref>{{Harvard citation text|Strang|2016}} p. 3, Β§ 1.1</ref><ref>{{Harvard citation text|Lay|Lay|McDonald|2016}} p. 28, ch. 1</ref><ref>{{Harvard citation text|Axler|2015}} p. 28, Β§ 2.3</ref><ref>{{Harvard citation text|nLab|2015}} Linear combinations.</ref> The concept of linear combinations is central to [[linear algebra]] and related fields of mathematics. Most of this article deals with linear combinations in the context of a [[vector space]] over a [[field (mathematics)|field]], with some generalizations given at the end of the article. == Definition== Let ''V'' be a [[vector space]] over the field ''K''. As usual, we call elements of ''V'' ''[[vector space|vector]]s'' and call elements of ''K'' ''[[scalar (mathematics)|scalars]]''. If '''v'''<sub>1</sub>,...,'''v'''<sub>''n''</sub> are vectors and ''a''<sub>1</sub>,...,''a''<sub>''n''</sub> are scalars, then the ''linear combination of those vectors with those scalars as coefficients'' is :<math>a_1 \mathbf v_1 + a_2 \mathbf v_2 + a_3 \mathbf v_3 + \cdots + a_n \mathbf v_n.</math> There is some ambiguity in the use of the term "linear combination" as to whether it refers to the expression or to its value. In most cases the value is emphasized, as in the assertion "the set of all linear combinations of '''v'''<sub>1</sub>,...,'''v'''<sub>''n''</sub> always forms a subspace". However, one could also say "two different linear combinations can have the same value" in which case the reference is to the expression. The subtle difference between these uses is the essence of the notion of [[linear dependence]]: a family ''F'' of vectors is linearly independent precisely if any linear combination of the vectors in ''F'' (as value) is uniquely so (as expression). In any case, even when viewed as expressions, all that matters about a linear combination is the coefficient of each '''v'''<sub>''i''</sub>; trivial modifications such as permuting the terms or adding terms with zero coefficient do not produce distinct linear combinations. In a given situation, ''K'' and ''V'' may be specified explicitly, or they may be obvious from context. In that case, we often speak of ''a linear combination of the vectors'' '''v'''<sub>1</sub>,...,'''v'''<sub>''n''</sub>, with the coefficients unspecified (except that they must belong to ''K''). Or, if ''S'' is a [[subset]] of ''V'', we may speak of ''a linear combination of vectors in S'', where both the coefficients and the vectors are unspecified, except that the vectors must belong to the set ''S'' (and the coefficients must belong to ''K''). Finally, we may speak simply of ''a linear combination'', where nothing is specified (except that the vectors must belong to ''V'' and the coefficients must belong to ''K''); in this case one is probably referring to the expression, since every vector in ''V'' is certainly the value of some linear combination. Note that by definition, a linear combination involves only [[finite set|finite]]ly many vectors (except as described in the {{section link||Generalizations}} section. However, the set ''S'' that the vectors are taken from (if one is mentioned) can still be [[Infinity|infinite]]; each individual linear combination will only involve finitely many vectors. Also, there is no reason that ''n'' cannot be [[0 (number)|zero]]; in that case, we declare by convention that the result of the linear combination is the [[zero vector]] in ''V''. == Examples and counterexamples == {{inline citations|section|date=August 2013}} === Euclidean vectors === Let the field ''K'' be the set '''R''' of [[real number]]s, and let the vector space ''V'' be the [[Euclidean space]] '''R'''<sup>3</sup>. Consider the vectors {{nowrap|1='''e'''<sub>1</sub> = (1,0,0)}}, {{nowrap|1='''e'''<sub>2</sub> = (0,1,0)}} and {{nowrap|1='''e'''<sub>3</sub> = (0,0,1)}}. Then ''any'' [[Euclidean vector|vector]] in '''R'''<sup>3</sup> is a linear combination of '''e'''<sub>1</sub>, '''e'''<sub>2</sub>, and '''e'''<sub>3</sub>. To see that this is so, take an arbitrary vector (''a''<sub>1</sub>,''a''<sub>2</sub>,''a''<sub>3</sub>) in '''R'''<sup>3</sup>, and write: :<math> \begin{align} ( a_1 , a_2 , a_3) & = ( a_1 ,0,0) + (0, a_2 ,0) + (0,0, a_3) \\[6pt] & = a_1 (1,0,0) + a_2 (0,1,0) + a_3 (0,0,1) \\[6pt] & = a_1 \mathbf e_1 + a_2 \mathbf e_2 + a_3 \mathbf e_3. \end{align} </math> === Functions === Let ''K'' be the set '''C''' of all [[complex number]]s, and let ''V'' be the set C<sub>'''C'''</sub>(''R'') of all [[continuous function]]s from the [[real line]] '''R''' to the [[complex plane]] '''C'''. Consider the vectors (functions) ''f'' and ''g'' defined by ''f''(''t'') := ''e''<sup>''it''</sup> and ''g''(''t'') := ''e''<sup>β''it''</sup>. (Here, ''e'' is the [[e (mathematical constant)|base of the natural logarithm]], about 2.71828..., and ''i'' is the [[imaginary unit]], a square root of β1.) Some linear combinations of ''f'' and ''g'' are: *<div style="vertical-align: 0%;display:inline;"><math> \cos t = \tfrac12 \, e^{i t} + \tfrac12 \, e^{-i t} </math></div> *<div style="vertical-align: 0%;display:inline;"><math> 2 \sin t = (-i) e^{i t} + (i) e^{-i t}. </math></div> On the other hand, the constant function 3 is ''not'' a linear combination of ''f'' and ''g''. To see this, suppose that 3 could be written as a linear combination of ''e''<sup>''it''</sup> and ''e''<sup>β''it''</sup>. This means that there would exist complex scalars ''a'' and ''b'' such that {{nowrap|1=''ae''<sup>''it''</sup> + ''be''<sup>β''it''</sup> = 3}} for all real numbers ''t''. Setting ''t'' = 0 and ''t'' = Ο gives the equations {{nowrap|1=''a'' + ''b'' = 3}} and {{nowrap|1=''a'' + ''b'' = β3}}, and clearly this cannot happen. See [[Euler's identity]]. === Polynomials === Let ''K'' be '''R''', '''C''', or any field, and let ''V'' be the set ''P'' of all [[polynomial]]s with coefficients taken from the field ''K''. Consider the vectors (polynomials) ''p''<sub>1</sub> := 1, {{nowrap|1=''p''<sub>2</sub> := ''x'' + 1}}, and {{nowrap|1=''p''<sub>3</sub> := ''x''<sup>2</sup> + ''x'' + 1}}. Is the polynomial ''x''<sup>2</sup> β 1 a linear combination of ''p''<sub>1</sub>, ''p''<sub>2</sub>, and ''p''<sub>3</sub>? To find out, consider an arbitrary linear combination of these vectors and try to see when it equals the desired vector ''x''<sup>2</sup> β 1. Picking arbitrary coefficients ''a''<sub>1</sub>, ''a''<sub>2</sub>, and ''a''<sub>3</sub>, we want :<math> a_1 (1) + a_2 ( x + 1) + a_3 ( x^2 + x + 1) = x^2 - 1. </math> Multiplying the polynomials out, this means :<math> ( a_1 ) + ( a_2 x + a_2) + ( a_3 x^2 + a_3 x + a_3) = x^2 - 1 </math> and collecting like powers of ''x'', we get :<math> a_3 x^2 + ( a_2 + a_3 ) x + ( a_1 + a_2 + a_3 ) = 1 x^2 + 0 x + (-1). </math> Two polynomials are equal [[if and only if]] their corresponding coefficients are equal, so we can conclude :<math> a_3 = 1, \quad a_2 + a_3 = 0, \quad a_1 + a_2 + a_3 = -1. </math> This [[system of linear equations]] can easily be solved. First, the first equation simply says that ''a''<sub>3</sub> is 1. Knowing that, we can solve the second equation for ''a''<sub>2</sub>, which comes out to β1. Finally, the last equation tells us that ''a''<sub>1</sub> is also β1. Therefore, the only possible way to get a linear combination is with these coefficients. Indeed, :<math> x^2 - 1 = -1 - ( x + 1) + ( x^2 + x + 1) = - p_1 - p_2 + p_3 </math> so ''x''<sup>2</sup> β 1 ''is'' a linear combination of ''p''<sub>1</sub>, ''p''<sub>2</sub>, and ''p''<sub>3</sub>. On the other hand, what about the polynomial ''x''<sup>3</sup> β 1? If we try to make this vector a linear combination of ''p''<sub>1</sub>, ''p''<sub>2</sub>, and ''p''<sub>3</sub>, then following the same process as before, we get the equation :<math> \begin{align} & 0 x^3 + a_3 x^2 + ( a_2 + a_3 ) x + ( a_1 + a_2 + a_3 ) \\[5pt] = {} & 1 x^3 + 0 x^2 + 0 x + (-1). \end{align} </math> However, when we set corresponding coefficients equal in this case, the equation for ''x''<sup>3</sup> is :<math> 0 = 1 </math> which is always false. Therefore, there is no way for this to work, and ''x''<sup>3</sup> β 1 is ''not'' a linear combination of ''p''<sub>1</sub>, ''p''<sub>2</sub>, and ''p''<sub>3</sub>. == The linear span == {{main article|Linear span}} Take an arbitrary field ''K'', an arbitrary vector space ''V'', and let '''v'''<sub>1</sub>,...,'''v'''<sub>''n''</sub> be vectors (in ''V''). It is interesting to consider the set of ''all'' linear combinations of these vectors. This set is called the ''[[linear span]]'' (or just ''span'') of the vectors, say ''S'' = {'''v'''<sub>1</sub>, ..., '''v'''<sub>''n''</sub>}. We write the span of ''S'' as span(''S'')<ref>{{Harvard citation text|Axler|2015}} pp. 29-30, Β§Β§ 2.5, 2.8</ref><ref>{{Harvard citation text|Katznelson|Katznelson|2008}} p. 9, Β§ 1.2.3</ref> or sp(''S''): :<math> \operatorname{span}( \mathbf v_1 ,\ldots, \mathbf v_n) := \{ a_1 \mathbf v_1 + \cdots + a_n \mathbf v_n : a_1 ,\ldots, a_n \in K \}. </math> == Linear independence == {{main|Linear independence}} Suppose that, for some sets of vectors '''v'''<sub>1</sub>,...,'''v'''<sub>''n''</sub>, a single vector can be written in two different ways as a linear combination of them: :<math>\mathbf v = \sum_i a_i \mathbf v_i = \sum_i b_i \mathbf v_i\text{ where } a_i \neq b_i.</math> This is equivalent, by subtracting these (<math>c_i := a_i - b_i</math>), to saying a non-trivial combination is zero:<ref>{{Harvard citation text|Axler|2015}} pp. 32-33, Β§Β§ 2.17, 2.19</ref><ref>{{Harvard citation text|Katznelson|Katznelson|2008}} p. 14, Β§ 1.3.2</ref> :<math>\mathbf 0 = \sum_i c_i \mathbf v_i.</math> If that is possible, then '''v'''<sub>1</sub>,...,'''v'''<sub>''n''</sub> are called ''[[linearly dependent]]''; otherwise, they are ''linearly independent''. Similarly, we can speak of linear dependence or independence of an arbitrary set ''S'' of vectors. If ''S'' is linearly independent and the span of ''S'' equals ''V'', then ''S'' is a [[basis (linear algebra)|basis]] for ''V''. == Affine, conical, and convex combinations == By restricting the coefficients used in linear combinations, one can define the related concepts of [[affine combination]], [[conical combination]], and [[convex combination]], and the associated notions of sets closed under these operations. {| class="wikitable" style="text-align: left;" |- ! Type of combination !! Restrictions on coefficients !! Name of set !! Model space |- | Linear combination || no restrictions || [[Vector subspace]] || <math>\mathbf{R}^n</math> |- | [[Affine combination]] || <math display="inline">\sum a_i = 1</math>|| [[Affine subspace]] || Affine [[hyperplane]] |- | [[Conical combination]] || <math>a_i \geq 0</math> || [[Convex cone]] || [[Quadrant (plane geometry)|Quadrant]], [[Octant (solid geometry)|octant]], or [[orthant]] |- | [[Convex combination]] || <math>a_i \geq 0</math> and <math display="inline">\sum a_i = 1</math>|| [[Convex set]] || [[Simplex]] |} Because these are more ''restricted'' operations, more subsets will be closed under them, so affine subsets, convex cones, and convex sets are ''generalizations'' of vector subspaces: a vector subspace is also an affine subspace, a convex cone, and a convex set, but a convex set need not be a vector subspace, affine, or a convex cone. These concepts often arise when one can take certain linear combinations of objects, but not any: for example, [[probability distribution]]s are closed under convex combination (they form a convex set), but not conical or affine combinations (or linear), and [[positive measure]]s are closed under conical combination but not affine or linear β hence one defines [[signed measure]]s as the linear closure. Linear and affine combinations can be defined over any field (or ring), but conical and convex combination require a notion of "positive", and hence can only be defined over an [[ordered field]] (or [[ordered ring]]), generally the real numbers. If one allows only [[scalar multiplication]], not addition, one obtains a (not necessarily convex) [[Cone (linear algebra)|cone]]; one often restricts the definition to only allowing multiplication by positive scalars. All of these concepts are usually defined as subsets of an ambient vector space (except for affine spaces, which are also considered as "vector spaces forgetting the origin"), rather than being axiomatized independently. == Operad theory == {{Main article|Operad theory}} More abstractly, in the language of [[operad theory]], one can consider vector spaces to be [[Algebra (ring theory)|algebras]] over the operad <math>\mathbf{R}^\infty</math> (the infinite [[Direct sum of modules|direct sum]], so only finitely many terms are non-zero; this corresponds to only taking finite sums), which parametrizes linear combinations: the vector <math>(2,3,-5,0,\dots)</math> for instance corresponds to the linear combination <math>2 \mathbf v_1 + 3 \mathbf v_2 - 5 \mathbf v_3 + 0 \mathbf v_4 + \cdots</math>. Similarly, one can consider affine combinations, conical combinations, and convex combinations to correspond to the sub-operads where the terms sum to 1, the terms are all non-negative, or both, respectively. Graphically, these are the infinite affine hyperplane, the infinite hyper-octant, and the infinite simplex. This formalizes what is meant by <math>\mathbf{R}^n</math> being or the standard simplex being model spaces, and such observations as that every bounded [[convex polytope]] is the image of a simplex. Here suboperads correspond to more restricted operations and thus more general theories. From this point of view, we can think of linear combinations as the most general sort of operation on a vector space β saying that a vector space is an algebra over the operad of linear combinations is precisely the statement that ''all possible'' algebraic operations in a vector space are linear combinations. The basic operations of addition and scalar multiplication, together with the existence of an [[additive identity]] and additive inverses, cannot be combined in any more complicated way than the generic linear combination: the basic operations are a [[generating set]] for the operad of all linear combinations. Ultimately, this fact lies at the heart of the usefulness of linear combinations in the study of vector spaces. == Generalizations== If ''V'' is a [[topological vector space]], then there may be a way to make sense of certain ''infinite'' linear combinations, using the topology of ''V''. For example, we might be able to speak of ''a''<sub>1</sub>'''v'''<sub>1</sub> + ''a''<sub>2</sub>'''v'''<sub>2</sub> + ''a''<sub>3</sub>'''v'''<sub>3</sub> + β―, going on forever. Such infinite linear combinations do not always make sense; we call them ''convergent'' when they do. Allowing more linear combinations in this case can also lead to a different concept of span, linear independence, and basis. The articles on the various flavors of topological vector spaces go into more detail about these. If ''K'' is a [[commutative ring]] instead of a field, then everything that has been said above about linear combinations generalizes to this case without change. The only difference is that we call spaces like this ''V'' [[module (mathematics)|modules]] instead of vector spaces. If ''K'' is a [[noncommutative ring]], then the concept still generalizes, with one caveat: since modules over noncommutative rings come in left and right versions, our linear combinations may also come in either of these versions, whatever is appropriate for the given module. This is simply a matter of doing scalar multiplication on the correct side. A more complicated twist comes when ''V'' is a [[bimodule]] over two rings, ''K''<sub>L</sub> and ''K''<sub>R</sub>. In that case, the most general linear combination looks like :<math> a_1 \mathbf v_1 b_1 + \cdots + a_n \mathbf v_n b_n </math> where ''a''<sub>1</sub>,...,''a''<sub>''n''</sub> belong to ''K''<sub>L</sub>, ''b''<sub>1</sub>,...,''b''<sub>''n''</sub> belong to ''K''<sub>R</sub>, and '''v'''<sub>1</sub>,β¦,'''v'''<sub>''n''</sub> belong to ''V''. == See also == * [[Weighted sum]] == Citations == {{reflist}} ==References== === Textbook === *{{Cite book|last=Axler|first=Sheldon Jay|title=Linear Algebra Done Right|series=Undergraduate Texts in Mathematics |publisher=[[Springer Science+Business Media | Springer]]|year=2015|isbn=978-3-319-11079-0|edition=3rd|location=|pages=|author-link=Sheldon Axler|doi=10.1007/978-3-319-11080-6}} * {{Cite book|last1=Katznelson|first1=Yitzhak|title=A (Terse) Introduction to Linear Algebra|last2=Katznelson|first2=Yonatan R.|publisher=[[American Mathematical Society]]|year=2008|isbn=978-0-8218-4419-9|author-link=Yitzhak Katznelson}} * {{Cite book|last1=Lay|first1=David C.|title=Linear Algebra and its Applications|last2=Lay|first2=Steven R.|last3=McDonald|first3=Judi J.|publisher=Pearson|year=2016|isbn=978-0-321-98238-4|edition=5th|location=|pages=}} * {{Cite book|last=Strang|first=Gilbert|title=Introduction to Linear Algebra|publisher=Wellesley Cambridge Press|year=2016|isbn=978-0-9802327-7-6|edition=5th|location=|pages=|author-link=Gilbert Strang}} === Web === * {{Cite web|date=27 October 2015|title=Linear Combinations|url=https://ncatlab.org/nlab/show/linear+combination|access-date=16 Feb 2021|website=[[nLab]]|ref=CITEREFnLab2015}} ==External links== * [https://www.khanacademy.org/math/linear-algebra/vectors_and_spaces/linear_combinations/v/linear-combinations-and-span Linear Combinations and Span: Understanding linear combinations and spans of vectors], khanacademy.org. {{linear algebra}} [[Category:Linear algebra]]
Edit summary
(Briefly describe your changes)
By publishing changes, you agree to the
Terms of Use
, and you irrevocably agree to release your contribution under the
CC BY-SA 4.0 License
and the
GFDL
. You agree that a hyperlink or URL is sufficient attribution under the Creative Commons license.
Cancel
Editing help
(opens in new window)
Pages transcluded onto the current version of this page
(
help
)
:
Template:Cite book
(
edit
)
Template:Cite web
(
edit
)
Template:Harvard citation text
(
edit
)
Template:Inline citations
(
edit
)
Template:Linear algebra
(
edit
)
Template:Main
(
edit
)
Template:Main article
(
edit
)
Template:Navbox
(
edit
)
Template:Nowrap
(
edit
)
Template:Redirects
(
edit
)
Template:Reflist
(
edit
)
Template:Section link
(
edit
)
Template:Short description
(
edit
)