$$ \newcommand{\qed}{\tag*{$\square$}} \newcommand{\span}{\operatorname{span}} \newcommand{\dim}{\operatorname{dim}} \newcommand{\rank}{\operatorname{rank}} \newcommand{\norm}[1]{\|#1\|} \newcommand{\grad}{\nabla} \newcommand{\prox}[1]{\operatorname{prox}_{#1}} \newcommand{\inner}[2]{\langle{#1}, {#2}\rangle} \newcommand{\mat}[1]{\mathcal{M}[#1]} \newcommand{\null}[1]{\operatorname{null} \left(#1\right)} \newcommand{\range}[1]{\operatorname{range} \left(#1\right)} \newcommand{\rowvec}[1]{\begin{bmatrix} #1 \end{bmatrix}^T} \newcommand{\Reals}{\mathbf{R}} \newcommand{\RR}{\mathbf{R}} \newcommand{\Complex}{\mathbf{C}} \newcommand{\Field}{\mathbf{F}} \newcommand{\Pb}{\operatorname{Pr}} \newcommand{\E}[1]{\operatorname{E}[#1]} \newcommand{\Var}[1]{\operatorname{Var}[#1]} \newcommand{\argmin}[2]{\underset{#1}{\operatorname{argmin}} {#2}} \newcommand{\optmin}[3]{ \begin{align*} & \underset{#1}{\text{minimize}} & & #2 \\ & \text{subject to} & & #3 \end{align*} } \newcommand{\optmax}[3]{ \begin{align*} & \underset{#1}{\text{maximize}} & & #2 \\ & \text{subject to} & & #3 \end{align*} } \newcommand{\optfind}[2]{ \begin{align*} & {\text{find}} & & #1 \\ & \text{subject to} & & #2 \end{align*} } $$
Until now, the role of vectors was limited to their membership in vector spaces. But the vectors inside a vector space play roles more expressive than membership — the contents of vector spaces can in some sense be generated by lists of vectors, if the vectors in the lists are chosen appropriately. It is our endeavor in this section to make the previous statement precise. We begin by defining vocabulary needed to do so. Note that whenever we refer to a list of vectors in this section, we make implicit the assumption that the vectors all belong to the same vector space.
Definition 3.1 A weighted sum of a list of vectors is called a linear combination. Such weighted sums have the form , where the coefficients are drawn from the field over which the vector space is defined; for our purposes, simply think of the as real or complex numbers.
If the vectors belong to the same vector space, then every linear combination of them also belongs to the vector space; if this is not immediately obvious to you, revisit the section on vector spaces and prove to yourself, in no more than a sentence, why this is the case.
Definition 3.2 The set generated by taking all linear combinations of the vectors is called the span of the list of vectors, and we write
The span of the empty list is defined to be the singleton set .
Spans connect lists of vectors to vector spaces: The span of a list of vectors is a vector space. In fact, the generated vector space is the the smallest subspace containing all the vectors in the list.
Proof. Let , where all the vectors belong to the same vector space. That is a subspace of is clear, for , is closed under addition, and is closed under scalar multiplication. That each of the belong to is also clear (simply take all but one of the coefficients in a linear combination of the vectors to be ). And is the smallest subspace containing because every vector space containing them must contain their every linear combination, for vector spaces are closed under addition and scalar multiplication.
Linear algebra is the study of linear maps on finite-dimensional vector spaces, whereas the study of infinite-dimensional is relegated to functional analysis (a subject for another year). We are now positioned to make precise what we mean by finitude.
Definition 3.3 A vector space is finite-dimensional if it is the span of a finite list of vectors.
Combining spans with the notion of linear independence will allow us define the concept of a basis, which will prove to be one of our most useful tools.
Definition 3.4 A list of vectors is linearly indepedent if the only way to write as a linear combination of them is to take every coefficient to be .
The definition implies that cannot be in any linearly independent list. Moreover no vector in a linearly independent list can be written as a linear combination of the others, i.e., no vector in such a list is in the span of the others. These implications are instances of the following theorem: A list of vectors is linearly independent if and only if every vector in its span can be written uniquely as a linear combination of the vectors in it.
Proof. If a list of vectors is linearly independent, then each vector in its span has a unique representation as a linear combination of the vectors in it: taking two linear-combination representations of a vector in the span and subtracting one from the other shows that the coefficients in both representations must be equal if the list is to be linearly independent. And if each vector in the span of a list of vectors has a unique linear-combination representation, then in particular the vector does as well and the list is therefore linearly independent.
Definition 3.5 A list of vectors is linearly depedent if it is not linearly independent.
In any linearly dependent list at least one vector is redundant: There is some vector for that is in the span of , and the span of the list is unchanged by removing . (If you do not immediately see why this is true, stop here and verify this fact with approximately three lines of proof.) This fact implies that every list of linearly independent vectors in a vector space is no longer than any list of vectors spanning .
We can now define the concept of a basis.
Definition 3.6 A list of vectors in a vector space is a basis of if (1) it is linearly independent and (2) it spans ; the vectors in a basis are called basis vectors.
A basis can be thought of as a compact description of a vector space: Every vector in by definition can be written uniquely as a linear combination of basis vectors. It does not make sense to speak of the basis for a vector space, for every vector space has infinitely many bases.
Here are some important facts about bases, the veracity of which you may prove using material already presented in this section: All bases of a vector space have the same length, every linearly independent list in can be extended to a basis for , and all lists spanning can be converted to a basis for by iteratively removing redundant vectors.
These facts allow us to define concretely the dimension of a vector space.
Definition 3.7 The dimension of a vector space , denoted by , is the length of any basis for it.
This definition agrees with our intuitive understanding of dimension: is of dimension , for the scalar is a basis for it, and is of dimension , for the vectors , and form a basis for it, and more generally is of dimension , for the list of vectors
is a basis for it. This basis is referred to as the standard basis for , and the indicator vector for the -th coordinate is by convention referred to as .
The dimension of a vector space is analagous to the cardinality of a set: If is a subspace of , then , and for subspaces and .
Prove some of the facts presented in sections 3.2 and 3.3.
Linear Algebra Done Right, by Sheldon Axler.