Bases and Dimension
- Linear Algebra 1.4
Basis of a Vector Space
For $S \subset V$, the following are equivalent:
- $S$ is linearly independent and spans $V$.
- Every nonzero vector $\b{v} \in V$ is an essentially unique linear combination of vectors in $S$.
- $S$ is a minimal spanning set of $V$, that is, $S$ spans $V$ but no proper subset of $S$ spans $V$.
- $S$ is a maximal linearly independent set, that is, $S$ is linearly independent but no proper superset of $S$ is linearly independent.
A set of vectors satisfying any of the above equivalent conditions is called a basis of $V$.
Proof.
We’ve seen that (1) and (2) are equivalent.
(1) $\Rightarrow$ (3): If some proper subset $S^\prime \subset S$ spans $V$, then any vector in $S \setminus S^\prime$ can be written as a linear combination of vectors in $S^\prime$, contradicting the linear independence of $S$.
(3) $\Rightarrow$ (1): If $S$ is not linearly independent, then some vector $\b{v} \in S$ can be written as a linear combination of other vectors in $S$. Then $S \setminus \Set{\b{v}}$ still spans $V$, contradicting the minimality of $S$.
(1) $\Rightarrow$ (4): If some proper superset $S^\prime \supset S$ is linearly independent, then any vector in $S^\prime \setminus S$ can be written as a linear combination of vectors in $S$, contradicting the linear independence of $S^\prime$.
(4) $\Rightarrow$ (1): If $S$ does not span $V$, then there exists a vector $\b{v} \in V \setminus \span(S)$. Then $S \cup \Set{\b{v}}$ is still linearly independent, contradicting the maximality of $S$.
Standard Basis
A finite set $S=\Set{\b{v}_1, \dots, \b{v}_n}\subset V$ is a basis of $V$ if and only if:
\[ V = \bigoplus_{i=1}^n \span \{ \b{v}_i \} \]
The most common example of a basis is the standard basis of $\mathbb{R}^n$. The $i$-th standard basis vector $\b{e}_i$ is the vector in $\mathbb{R}^n$ with a $1$ in the $i$-th coordinate and $0$ in all other coordinates.
\[ e_1 = (1, 0, \dots, 0) \nl e_2 = (0, 1, \dots, 0) \nl \vdots \nl e_n = (0, 0, \dots, 1) \]
The set $\Set{\b{e}_1, \dots, \b{e}_n}$ is a standard basis of $\mathbb{R}^n$.
Existence of a Basis
Now, let’s prove that every nontrivial vector space has a basis.
Let $V$ be a nontrivial vector space and $I\subset V$ be a linearly independent set. Let $S \supset I$ be a spanning set of $V$; $\span(S) = V$. Then there exists a subset basis $\mathcal{B}$ for $V$ which $I \subseteq \mathcal{B} \subseteq S$. In particular,
- Any vector space, except the trivial vector space, has a basis.
- Any linearly independent set in $V$ is contained in some basis of $V$.
- Any spanning set of $V$ contains some basis of $V$.
Proof.
Consider the collection $\mathcal{A} = \Set{ A \subseteq S | I \subseteq A, A \text{ is linearly independent} }$. Then $\mathcal{A}$ is nonempty since $I \in \mathcal{A}$. Consider the partial order on $\mathcal{A}$ given by set inclusion. Let $\mathcal{C} \subseteq \mathcal{A}$ be a chain in $\mathcal{A}$ and $U = \bigcup_{A \in \mathcal{C}} A$. $U$ is an upper bound of $\mathcal{C}$ since $I \subseteq U$ and $U$ is linearly independent. Then according to Zorn’s lemma, $\mathcal{A}$ has a maximal element $\mathcal{B}$, which is linearly independent.
Now let’s show that $\mathcal{B}$ is a basis of $V$. By the definition of $\mathcal{A}$, we have $\mathcal{B} \subseteq S$, so $\span(\mathcal{B}) \subseteq \span(S)$. Also, if any $\b{s}\in S$ is $\b{s} \notin \span(\mathcal{B})$, then $\mathcal{B} \cup \Set{\b{s}}$ is linearly independent, contradicting the maximality of $\mathcal{B}$. Hence, $\b{s} \in \span(\mathcal{B})$ for all $\b{s} \in S$, so $\span(S) \subseteq \span(\mathcal{B})$. Therefore, $\span(\mathcal{B}) = \span(S) = V$, so $\mathcal{B}$ is a basis of $V$.
We can show that any subspace of a vector space has a complementary subspace using this theorem, which was introduced here.
Dimension of a Vector Space
Replacement Theorem
If $V$ be a vector space and assume that the vectors $\b{v}_1,\dots,\b{v}_n$ are linearly independent and the vectors $\b{s}_1,\dots,\b{s}_m$ span $V$. Then $n \leq m$.
Proof.
Suppose $n > m$. Denote $S = \Set{ \b{s}_1, \dots, \b{s}_m }$, then $\b{v}_1\in V=\span(S)$, so:
\[ \b{v}_1 = \sum_{i=1}^m a_i \b{s}_i \]
with at least one coefficient, WLOG $a_1 \neq 0$. Then we can express $\b{s}_1$ as a linear combination of $\b{v}_1$ and the other $\b{s}_i$’s:
\[ \b{s}_1 = a_1^{-1} \left( \b{v}_1 - \sum_{i=2}^m a_i \b{s}_i \right) \]
Therefore, $\b{s}_k \in \span((S\cup \Set{\b{v}_1}) \setminus \Set{\b{s}_k})$. Since $\span(S\cup \Set{\b{v}_1}) = V$, we have $\span((S\cup \Set{\b{v}_1}) \setminus \Set{\b{s}_k}) = V$. Then we can repeat the above process for $\b{v}_2, \cdots$ until we get a spanning set without vectors in $S$.
\[ \b{v}_j = \sum_{i=1}^{j-1} a_i \b{v}_i + \sum_{i=j}^m a_i \b{s}_i \]
If $\forall i \geq j, a_i = 0$, then $\b{v}_j \in \span(\Set{\b{v}_1, \dots, \b{v}_{j-1}})$, contradicting the linear independence of $\b{v}_1,\dots,\b{v}_n$. Therefore, there exists some $i \geq j$ such that $a_i \neq 0$, WLOG $a_j \neq 0$, which guarantees the repeatability of the above process. Finally, we get a spanning set $\Set{\b{v}_1, \dots, \b{v}_m}$ of $V$, contradicting the linear independence. Thus, $n \leq m$.
By the corollary of the replacement theorem, if any vector space has a finite spanning set, then all its bases have the same number of vectors. Suppose that $V$ has two bases $\mathcal{B}$ and $\mathcal{C}$. Then $\mathcal{B}$ is a linearly independent set and $\mathcal{C}$ is a spanning set, so $|\mathcal{B}| \leq |\mathcal{C}|$. Similarly, $\mathcal{C}$ is a linearly independent set and $\mathcal{B}$ is a spanning set, so $|\mathcal{C}| \leq |\mathcal{B}|$. Therefore, $|\mathcal{B}| = |\mathcal{C}|$.
Infinite-Dimensional Vector Spaces
A vector space is called finite-dimensional if it has a finite spanning set, and infinite-dimensional otherwise. For infinite-dimensional cases, we can also show that all bases have the same cardinality. Let $\mathcal{B} = \Set{ \b{b}_i | i \in B}$ and $\mathcal{C} = \Set{ \b{c}_j | j \in C}$ be two bases of $V$. Then any vector $\b{c} \in \mathcal{C}$ can be written as a finite linear combination of vectors in $\mathcal{B}$, where all of the coefficients are nonzero, say:
\[ \b{c} = \sum_{i \in U_\b{c}} r_i \b{b}_i \]
But because $\mathcal{C}$ is a basis, we must have:
\[ \bigcup_{\b{c} \in \mathcal{C}} U_\b{c} = B \]
for if the vectors in $\mathcal{C}$ can be expressed as finite linear combinations of a proper subset $\mathcal{B}^\prime \subset \mathcal{B}$, then $\mathcal{B}^\prime$ would also span $V$, which is contradictory. Then since $U_\b{c} < \aleph_0$ for all $\b{c} \in \mathcal{C}$, we have:
\[ |\mathcal{B}| = |B| \le \aleph_0 \cdot |\mathcal{C}| = |\mathcal{C}| \]
By reversing the role of $\mathcal{B}$ and $\mathcal{C}$, we also have $|\mathcal{C}| \le |\mathcal{B}|$, so $|\mathcal{B}| = |\mathcal{C}|$.
The common cardinality of all bases of a vector space $V$ is called the dimension of $V$, denoted by $\dim(V)$. If $\dim(V) = \kappa$ for some cardinal $\kappa$, then we say that $V$ is $\kappa$-dimensional. A vector space is finite-dimensional if and only if its dimension is finite, and infinite-dimensional otherwise. We can easily verify the following properties of dimension:
- If $\dim(V)=n<\infty$, then any finite generating set of $V$ has at least $n$ vectors, and any linearly independent set in $V$ has at most $n$ vectors. If any of these sets has exactly $n$ vectors, then it is a basis for $V$.
- Every linearly independent subset of $V$ can be extended to a basis for $V$.
- If $W$ is a subspace of $V$, then $\dim(W) \leq \dim(V)$, and if $\dim(W) = \dim(V) < \infty$, then $W = V$.
Let’s see some examples of dimensions of vector spaces. For any field $F$,
- $\dim( \{ \b{0} \}) = 0$
- $\dim(F^n) = n$
- $\dim(\mathcal{M}_{m \times n}(F)) = mn$
- $\dim(\mathrm{Sym}_n(F)) = \frac{n(n+1)}{2}$
- $\dim(\mathrm{Skew}_n(F)) = \frac{n(n-1)}{2}$
- $\dim(\mathcal{P}_n(F)) = n+1$
- $\dim(F[x]) = \aleph_0$
- $\dim(\mathcal{F}(X, F)) = |F|^{|X|}$
Specifying the base field is important when determining the dimension of a vector space. For example, $\mathbb{C}^n$ can be regarded as a vector space over $\mathbb{C}$ or $\mathbb{R}$, and we have $\dim_\mathbb{C}(\mathbb{C}^n) = n$ but $\dim_\mathbb{R}(\mathbb{C}^n) = 2n$.
Also, let’s see the following theorems. Let $V$ be a vector space.
- If $\mathcal{B}$ is a basis of $V$ and $\mathcal{B} = \mathcal{B}_1 \sqcup \mathcal{B}_2$, then
\[ V = \span(\mathcal{B}_1) \oplus \span(\mathcal{B}_2) \]
- Let $V = S_1 \oplus S_2$ and $\mathcal{B}_i$ be a basis of $S_i$, then $\mathcal{B}_1 \sqcup \mathcal{B}_2$ is a basis of $V$.
- Let $S_1$ and $S_2$ be subspaces of vector space $V$. Then
\[ \dim(S_1) + \dim(S_2) = \dim(S_1 + S_2) + \dim(S_1 \cap S_2) \]
In particular, if they are complementary,
\[ \dim(S_1 \oplus S_2) = \dim(S_1) + \dim(S_2) \]
We omit the proofs here.