Vector space: Difference between revisions

From Citizendium
Jump to navigation Jump to search
imported>Paul Wormer
mNo edit summary
 
(9 intermediate revisions by 5 users not shown)
Line 1: Line 1:
{{subpages}}
{{subpages}}


'''Vector spaces''', also known as '''linear spaces''',  are an abstract [[mathematics|mathematical]] construct with many important applications in the  
A '''vector space''', also known as '''linear space''',  is an abstract [[mathematics|mathematical]] construct with many important applications in the  
[[natural science]]s, in particular in [[physics]] and numerous areas of mathematics.
[[natural science]]s, in particular in [[physics]] and numerous areas of mathematics.
Some vector spaces make sense somewhat intuitively, such as the space of 3D [[vector]]s in standard [[Euclidean space]],
Some vector spaces make sense somewhat intuitively, such as the space of 2D [[vector]]s in standard [[Euclidean plane]], and the language that we use when talking about these intuitive spaces has been taken to describe the more abstract notion as well.  For example, we know how to add vectors and multiply them by [[real number]]s ([[scalar]]s) in <font style = "vertical-align: 20%"><math>\mathbb{R}^3</math></font>,
and the language that we use when talking about these intuitive spaces has been taken to describe the more
abstract notion as well.  For example, we know how to add vectors and multiply them by [[real number]]s ([[scalar]]s) in
<math>\mathbb{R}^3</math>,
and these notions of vector addition and scalar multiplication are defined in a more general sense (as we will see below).
and these notions of vector addition and scalar multiplication are defined in a more general sense (as we will see below).


Vector spaces are important because many different mathematical objects that at first glance seem unrelated
Vector spaces are important because many different mathematical objects that at first glance seem unrelated in fact share a common structure.  By defining this structure and proving things about it in general, we are then able to apply these results to each specific case without having to re-prove them each time. Besides vectors in <font style = "vertical-align: 20%"><math>\mathbb{R}^3</math></font> that are relatively easy to visualize, we can make a vector space out of <font style = "vertical-align: 17%"><math>\mathbb{R}^n</math></font> for any [[natural number]] ''n''; or the [[complex number|complex plane]] or powers of it, <font style = "vertical-align: 13%"><math>\mathbb{C}^n</math></font>; or [[polynomial]]s of degree ''n''.
in fact share a common structure.  By defining this structure and proving things about it in general, we are
then able to apply these results to each specific case without having to re-prove them each time.
Besides vectors in <math>\mathbb{R}^3</math> that are relatively easy to visualize, we can make a vector space
out of <math>\mathbb{R}^n</math> for any [[natural number]] <math>n</math>; or the [[complex number|complex plane]]
or powers of it, <math>\mathbb{C}^n</math>; or [[polynomial]]s of degree <math>n</math>.


No matter what vector space you have to work with though, it is often useful to keep a picture of either 2D or 3D space
Analyzing the structure of vector spaces in abstraction is also important for understanding which properties of a particular space follow solely from it having the structure of a vector space, and which require imposing additional structure on top of the vector space structure.  For instance, vectors in every vector space can always be uniquely identified by assigning them a set of coordinatesHowever, the useful notion of the angle between vectors in <math>\mathbb{R}^3</math> cannot be defined solely in terms of the vector space structure; it requires imposing the additional structure given an [[inner product]] on the space.  Compartmentalizing mathematical information in this way can greatly aid mathematical intuition.
in mindThis helps when thinking of things such as [[orthogonal]] polynomials or [[matrix|matrices]].


No matter what vector space you have to work with though, it is often useful to keep a picture of either 2D or 3D space in mind.  This helps when thinking of things such as [[orthogonal]] polynomials or [[matrix|matrices]].


==Definition==
==Definition==
Line 35: Line 27:
;1. <math>V</math> is closed under addition
;1. <math>V</math> is closed under addition
:The vector <math>\vec{u}+\vec{v}</math> is also an element of <math>V</math>.  This is automatically satisfied when the addition operation is defined as being [[injective]] as it was above.  Care must be taken however if <math>V</math> is a subset of some larger set <math>W</math> and <math>+:\,\,V\times V\to W</math>, as is often the case when looking at [[subspaces]].
:The vector <math>\vec{u}+\vec{v}</math> is also an element of <math>V</math>.  This is automatically satisfied when the addition operation is defined as being [[injective]] as it was above.  Care must be taken however if <math>V</math> is a subset of some larger set <math>W</math> and <math>+:\,\,V\times V\to W</math>, as is often the case when looking at [[subspaces]].
;2. Addition is commutative
;2. Addition is [[commutative]]
:The order in which two vectors are added does not affect the result, <math>\vec{u}+\vec{v}=\vec{v}+\vec{u}</math>.
:The order in which two vectors are added does not affect the result, <math>\vec{u}+\vec{v}=\vec{v}+\vec{u}</math>.
;3. Addition is associative
;3. Addition is [[associative]]
:<math>\vec{u}+(\vec{v}+\vec{w})=(\vec{u}+\vec{v})+\vec{w}</math>.  This means that even though addition is strictly defined as a [[binary operation]], the object <math>\vec{u}+\vec{v}+\vec{w}</math> is well defined.
:<math>\vec{u}+(\vec{v}+\vec{w})=(\vec{u}+\vec{v})+\vec{w}</math>.  This means that even though addition is strictly defined as a [[binary operation]], the object <math>\vec{u}+\vec{v}+\vec{w}</math> is well defined.
;4. An additive identity exists in <math>V</math>
;4. An additive identity exists in <math>V</math>
Line 47: Line 39:
;7. Scalar multiplication is distributive over addition in <math>F</math>
;7. Scalar multiplication is distributive over addition in <math>F</math>
:<math>(a+b)\vec{u}=a\vec{u}+b\vec{u}</math>.  It is important to note that the addition occurring on the left-hand side of this equality is a 'different operation' from the addition on the right-hand side.  While the latter is vector addition as defined above, the former is the addition operation defined on the field <math>F</math>.
:<math>(a+b)\vec{u}=a\vec{u}+b\vec{u}</math>.  It is important to note that the addition occurring on the left-hand side of this equality is a 'different operation' from the addition on the right-hand side.  While the latter is vector addition as defined above, the former is the addition operation defined on the field <math>F</math>.
;8. Vector addition is distributive over scalar multiplication
;8. Vector addition is [[distributive]] over scalar multiplication
:<math>a(\vec{u}+\vec{v})=a\vec{u}+a\vec{v}</math>.  In this case vector addition takes place on both sides of the equality.
:<math>a(\vec{u}+\vec{v})=a\vec{u}+a\vec{v}</math>.  In this case vector addition takes place on both sides of the equality.
;9. Scalar multiplication is associative
;9. Scalar multiplication is associative
Line 72: Line 64:
==Some important theorems==
==Some important theorems==
===Linear dependence===
===Linear dependence===
{{Main|Linear independence}}
A system of ''p'' ( &ge; 1 ) vectors  <math>\vec{u}_1, \dots, \vec{u}_p </math> of a vector space ''V'' is called ''linearly dependent'' if there exist coefficients (elements in ''F'' ) ''a''<sub>1</sub>, ..., ''a''<sub>''p''</sub> not all zero, such that the linear combination is the zero vector in ''V'',
A system of ''p'' ( &ge; 1 ) vectors  <math>\vec{u}_1, \dots, \vec{u}_p </math> of a vector space ''V'' is called ''linearly dependent'' if there exist coefficients (elements in ''F'' ) ''a''<sub>1</sub>, ..., ''a''<sub>''p''</sub> not all zero, such that the linear combination is the zero vector in ''V'',
:<math>
:<math>
Line 104: Line 97:


==Examples of vector spaces==
==Examples of vector spaces==
===Sequences===
The set of all sequences {x<sub>1</sub>, x<sub>2</sub>, &hellip;, x<sub>n</sub>} of ''n'' elements of a [[Field theory (mathematics)|field]], in particular, the real numbers.
Except for the [[Euclidean plane]], the best known vector space is the space <font style = "vertical-align: 17%"><math>\mathbb{R}^n</math></font>.  &nbsp;For integral finite ''n'' > 0 this space can be represented as columns (stacks) of ''n'' real numbers. In order to make the discussion concrete we consider the case ''n'' = 4. It will be clear how the rules apply to general finite ''n''.
====Addition====
:<math>
\begin{pmatrix}
x_1\\x_2\\x_3\\x_4\\
\end{pmatrix}
+
\begin{pmatrix}
y_1\\y_2\\y_3\\y_4\\
\end{pmatrix}
=
\begin{pmatrix}
x_1+y_1\\x_2+y_2\\x_3+y_3\\x_4+y_4\\
\end{pmatrix}
</math>
Because ''x''<sub>''k''</sub> and ''y''<sub>''k''</sub> are real numbers, ''x''<sub>''k''</sub>+''y''<sub>''k''</sub> is a well-defined real number.
====Negative vector====
:<math>
\begin{pmatrix}
x_1\\x_2\\x_3\\x_4\\
\end{pmatrix}
+
\begin{pmatrix}
-x_1\\-x_2\\-x_3\\-x_4\\
\end{pmatrix}
=
\begin{pmatrix}
x_1-x_1\\x_2-x_2\\x_3-x_3\\x_4-x_4\\
\end{pmatrix}
=
\begin{pmatrix}
0\\0\\0\\0\\
\end{pmatrix}
</math>
====Zero vector====
:<math>
\begin{pmatrix}
x_1\\x_2\\x_3\\x_4\\
\end{pmatrix}
+
\begin{pmatrix}
0\\0\\0\\0\\
\end{pmatrix}
=
\begin{pmatrix}
x_1\\x_2\\x_3\\x_4\\
\end{pmatrix}
</math>
====Multiplication by real number====
:<math>
a \begin{pmatrix}
x_1\\x_2\\x_3\\x_4\\
\end{pmatrix}
=
\begin{pmatrix}
a\,x_1\\a\,x_2\\a\,x_3\\a\,x_4\\
\end{pmatrix}
</math>
Because ''a'' and ''x''<sub>''k''</sub>  are real numbers, ''a''&thinsp;''x''<sub>''k''</sub> is well-defined and real.
The reader may easily convince him/herself,  using the known properties of real numbers, that these columns of real numbers satisfy the postulates of a vector space. Its dimension is at least 4, because the following 4 vectors are linearly independent,
:<math>
\begin{pmatrix}
1\\0\\0\\0\\
\end{pmatrix}, \quad
\begin{pmatrix}
0\\1\\0\\0\\
\end{pmatrix}, \quad\begin{pmatrix}
0\\0\\1\\0\\
\end{pmatrix}, \quad\begin{pmatrix}
0\\0\\0\\1\\
\end{pmatrix}.  \qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad(1)
</math>
Indeed, assume that one or more of the coefficients (real numbers) ''a''<sub>''k''</sub> is not equal to zero, then the equation
:<math>
a_1\begin{pmatrix}
1\\0\\0\\0\\
\end{pmatrix}
+
a_2\begin{pmatrix}
0\\1\\0\\0\\
\end{pmatrix}
+
a_3\begin{pmatrix}
0\\0\\1\\0\\
\end{pmatrix}
+
a_4\begin{pmatrix}
0\\0\\0\\1\\
\end{pmatrix}
=
\begin{pmatrix}
a_1\\a_2\\a_3\\a_4\\
\end{pmatrix} =
\begin{pmatrix}
0\\0\\0\\0\\
\end{pmatrix}.
</math>
leads to all four ''a''&prime;s are zero (two vectors are equal if and only if their corresponding elements are equal). This is in contradiction to the assumption that one or more of the coefficients  ''a''<sub>''k''</sub> is not equal to zero.
The set (1) is maximally linearly independent because any non-zero vector can be expressed in the four vectors,
:<math>
\begin{pmatrix}
x_1\\x_2\\x_3\\x_4\\
\end{pmatrix}
=
x_1\begin{pmatrix}
1\\0\\0\\0\\
\end{pmatrix}
+
x_2\begin{pmatrix}
0\\1\\0\\0\\
\end{pmatrix}
+
x_3\begin{pmatrix}
0\\0\\1\\0\\
\end{pmatrix}
+
x_4\begin{pmatrix}
0\\0\\0\\1\\
\end{pmatrix}
\quad\Longrightarrow\quad
-1\begin{pmatrix}
x_1\\x_2\\x_3\\x_4\\
\end{pmatrix}
+
x_1\begin{pmatrix}
1\\0\\0\\0\\
\end{pmatrix}
+
x_2\begin{pmatrix}
0\\1\\0\\0\\
\end{pmatrix}
+
x_3\begin{pmatrix}
0\\0\\1\\0\\
\end{pmatrix}
+
x_4\begin{pmatrix}
0\\0\\0\\1\\
\end{pmatrix} =
\begin{pmatrix}
0\\0\\0\\0\\
\end{pmatrix}.
</math>
The equation on the right is a valid equation between five vectors that do not have a prefactor zero and yet give the zero vector. Hence it is not possible to find a fifth vector linearly independent of the vectors (1): any five vectors form a linearly dependent set. In other words, the four vectors in Eq. (1) form a ''basis'' of the vector space <font style = "vertical-align: 17%"><math>\mathbb{R}^n</math></font>.
===Polynomials===
The set of all polynomials of ''n'' variables {x<sub>i</sub>} with various coefficients {''&alpha;<sub>i</sub>''} from a field. Given polynomials like:
:<math> f = \sum_{i=1}^n \ \alpha_i x_i^n \ , </math>
and
:<math>g = \sum_{i=1}^n \ \beta_i x_i^n \ , </math>
it is clear that the various operations above are directly represented by a mapping:
:<math> f\ \rightarrow \ \{\alpha_1, \ \ldots \ , \ \alpha_n \} ; \ g \ \rightarrow \ \{\beta_1, \ \ldots \ , \ \beta_n \} \ ,  </math>
with the various powers of {x<sub>i</sub>} serving as place markers, so all the operations surveyed above for sequences apply equally here.
===Function spaces===
Consider the field ℝ of real numbers and ''I'' an interval in ℝ. The set ''C(I)'' of all real valued continuous functions on ''I'', the set ''D(I)'' of all real differentiable functions and the set ''A(I)'' of all real analytic function on ''I'' are linear spaces contained in the linear space of all real valued functions defined on ''I''.


==Applications of vector spaces==
==Applications of vector spaces==
Line 112: Line 267:


===Differential equations===
===Differential equations===
[[Category:Suggestion Bot Tag]]

Latest revision as of 12:01, 4 November 2024

This article is developing and not approved.
Main Article
Discussion
Related Articles  [?]
Bibliography  [?]
External Links  [?]
Citable Version  [?]
 
This editable Main Article is under development and subject to a disclaimer.

A vector space, also known as a linear space, is an abstract mathematical construct with many important applications in the natural sciences, in particular in physics and numerous areas of mathematics. Some vector spaces make sense somewhat intuitively, such as the space of 2D vectors in standard Euclidean plane, and the language that we use when talking about these intuitive spaces has been taken to describe the more abstract notion as well. For example, we know how to add vectors and multiply them by real numbers (scalars) in , and these notions of vector addition and scalar multiplication are defined in a more general sense (as we will see below).

Vector spaces are important because many different mathematical objects that at first glance seem unrelated in fact share a common structure. By defining this structure and proving things about it in general, we are then able to apply these results to each specific case without having to re-prove them each time. Besides vectors in that are relatively easy to visualize, we can make a vector space out of for any natural number n; or the complex plane or powers of it, ; or polynomials of degree n.

Analyzing the structure of vector spaces in abstraction is also important for understanding which properties of a particular space follow solely from it having the structure of a vector space, and which require imposing additional structure on top of the vector space structure. For instance, vectors in every vector space can always be uniquely identified by assigning them a set of coordinates. However, the useful notion of the angle between vectors in cannot be defined solely in terms of the vector space structure; it requires imposing the additional structure given an inner product on the space. Compartmentalizing mathematical information in this way can greatly aid mathematical intuition.

No matter what vector space you have to work with though, it is often useful to keep a picture of either 2D or 3D space in mind. This helps when thinking of things such as orthogonal polynomials or matrices.

Definition

A vector space over a field is a set that satisfies certain axioms (see below) and which is equipped with two operations, vector addition and scalar multiplication. Vector addition is defined as a map

that takes the ordered pair to the vector . Here represents the Cartesian product between sets. Scalar multiplication is defined in a similar way, as a map

that takes the ordered pair to the vector . Note that frequently the dot representing scalar multiplication is omitted, the result being written simply as instead. This is especially common when an inner product will also be defined on the vector space, with the dot then representing the inner product between two vectors. It is important to keep in mind the distinction between scalar multiplication, which multiplies one vector by a scalar, and an inner or scalar product, that combined two vectors to yield a scalar.

Axioms of a vector space

Let be a set, , , and elements of that set, and and scalar elements of a field, . Then is a vector space if the following axioms hold true for all choices of

1. is closed under addition
The vector is also an element of . This is automatically satisfied when the addition operation is defined as being injective as it was above. Care must be taken however if is a subset of some larger set and , as is often the case when looking at subspaces.
2. Addition is commutative
The order in which two vectors are added does not affect the result, .
3. Addition is associative
. This means that even though addition is strictly defined as a binary operation, the object is well defined.
4. An additive identity exists in
Labeled , the additive identity or zero vector satisfies .
5. The additive inverse exists in
A vector can be found such that .
6. is closed under scalar multiplication
The vector is itself an element of .
7. Scalar multiplication is distributive over addition in
. It is important to note that the addition occurring on the left-hand side of this equality is a 'different operation' from the addition on the right-hand side. While the latter is vector addition as defined above, the former is the addition operation defined on the field .
8. Vector addition is distributive over scalar multiplication
. In this case vector addition takes place on both sides of the equality.
9. Scalar multiplication is associative
. This means that the algebraic structure of the underlying field is preserved. Note that the left-hand side of this equality contains two subsequent applications of the scalar multiplication defined above, while the right-hand side contains one scalar multiplication as defined in (that of ), followed by scalar multiplication with the vector .
10. The multiplicative identity of provides a scalar multiplicative identity
, where is the multiplicative identity of the field .

Properties 1 - 5 state that a vector space is an Abelian group with addition as group operation.

These axioms can be expressed concisely in mathematical notation as follows:

Some important theorems

Linear dependence

For more information, see: Linear independence.

A system of p ( ≥ 1 ) vectors of a vector space V is called linearly dependent if there exist coefficients (elements in F ) a1, ..., ap not all zero, such that the linear combination is the zero vector in V,

Otherwise, the vectors are called linearly independent. A single vector not equal to the zero vector is obviously linearly independent.


If all a1, ..., ap are zero (in F ) then     If the set is linearly independent then the relation   implies that all a1, ..., ap are zero. Hence a set of p vectors in V is linearly independent if

Every set of vectors containing the zero vector is linearly dependent.

A system of linearly independent vectors remains linearly independent if some vectors are omitted from the system. For, let a subset of the first q vectors , with q < p, be linearly dependent then one or more coefficients not equal to zero can be found while the following is true

Add to the left- and right-hand side of this expression     and we get a contradiction.

Dimension

In general there are infinitely many linearly independent vectors in a vector space. When the maximum number of linearly independent vectors is finite, say n, the vector space is called of finite dimension n. Otherwise the space is called infinite-dimensional. If V′ is a linear subspace of the n-dimensional space V (all elements of V′ belong simultaneously to V ), and V′ contains a set B of m linearly independent vectors then m < n, because B belongs to the n-dimensional space V. It follows that m is finite and that all subspaces of finite-dimensional spaces are finite-dimensional. If m is the maximum number of linearly independent vectors in V′ then this subspace is of dimension m < n. For finite n it can be shown that V′ coincides with V (is an "improper" subspace) if and only if n = m.

Examples of vector spaces

Sequences

The set of all sequences {x1, x2, …, xn} of n elements of a field, in particular, the real numbers. Except for the Euclidean plane, the best known vector space is the space .  For integral finite n > 0 this space can be represented as columns (stacks) of n real numbers. In order to make the discussion concrete we consider the case n = 4. It will be clear how the rules apply to general finite n.

Addition

Because xk and yk are real numbers, xk+yk is a well-defined real number.

Negative vector

Zero vector

Multiplication by real number

Because a and xk are real numbers, axk is well-defined and real.

The reader may easily convince him/herself, using the known properties of real numbers, that these columns of real numbers satisfy the postulates of a vector space. Its dimension is at least 4, because the following 4 vectors are linearly independent,

Indeed, assume that one or more of the coefficients (real numbers) ak is not equal to zero, then the equation

leads to all four a′s are zero (two vectors are equal if and only if their corresponding elements are equal). This is in contradiction to the assumption that one or more of the coefficients ak is not equal to zero.

The set (1) is maximally linearly independent because any non-zero vector can be expressed in the four vectors,

The equation on the right is a valid equation between five vectors that do not have a prefactor zero and yet give the zero vector. Hence it is not possible to find a fifth vector linearly independent of the vectors (1): any five vectors form a linearly dependent set. In other words, the four vectors in Eq. (1) form a basis of the vector space .

Polynomials

The set of all polynomials of n variables {xi} with various coefficients {αi} from a field. Given polynomials like:

and

it is clear that the various operations above are directly represented by a mapping:

with the various powers of {xi} serving as place markers, so all the operations surveyed above for sequences apply equally here.

Function spaces

Consider the field ℝ of real numbers and I an interval in ℝ. The set C(I) of all real valued continuous functions on I, the set D(I) of all real differentiable functions and the set A(I) of all real analytic function on I are linear spaces contained in the linear space of all real valued functions defined on I.

Applications of vector spaces

Classical mechanics

Quantum mechanics

Differential equations