We have learned about linear algebra in a concrete way, such as solving linear equations, matrix operations, etc. In this chapter, we will learn about the abstract concept of linear spaces, which is a generalization of vectors and matrices. First of all, you may ask, why do we need to learn such abstract concepts? check the following example. We have learned about vectors at least in high school, like displacement, velocity, force, etc. actually, when we say vectors, our first impression is that they must have two or three components, which represent a point in a plane or in space. but after we learned about the linear algebra, we learned that something like is also a vector, even though we cannot imagine what its physical picture in the real world.
So, basically, the change of the concept of vector to a arbitrary number of components is a generalization or a abstraction to the original concept. so that we can handle more with the new definition. but at the cost of losing some physical meaning. In this lecture, we will repeat this process, to further abstract the concept of vector, so that we can figure out some shared, common, or general properties of these.
let us recall the process of the abstraction from to , here, I will give a more rigorous definition, as the first step to get you familiar with the algebraic structure.
Recall that if given a structure like , is this a vector? No, definitely not. actually, it is just a tuple, which is an ordered finite sequence or list of elements. btw you can construct it with ordered pairs, anyway. Although we usually by default treat it as a vector, that because we can define a addition and scalar multiplication on it, in a very natural way. but with the structure of a tuple only, we cannot do any operation on it, unless you define some in advance. for example, if you try to run
in Python, it will return (1, 2, 3, 1, 2, 3), and if you try to write on your paper, the reader will think it is by default. that because for the programming languages, the + operator is overloaded to concatenate two tuples. but for the math, especially in the Euclid space, the + operator is defined to add their elements one by one. so you will understand that it is important to declare or at least to know what a symbol exactly means before using it, otherwise you will get confused. so, the tuple is not a vector, but we can define a vector structure on it, and then it becomes a vector.
Application of Vectors
Then, let's return back to the topic of vectors. for example, we know that force can be decomposed into two orthogonal components, and we use this to analyze mechanics problems. Why can we do that? because the force is a vector. but similarly, a sine signal
can be decomposed into two components and , we usually use a constellation to represent this decomposition. Have you ever noticed the similarity between the two cases? therefore, our goal today is to come up with a structure to deal with all the similar cases.
From these two examples above, looking like a list is neither a sufficient condition for being a vector (because addition on lists isn't necessarily vector addition), nor is it a necessary condition (because even if something doesn't look like a list, like a function or a polynomial, it can still behave exactly like a vector).So, Finally, what is a vector? how should we define what a vector is?
Here we will first introduce the concept of field, which are the basic algebraic structures that we will use to define linear spaces.
Introduction to Field
Before we start talking about the vector itself, let's begin from a more basic concept. Our story will starts with a set, the fundamental building block of mathematics. Suppose there is a set . it is too trivial(boring) just like the tuple we mentioned above. with a set only we can seldom do anything.
So, we want to define a binary operation on the set , which is a function that maps two elements of the set to the set itself.
A binary operation on a set is a function
Here is some examples of binary operations:
- defined on , (is a binary operation?)
- defined on sets are also binary operation. (are binary operations?)
- defined on , (is a binary operation?)
the we can define¹ a field is a set with two binary operations, addition and multiplication, that satisfy certain properties. Note that the name "addition" and "multiplication" are just names, they do not necessarily mean the same as the addition and multiplication of real numbers. The important thing is that these operations satisfy certain properties.then the "addition" and "multiplication" satisfies' the following properties.
A field is a set with two binary operations, and , such that:
- Associativity of addition
-
- Commutativity of addition
-
- Additive identity element²
-
- Inverse element for addition
-
- Associativity of multiplication
-
- Commutativity of multiplication
-
- Multiplicative identity element
-
- Inverse element for multiplication
-
- Distributivity of multiplication over addition
-
Actually, we call the structure that satisfied the first 4 properties above an abelian group. So, if we define the abelian group, then a more simple definition can be given as,
the most common example of a field is the rational numbers, you can easily verify that the rational numbers satisfy all the properties of a field. The real numbers are also a field, and so are the complex numbers.
- (rational numbers)
- (real numbers)
- (complex numbers)
Meanwhile, the integers are not a field, because they do not have multiplicative inverses for all non-zero elements (for example ), the natural numbers are not even a group [Task: search for the definition of groups, and tell why.].
Vector Space
Now We're ready to define what a vector space is.
A vector space over , consists of a set (whose elements are called vectors) and a field (whose elements are called scalars), together with two operations:
- Vector addition
- Scalar multiplication³
These operations must satisfy the following axioms. For vectors and scalars :
-
Associativity:
-
Commutativity:
-
Identity element: There exists such that for all
-
Inverse element: For every , there exists such that
-
Associativity:
-
Identity: ⁴
-
Distributivity over vector addition:
-
Distributivity over scalar addition:
⁵ Note that these eight axioms completely characterize what we mean by a vector space. If a set with operations satisfies these axioms over some field , then we call it a vector space over .
For now, we can eventually answer the question, what is a vector? the answer is simple but abstract: A vector is an element of a vector space. And the vector space is defined by the eight axioms above.
Now let's look at some concrete examples to see how this abstract definition applies to familiar and not-so-familiar cases.
The most familiar example is
over the field . Here:
- Vector addition:
- Scalar multiplication:
- Additive identity:
- Additive inverse:
You can verify that all eight axioms are satisfied.
Let be the set of all polynomials of degree at most with real coefficients:
Here:
- Vector addition:
- Scalar multiplication:
- Additive identity:
- Additive inverse:
Notice how polynomials, which don't "look like" vectors in the geometric sense, still satisfy all the vector space axioms!
Let be the set of all functions mapping a nonempty set to . We define
- Addition:
- Scalar multiplication:
- additive identity: const function for all
- additive inverse:
then is a vector space over .
This also forms a vector space over . Functions can be thought of as "vectors" in a very abstract sense, where the "components" are the function values at each point in the domain. and we can restric this function vector space to a subset of functions, while still satisfying the vector space axioms.
Let be the set of all continuous functions defined on an interval . The structure is defined basically the same as in , but with the additional property that the functions are continuous over the interval .
by the continuous functions' properties, we know that the sum of two continuous functions is also continuous, and the scalar multiplication of a continuous function is also continuous. So this set also forms a vector space over .
Furthermore, let be the set of all continuously differentiable functions⁶ . The structure is defined similarly, but with the additional property that the functions have continuous first derivatives. similarly, we can define as the set of all functions with continuous derivatives up to order .
They are all vector spaces over . the proof is left to the reader as exercises.
This is the example we started with!
The set of all solutions to the differential equation
is a vector space over . where are known coefficients. The addition and scalar multiplication are defined the same as in .
Some useful properties of this vector space can be derived from the definition(axioms) directly. and these properties should be proved before we use them. they are obviously true, but quite tricky to prove.
Suppose is a vector space. Let . Suppose and are additive inverses of . Then
∎
By the uniqueness of the additive identity, the notation is well-defined as the unique additive inverse of . and we can define the substraction operation as .
Let . By the definition of scalar multiplication, we have: Suppose is the additive inverse of , so that . Then we have:
∎
Let and be the additive identity. Then: By the definition of additive identity, we have: Therefore,
∎
This equation says that , when added to , gives . Thus is the additive inverse of , as desired. ∎
Subspaces
Now that we understand vector spaces, let's talk about subspaces. A subspace is essentially a "vector space within a vector space."
A subset over the same is a subspace of if and only if:
- is closed under vector addition:
- is closed under scalar multiplication:
Note that if these three conditions are satisfied, then automatically inherits all the vector space axioms from , so is itself a vector space. meanwhile, if is a subset of but does not satisfy these conditions, it won't be a subspace.
In , any line passing through the origin forms a subspace. For instance:
You can verify:
- If , then
- If and , then
Similarly, any plane or line through the origin in is a subspace. but note that those that do not pass through the origin are not subspaces.
In the vector space of , the set of even/odd functions forms a subspace.
- It contains the function.
- If and are even/odd, then is also even/odd.
- If is even/odd and , then is also even/odd.
The set of all continuous functions on an interval forms a subspace of the vector space of all functions .
- constant value function is in
- If , then is also continuous, so
- If and , then is also continuous, so
The set of all solutions to a linear homogeneous differential equation in forms a subspace of the vector space of functions.
- The zero function is a solution (the trivial solution).
- If and are solutions, then is also a solution.
- If is a solution and , then is also a solution.
Sum of Subspaces
Now let's talk about the sum of two subspaces. Given two subspaces and of a vector space , their sum, denoted , is defined as:
The sum of two subspaces and of a vector space is the set:
but note that is not the same as . The sum is a vector space itself, and it contains all possible sums of vectors from and whereas just combines the elements of both subspaces so it will not must be a vector space.
Suppose is the subspace of all vectors of the form in , and is the subspace of all vectors of the form . Then:
forms the plane while is just the union of -axis and -axis, which is not a vector space.
The reader can verify that contains the additive identity and is closed under addition and scalar multiplication. Thus it is a subspace of .
The subspaces are all contained in (to see this, consider sums where all except one of the 's are ). Conversely, every subspace of containing contains (because subspaces must contain all finite sums of their elements). Thus is the smallest subspace of containing . ∎