# Advanced Mathematics for Engineers and Scientists/Vector Spaces: Mathematic Playgrounds

## Contents

## Vector Spaces: Mathematic Playgrounds[edit]

The study of partial differential equations requires a clear definition of what kind of numbers are being dealt with and in what way. PDEs are normally studied in certain kinds of vector spaces, which have a number of properties and rules associated with them which make possible the analysis and unifies many notions.

### The Real Field[edit]

A **field** is a set that is bundled with two operations on the set called addition and multiplication which obey certain rules, called axioms. The letter will be used to represent the field, and from definition a field requires the following ( and are in ):

**Closure under addition and multiplication**: the addition and multiplication of field members produces members of the same field.**Addition and multiplication are associative**: and .**Addition and multiplication are commutative**: and .**Addition and multiplication are distributive**: and .**Existence of additive identity**: there is an element in notated 0, sometimes called the sum of no numbers, such that .**Existence of multiplicative identity**: there is an element in notated 1 different from 0, sometimes called the product of no numbers, such that .**Existence of additive inverse**: there is an element in associated with notated such that .**Existence of multiplicative inverse**: there is an element in associated with (if is nonzero), notated such that .

These are called the field axioms. The field that we deal with, by far the most common one, is the real field. The set associated with the real field is the set of real numbers, and addition and multiplication are the familiar operations that everyone knows about.

Another example of a set that can form a field is the set of rational numbers, numbers which are expressible as the ratio of two integers. An example of a common set that **doesn't** form a field is the set of integers: there generally is no multiplicative inverse since the reciprocal of an integer generally is not an integer.

Note that when we say that an object is *in* , what is meant is that the object is a member of the set associated in the field and that it complies with the field axioms.

### The Vector[edit]

Most non-mathematics students are taught that vectors are ordered groups ("tuples") of quantities. This is not complete, vectors are a lot more general than that. Informally, **a vector is defined as an object that can be scaled and added with other vectors**. This will be made more specific soon.

Examples of vectors:

- The real numbers.
- Pairs, triples, etc of real numbers.
- Polynomials.
- Most functions.

Examples of objects that are not vectors:

- Members of the extended real numbers. Specifically, the infinity and negative infinity elements neither scale nor add.
- The integers, at least when scaled by real numbers (since the result will not necessarily be an integer).

An interesting (read: confusing) fact to note is that, by the definition above, matrices and even tensors qualify as vectors since they can be scaled or added, even though these objects are considered generalizations of more "conventional" vectors, and calling a tensor a vector will lead to confusion.

### The Vector Space[edit]

A **vector space** can be thought of as a generalization of a field.

Letting represent some field, a vector space over is a set of vectors bundled with two operations called vector addition and scalar multiplication, notated:

- Vector addition: , where .
- Scalar multiplication: , where and .

The members of are called **vectors**, and the members of the field associated with are called **scalars**. Note that these operations imply *closure* (see the first field axiom), so that it does not have to be explicitly stated. **Note also that this is essentially where a vector is defined: objects that can be added and scaled**. The vector space must comply with the following axioms ( and are in ; and are in ):

**Addition is associative**: .**Addition is commutative**: .**Scalar multiplication is distributive over vector addition**: .**Scalar multiplication is distributive over field addition**: .**Scalar and field multiplication are compatible**: .**Existence of additive identity**: there is an element in notated 0 such that .**Existence of additive inverse**: there is an element in associated with notated such that .**Existence of multiplicative identity**: there is an element in notated 1 different from 0 such that .

An example of a vector space is one where polynomials are vectors over the real field. An example of a space that is not a vector space is one where vectors are rational numbers over the real field, since scalar multiplication can lead to vectors that are not rational (implied closure under scalar multiplication is violated).

By analogy with linear functions, vectors are linear by nature, hence a vector space is also called a **linear space**. The name "linear vector space" is also used, but this is somewhat redundant since there is no such thing as a nonlinear vector space. It's now worth mentioning an important quantity called a *linear combination* (not part of the definition of a vector space, but important):

where is a sequence of field members and is a sequence of vectors. The fact that a vector can be formed by a linear combination of other vectors is much of the essence of the vector field.

Note that a field over itself qualifies as a vector space. Fields of real numbers and other familiar objects are sometimes called spaces, since distance and other useful concepts apply.

The definition of a vector space is quite general. Note that, for example, there is no mention of any kind of product between vectors, nor is there a notion of the "length" of a vector. The vector space as defined above is very primitive, but it's a starting point: through various extensions, specific vector spaces can have a lot of nice properties and other features that make our playgrounds fun and comfortable. We'll discuss bases (plural of basis) and then take on some specific vector spaces.

#### The Basis[edit]

A nonempty subset of is called a **linear subspace** of if is itself a vector space. The requirement that be a vector space can be safely made specific by saying that is closed under vector addition and scalar multiplication, since the rest of the vector space properties are inherited.

The **linear span** of a set of vectors in may then be defined as:

Where . The span is the intersection over all choices of . This concept may be extended so that is not necessarily finite. *The span of is the intersection of all of the linear subspaces of *.

Now, think of what happens if a vector is removed from the set . Does the span change? Not necessarily, it may be possible that the remaining vectors in the span are sufficient to "fill in" for the missing vector through **linear combination** of the remaining vectors.

Let be a subset of . If the span of is the same as the span of , and if removing a vector from necessarily changes its span, then the set of vectors is called a **basis** of , and the vectors of are called **linearly independent**. It can be proven that a basis can be constructed for *every* vector space.

Note that the basis is not unique. This obscure definition of a basis is convenient because it is very broad, it is worth understanding fully. An important property of a vector space is that it necessarily has a basis, and that any vector in the space may be written in terms of a linear combination of the members of the basis.

A more understandable (though less elemental) explanation is provided: for a vector space over the field , the vectors form a **basis** of (where are in and the following are in ), satisfying the following properties:

- The basis vectors are linearly independent: if

- then
- without exception.

- The basis vectors span : for some given in , it is possible to choose so that

The basis vectors of a vector space are usually notated with as .

### Euclidean n-Space[edit]

As most students are familiar with Euclidean n-space, this section serves more of an example than anything else.

Let be the field of real numbers, then the *vector space* over is defined to be the space of n-tuples of members of . In other words, more clearly:

- If , then , where are the vectors of the vector space .

These vectors are called *n-dimensional coordinates*, and the vector space is called the **real coordinate space**; note that coordinates (unlike more general vectors) are often notated in boldface, or else with an arrow over the letter. They are also called spatial vectors, geometric vectors, just "vectors" if the context allows, and sometimes "points" as well, though some authors refuse to consider points as vectors, attributing a "fixed" sense to points so that points can't be added, scaled, or otherwise messed with. Part of the reason for this is that it allows one to say that some vector space is bound to a point, the point being called the origin.

The Euclidean n-space is the special real coordinate n-space with some additional **structure** defined which (finally) gives rise to the geometric notion of (specifically these) vectors.

To begin with, an inner product is first defined, notated with either angle braces or a dot:

This quantity, which turns two vectors into a scalar (a member of ) doesn't have a great deal of geometric meaning until some more structure is defined. In a coordinate space, the dot notation is favored, and this product is often called the "dot product", especially when or . The definition of this inner product qualifies as an *inner product space*.

Next comes the norm, in terms of the inner product:

The notation involving single pipes around the letter x is common, again, when or , due to analogy with absolute value, for real and especially complex numbers. For a coordinate space, the norm is often called the length of . This quickly leads to the notion of the distance between two vectors:

Which is simply the length of the vector "from" to .

Finally, the angle between and is defined through, for ,

The *motivation* for this definition of angle, valid for any , comes from the fact that one can prove that the literal measurable angle between two vectors in satisfies the above (the norm is motivated similarly). Discussing these 2D angles and distances of course mandates making precise the notion of a vector as an "arrow" (ie, correlating vectors to things you can draw on a sheet of paper), but that would get involved and most are already subconsciously familiar with this and it's not the point of this introduction.

This completes the definition of . A thorough introduction to Euclidean space isn't very fitting in a text on Partial Differential equations, it is included so that one can see how a familiar vector space can be constructed ground-up through extensions called "structure".

### Banach Spaces[edit]

Banach spaces are more general than Euclidean space, and they begin our departure from vectors as geometric objects into vectors as toys in the crazy world of functional analysis.

To be terse, a Banach space is defined as any **complete normed vector space**. The details follow.

#### The Inner Product[edit]

The inner product is a vector operation which results in a scalar. The vectors are members of a vector space , and the scalar is a member of the field associated with . A vector space on which an inner product is defined is said to be "equipped" with an inner product, and the space is an **inner product space**. The inner product of and is usually notated .

A truly general definition of the inner product would be long. Normally, if the vectors are real or complex in nature (eg, complex coordinates or real valued functions), the inner product must satisfy the following axioms:

**Distributive in the first variable**: .**Associative in the first variable**: .**Nondegeneracy and nonnegativity**: , equality will hold only when .**Conjugate symmetry**: .

Note that if the space is real, the last requirement (the overbar indicates complex conjugation) simplifies to , and then the first two axioms extend to the second variable.

A desirable property of an inner product is some kind of orthogonality. Two nonzero vectors are said to be orthogonal only if their inner product is zero. Remember that we're talking about vectors in general, not specifically Euclidean.

Inner products are by no means unique, good definitions are what add quality to specific spaces. The Euclidean inner product, for example, defines the Euclidean distance and angle, quantities which form the foundation of Euclidean geometry.

#### The Norm[edit]

The norm is usually, though not universally, defined in terms of the inner product, which is why the inner product was discussed first (to be technically correct, a Banach space doesn't necessarily need to have an inner product). The norm is an operation (notated with double pipes) which takes one vector and generates one scalar, necessarily satisfying the following axioms:

**Scalability**: .**The triangle inequality**: .**Nonnegativity**: , equality only when .

The fact that can be proven from the first two statements above.

Definition requires that *only* when (compare this to the inner product, which can be zero even if fed nonzero vectors); if this condition is relaxed so that is possible for nonzero vectors, the resulting operation is called a seminorm.

The **distance** between two vectors and is a useful quantity which is defined in terms of the norm:

The distance is often called the **metric**, and a vector space equipped with a distance is called a **metric space**.

#### Completeness[edit]

As stated before, a Banach space is defined as a **complete normed vector space**. The norm was described above, so that all that is left to establish the definition of a Banach space is **completeness**.

Consider a sequence of vectors in a vector space . This sequence of vectors is called a **Cauchy sequence** if these vectors "tend" toward some "destination" vector, as shown in the pictures at right. Stated precisely, a sequence is a Cauchy sequence if it is always possible to make the distance *arbitrarily small* by picking larger values of and .

The limit of a Cauchy sequence is:

A vector space is called complete if every Cauchy sequence has a limit that is also in . A Banach space is, finally, a vector space equipped with a norm that is complete. Note that completeness implies existence of distance, which means that every Banach space is a metric space.

An example of a vector space that is complete is Euclidean n-space. An example of a vector space that isn't complete is the space of rational numbers over rational numbers: it is possible to form a sequence of rational numbers which limit to an irrational number.

#### Hilbert spaces[edit]

Note that the inner product was defined above but not subsequently used in the definition of a Banach space. Indeed, a Banach space must have a norm but doesn't necessarily need to have an inner product. However, if the norm in a Banach space is defined through the inner product by

then the resulting special Banach space is called a **Hilbert space**. Hilbert spaces are important in the study of partial differential equations (some relevance finally!) because many theorems and important results are valid only in Hilbert spaces.