4.1. Vector Spaces#

4.1.1. Algebraic Structures#

In mathematics, the term algebraic structure refers to an arbitrary set with one or more operations defined on it.

  • Simpler algebraic structures include groups, rings, and fields.

  • More complex algebraic structures like vector spaces are built on top of the simpler structures.

We will develop the notion of vector spaces as a progression of these algebraic structures.

4.1.1.1. Groups#

A group is a set with a single binary operation. It is one of the simplest algebraic structures.

Definition 4.1 (Group)

Let G be a set. Let :G×GG be a binary operation defined on G mapping (g1,g2)(g1,g2) and denoted as

(g1,g2)g1g2.

If the binary operation satisfies the following properties:

  1. [Closure] The set G is closed under the binary operation . i.e.

    g1,g2G,g1g2G.
  2. [Associativity] For every g1,g2,g3G

    g1(g2g3)=(g1g2)g3
  3. [Identity element] There exists an element eG such that

    ge=eg=ggG
  4. [Inverse element] For every gG there exists an element g1G such that

    gg1=g1g=e

then the set G together with the operator denoted as (G,) is known as a group.

Above properties are known as group axioms. Note that commutativity is not a requirement of a group.

Remark 4.1

Frequently, the group operation is the regular mathematical addition. In those cases, we write g1g2 as g1+g2. Otherwise, we will write g1g2 as g1g2.

Often, we may simply write a group (G,) as G when the underlying operation is clear from the context.

4.1.1.2. Commutative groups#

A commutative group is a richer structure than a group. Its elements also satisfy the commutativity property.

Definition 4.2 (Commutative group)

Let (G,) be a group such that its binary operation satisfies an additional property:

  1. [Commutativity] For every g1,g2G

    g1g2=g2g1

Then (G,) is known as a commutative group or an Abelian group.

4.1.1.3. Rings#

A ring is a set with two binary operations defined over it with some properties as described below.

Definition 4.3 (Associative ring)

Let R be a set with two binary operations +:R×RR (addition) mapping (r1,r2)r1+r2 and :R×RR (multiplication) mapping (r1,r2)r1r2 such that (R,+,) satisfies following properties:

  1. (R,+) is a commutative group.

  2. R is closed under multiplication.

    r1r2Rr1,r2R
  3. Multiplication is associative.

    r1(r2r3)=(r1r2)r3r1,r2,r3R
  4. Multiplication distributes over addition.

    r1(r2+r3)=(r1r2)+(r1r3)r1,r2,r3R(r1+r2)r3=(r1r3)+(r2r3)r1,r2,r3R

Then, (R,+,) is known as an associative ring. We denote the identity element for + as 0 and call it additive identity.

Remark 4.2

In the sequel we will write r1r2 as r1r2. We may simply write a ring (R,+,) as R when the underlying operations +, are clear from the context.

There is a hierarchy of ring like structures. In particular we mention:

  1. Associative ring with identity

  2. Field

Definition 4.4 (Associative ring with identity)

Let (R,+,) be an associate ring with the property that there exists an element 1R (known as multiplicative identity) such that

1r=r1=rrR.

Then (R,+,) is known as an associative ring with identity.

4.1.1.4. Fields#

Field is the richest algebraic structure on one set with two operations.

Definition 4.5 (Field)

Let F be a set with two binary operations +:F×FF (addition) mapping (x1,x2)x1+x2 and :F×FF (multiplication) mapping (x1,x2)x1x2 such that (F,+,) satisfies following properties:

  1. (F,+) is a commutative group (with additive identity as 0F).

  2. (F{0},) is a commutative group (with multiplicative identity as 1F).

  3. Multiplication distributes over addition:

    α(β+γ)=(αβ)+(αγ)α,β,γF

Then (F,+,) is known as a field.

The definition above implies a number of properties. For any a,b,cF, we have:

  • Closure: a+bF, abF.

  • Associativity of addition and multiplication: a+(b+c)=(a+b)+c and a(bc)=(ab)c.

  • Commutativity of addition and multiplication: a+b=b+a and ab=ba.

  • Additive identity: 0+a=a+0=a.

  • Multiplicative identity: 1a=a1=a.

  • Additive inverses: a+(a)=(a)+a=0.

  • Multiplicative inverses for every a0: aa1=a1a=1.

  • Distributivity: a(b+c)=(ab)+(ac).

Example 4.1 (Examples of fields)

  • The set of real numbers R is a field.

  • The set of complex numbers C is a field.

  • The Galois field GF-2 is the the set {0,1} with modulo-2 additions and multiplications.

4.1.2. Vector Spaces#

We are now ready to define a vector space. A vector space involves two sets. One set V contains the vectors. The other set F (a field) contains scalars which are used to scale the vectors.

Definition 4.6 (Vector space)

A set V is called a vector space over the field F (or a F-vector space) if there exist two mappings:

  • Vector addition: +:V×VV mapping (v1,v2)v1+v2 for any v1,v2V to another vector in V denoted by v1+v2

  • Scalar multiplication: :F×VV mapping (α,v)αv with αF and vV to another vector in V denoted by αv or just αv

which satisfy following requirements:

  1. (V,+) is a commutative group.

  2. Scalar multiplication distributes over vector addition +:

    α(v1+v2)=αv1+αv2αF;v1,v2V.
  3. Addition in F distributes over scalar multiplication :

    (α+β)v=(αv)+(βv)α,βF;vV.
  4. Multiplication in F commutes over scalar multiplication:

    (αβ)v=α(βv)=β(αv)=(βα)vα,βF;vV.
  5. Scalar multiplication from multiplicative identity 1F satisfies the following:

    1v=vvV.

Some remarks are in order:

  • V as defined above is also known as a F vector space.

  • Elements of V are known as vectors.

  • Elements of F are known as scalars.

  • There are two 0s involved: 0F and 0V. It should be clear from context which 0 is being referred to.

  • 0V is known as the zero vector.

  • All vectors in V{0} are non-zero vectors.

  • We will typically denote elements of F by α,β,.

  • We will typically denote elements of V by v1,v2,.

We quickly look at some vector spaces which will appear again and again in our discussions.

4.1.2.1. N-Tuples#

Example 4.2 (n-tuples as a vector space)

Let F be some field. The set of all n-tuples (a1,a2,,an) with a1,a2,,anF is denoted as Fn. This is a vector space over F with the operations of coordinate-wise addition and scalar multiplication.

Let u,vFn with

u=(u1,,un)

and

v=(v1,,vn).

Vector addition is defined as

u+v(u1+v1,,un+vn).

The zero vector is defined as:

0=(0,,0).

Let cF. Scalar multiplication is defined as

cu(cu1,,cun).

u,v are called equal if u1=v1,,un=vn.

In matrix notation, vectors in Fn can also be written as row vectors:

u=[u1un]

or column vectors:

u=[u1un].

We next verify that Fn equipped with the vector addition and scalar multiplication defined above satisfies all the properties of a vector space. Let us assume x,y,zFn and α,β,γF.

We can now verify all the properties of the vector space:

  1. Fn is closed under vector addition.

    x+y=(x1+y1,,xn+yn).

    Since xi+yiF, hence x+yFn.

  2. Vector addition is associative.

    x+(y+z)=(x1,,xn)+((y1,,yn)+(z1,,zn))=(x1,,xn)+(y1+z1,,yn+zn)=(x1+(y1+z1),,xn+(yn+zn))=((x1+y1)+z1,,(xn+yn)+zn)=(x1+y1,,xn+yn)+(z1,,zn)=((x1,,xn)+(y1,,yn))+(z1,,zn)=(x+y)+z.
  3. Vector addition is commutative.

    x+y=(x1+y1,,xn+yn)=(y1+x1,,yn+xn)=y+x.
  4. Additive identity:

    x+0=(x1,,xn)+(0,,0)=(x1,,xn)=x.

    Since commutativity is already established, hence:

    0+x=x+0=x.

    Thus, 0 is the additive identity element.

  5. Additive inverse: Let

    u=(x1,,xn).

    Then,

    x+u=(x1,,xn)+(x1,,xn)=(x1+(x1),,xn+(xn))=(0,,0)=0.

    By commutativity u+x=x+u=0. Thus, every vector has an additive inverse. We can write u=x.

  6. Scalar multiplication distributes over vector addition.

    α(x+y)=α(x1+y1,,xn+yn)=(α(x1+y1),,α(xn+yn))=(αx1+αy1,,αxn+αyn)=(αx1,,αxn)+(αy1,,αyn)=α(x1,,xn)+α(y1,,yn)=αx+αy.
  7. Addition in F distributes over scalar multiplication.

    (α+β)x=(α+β)(x1,,xn)=((α+β)x1,,(α+β)xn)=(αx1+βx1,,αxn+βxn)=(αx1,,αxn)+(βx1,,βxn)=αx+βx.
  8. Multiplication commutes over scalar multiplication:

    (αβ)x=(αβ)(x1,,xn)=((αβ)x1,,(αβ)xn)=(α(βx1),,α(βxn))=α(βx1,,βxn)=α(β(x1,,xn))=α(βx).

    Similarly,

    (αβ)x=(βα)x=β(αx).
  9. Scalar multiplication from multiplicative identity of F.

    1x=1(x1,,xn)=(x1,,xn)=x.

Thus, Fn supports all the properties of a vector space. It is indeed a vector space.

Example 4.3 (A field is a vector space)

A field of scalars F is a vector space over F in its own right. One way to see this is by treating the scalar as a tuple with one value (i.e., 1-tuple). So, we can write it as F1.

The real line R is a vector space in its own right as R1. Similarly, the complex plane C is a vector space.

More formally, let V=F. We will call the element of V as vectors and elements of F as scalars.

  1. For every xF, let x be corresponding vector in V with x=x.

  2. Define vector addition on V as as the corresponding addition in the field F:

    x+y=x+yx,yV.
  3. Define scalar multiplication on V as the corresponding multiplication in the field F:

    αx=αxαFandxV.
  4. Define zero vector as 0F: 0=0.

We can now verify all the properties of the vector space:

  1. V is closed under vector addition. Let x+yV. Then, x+y=x+y. Since x+yF, hence x+yV.

  2. Vector addition is associative. For every x,y,zV:

    x+(y+z)=x+(y+z)=(x+y)+z=(x+y)+z.
  3. Vector addition is commutative.

    x+y=x+y=y+x=y+x.
  4. For every xV,

x+0=x+0=x=x=x=0+x=0+x.

Thus, 0 is the additive identity element.

  1. Let xV. Define y=x. Since xF, hence yV. Now,

    x+y=x+(x)=0=0.

    Thus, every vector has an additive inverse.

  2. Scalar multiplication distributes over vector addition. Let αF and x,yV.

    α(x+y)=α(x+y)=αx+αy=αx+αy.
  3. Addition in F distributes over scalar multiplication. Let α,βF and xV.

    (α+β)x=(α+β)x=αx+βx=αx+βx.
  4. Multiplication commutes over scalar multiplication:

    (αβ)x=(αβ)x=α(βx)=α(βx).
  5. Scalar multiplication from multiplicative identity of F.

    1x=1x=x=x.

Thus, V=F satisfies all requirements of being a vector space. It is indeed a vector space in its own right.

4.1.2.2. Matrices#

Example 4.4 (Matrices)

Let F be some field. A matrix is an array of the form

[a11a12a1Na21a22a2NaM1aM2aMN]

with M rows and N columns where aijF.

The set of these matrices is denoted as FM×N which is a vector space with operations of matrix addition and scalar multiplication.

Let A,BFM×N. Matrix addition is defined by

(A+B)ijAij+Bij.

Let cF. Scalar multiplication is defined by

(cA)ijcAij.

4.1.2.3. Polynomials#

Example 4.5 (Polynomials)

Let F[x] denote the set of all polynomials with coefficients drawn from field F. i.e. if f(x)F[x], then it can be written as

f(x)=anxn+an1xn1++a1x+a0

where aiF.

The set F[x] is a vector space with usual operations of addition and scalar multiplication

f(x)+g(x)=(an+bn)xn++(a1+b1)x+(a0+b0).
cf(x)=canxn++ca1x+ca0.

4.1.2.4. Vector Space Identities#

Some useful identities are presented here.

Theorem 4.1 (Uniqueness of additive identity)

The 0 vector in a vector space V is unique.

Proof. Assume that there is another vector o satisfying

v+o=o+v=vvV.

Then, in particular, it satisfies: 0+o=0. At the same time, we have o=o+0. Thus,

0=0+o=o+0=o.

Thus, 0 is unique.

Theorem 4.2 (Cancellation law)

Let V be an F vector space. Let x,y,z be some vectors in V such that x+z=y+z. Then x=y.

Proof. Since zV, there exists an additive inverse vV such that z+v=0.

Now

x=x+0=x+(z+v)=(x+z)+v=(y+z)+v=y+(z+v)=y+0=y.

Corollary 4.1 (Uniqueness of additive inverse)

The additive inverse of a vector x in V is unique.

Proof. Let y and z be additive inverses of x. Then, we have:

x+y=0=x+z.

By cancellation law, y=z.

Thus, the additive inverse of a vector is unique.

Theorem 4.3

In a vector space V the following statements are true

  1. 0x=0xV.

  2. a0=0aF.

  3. (a)x=(ax)=a(x)aF and xV.

Proof. (1) 0x=0xV.

By distributive law:

0x=(0+0)x=0x+0x.

By existence of zero vector:

0x=0+0x.

Thus,

0x=0x+0x=0+0x.

Now, by cancellation law:

0x=0.

(2) a0=0aF.

Due to additive identity and distribution law:

a0=a(0+0)=a0+a0.

Due to additive identity:

a0=0+a0.

Combining

a0+a0=0+a0.

Now applying the cancellation law, we get:

a0=0.

(3) (a)x=(ax)=a(x)aF and xV.

We start with

0=0x=(a+(a))x=ax+(a)x.

Thus, (a)x is the additive inverse of ax. Thus, (a)x=(ax).

Similarly,

0=a0=a(x+(x))=ax+a(x).

Thus, a(x) is also additive inverse of ax. Thus, a(x)=(ax).

4.1.3. Linear Independence#

Definition 4.7 (Linear combination)

A linear combination of two vectors v1,v2V is defined as

αv1+βv2

where α,βF.

A linear combination of p vectors v1,,vpV is defined as

i=1pαivi

where αiF.

Note that a linear combination always consists of a finite number of vectors.

Definition 4.8 (Linear combination vector)

Let V be a vector space and let S be a nonempty subset of V. A vector vV is called a linear combination of vectors of S if there exist a finite number of vectors s1,s2,,snS and scalars a1,,aN in F such that

v=a1s1+a2s2+ansn.

We also say that v is a linear combination of s1,s2,,sn and a1,a2,,an are the coefficients of linear combination.

Definition 4.9 (Trivial linear combination)

A linear combination a1s1+a2s2+ansn is called trivial if a1=a2==an=0.

0 is a trivial linear combination of any subset of V as:

0=0s1+0s2+0sn.

A linear combination may refer to the expression itself or its value. e.g. two different linear combinations may have same value.

Definition 4.10 (Linear dependence)

A finite set of non-zero vectors {v1,,vp}V is called linearly dependent if there exist α1,,αpF not all 0 such that

i=1pαivi=0.

Since αi are not all zero, hence, it is a non-trivial combination.

Definition 4.11 (Linearly dependent set)

A set SV is called linearly dependent if there exist a finite number of distinct vectors u1,u2,,unS and scalars a1,a2,,anF not all zero, such that

a1u1+a2u2++anun=0.

In other words, there exists a non-trivial linear combination which equals 0.

Definition 4.12 (Linearly independent set)

A set SV is called linearly independent if it is not linearly dependent.

Definition 4.13 (Linearly independent vectors)

More specifically a finite set of non-zero vectors {v1,,vn}V is called linearly independent if

i=1nαivi=0αi=01in.

In other words, the only linear combination giving us 0 vector is the trivial linear combination.

Example 4.6 (Examples of linearly dependent and independent sets)

  • The empty set is linearly independent.

  • A set of a single non-zero vector {v} is always linearly independent. Prove!

  • If two vectors are linearly dependent, we say that they are collinear.

  • Alternatively if two vectors are linearly independent, we say that they are not collinear.

  • If a set {v1,,vp} is linearly independent, then any subset of it will be linearly independent. Prove!

  • Adding another vector v to the set may make it linearly dependent. When?

  • It is possible to have an infinite set to be linearly independent. Consider the set of polynomials {1,x,x2,x3,}. This set is infinite, yet linearly independent.

Theorem 4.4

Let V be a vector space. Let S1S2V. If S1 is linearly dependent, then S2 is linearly dependent.

Corollary 4.2

Let V be a vector space. Let S1S2V. If S2 is linearly independent, then S1 is linearly independent.

4.1.4. Span#

Vectors can be combined to form other vectors. It makes sense to consider the set of all vectors which can be created by combining a given set of vectors.

Definition 4.14 (Span)

Let SV be a subset of vectors. The span of S denoted as S or spanS is the set of all possible linear combinations of vectors belonging to S.

spanS=S{vV|v=i=1pαivifor someviS;αiF;pN}

For convenience we define span={0}.

Span of a finite set of vectors {v1,,vp} is denoted by v1,,vp.

v1,,vp={i=1pαivi|αiF}.

We say that a set of vectors SV spans V if S=V.

Proposition 4.1

Let SV, then span(S)V.

Definition 4.15 (Spanning a vector space)

Let SV. We say that S spans (or generates) V if

S=V.

In this case we also say that vectors of S span (or generate) V.

Theorem 4.5

Let S be a linearly independent subset of a vector space V and let vVS. Then S{v} is linearly dependent if and only if vspan(S).

4.1.5. Basis#

Definition 4.16

A set of linearly independent vectors B is called a basis of V if B=V; i.e., B spans V.

Example 4.7 (Basis examples)

  • Since span()={0} and is linearly independent, is a basis for the zero vector space {0}.

  • The basis {e1,,eN} with e1=(1,0,,0), e2=(0,1,,0), , eN=(0,0,,1), is called the standard basis for FN.

  • The set {1,x,x2,x3,} is the standard basis for F[x]. Indeed, an infinite basis. Note that though the basis itself is infinite, yet every polynomial pF[x] is a linear combination of finite number of elements from the basis.

We review some properties of bases.

Theorem 4.6 (Unique representation)

Let V be a vector space and B={v1,v2,,vn} be a subset of V. Then B is a basis for V if and only if each vV can be uniquely expressed as a linear combination of vectors of B. In other words:

v=a1v1+a2v2++anvn

for unique scalars a1,,an.

This theorem states that a basis B provides a unique representation to each vector vV where the representation is defined as the n-tuple (a1,a2,,an).

Proof. Assume that each vV can be expressed uniquely as a linear combination of vectors of B.

  1. Then 0 is a unique linear combination of B.

  2. But 0=0v1++0vn.

  3. Thus, B is linearly independent.

  4. Also, B spans V since every vector can be expressed as a linear combination.

  5. Thus, B is a basis.

Assume that B is a basis for V.

  1. Then, B is linearly independent and B spans V.

  2. Let vV.

  3. Let v=i=1naivi and v=i=1nbivi be two different representations of v in B.

  4. Then,

    i=1naivi=i=1nbivii=1n(aibi)vi=0.
  5. But B is linearly independent.

  6. Hence, aibi=0 for i=1,,n.

  7. Thus, ai=bi for i=1,,n.

  8. Hence, every vV has a unique representation as a linear combination of B.

If the basis is infinite, then the above theorem needs to be modified as follows:

Theorem 4.7 (Unique representation for infinite basis)

Let V be a vector space and B be a subset of V. Then B is a basis for V if and only if each vV can be uniquely expressed as a linear combination of vectors of B:

v=a1v1+a2v2++anvn

for unique scalars a1,,an and unique vectors v1,v2,vnB.

Theorem 4.8

If a vector space V is spanned by a finite set S, then some subset of S is a basis for V. Hence V has a finite basis.

Theorem 4.9 (Replacement theorem)

Let V be a vector space that is spanned by a set G containing exactly n vectors. Let L be a linearly independent subset of V containing exactly m vectors.

Then mn and there exists a subset H of G containing exactly nm vectors such that LH spans V.

Corollary 4.3

Let V be a vector space having a finite basis. Then, every basis for V contains the same number of vectors.

4.1.6. Dimension#

Definition 4.17 (Dimension of vector space)

A vector space V is called finite-dimensional if it has a basis consisting of a finite number of vectors. This unique number of vectors in any basis B of the vector space V is called the dimension or dimensionality of the vector space. It is denoted as dimV. We say:

dimV|B|

If V is not finite-dimensional, then we say that V is infinite-dimensional.

Example 4.8 (Vector space dimensions)

  • Dimension of FN is N.

  • Dimension of FM×N is MN.

  • The vector space of polynomials F[x] is infinite dimensional.

Proposition 4.2

Let V be a vector space with dimension n.

  1. Any finite spanning set for V contains at least n vectors, and a spanning set that contains exactly n vectors is a basis for V.

  2. Any linearly independent subset of V that contains exactly n vectors is a basis for V.

  3. Every linearly independent subset of V can be extended to a basis for V.

4.1.7. Ordered Basis#

Definition 4.18 (Ordered basis)

For a finite dimensional vector space V, an ordered basis is a basis for V with a specific order. In other words, it is a finite sequence (or tuple) of linearly independent vectors in V that spans V.

Typically, we will write an ordered basis as B={v1,v2,,vn} and assume that the basis vectors are ordered in the order they appear.

With the help of an ordered basis, we can define a coordinate vector.

Definition 4.19 (Coordinate vector)

Let B={v1,,vn} be an ordered basis for V. For any xV, let α1,,αn be unique scalars such that

x=i=1nαivi.

The coordinate vector of x relative to B is defined as

[x]B=[α1αn].

4.1.8. Subspaces#

Definition 4.20 (Subspace)

Let W be a subset of V. Then W is called a subspace if W is a vector space in its own right under the same vector addition + and scalar multiplication operations. i.e.

+:W×WW(w1,w2)w1+w2w1,w2W
:F×WW(α,w)αwαwαF;wW

are defined by restricting +:V×VV and :V×VV to W and W is closed under these operations.

Example 4.9 (Trivial subspaces)

  • V is a subspace of V.

  • {0} is a subspace of any V.

Theorem 4.10 (Subspace characterization)

A subset WV is a subspace of V if and only if

  • 0W.

  • x+yW whenever x,yW.

  • αxW whenever αF and xW.

In other words, a subset of V is a subspace if and only if it contains the zero vector and is closed under vector addition and scalar multiplication.

Example 4.10 (Symmetric matrices)

A matrix MFM×N is symmetric if

MT=M.

The set of symmetric matrices forms a subspace of set of all M×N matrices.

Example 4.11 (Diagonal matrices)

A matrix M is called a diagonal if Mij=0 whenever ij.

The set of diagonal matrices is a subspace of FM×N.

Definition 4.21 (Proper subspace)

A subspace W of V is called a proper subspace if VW.

In other words, there are vectors in V which are not included in W.

Theorem 4.11 (Intersection of subspaces)

Any intersection of subspaces of a vector space V is a subspace of V.

Proof. Let {Wi}iI be a family of subspaces of a vector space V indexed by I.

Let the intersection of the subspaces be:

W=iIWi.

[Zero vector]

  1. Since 0Wi for every iI, hence 0W.

[Vector addition]

  1. Let x,yWi for every iI.

  2. Then, x+yWi for every iI since Wi are closed under vector addition.

  3. Then, x+yW.

  4. Thus, W is closed under vector addition.

[Scalar multiplication]

  1. Let xWi for every iI and let tF.

  2. Then, txWi for ever iI since Wi are closed under scalar multiplication.

  3. Thus, txW.

  4. Thus, W is closed under scalar multiplication.

Thus, W is a subspace of V.

We note that a union of subspaces is not necessarily a subspace, since its not closed under addition.

Theorem 4.12 (Span is a subspace)

Let V be a vector space. The span of a set SV given by spanS is a subspace of V. Moreover, any subspace of V that contains S must also contain the span of S.

Proof. If S=, then by convention, spanS={0}. This happens to be the trivial subspace.

We now consider S to be nonempty (possibly infinite).

[Zero vector]

  1. Since S is nonempty, hence there exists some vS.

  2. Then, 0v=0spanS as it is a linear combination of elements of S.

[Scalar multiplication]

  1. Let vspanS.

  2. Then, v=i=1ptivi where viS, tiF, pN.

  3. But then for any tF:

    tv=t(i=1ptivi)=i=1p(tti)vi.
  4. Thus, tv is also a linear combination of elements of S.

  5. Thus, spanS is closed under scalar multiplication.

[Vector addition]

  1. Let x,yspanS.

  2. Let x=i=1prixi and y=i=1qtiyi be linear combinations of elements of S.

  3. Then,

    x+y=i=1prixi+i=1qtiyi

    is also a linear combination of (up to p+q) elements of S.

  4. Thus, x+yspanS.

  5. Thus, spanS is closed under vector addition.

Combining, spanS is a linear subspace.

Let W be a subspace such that SW.

  1. Let vspanS.

  2. Then, v=i=1ptivi is a linear combination of elements of S.

  3. But W is closed under linear combinations and SW.

  4. Thus, vW.

  5. Thus, spanSW.

This theorem is quite useful. It allows us to construct subspaces from a given basis.

Let B be a basis of an n dimensional space V. There are n vectors in B. We can create 2n distinct subsets of B. Thus we can easily construct 2n distinct subspaces of V.

Choosing some other basis lets us construct another set of subspaces.

An n-dimensional vector space has infinite number of bases. Correspondingly, there are infinite possible subspaces.

Definition 4.22 (Partial order on subspaces)

Let V be a vector space. If W1 and W2 are two subspaces of V then we say that W1 is smaller than W2 if W1W2; i.e., W1 is a proper subset of W2. We denote this partial order by W1W2.

If W1W2, we denote this is W1W2.

Theorem 4.13 (Span as the smallest subspace)

Let V be a vector space. Let v1,,vp be some arbitrary vectors in V. Let S={v1,,vp}. Let the span of S be denoted as

U=spanS=v1,,vp.

Let W be the smallest subspace containing the vectors in S. Then

W=U.

i.e. W is same as the span of {v1,,vp}. In other words, spanS is the smallest subspace containing S.

Proof. Since W contains S, hence spanSW due to Theorem 4.12. spanS is also a linear subspace. So W=spanS must hold.

If the vector space V is not finite dimensional, then its subspaces may or may not be finite dimensional. E.g., if W is a span of a fixed p number of vectors from V, then it is finite dimensional with dimWp irrespective of whether V is finite dimensional or not.

However, if V is finite dimensional, then its subspaces must be finite dimensional.

Theorem 4.14 (Subspaces of a finite dimensional vector space)

Let W be a subspace of a finite-dimensional vector space V. Then W is finite dimensional and

dimWdimV.

Moreover, if

dimW=dimV,

then W=V.

Proof. If V is finite dimensional, then it has a finite basis of say n basis vectors. Since WV hence the basis spans W too. Thus, dimension of W cannot be greater than n.

Now, assume that

n=dimW=dimV.
  1. Let B be a basis for W.

  2. Then B spans W and contains n linearly independent vectors.

  3. Since WV, hence BV.

  4. But n=dimV.

  5. Hence, any set of n linearly independent vectors must be a basis for V too.

  6. Thus, B is a basis for V.

  7. Thus, V=W=spanB.

Corollary 4.4

If W is a subspace for a finite-dimensional vector space V then any basis for W can be extended to a basis for V.

Definition 4.23 (Subspace codimension)

Let V be a finite dimensional vector space and W be a subspace of V. The codimension of W is defined as

codimW=dimVdimW.

4.1.9. Direct Sum of Vector Spaces#

Consider two vector spaces V and W over same scalar field F. Often, it is useful to consider the Cartesian product set V×W and add a vector space structure to it. One common example is V=Rn, W=Rm and V×W=Rn+m.

Definition 4.24 (Direct sum of vector spaces)

Let V and W be two different vector spaces over some field F. A vector space structure over the field F can be introduced to the set V×W as per the following definitions.

  1. Let 0vV and 0wW be the additive identities of V and W respectively. The additive identity of V×W is given by

    0=(0v,0w).
  2. [Vector addition] Let (x,u) and (y,v) be in V×W. Then, their sum is defined as:

    (x,u)+(y,v)(x+y,u+v).
  3. [Scalar multiplication] Let (x,u)V×W and αF. Then, the scalar multiplication is defined as:

    α(x,u)(αx,αu).

V×W equipped with these definitions is a vector space over F in its own right and is called the direct sum of V and W denoted by VW.

Note

Some authors prefer to call VW as direct product (Cartesian product with vector space structure) and reserve the term direct sum for the sums within the same vector space (Definition 4.29 below).

For them, external direct sum is a direct product and internal direct sum is a direct sum.

Example 4.12

With the definition of direct sum given above, we have:

F2=FF

or more generally,

Fn=FFF=i=1nF.

Specifically,

Rn=i=1nR.

Also,

Rn+m=Rn×Rm.

4.1.10. Sets in Vector Spaces#

4.1.10.1. Set Arithmetic#

Definition 4.25 (Arithmetic on sets)

Let C, D be subsets of an F vector space V. Let zV. Let λF. Let ΛF.

The addition of sets is defined as:

C+D{x+y|xC,yD}.

The subtraction of sets is defined as:

CD{xy|xC,yD}.

Addition of a set with a vector is defined as:

z+C{z+x|xC}={z}+C.

Subtraction of set with a vector is defined as:

Cz{xz|xC}=C{z}.

Scalar multiplication of a set with a scalar is defined as:

λC{λx|xC}.

Multiplication of a set of scalars with a set of vectors is defined as:

ΛCλΛλC.

Multiplication of a set of scalars with a vector is defined as:

ΛzΛ{z}={λz|λΛ}.

4.1.10.2. Symmetry#

Definition 4.26 (Symmetric reflection)

The symmetric reflection of a set C across the origin is given by C=(1)C.

Definition 4.27 (Symmetric set)

A set C is said to be symmetric if C=C holds true.

If a nonempty set is symmetric, it must contain 0 since 0=0.

4.1.10.3. Set Arithmetic Properties#

Theorem 4.15 (Set vector arithmetic identities)

Let C, D be subsets of an F vector space V. Let zV. Let λF. Let ΛF.

C=(C+z)z.
CDC+zD+z.

Proof. We show that C=(C+z)z.

  1. Let xC.

  2. Then, x=(x+z)z.

  3. Then, x+zC+z.

  4. Then, (x+z)z(C+z)z.

  5. Thus, C(C+z)z.

  6. Conversely, let x(C+z)z.

  7. Then, x+zC+z.

  8. But then, xC.

  9. Thus, (C+z)zC.

  10. Combining C=(C+z)z.

We show that CDC+zD+z.

Assume CD.

  1. Let xC+z.

  2. Then, xzC.

  3. Then, xzD since CD.

  4. Then, xD+z.

  5. Thus, C+zD+z.

Now, assume C+zD+z.

  1. Let xC.

  2. Then, x+zC+z.

  3. Then, x+zD+z since C+zD+z.

  4. Then, xD.

  5. Thus, CD.

Theorem 4.16 (Properties of set arithmetic)

Let V be a vector space over a field F.

Let C,D,EV and α,βF.

  1. Set addition is commutative:

    C+D=D+C.
  2. Set addition is associative:

    C+(D+E)=(C+D)+E.
  3. Scalar multiplication with a set commutes with multiplication in F:

    α(βC)=(αβ)C=(βα)C=β(αC).
  4. Scalar multiplication distributes over set addition:

    α(C+D)=(αC)+(αD).
  5. The set {0} is the identity element for set addition:

    C+{0}={0}+C=C.
  6. The set F=C+(C) is symmetric and 0F.

  7. (α+β)CαC+βC.

Proof. We shall prove some of the properties.

[Commutativity]

  1. Let uC+D.

  2. Then, there exists, xC and yD such that u=x+y.

  3. But then, u=y+x since vector addition is commutative.

  4. Thus, uD+C.

  5. Thus, C+DD+C.

  6. Similarly, D+CC+D.

[Associativity]

  1. Let aC+(D+E).

  2. Then, there exists cC and bD+E such that a=c+b.

  3. Then, there exists dD and eE such that b=d+e.

  4. Thus, a=c+(d+e).

  5. But vector addition is associative.

  6. Thus, a=(c+d)+e.

  7. Then, c+dC+D.

  8. Thus, a(C+D)+E.

  9. Thus, C+(D+E)(C+D)+E.

  10. Similarly reasoning shows that (C+D)+EC+(D+E).

(α+β)CαC+βC

  1. Let x(α+β)C.

  2. Then, there exists yC such that x=(α+β)y.

  3. Then, x=αy+βy.

  4. Then, αyαC and βyβC.

  5. Thus, xαC+βC.

  6. Thus, (α+β)CαC+βC.

Additive inverses don’t exist for sets containing more than one element.

Theorem 4.17 (Distributive law for set sum over intersection)

Let V be a vector space over a field F.

Let C,D,EV. Then

(CD)+E(C+E)(D+E).

Proof. We show that (CD)+E(C+E)(D+E).

  1. Let v(CD)+E.

  2. Then, v=x+y such that x(CD) and yE.

  3. Then, xC, xD and yE.

  4. Thus, x+yC+E and y+yD+E.

  5. Thus, v=x+y(C+E)(D+E).

  6. Thus, (CD)+E(C+E)(D+E).

We mention that (C+E)(D+E)(CD)+E in general. Here is an example.

  1. Consider V=R2.

  2. Let C to be the x-axis.

  3. Let D to be the y-axis.

  4. Let E to be the line {(x,y)|x=y}.

  5. Both C+E and D+E are the whole plane R2.

  6. Thus, (C+E)(D+E)=R2.

  7. But CD={0}.

  8. Thus, (CD)+E=E.

  9. Clearly, (C+E)(D+E)(CD)+E.

4.1.10.4. Direct Sums#

Definition 4.28 ((External) direct sum over two different vector spaces)

Let V and W be vector spaces over F. Let CV and DW. Then, the direct sum of C and D is defined as

CD{x=(y,z)|yC,zD}.

CD is a subset of the (direct sum) vector space VW.

Definition 4.29 ((Internal) direct sum of sets in same vector space)

Let C,D be subsets of V. If each vector xC+D can be expressed uniquely in the form

x=y+z,yC,zD;

then the sum C+D is called a direct sum of C and D and is denoted by CD.

The key idea behind a direct sum is that it enables a unique decomposition of a vector into its components belonging to individual sets.

Theorem 4.18

Two subsets C,D of V are eligible for a direct sum if and only if the symmetric sets CC and DD have only the zero vector 0 in common.

If two subspaces of V have only 0 in common, then any vector in their sum can be decomposed uniquely between the two subspaces.

4.1.11. Sums of Subspaces#

Definition 4.30 ((Internal) direct sum of subspaces)

If W1 and W2 are two subspaces of V such that W1W2={0}, then the direct sum of W1 and W2 is defined as:

W1W2W1+W2={x1+x2|x1W1,x2W2}.

If, in particular, V=W1+W2, we say that V is a direct sum of its subspaces W1 and W2 and write as V=W1W2.

Example 4.13 (Vector space as direct sum of spans of basis vectors)

The spans of basis vectors have only 0 in common. They indeed offer a unique decomposition of any vector in the space. Thus, for a finite dimensional space V with a basis B={v1,vn}

V=i=1nspan{vi}.

Theorem 4.19 (Dimension of sum of subspaces)

Let W1 and W2 be finite dimensional subspaces of a vector space V. Then

dim(W1+W2)=dimW1+dimW2dim(W1W2).

Theorem 4.20 (Dimension of direct sum of subspaces)

Let W1 and W2 be finite dimensional subspaces of a vector space V such that W1W2={0}. Then

dim(W1W2)=dimW1+dimW2.

Proof. We note that W1W2={0} implies that dim(W1W2)=0. Rest follows from the previous result Theorem 4.19.

4.1.12. Real Vector Spaces#

Definition 4.31 (Real vector space)

A real vector space V is a vector space defined over the real field R.

In other words, the scalars come from the field of real numbers.

There are several features associated with the real field.