Jump to content

Solutions To Mathematics Textbooks/Algebra (9780132413770)/Chapter 3

From Wikibooks, open books for an open world

Exercise 1.2

[edit | edit source]

Using an "educated guess" one observes that . With this, it is easy to see that , , and .

Exercise 1.3

[edit | edit source]

, as all the coefficients divisible by 7 reduce to 0.

Exercise 1.10

[edit | edit source]

Let us denote the matrices (appearing in the same order as in the book) by . We need to check the following:

  • is a group with matrix addition and as the identity. We get the following addition table
0 1 A B
0 0 1 A B
1 1 0 B A
A A B 0 1
B B A 1 0

So we see that the elements with addition form an abelian group with as the identity.

  • is a group with as the identity. Again we have the multiplication table
1 A B
1 1 A B
A A B 1
B B 1 A

Again, we see that the elements with matrix multiplication form an abelian group with as the identity.

  • The distributive law follows from the distributive law for matrices in general.

Exercise 1.11

[edit | edit source]

Writing out a product and sum of two elements from the given set, and noticing that the coefficients of both elements, and thus their sums and products are in , implies that the sum and product are in the set. To see that each non-zero element has an inverse in the -operation is trivial. To see the same for the product operation, write the equations coming from the condition , where is a known element from the set and a candidate for its inverse with unknown coefficients as a linear system. By Corollary 3.2.8 this system has a solution. The distribution law is immediate.

Exercise 2.2

[edit | edit source]

a) The space of symmetric matrices is a vector space, since the sum of two symmetric matrices is a symmetric matrix, and a scaling of a symmetric matrix by a scalar is symmetric.

b) The space of invertible matrices is not a vector space, since it does not contain the zero matrix.

c) The space of upper triangular matrices is also a vector space by similar reasoning as used in part a).

Exercise 3.1

[edit | edit source]

One possible basis for the space of symmetric matrices is for example the matrices for that have zeros everywhere but in the and entry. There are such matrices, and they are linearly independent, since no two such matrices have ones in the same entry. Furthermore, the matrices are symmetric and clearly any symmetric matrix can be written as a linear combination of .

Exercise 3.7

[edit | edit source]

Let be coefficients such that

. (1)

The matrix has as the th column the vector where is the th element of the vector . Denote , so that the (1) implies together with the condition that the vectors form a basis that for all . So then we must have for all . This implies that for all , but since the vectors form a basis, we must have for all .

Exercise 3.8

[edit | edit source]

Let be the matrix with the vectors as column vectors. Let . Then is equivalent to saying that is a linear combination of the vectors . By Theorem 1.2.21, has a unique solution if and only if is invertible.

In particular, this means that also has a unique solution . This shows that 1) span the space and 2) are linearly independent.

Exercise 4.2

[edit | edit source]

a) .

b) .

c) or .

Exercise 4.3

[edit | edit source]

The given operations correspond to row operations on matrices. By Theorem 1.2.16, any matrix that is invertible can be reduced to the identity using such operations. In Exercise 3.8 we proved that the columns of a matrix form a basis if and only if the matrix is invertible.

Exercise 4.4

[edit | edit source]

a) Any basis in corresponds to a matrix that is invertible, i.e., an element of . On the other hand the column vectors of any element from form basis vectors for .

b) For we have that there are in total matrices in of which we have to count the ones that are not invertible. Considering the columns of a matrix in , we have

  • first columns that are not the column vector, and scalings of the first column with a value other than zero.
  • If the first column is , the second column can be chosen in ways such that it is not also the vector.
  • If the second column is , the first column can also be chosen in ways such that it is not the vector.
  • There is exactly one matrix with both columns .

Combining these facts, we get that there are invertible matrices in .

For we want to compute the number of matrices in with determinant equal to 1. In there are equally many elements with determinant 1, 2, 3, etc.Therefore, the number of elements in is the number of elements with determinant 1 times . From the previous calculation we thus get that the number of elements in is .

Exercise 4.5

[edit | edit source]

a) The key for finding the number of subspaces is to find the number of linearly independent vectors in .

  • Subspaces of dimension 0: 1.
  • Subspaces of dimension 1: Each subspace is spanned by a nonzero vector of the form with . There are such vectors. For any such given vector, there are nonzero scalings with a scalar in . Hence, the number of linearly independent vectors is . Each such vector spans a subspace that is different from the subspaces spanned by the others.
  • Subspaces of dimension 2: Let be some maximal collection of linearly independent vectors of . We know that , and any two vectors from span a two-dimensional subspace of . We can choose two vectors from in ways, but this is not the number of two-dimensional subspaces of . Indeed, say we choose such that . Then is a subspace of containing points and linearly independent vectors. As , this means contains some vector . The number of pairs of linearly independent vectors in is , and hence the number of two-dimensional subspaces of is . Another way to arriving the same conclusion is as follows: Let be a subspace of of dimension 2. Then, is spanned by two linearly independent vectors, and there is a vector such that . In other words, the vectors in are linearly independent of . We know that there are linearly independent vectors in , so whenever we choose one of such vectors, we are left with a subspace of dimension 2 that does not contain the chosen vector (but contains all the others). Hence, there are also subspaces of dimension 2.
  • Subspaces of dimension 3: 1.

b) The case of can be generalised from the previous case:

  • Subspaces of dimeansion 0: 1.
  • Subspaces of dimension 1: The number of linearly independent vectors can be calculated similarly as in a), and we get .
  • Subspaces of dimension 2: Similarly as in the case of , we have that the number of two-dimensional subspaces is .
  • Subspaces of dimension 3: Similarly as in the case of , we have that for each three-dimensional subspace we have a one-dimensional subspace "left over". Therefore the number of three-dimensional subspaces is .
  • Subspaces of dimension 4: 1.

Exercise 5.1

[edit | edit source]

Let be the space of symmetric and the space of skew-symmetric matrices. It is clear that and and that contains only the zero matrix, so the spaces are independent. By Proposition 3.6.4 b), we have and so by Proposition 3.4.23, .

Exercise 5.2

[edit | edit source]

The condition introduces a linear dependency between the elements of the matrix. Therefore, we have , and thus any one-dimensional subspace of that is independent of suffices. For example we can take as the span of the matrix for which the top-left corner element is 1 and the rest are 0. Then .

Exercise 6.1

[edit | edit source]

The given vectors span the set of sequences that are constant apart from a finite set of indices.

Exercise M.3

[edit | edit source]

a) Let and , and . Then we have also . The coefficients are linear in the coefficients , so

we can solve as follows

Setting each to zero yields a system of equations

.

By Corollary 1.2.14 this system has a solution where at least one of the coefficients is non-zero, so there is a polynomial that is not identically zero, but for every .

b) We can solve for example using similar approach to part a).

c) Let be a polynomial of degree and polynomial of degree , so that and . Let be a polynomial with unknown coefficients . In order to have this polynomial vanish at , we have to solve the equations that set the coefficient of to 0 for each in the polynomial . These equations are linear in , and there are of them. On the other hand, there are variables , so by Corollary 1.2.14 the linear system has a non-zero solution. Note that in part a) we restricted the degree of the polynomial to 2, and thus did not end up with as many equations as in this proof.