Linear Algebra/Combining Subspaces/Solutions
Solutions[edit  edit source]
 This exercise is recommended for all readers.
 Problem 1
Decide if is the direct sum of each and .
 ,
 ,
 ,
 ,
 Answer
With each of these we can apply Lemma 4.15.
 Yes. The plane is the sum of this and because for any scalars and
 Yes. To see that any vector in the plane is a combination of vectors from these parts, consider this relationship.
 Yes. Each vector in the plane is a sum in this way
 No. The intersection is not trivial.
 No. These are not subspaces.
 This exercise is recommended for all readers.
 Problem 2
Show that is the direct sum of the plane with each of these.
 the axis
 the line
 Answer
With each of these we can use Lemma 4.15.
 Any vector in can be decomposed as this sum.
 Any vector in can be decomposed as
 Problem 3
Is the direct sum of and ?
 Answer
It is. Showing that these two are subspaces is routine. To see that the space is the direct sum of these two, just note that each member of has the unique decomposition .
 This exercise is recommended for all readers.
 Problem 4
In , the even polynomials are the members of this set
and the odd polynomials are the members of this set.
Show that these are complementary subspaces.
 Answer
To show that they are subspaces is routine. We will argue they are complements with Lemma 4.15. The intersection is trivial because the only polynomial satisfying both conditions and is the zero polynomial. To see that the entire space is the sum of the subspaces , note that the polynomials , , , etc., are in and also note that the polynomials , , etc., are in . Hence any member of is a combination of members of and .
 Problem 5
Which of these subspaces of
: the axis, :the axis, :the axis,
:the plane , :the plane
can be combined to
 sum to ?
 direct sum to ?
 Answer
Each of these is .

These are broken into lines for legibility.
, , , ,
, , ,
, , ,
, ,
,
, ,
, ,
, ,
 , , , ,
 This exercise is recommended for all readers.
 Problem 6
Show that .
 Answer
Clearly each is a subspace. The bases for the subspaces, when concatenated, form a basis for the whole space.
 Problem 7
What is if ?
 Answer
It is .
 Problem 8
Does Example 4.5 generalize? That is, is this true or false:if a vector space has a basis then it is the direct sum of the spans of the onedimensional subspaces ?
 Answer
True by Lemma 4.8.
 Problem 9
Can be decomposed as a direct sum in two different ways? Can ?
 Answer
Two distinct direct sum decompositions of are easy to find. Two such are and , and also and . (Many more are possible, for example and its trivial subspace.)
In contrast, any partition of 's singlevector basis will give one basis with no elements and another with a single element. Thus any decomposition involves and its trivial subspace.
 Problem 10
This exercise makes the notation of writing "" between sets more natural. Prove that, where are subspaces of a vector space,
and so the sum of subspaces is the subspace of all sums.
 Answer
Set inclusion one way is easy: is a subset of because each is a sum of vectors from the union.
For the other inclusion, to any linear combination of vectors from the union apply commutativity of vector addition to put vectors from first, followed by vectors from , etc. Add the vectors from to get a , add the vectors from to get a , etc. The result has the desired form.
 Problem 11
(Refer to Example 4.19. This exercise shows that the requirement that pariwise intersections be trivial is genuinely stronger than the requirement only that the intersection of all of the subspaces be trivial.) Give a vector space and three subspaces , , and such that the space is the sum of the subspaces, the intersection of all three subspaces is trivial, but the pairwise intersections , , and are nontrivial.
 Answer
One example is to take the space to be , and to take the subspaces to be the plane, the plane, and the plane.
 This exercise is recommended for all readers.
 Problem 12
Prove that if then is trivial whenever . This shows that the first half of the proof of Lemma 4.15 extends to the case of more than two subspaces. (Example 4.19 shows that this implication does not reverse; the other half does not extend.)
 Answer
Of course, the zero vector is in all of the subspaces, so the intersection contains at least that one vector. By the definition of direct sum the set is independent and so no nonzero vector of is a multiple of a member of , when . In particular, no nonzero vector from equals a member of .
 Problem 13
Recall that no linearly independent set contains the zero vector. Can an independent set of subspaces contain the trivial subspace?
 Answer
It can contain a trivial subspace; this set of subspaces of is independent: . No nonzero vector from the trivial space is a multiple of a from the axis, simply because the trivial space has no nonzero vectors to be candidates for such a multiple (and also no nonzero vector from the axis is a multiple of the zero vector from the trivial subspace).
 This exercise is recommended for all readers.
 Problem 14
Does every subspace have a complement?
 Answer
Yes. For any subspace of a vector space we can take any basis for that subspace and extend it to a basis for the whole space. Then the complement of the original subspace has this for a basis: .
 This exercise is recommended for all readers.
 Problem 15
Let be subspaces of a vector space.
 Assume that the set spans , and that the set spans . Can span ? Must it?
 Assume that is a linearly independent subset of and that is a linearly independent subset of . Can be a linearly independent subset of ? Must it?
 Answer
 It must. Any member of can be written where and . As spans , the vector is a combination of members of . Similarly is a combination of members of .
 An easy way to see that it can be linearly independent is to take each to be the empty set. On the other hand, in the space , if and and and , then their union is not independent.
 Problem 16
When a vector space is decomposed as a direct sum, the dimensions of the subspaces add to the dimension of the space. The situation with a space that is given as the sum of its subspaces is not as simple. This exercise considers the twosubspace special case.
 For these subspaces of find , , , and .
 Suppose that and are subspaces of a vector space. Suppose that the sequence is a basis for . Finally, suppose that the prior sequence has been expanded to give a sequence that is a basis for , and a sequence that is a basis for . Prove that this sequence
 Conclude that .
 Let and be eightdimensional subspaces of a tendimensional space. List all values possible for .
 Answer
 The intersection and sum are
 We write for the basis for , we write for the basis for , we write for the basis for , and we write for the basis under consideration.
To see that that spans , observe that any vector from can be written as a linear combination of the vectors in , simply by expressing in terms of and expressing in terms of .
We finish by showing that is linearly independent. Consider
 Just count the basis vectors in the prior item:, and , and , and .
 We know that . Because , we know that must have dimension greater than that of , that is, must have dimension eight, nine, or ten. Substituting gives us three possibilities or or . Thus must be either eight, seven, or six. (Giving examples to show that each of these three cases is possible is easy, for instance in .)
 Problem 17
Let and for each index suppose that is a linearly independent subset of . Prove that the union of the 's is linearly independent.
 Answer
Expand each to a basis for . The concatenation of those bases is a basis for and thus its members form a linearly independent set. But the union is a subset of that linearly independent set, and thus is itself linearly independent.
 Problem 18
A matrix is symmetric if for each pair of indices and , the entry equals the entry. A matrix is antisymmetric if each entry is the negative of the entry.
 Give a symmetric matrix and an antisymmetric matrix. (Remark. For the second one, be careful about the entries on the diagional.)
 What is the relationship between a square symmetric matrix and its transpose? Between a square antisymmetric matrix and its transpose?
 Show that is the direct sum of the space of symmetric matrices and the space of antisymmetric matrices.
 Answer
 Two such are these.
 A square symmetric matrix equals its transpose. A square antisymmetric matrix equals the negative of its transpose.
 Showing that the two sets are subspaces is easy. Suppose that .To express as a sum of a symmetric and an antisymmetric matrix, we observe that
 Problem 19
Let be subspaces of a vector space. Prove that . Does the inclusion reverse?
 Answer
Assume that . Then where and . Note that and, as a subspace is closed under addition, . Thus .
This example proves that the inclusion may be strict: in take to be the axis, take to be the axis, and take to be the line . Then and are trivial and so their sum is trivial. But is all of so is the axis.
 Problem 20
The example of the axis and the axis in shows that does not imply that . Can and happen?
 Answer
It happens when at least one of is trivial. But that is the only way it can happen.
To prove this, assume that both are nontrivial, select nonzero vectors from each, and consider . This sum is not in because would imply that is in , which violates the assumption of the independence of the subspaces. Similarly, is not in . Thus there is an element of that is not in .
 This exercise is recommended for all readers.
 Problem 21
Our model for complementary subspaces, the axis and the axis in , has one property not used here. Where is a subspace of we define the orthogonal complement of to be
(read " perp").
 Find the orthocomplement of the axis in .
 Find the orthocomplement of the axis in .
 Find the orthocomplement of the plane in .
 Show that the orthocomplement of a subspace is a subspace.
 Show that if is the orthocomplement of then is the orthocomplement of .
 Prove that a subspace and its orthocomplement have a trivial intersection.
 Conclude that for any and subspace we have that .
 Show that equals the dimension of the enclosing space.
 Answer
 The set
 The plane.
 The axis.
 Assume that is a subspace of some . Because contains the zero vector, since that vector is perpendicular to everything, we need only show that the orthocomplement is closed under linear combinations of two elements. If then and for all . Thus for all and so is closed under linear combinations.
 The only vector orthogonal to itself is the zero vector.
 This is immediate.
 To prove that the dimensions add, it suffices by Corollary 4.13 and Lemma 4.15 to show that is the trivial subspace . But this is one of the prior items in this problem.
 This exercise is recommended for all readers.
 Problem 22
Consider Corollary 4.13. Does it work both ways— that is, supposing that , is if and only if ?
 Answer
Yes. The lefttoright implication is Corollary 4.13. For the other direction, assume that . Let be bases for . As is the sum of the subspaces, any can be written and expressing each as a combination of vectors from the associated basis shows that the concatenation spans . Now, that concatenation has members, and so it is a spanning set of size . The concatenation is therefore a basis for . Thus is the direct sum.
 Problem 23
We know that if then there is a basis for that splits into a basis for and a basis for . Can we make the stronger statement that every basis for splits into a basis for and a basis for ?
 Answer
No. The standard basis for does not split into bases for the complementary subspaces the line and the line .
 Problem 24
We can ask about the algebra of the "" operation.
 Is it commutative; is ?
 Is it associative; is ?
 Let be a subspace of some vector space. Show that .
 Must there be an identity element, a subspace such that for all subspaces ?
 Does leftcancelation hold:if then ? Right cancelation?
 Answer
 Yes, for all subspaces because each side is the span of .
 This one is similar to the prior one— each side of that equation is the span of .
 Because this is an equality between sets, we can show that it holds by mutual inclusion. Clearly . For just recall that every subset is closed under addition so any sum of the form is in .
 In each vector space, the identity element with respect to subspace addition is the trivial subspace.
 Neither of left or right cancelation needs to hold. For an example, in take to be the plane, take to be the axis, and take to be the axis.
 Problem 25
Consider the algebraic properties of the direct sum operation.
 Does direct sum commute: does imply that ?
 Prove that direct sum is associative:.
 Show that is the direct sum of the three axes (the relevance here is that by the previous item, we needn't specify which two of the three axes are combined first).
 Does the direct sum operation leftcancel:does imply ? Does it rightcancel?
 There is an identity element with respect to this operation. Find it.
 Do some, or all, subspaces have inverses with respect to this operation:is there a subspace of some vector space such that there is a subspace with the property that equals the identity element from the prior item?
 Answer
 They are equal because for each, is the direct sum if and only if each can be written in a unique way as a sum and .
 They are equal because for each, is the direct sum if and only if each can be written in a unique way as a sum of a vector from each and .
 Any vector in can be decomposed uniquely into the sum of a vector from each axis.
 No. For an example, in take to be the axis, take to be the axis, and take to be the line .
 In any vector space the trivial subspace acts as the identity element with respect to direct sum.
 In any vector space, only the trivial subspace has a directsum inverse (namely, itself). One way to see this is that dimensions add, and so increase.