Linear Equations
Theorem 2.2.2 also gives a useful way to describe the solutions to a system
of linear equations. There is a related system
called the associated homogeneous system, obtained from the original system by replacing all the constants by zeros. Suppose is a solution to and is a solution to (that is and ). Then is another solution to . Indeed, Theorem 2.2.2 gives
This observation has a useful converse.
Theorem :
Suppose is any particular solution to the system of linear equations. Then every solution to has the form
for some solution of the associated homogeneous system .
Proof:
Suppose is also a solution to , so that . Write . Then and, using Theorem 2.2.2, we compute
Hence is a solution to the associated homogeneous system .
Note that gaussian elimination provides one such representation.
Example :
Express every solution to the following system as the sum of a specific solution plus a solution to the associated homogeneous system.
Solution:
Gaussian elimination gives , , , and where and are arbitrary parameters. Hence the general solution can be written
Thus
is a particular solution (where ), and
gives all
solutions to the associated homogeneous system. (To see why this is so,
carry out the gaussian elimination again but with all the constants set
equal to zero.)
The following useful result is included with no proof.
Theorem :
The Dot Product
Definition 2.5 is not always the easiest way to compute a matrix-vector product because it requires that the columns of be explicitly identified. There is another way to find such a product which uses the matrix as a whole with no reference to its columns, and hence is useful in practice. The method depends on the following notion.
If and are two ordered -tuples, their is defined to be the number
obtained by multiplying corresponding entries and adding the results.
To see how this relates to matrix products, let denote a matrix and let be a -vector. Writing
in the notation of Section 2.1, we compute
From this we see that each entry of is the dot product of the corresponding row of with . This computation goes through in general, and we record the result in Theorem 2.2.5.
Theorem 2.2.5 Dot Product Rule
Let be an matrix and let be an -vector. Then each entry of the vector is the dot product of the corresponding row of with .
This result is used extensively throughout linear algebra.
If is and is an -vector, the computation of by the dot product rule is simpler than using Definition 2.5 because the computation can be carried out directly with no explicit reference to the columns of (as in Definition 2.5. The first entry of is the dot product of row 1 of with . In hand calculations this is computed by going across row one of , going down the column , multiplying corresponding entries, and adding the results. The other entries of are computed in the same way using the other rows of with the column .
In general, compute entry of as follows (see the diagram):
Go across row of and down column , multiply corresponding entries, and add the results.
As an illustration, we rework Example 2.2.2 using the dot product rule instead of Definition.
Example :
If
and , compute .
Solution:
The entries of are the dot products of the rows of with :
Of course, this agrees with the outcome in above Example
Example :
Write the following system of linear equations in the form .
Solution:
Write , , and . Then the dot product rule gives , so the entries of are the left sides of the equations in the linear system. Hence the system becomes because matrices are equal if and only corresponding entries are equal.
Example :
If is the zero matrix, then for each -vector .
Solution:
For each , entry of is the dot product of row of with , and this is zero because row of consists of zeros.
The Identity Matrix
The first few identity matrices are
In Example 2.2.6 we showed that for each -vector using Definition 2.5. The following result shows that this holds in general, and is the reason for the name.
Example :
For each we have for each -vector in .
Solution:
We verify the case . Given the -vector
the dot product rule gives
In general, because entry of is the dot product of row of with , and row of has in position and zeros elsewhere.
Example :
Let be any matrix with columns . If denotes column of the identity matrix , then for each .
Solution:
Write
where , but for all . Then Theorem 2.2.5 gives
Example 2.2.12will be referred to later; for now we use it to prove:
Theorem :
Let and be matrices. If for all in , then .
Proof:
Write and and in terms of their columns. It is enough to show that holds for all . But we are assuming that , which gives by Example 2.2.12.
We have introduced matrix-vector multiplication as a new way to think about systems of linear equations. But it has several other uses as well. It turns out that many geometric operations can be described using matrix multiplication, and we now investigate how this happens. As a bonus, this description provides a geometric “picture” of a matrix by revealing the effect on a vector when it is multiplied by . This “geometric view” of matrices is a fundamental tool in understanding them.
Matrix Multiplication
In Section 2.2 matrix-vector products were introduced. If is an matrix, the product was defined for any -column in as follows: If where the are the columns of , and if ,
Definition 2.5 reads
(2.5)
This was motivated as a way of describing systems of linear equations with coefficient matrix . Indeed every such system has the form where is the column of constants.
In this section we extend this matrix-vector multiplication to a way of multiplying matrices in general, and then investigate matrix algebra for its own sake. While it shares several properties of ordinary arithmetic, it will soon become clear that matrix arithmetic is different in a number of ways.
Matrix Multiplication
Thus the product matrix is given in terms of its columns : Column of is the matrix-vector product of and the corresponding column of . Note that each such product makes sense by Definition 2.5 because is and each is in (since has rows). Note also that if is a column matrix, this definition reduces to Definition 2.5 for matrix-vector multiplication.
Given matrices and , Definition 2.9 and the above computation give
for all in . We record this for reference.
Theorem :
Let be an matrix and let be an matrix. Then the product matrix is and satisfies
Here is an example of how to compute the product of two matrices using Definition 2.9.
Example :
Compute if
and
.
Solution:
The columns of are
and , so Definition 2.5 gives
Hence Definition 2.9 above gives .
While Definition 2.9 is important, there is another way to compute the matrix product that gives a way to calculate each individual entry. In Section 2.2 we defined the dot product of two -tuples to be the sum of the products of corresponding entries. We went on to show (Theorem 2.2.5) that if is an matrix and is an -vector, then entry of the product is the dot product of row of with . This observation was called the “dot product rule” for matrix-vector multiplication, and the next theorem shows that it extends to matrix multiplication in general.
Dot Product Rule
product of row of with column of .
Proof:
Write in terms of its columns. Then is column of for each . Hence the -entry of is entry of , which is the dot product of row of with . This proves the theorem.
Thus to compute the -entry of , proceed as follows (see the diagram):
Go across row of , and down column of , multiply corresponding entries, and add the results.
No comments:
Post a Comment