|
|||||
Matrices provide a convenient way to describe linear equations. Thus if you
take the coefficients of your unknowns in some standard order as the row elements
of your matrix you can define a matrix of coefficients for any set of equations. We can then write the original equations as the single matrix equation Using the definition of matrix multiplication, which is: taking the dot products of the rows of the first matrix with the columns (here single column) of the second to produce the corresponding elements of the product, you should verify that this matrix equation is exactly the same as our original three equations. The process of Gaussian elimination can be applied in this matrix form here.
The rules are In this form such operations are called "elementary row operations" and Gaussian elimination is called row reduction. What you do here is perform enough of operation 2 to form 0's in the matrix
on one side of the main diagonal. When this is done you can determine one unknown
and then substitute successively to find the others. The n dimensional matrix whose diagonal elements are 1 and off diagonal elements are 0 is called the n dimensional identity matrix, and is written as I usually without any indication of what its size is, unless that can cause confusion, in which case it is written as In. It has the property that its matrix product with any matrix M of the same dimension is M itself, and its operation on any n dimensional vector v is v itself. Thus if you start with the matrix equation Mv = r, and row reduce
to find another representation of the same set of equations for which M has been
reduced to the identity matrix I, you have Iv = r' where r' is the
result of the same row operations on the right side of the equation as those that
reduced M to I, |