Ch 2.2 - The Inverse of a Matrix
Understanding Matrix Operations and Efficiency
Equality of Transposed Products:
The quantities and are equal, as indicated by Theorem 3(d).
Scalar Product (Dot Product) vs. Outer Product:
For a column vector :
The scalar product . This results in a matrix, usually written without brackets (a scalar).
Example from transcript (assuming for calculation of 34): If , then .
The outer product .
Example from transcript (assuming for calculation of matrix): If , then .
Matrix Product Definition and Undefined Operations:
The product is not defined if the number of columns in does not match the number of rows in . For instance, if is and is a column vector, would be a row vector. For the product to be defined, would need to have 2 rows, implying is a 1x2 row vector not a column vector which is contradictory to typical conventions for . The exact dimensions of and that lead to this undefined product were not explicitly stated beyond
x does not have two rows to match the two columns of A^T.
Computational Efficiency of Matrix Multiplication:
When computing , it is generally more efficient to compute it as .
Comparison for a matrix A and vector x:
Method 1: .
Computing requires multiplications (4 for each of the 4 entries in the resulting vector).
Computing then requires another multiplications.
Total: multiplications.
Method 2: .
Computing requires multiplications (4 for each of the 16 entries in the matrix ).
Computing then requires another multiplications.
Total: multiplications.
Conclusion: Computing is significantly faster than computing .
Properties of Matrix Products (AB) with Identical Rows/Columns in A:
If all columns of matrix A are identical, then all columns of the product AB are also identical.
If all rows of matrix A are identical (i.e., for all ), then all rows of the product AB are also identical ().
If all entries in A are the same (implying both identical rows and identical columns), then all entries in AB will also be the same.
The Inverse of a Matrix
Matrix Analogue of a Reciprocal:
The concept of a matrix inverse () is analogous to the multiplicative inverse (reciprocal) of a non-zero real number (e.g., ).
For real numbers, the inverse satisfies and .
Key Distinctions for Matrices:
Two-Sided Definition: Because matrix multiplication is generally not commutative (), the matrix generalization requires both equations to be satisfied: and .
No Division Notation: Slanted-line notation (e.g., ) is avoided for matrices.
Square Matrices Only: A full generalization of the inverse is possible only for square matrices ().
Definition of an Invertible Matrix:
An matrix A is invertible (or nonsingular) if there exists an matrix C such that and , where is the identity matrix.
C is called an inverse of A.
Uniqueness of the Inverse: The inverse C is uniquely determined by A.
Proof: If B were another inverse of A, then . Thus, .
The unique inverse is denoted as , so the defining equations are and .
A matrix that is not invertible is called a singular matrix.
Example 1: Verifying an Inverse for a Matrix:
Given and .
.
.
Since both conditions are met, .
Inverse of a Matrix
Theorem 4: Formula for the Inverse of a Matrix:
Let .
If , then A is invertible and its inverse is given by:
If , then A is not invertible.
Determinant of a Matrix:
The quantity is called the determinant of A, denoted as .
Theorem 4 implies that a matrix A is invertible if and only if .
Example 2: Finding the Inverse of a Matrix:
Find the inverse of .
First, calculate the determinant: .
Since , A is invertible.
Using Theorem 4:
Solving Linear Systems with Invertible Matrices
Importance of Invertible Matrices:
They are essential for algebraic calculations and formula derivations in linear algebra.
They can provide insights into mathematical models of real-life situations.
Theorem 5: Unique Solution for :
If A is an invertible matrix, then for any vector in , the matrix equation has a unique solution given by .
Proof of Existence: Substituting for in the equation: . Thus, is a solution.
Proof of Uniqueness: Assume is any solution such that . Multiplying both sides by on the left:
Consequently, any solution must be equal to , proving uniqueness.
Practical Application: Elastic Beam Deflection (Example 3)
Scenario: A horizontal elastic beam is supported at its ends and subjected to forces at three points (1, 2, 3), causing deflections.
Variables:
: A vector listing the forces at the three points.
: A vector listing the amounts of deflection at the three points.
Hooke's Law Relationship:
D: The flexibility matrix.
: The stiffness matrix.
Physical Significance of Columns of D (Flexibility Matrix):
Express as , where are the standard basis vectors (columns of the identity matrix).
Interpreting : The vector represents a unit force applied downward at point 1, with zero force at the other two points.
First column of D (): Contains the beam deflections that result from applying a unit force at point 1 (and zero force at points 2 and 3).
Similarly, the second and third columns of D list the deflections caused by a unit force at points 2 and 3, respectively.
Physical Significance of Columns of (Stiffness Matrix):
The inverse equation is , which computes the force vector required to produce a given deflection vector (i.e., this matrix describes the beam's stiffness).
Express as .
Interpreting as a deflection vector: The vector now represents a unit deflection at point 1, with zero deflections at the other two points.
First column of (): Lists the forces that must be applied at the three points to produce a unit deflection at point 1 and zero deflections at points 2 and 3.
Similarly, columns 2 and 3 of list the forces required to produce unit deflections at points 2 and 3, respectively.
Note on Forces: To achieve specific deflections, some forces in these columns might be negative (indicating an upward force).
Units: If flexibility is measured in
inches of deflection per pound of load, then stiffness matrix entries are given inpounds of load per inch of deflection.
Practicalities of Using to Solve
Computational Efficiency: The formula is rarely used for numerical computations of for large matrices.
Row reduction of the augmented matrix is almost always faster and generally more accurate (as it can minimize rounding errors).
Exception: The case can be an exception, where mental calculation of might make using the formula quicker.
Example 4: Solving a System Using
System:
3x1 + 4x2 = 3 \
5x1 + 6x2 = 7This is equivalent to where and .
From Example 2, we found .
Solution:
\mathbf{x} = A^{-1}\mathbf{b} = [[-3, 2], [5/2, -3/2]] [[3], [7]] \
= [[(-3)(3) + (2)(7)], [(5/2)(3) + (-3/2)(7)]] \
= [[-9 + 14], [15/2 - 21/2]] \
= [[5], [-6/2]] = [[5], [-3]]So, and .
Properties of Invertible Matrices
Theorem 6: Important Facts about Invertible Matrices:
a. Inverse of an Inverse: If A is an invertible matrix, then its inverse, , is also invertible, and .
Proof: By definition, and . These equations satisfy the conditions for A to be the inverse of .
b. Inverse of a Product: If A and B are invertible matrices, then their product AB is also invertible. The inverse of AB is the product of their inverses in reverse order:
Proof: We need to show that and :
A similar calculation shows .
c. Inverse of a Transpose: If A is an invertible matrix, then its transpose, , is also invertible. The inverse of is the transpose of :
Proof: Using Theorem 3(d) (which states ):
And also:
Thus, is invertible, and its inverse is .
Generalization of Theorem 6(b):
The product of any number of invertible matrices is invertible, and its inverse is the product of their inverses in reverse order. For example, .
Role of Definitions in Proofs: Proofs rigorously demonstrate that a proposed inverse (or other property) satisfies the formal definition. For example, showing is the inverse of AB means showing it satisfies the definition of an inverse with AB.
Elementary Matrices and Computing
Connection to Row Operations:
A significant connection exists between invertible matrices and elementary row operations.
An invertible matrix A is row equivalent to the identity matrix .
This relationship provides a systematic method for finding .
Definition of an Elementary Matrix:
An elementary matrix is a matrix obtained by performing a single elementary row operation on an identity matrix ().
There are three types of elementary matrices, corresponding to the three elementary row operations:
Row Replacement: Adding a multiple of one row to another.
Row Interchange: Swapping two rows.
Row Scaling: Multiplying a row by a nonzero scalar.
Example 5: Elementary Matrices and Row Operations:
Let .
(obtained by on ).
performs the operation on A.
(obtained by on ).
performs the operation on A.
(obtained by on ).
performs the operation on A.
General Fact: If an elementary row operation is performed on an matrix A, the resulting matrix can be written as , where E is the elementary matrix created by performing the same row operation on .
Invertibility of Elementary Matrices:
Every elementary matrix E is invertible.
This is because elementary row operations are reversible (Section 1.1).
The inverse of an elementary matrix E is simply the elementary matrix of the same type that performs the reverse operation, transforming E back into the identity matrix I.
Example 6: Finding the Inverse of an Elementary Matrix:
Given (which adds -4 times row 1 to row 3 of ).
To reverse this operation and transform back into , one must add +4 times row 1 to row 3.
Therefore, the inverse is .
The Algorithm for Finding
Theorem 7: Invertibility and Row Equivalence to Identity:
An matrix A is invertible if and only if A is row equivalent to the identity matrix ().
Furthermore, if A is invertible, any sequence of elementary row operations that reduces A to will also transform into .
Proof of Theorem 7:
Part 1: If A is invertible, then (A is row equivalent to ).
By Theorem 5, if A is invertible, the equation has a solution for every .
This implies that A has a pivot position in every row (Theorem 4 in Section 1.4).
Since A is a square matrix, having n pivot positions in n rows means all n pivot positions must lie on the main diagonal.
Therefore, the reduced echelon form of A must be , meaning .
Part 2: If , then A is invertible.
If , it means A can be transformed into by a sequence of elementary row operations.
Each elementary row operation corresponds to left-multiplication by an elementary matrix. So, there exist elementary matrices such that:
Since each elementary matrix is invertible, their product is also invertible (by the generalization of Theorem 6b).
Let . Then . This directly implies that A is invertible, and its inverse is (since multiplying by on the left yields the identity, must be A).
More specifically, from , we multiply both sides on the left by to get . Then, implies .
This means that is precisely the matrix obtained by applying the same sequence of elementary operations to (because ).
Algorithm for Finding :
To find the inverse of an matrix A, form the augmented matrix by placing A and the identity matrix side-by-side.
Perform row operations to reduce this augmented matrix.
If A is row equivalent to : The augmented matrix will transform from to . The matrix on the right side will be .
If A is not row equivalent to : If, during row reduction, a row of zeros appears on the left side (where A initially was), then A is not invertible, and no inverse exists.
Example 7: Finding the Inverse of a Matrix:
Find the inverse of , if it exists.
Form the augmented matrix :
Perform row operations:
:
:
:
:
:
Since A is row equivalent to I, A is invertible. The inverse is:
Checking the Answer:
It's good practice to verify the calculated inverse by checking if (or ).
Note: If A is known to be invertible and you derive a matrix C such that , then C must be . It is not strictly necessary to also check in this context of computation, as the algorithm guarantees that if a matrix on the right emerges, it is the unique inverse.
Another View of Matrix Inversion (Solving Multiple Systems Simultaneously):
Finding by row reducing can be viewed as simultaneously solving separate matrix equations:
where are the columns of the identity matrix . The augmented columns for these systems are simply the columns of , forming .The property and the definition of matrix multiplication demonstrate that the columns of are precisely the solutions to these systems.
Practical Use: This perspective is valuable if an applied problem only requires finding one or two specific columns of . In such cases, only the corresponding systems need to be solved, rather than computing the full inverse.