Matrix Inverses and Elementary Matrices
An inverse matrix undoes a linear process. Elementary matrices make this idea concrete: every row operation is the same as multiplying by a simple invertible matrix. This connects row reduction, solving systems, and algebraic invertibility into one framework.
The main point is not merely that some square matrices have formulas for inverses. The deeper point is that invertibility is an equivalence of many ideas: no information is lost, the system has exactly one solution for every , the columns form a basis, the determinant is nonzero, and row reduction reaches the identity. Elementary matrices are the bridge between the algorithm and these structural statements.
Definitions
A square matrix is invertible if there is a matrix such that
If no such matrix exists, is singular.
An elementary matrix is obtained by performing one elementary row operation on an identity matrix. Left multiplication by an elementary matrix performs the same row operation on any compatible matrix. For example, if is created by swapping rows and of , then swaps rows and of .
For a square matrix , the augmented matrix
can be row-reduced to compute . If the left side becomes , the right side becomes . If the left side cannot become , then is not invertible.
The inverse of a matrix has the special formula
provided .
Key results
The inverse of an invertible matrix is unique. If and both satisfy the inverse equations for , then
Products of invertible matrices are invertible, and the inverse reverses order:
The order reversal is the same phenomenon as function composition: to undo "first , then ," one must undo first and second.
Elementary matrices are invertible. Their inverses are also elementary matrices: swap the same rows again, multiply by the reciprocal scalar, or add the opposite multiple of one row to another. If row operations reduce to , then
Thus
This explains why applying the same operations to produces .
For a square matrix , the following are equivalent:
- is invertible.
- The reduced row echelon form of is .
- has only the trivial solution.
- has a unique solution for every .
- The columns of form a basis of .
These equivalences are often called the invertible matrix theorem in an introductory course.
Visual
| Row operation | Elementary matrix action | Inverse operation |
|---|---|---|
| Swap and | swap rows of | swap the same rows |
| Scale by | scale row of | scale by |
| Replace by | add times row to row | add times row to row |
Worked example 1: Compute an inverse by row reduction
Problem: compute the inverse of
Step 1: augment with the identity.
Step 2: eliminate below the first pivot with .
Step 3: eliminate above the second pivot with .
Thus
Step 4: check by multiplication.
The checked answer is correct.
Worked example 2: Use an inverse to solve a system
Problem: solve
Step 1: write the system as :
Step 2: use the inverse from the previous example.
Step 3: multiply.
Checked answer: . Substitution gives and .
Code
import numpy as np
A = np.array([[1, 2],
[3, 7]], dtype=float)
b = np.array([5, 17], dtype=float)
A_inv = np.linalg.inv(A)
x = A_inv @ b
print(A_inv)
print(x)
print(A @ A_inv)
print(np.allclose(A @ x, b))
In numerical work, directly forming an inverse is usually less stable and less efficient than solving with a factorization. The inverse is still conceptually important, especially for proving structural facts.
Common pitfalls
- Writing . Matrix inversion is not entrywise division.
- Reversing the inverse product rule incorrectly. The correct identity is .
- Assuming every square matrix has an inverse. A square matrix can be singular.
- Row-reducing only and forgetting to apply the same operations to the identity side of .
- Using the inverse formula for larger matrices.
- Concluding that a matrix is invertible just because all entries are nonzero. Pivot structure, not entrywise nonzero status, controls invertibility.
When computing an inverse by row reduction, every row operation should be interpreted as left multiplication by an elementary matrix. This perspective explains why the method works: the sequence of operations that turns into is exactly the sequence whose product is . The right side of records that product as it is built.
The inverse is best understood as a structural guarantee, not always as a computational tool. If is invertible, then has the unique solution . That formula proves uniqueness and dependence on . In numerical computation, however, solving the system directly with a factorization is usually better than forming explicitly.
A quick singularity check is to look for dependence among columns or rows. If one column is a scalar multiple or linear combination of others, then the matrix cannot be invertible. Row reduction turns this observation into pivots: a missing pivot means one direction in the domain is collapsed, so the process cannot be reversed.
For products, imagine undoing actions. If a vector is first transformed by and then by , the combined action is . To reverse it, undo first and then , so the inverse is . This mental model is often safer than memorizing the formula alone.
It is also worth distinguishing left inverses and right inverses in rectangular settings. A square matrix inverse satisfies both and . For a non-square matrix, one may have a matrix that reverses the action on one side but not the other. This connects directly to one-to-one and onto behavior: full column rank supports left-inverse behavior, while full row rank supports right-inverse behavior. The square invertible case is the special situation where both happen at once.
Elementary matrices give a compact way to reason about algorithms. If elimination reduces to an upper triangular matrix , then a product of elementary matrices has transformed into . Rearranging that statement is the beginning of LU factorization. Thus inverse theory is not isolated from numerical methods; it is the algebraic background behind practical solvers.
When checking an inverse candidate , multiply on both sides if the problem is theoretical:
For square matrices over ordinary finite-dimensional spaces, one side actually implies the other, but checking both sides is a good habit in introductory work because it reveals order mistakes. If and are not even both defined, then the proposed inverse cannot be a two-sided inverse.
The inverse also clarifies equations with transformed variables. If and is invertible, then no information has been lost: . If is singular, two different inputs may produce the same output, or some outputs may be unreachable. Thus invertibility is both an algebraic property and an information-preservation property.
In proofs, it is often better to use the defining inverse equations than to compute the inverse. For example, to show that an equation has at most one solution, suppose and multiply by to get . This kind of argument is the real reason inverse notation is powerful.
Before using an inverse in a solution, ask what it proves: existence, uniqueness, reversibility, or a concrete formula. Those are related but distinct purposes, and separating them keeps both computations and proofs cleaner.
When the purpose is only to solve one system, a direct solve is usually the better computational expression of the same idea.