Inverses, Pseudoinverses
Tags |
---|
Inverses
Sufficient conditions (cheat sheet)
Here are some sufficient conditions
- eigenbasis with non-zero eigenvalues
- Positive definite + symmetric (symmetric guarantees eigenbasis, positive definite guarantees that all eigenvalues are positive)
- No null space
True inverse
Let’s consider a matrix . If , then it is invertible. Define as the inverse of the matrix. This maps anything in output land to one distinct element in input land.
If an inverse exists, then
- this only applies for square matrices; rectangular matrices can have left inverses or right inverses but not both
- "nonsingular" and "invertible" are the same
Properties
-
-
- this has deep implications in dual spaces
Singular Matrix
A square matrix won't have an inverse if you can find a non-trivial vector such that
Why? Suppose that there exists . Therefore, (the last one is by the definition of matrices/linear maps. Therefore, there exists at least one such that
Geometrically, a singular matrix crushes a certain vector to zero, and it's impossible to recover.
Deriving inverse matrices
You can do gaussian-jordan elimination: start by mating the matrix with the identity matrix, then performing row operations until you you get the identity matrix on one side and the inverse matrix on the other.
- the proof that this is valid involves recognizing that every row operation is akin to performing an elementary linear transformation. composing these transformations result in the creation of the identity matrix
Or, you can solve a system of linear equations by filling with unknown variables and setting up
Rank and invertibility
A matrix is left-invertible if it has full column rank (every input maps to unique output; thnk about what the columns mean)
A matrix is right invertible if it has full row rank (corresponding to a column space that spans the entire output, because a full row rank has the same dimension as the codomain.
Pseudoinverse Formulation
If you have as an matrix, if there is some such that , how do you find the solution ? Well, we can find an approximation of that solution.
Pseudoinverse
But now let’s consider the matrix such that and perhaps .
Let’s figure out the first problem. We know that we can separate out the input space into and . Anything in is a lost cause. But can we potentially find an inverse that goes between and ? This is possible because transformations between these two spaces is injective and surjective.
Let’s define as a transformation between and . This is fully invertible. Note that these two spaces must have the same number of basis vectors, which means that if is square, there defintiely is some stuff in . We’ll deal with that too.
Now, let’s define the pseudoinverse
as such that
- if , then .
- If where and , then . This is a very elaborate way of saying that we ignore the part of the range that isn’t reachable by .
So, let’s get intuition. If you have anything that you want to invert, the will project it down into a lower subspace that is within the range of , and then compute the inverse. The projection is important. If you know that has rank in space of size , then the projection goes into a subspace of dimension .
Computing the pseudoinverse
Let’s first consult how we might compute the inverse of a matrix. If we can decompose (SVD) and it’s invertible, then the inverse is just because V and U are orthogonal and therefore easy inverses. The sigma is non-zero diagonal (otherwise isn’t invertible), so we just take the inverse of each element.
To compute the pseudoinverse, you do the exact thing, except that there will be zeros in the . Just leave those alone and invert the rest.
Properties
Here are the main properties
- (this makes sense, if you think about it)
- (this one is a little weird)
- (also a little weird, but interesting. It means that the row space is the same as the column space of the inverse, and it has the same rank as the projections)
-
And here are some more properties that should make sense. If we let , then we have
- (think about why this makes sense)
- (intuitively, the range of is the domain of , so is the identity operator for all things in the range of , and a "compression" for all things outside of the range of . In this manner, it does the same thing as .
-
-
- if is invertible, then it is obvious that
Now, is the same as , which means that it is a projection onto the range of . This makes sense, considering what the pseudoinverse does with .
is just , which you can imagine as the projection onto the range of (i.e. the row space)