User:IssaRice/Linear algebra/Linear transformation vs matrix views
If you've gone through linear algebra a couple of times, once via the matrix-based way and once via the linear maps-based way, then you should know that certain adjectives are applied to both matrices and to linear maps. For instance, we might talk about an injective matrix and also talk about an injective linear map. But it is almost never discussed (to my knowledge) whether these definitions actually correspond to each other in any way, or what the precise correspondence is.
Given an matrix we can define a linear map by .
Given a linear map , it is not immediately possible to get a corresponding matrix. We must choose some basis for and a basis for . Then we can get a matrix by setting the th column to be written in the basis .
We would hope that any property we attribute to a linear map is invariant of the matrix we use to represent it. For instance if is called "injective" then it should be injective regardless of what matrix we use. Similarly given any matrix that is injective, any of the possible linear maps that that matrix represents should be injective.
Examples of other properties like this: injective, surjective, bijective, rank, diagonalizable
On the other hand, a property like "the sum of the columns is equal to such and such" is not invariant
I think the root of the confusion is that for these invariant properties, it is possible to define them given either the matrix or the map. So then there are two definitions floating around, and i don't see people showing them equivalent in general.
we can think of a linear map as an equivalence class of matrices. or we can think of a matrix as an equivalence class of linear maps. then we can phrase these invariance results as basically saying that these properties are well-defined. the difference seems to be that here we want to show equivalence, so we need to do it in both directions (?).
the actual proofs are pretty tedious (is my guess)
let's run with injectivity as an example.
Definition. A linear map is injective iff implies for all .
Definition. An matrix is injective iff implies for all .
We want to say that these are basically the same thing. How do we express that? Some ideas:
(1) if is injective and are any bases, then is injective
(2) if is injective and are any bases, then for any such that we have that is injective
I think these can be combined into:
(3) for all such that : T injective iff A injective.
Potential proof strategy that might not be so tedious: basically imagine that A is always written in the standard basis. Then consider where are the standard bases in R^n and R^m. Then . Then somehow we can use the fact that change of coordinate matrices are invertible.
What about properties like "normal", "self-adjoint", "isometry"? what about trace, determinant? sum of the first column?
It seems like for "normal", we don't have this for every choice of bases. I think the correct statement is:
for all and all orthonormal such that : T normal iff A normal.
In particular, if we're working in , then the standard basis is orthonormal, so is normal iff is normal, where is the matrix of with respect to the standard basis. This is why you almost never see the distinction between "normal linear map" and "normal matrix". but as soon as you start using some other basis that isn't orthonormal, the distinction becomes important. Another way to say this is that when we say a matrix is normal, we're automatically assuming we're using the standard basis (or at least an orthonormal basis), because it is possible to adversarially select a basis under which the corresponding linear map is not normal.
given a property P of matrices, we can say that P is preserved under change of coordinates iff for every invertible matrices
this question discusses some of these: https://math.stackexchange.com/questions/3033378/what-properties-of-a-linear-map-can-be-determined-from-its-matrix