Fun with polynomials and linear algebra; or, slight abstract nonsense
Posted
This is mostly a bunch of notes to myself (with some slight expansion) and is a combination/extension/simplification of theorems/ideas/constructions from a bunch of texts, including Wistbauer’s “Foundations of Module and Ring Theory” and Fuhrmann’s “A Polynomial Approach to Linear Algebra”, along with others that at this point I don’t recall.
While most of the things here are pretty standard (indeed, many of these are just slightly generalized definitions given in introductions to linear algebra over arbitrary fields!) there is some stuff here that might be weird and surprising unless you’re already quite well-versed in the language of module theory. (We won’t be discussing modules here, though, even if many statements have natural module-theoretic generalizations.)
This document/post is more of an exercise to see just how many “standard” constructions can be done purely in the linear-algebraic language. In a sense, this document is an “any% speedrun” of some theorems (Chinese remainder theorem, e.g.) and tidbits that most of us associate with different parts of math. It will also have some small exercises littered throughout, which are fun one-or-two-liners for those interested.
Anyways, onwards.
Basic facts
Let be a vector space over a field in what follows. If is a vector space over the same field, we say
if there exists any invertible linear map between and . It is pretty easy to see that this is an equivalence relation.
Given a linear map note that is injective (one-to-one) only when implies that (why?). Equivalently, it is when the kernel is trivial, i.e., . These are, of course, identical statements, but the latter is sometimes nicer when dealing with quotients, which we define later.
Now, let be a subspace of , that is and is itself a vector space. We say that is finite-dimensional and that is the dimension of the subspace if
That is, if there exists some invertible linear map . (For those familiar with linear algebra, this is identical to saying that that there exists a finite number of which span the space and are linearly-independent; this is easy to see via the map with .)
Note that is unique (if it exists) so does not depend on the mapping. This means that if we have another subspace with , then since .
In general, given two finite-dimensional vector spaces and , then exactly when . The former often conveys slightly more information in practice since the mappings between and are fairly natural when constructed, but if we only know that then the map may be fairly arbitrary and potentially complicated (though it will be, of course, linear).
It is also pretty easy to see from this definition that when and are finite-dimensional. (Consider the “obvious” map and similarly for .)
One useful observation is that if with both quantities defined, then any linear map that is either injective or surjective (onto) must also be invertible. The “non-gentleman’s proof”1 in the injective case is by noting that any basis in is a basis in if is injective. (Note that this is not true in the non-finite case; why?) This is helpful since, as we will soon see, it allows us to take “natural” maps which are easily seen to be injective and immediately show that they are invertible by dimension-counting.
Finally, it’s useful to note that for some linear map , the preimage of this map can be written as for any satisfying .
Quotients by spaces
Given some subspace then (much like in standard abstract algebra) we can set
where is defined elementwise; i.e., . The notion of a coset for vector spaces is identical to the one of Abelian groups, except that we work with subspaces which are the natural analogue to subgroups in the standard setting.
Note that is itself a vector space under the standard operations (again, applied elementwise) with the zero element equivalent to . (In what follows, it will be worth it to use as and vice versa, but the “type” of zero should be clear from context.) To see that is a vector space, let and be any “vectors” in and let be any elements of the field which is over, then
As an exercise it is worth parsing the left hand side very carefully. This means that, like with any vector space, there is a notion of dimension, when the space is finite-dimensional.
Now, let be subspaces of with and defined, then note that if
and , then . Since and are finite then is equivalent to .
The proof of this is pretty simple: let giving . Then implies that , so since and so is injective. But an injective linear map between two spaces of the same dimension is invertible. Take then note that . Finally, applying to both sides gives , so , which means that there is some such that . (Another way of stating this is that defines a natural projection from to via .)
Polynomial fun
Consider the vector space of polynomials over some field, denote it . (It may be useful to consider instead to be the set of polynomials of degree for some , in which case would have finite dimension, but what we’ll talk about here holds in general, even when is not finite-dimensional.) Let be a polynomial over the same field and set to be the set of polynomials divisible by .
Note that is a perfectly reasonable subspace of since adding polynomials divisible by and multiplying them by constants does not change the fact that they are divisible by .
This view leads to a bunch of nice results.
For example, it is reasonable to ask what vector subspaces correspond to for polynomials and ? This is just the set of polynomials that can be written as for , by definition of and . Using Bezout’s lemma, this is the same as where is the (polynomial) greatest common divisor of and . Similarly, you should work out what is. (If you’re familiar with ideals, this is, for now, just a rewriting of the standard theorems there, hence the use of parenthesis notation!)
Additionally, note that the set of vectors in modulo (in the linear algebraic sense) is exactly the same as the set of polynomials modulo (in the polynomial sense). We will show that, indeed, quotienting by the space is the same as quotienting general polynomials by the polynomial , in that
where is the set of polynomials with degree . That is, the set of vectors in , modulo (or, equivalently, polynomials modulo the polynomial ) is identical to the set of polynomials of degree . From our little structure theorem above, this immediately shows that the polynomials decompose uniquely into and a remainder part in :
In this sense, the vector space captures a lot of the structure of the polynomial .
Proof
Ok, to start, note that a polynomial if and only if for some polynomial in , or, equivalently, for some polynomials . This should already be quite reminiscent of the division theorem.
It is not hard to show that for nonzero. In fact, this is the only polynomial fact we need to prove the theorem: let be a polynomial with , then for any polynomial . It is easy to set such that (why?) so we may assume in general that by applying this observation repeatedly. This implies that the monomials span so , as required. It is easy to see that by comparing degrees, which gives the final claim that
Of course, this exactly matches the degree bound for polynomial division and therefore the dimension of hence . Finally, since (why?) and we have the result by the mini-structure theorem above.
Abstract nonsense-ish Chinese remainder theorem
Now, we can get to a result that generalizes the case of ideals.
An interesting observation from Wisbauer 1991’s textbook is the following very lovely construction (let’s say over vector spaces, but it holds in general over modules with a slightly different condition).
Let be vector subspaces of such that is finite-dimensional, then
exactly when
This just follows from our definition of “dimension” and the surrounding discussion.
In other words, it is possible to decompose a quotient over an intersection of subspaces into a product of individual quotients with each subspace, exactly when the dimensions match.
Canonical map
In fact, the dimension equality implies the canonical map is also invertible. To see this, let be the map for the th component, such that is the “obvious” map. That is,
for . Now, note that where and , where is the canonical map and is the (induced!) injective map . (You should check that this is, indeed, injective.) It is clear that , and the dimension equality then tells us that the space maps from, , has the same dimension as the output space and hence is not just injective, but invertible.
As a special case, if is finite-dimensional and , then we get that there is a natural invertible map
exactly when .
Back to the theorem
After all of this, we might ask: why is this the Chinese remainder theorem?
As before, let be the vector space of polynomials over some field and let be polynomials with coefficients in that are mutually coprime, such that a least-common multiple of them is . Then recall that is exactly the set of remainders modulo (in the polynomial sense).
By the division argument above, we have that while is where is any least-common multiple of the . So, again from above, , so the left hand side and right hand side match and we can decompose any polynomial modulo into the polynomial modulo each uniquely and vice versa, which is exactly the Chinese remainder theorem.
Indeed, the map above lets us do the decomposition in a straightforward way since it tells us how the two spaces are related via the (and ).
Given the product description it is natural to consider elements of the form for some index with . Since any element in can be written as the sum of elements of this form, finding an inverse for this type of element suffices to find inverses for any element because (and therefore ) is linear.
Now, since is invertible, let for some and, by definition of we have . But note that for each , by definition, so . Since we know that where then this tells us exactly what we need to find the inverse: find some polynomial such that has the same remainder as (modulo ) but is a multiple of the polynomial , then take the remainder over the big polynomial .
If we want to go back to “polynomial land” then this follows directly from Bezout’s lemma using the fact that and are coprime. Of course, we can give a linear-algebraic description here too, using the definitions and discussion above regarding the equivalence between polynomials, greatest common divisors, and the vector subspaces generated by them.
To see this, note that we have , so any can be written in the form with and . (Again, this is Bezout’s lemma in a funny hat, but the point is that it’s all just vector spaces once you have the equivalence.) It is then pretty immediate that (a) , (b) by definition is a multiple of since , and, finally, (c) we may take as required.
The basis view
In the case that is finite-dimensional, the equivalence has a very simple interpretation in terms of stacked matrices.
For now, take the subspaces to satisfy , as it will make our lives slightly easier.
Since is finite dimensional, say of dimension , then so we have an invertible map . This also means that is of finite dimension (why?), call it . Now let be any linear map (representable as a matrix of size ) satisfying . In other words, the nullspace of is in the basis given by . Then the condition
is exactly the condition that the dimensions of the matrices add up: , and the condition that is similarly the condition that they share no elements that are mapped to zero simultaneously. Equivalently, it is the condition that the stacked matrix
has or is, in other words, invertible.
There’s all sorts of additional simple consequences of this view. For example, this means that the rows of the must all be linearly independent, or, given a system with outputs over , any linear measurements can reconstruct the output unambiguously.
As a final exercise: what’s the interpretation of and the dimension conditions in the case that the subspaces do not satisfy ?
Footnotes
-
A “gentleman’s proof” is one which is not basis-dependent. There is such a proof of this statement by noting that a proper subspace has at least one (nonzero) element and this element generates a subspace of dimension at least 1 making so since is injective. ↩