TLDR; a geometric proof of the determinant as the “volume scaling factor”, equal to the product of the eigenvalues, without the familiar (and confusing) sign-flipping arithmetic.
Why do defective matrices have eigenvalues?
The eigensystem of an N×N square matrix A can be read right-to-left;A=Q−1DQ
An eigenvalue lacks a true eigenvector when A shears its victim off-axis:A|ui⟩−λi|ui⟩≠|0⟩
Why is the determinant the product of eigenvalues?
The determinant is the volume scaling factor of A, equal to the product of A’s eigenvaluesdet(A)=∏iλi
The generalized eigensystem has a property that the regular eigensystem lacks; orthogonality. While many eigensystems are naturally orthogonal (e.g., when A=A†), some eigensytems have linear dependent eigenvectors. But the Schur decomposition guarantees that the generalized eigensystem can always use orthogonal eigenvectorsA=Q†UQ
The Schur decomposition has been proven to exist for any square matrix, so we can use the Schur decomposition to explain why a matrix scales volume proportional to the product of its eigenvalues. The key insight is that a shear transformation preserves volume. A parallelogram’s area depends only on its base and height, not the shear angle. Similarly, any 3D shape can be partitioned into parallel planes, and any shear between those planes doesn’t change the total volume (e.g., a stack of playing cards). The matrix A will stretch its generalized eigenvector |ui⟩ by λi and also shear the vector by some unknown |τ⟩. But the shear by |τ⟩ doesn’t affect the volume along |ui⟩; the volume is only changed by λi. The Schur decomposition provides an orthogonal basis to independently cover every axis in space; thus, all volume transformation is exhaustively described by the eigenvalues alone — the shears preserve volume!
Imagine the unit hyper-cube: a vector from zero to each standard coordinate. The vectors describing this cube are encoded by the columns of identity matrix I. Thus, AI=A is the result of A acting on the unit hyper-cube; the unit hyper-cube started with unit volume and was stretched into A. Given the Schur decomposition of A, the eigenvalues along the diagonal of U stretched each axis — scaling the overall volume — while the off-diagonal shears do not change the volume. Since the final volume is the same with or without the shears, the new volume must be the same as the un-sheared hyper-rectangle. Thus, if we define the determinant as the volume-scaling factor of a square matrix, then the determinant is the product of eigenvaluesdet(A)=∏iλi
Leave determinant arithmetic to the machines!
Linear algebra and eigensystems are too important for students to get hung-up on determinant arithmetic. The Leibniz formula is taught extensively in school and never used in real applications because it does not scale (requiring O(n!) operations). And because of Leibniz’s arcane arithmetic, one can earn a PhD in particle physics but ultimately understand the determinant from a YouTube video. Finally, most algorithms for calculating the determinant simply solve the eigensystem and multiply the eigenvalues, skipping the Leibniz formula entirely. The characteristic polynomial is not required to define nor find eigenvalues; it is merely a useful result of the Leibniz formula! Why not define eigenvalues via Eq. 4 and the Shur decomposition and teach the determinant as “the product of the eigenvalues — the volume scaling factor”?
Of course, the fact that the Leibniz formula calculates the product of eigenvalues is certainly no accident, and likely indicates that Hamilton and Grassman were onto something with their exterior algebra. But that is a question for scholars to ponder — an engineer is happy with the simplest tool that works. Yet most engineers don’t understand the determinant — one of the most important concepts in all of linear algebra! It’s time for humanity to step up our determinant game.
This work is similar “Down with Determinants” by Sheldon Axler, perhaps the first to define the determinant as the product of eigenvalues. For more detail on eigensystems, determinants, and bra-ket notation for linear algebra see my paper on linear algebra.