Matrices & Determinants
Matrix operations, determinant formulas, Cramer's rule, and invertibility
3.1 Matrix Operations
A matrix $A \in \mathbb{F}^{m \times n}$ is a rectangular array of scalars with $m$ rows and $n$ columns. The entry in row $i$ and column$j$ is denoted $a_{ij}$ or $(A)_{ij}$.
Fundamental Operations
- Addition: $(A + B)_{ij} = a_{ij} + b_{ij}$ (same dimensions required)
- Scalar multiplication: $(\alpha A)_{ij} = \alpha \cdot a_{ij}$
- Matrix multiplication: $(AB)_{ij} = \sum_{k=1}^{n} a_{ik} b_{kj}$ where $A \in \mathbb{F}^{m \times n}$, $B \in \mathbb{F}^{n \times p}$
- Transpose: $(A^T)_{ij} = a_{ji}$
- Conjugate transpose: $(A^*)_{ij} = \overline{a_{ji}}$
Matrix multiplication is associative ($(AB)C = A(BC)$) and distributes over addition, but is generally not commutative: $AB \neq BA$ in general. This non-commutativity is one of the most important features distinguishing linear algebra from scalar arithmetic.
Special Matrices
- Identity: $I_n$ with $(I_n)_{ij} = \delta_{ij}$ (Kronecker delta)
- Diagonal: $a_{ij} = 0$ for $i \neq j$
- Symmetric: $A^T = A$
- Skew-symmetric: $A^T = -A$
- Orthogonal: $A^T A = I$ (equivalently, $A^{-1} = A^T$)
- Unitary: $A^* A = I$ (complex analog of orthogonal)
- Hermitian: $A^* = A$ (complex analog of symmetric)
- Nilpotent: $A^k = 0$ for some $k$
- Idempotent: $A^2 = A$ (projection matrices)
Derivation 1: $(AB)^T = B^T A^T$
We verify the transpose of a product reverses the order:
This "reversal rule" extends to products of any length: $(A_1 A_2 \cdots A_k)^T = A_k^T \cdots A_2^T A_1^T$.
3.2 The Determinant
The determinant is a scalar-valued function on square matrices that encodes essential geometric and algebraic information. It measures the signed volume scaling factor of the linear transformation represented by the matrix.
Leibniz Formula
For an $n \times n$ matrix $A$:
where the sum is over all permutations $\sigma$ of $\{1, \ldots, n\}$ and$\text{sgn}(\sigma) = (-1)^{\text{inv}(\sigma)}$ is the sign of the permutation. This sum has $n!$ terms, making it impractical for computation but theoretically fundamental.
Derivation 2: Cofactor Expansion
The cofactor expansion (or Laplace expansion) along row $i$ gives:
where $M_{ij}$ is the minorβthe determinant of the$(n-1) \times (n-1)$ matrix obtained by deleting row $i$ and column $j$. The quantity $C_{ij} = (-1)^{i+j} M_{ij}$ is the cofactor.
To derive this from the Leibniz formula, we group the permutations by the value $\sigma(i) = j$. For each fixed $j$, the permutations with $\sigma(i) = j$ are in bijection with permutations of $\{1,\ldots,n\} \setminus \{j\}$, and the sign picks up a factor of$(-1)^{i+j}$ from the transpositions needed to move $j$ to position $i$.
For 2x2 and 3x3 Matrices
3.3 Properties of Determinants
Fundamental Properties
- P1 (Multiplicativity): $\det(AB) = \det(A)\det(B)$
- P2 (Transpose): $\det(A^T) = \det(A)$
- P3 (Scaling): $\det(\alpha A) = \alpha^n \det(A)$ for $A \in \mathbb{F}^{n \times n}$
- P4 (Inverse): $\det(A^{-1}) = \frac{1}{\det(A)}$
- P5 (Triangular): $\det(A) = \prod_{i=1}^{n} a_{ii}$ for triangular $A$
- P6 (Row swap): Swapping two rows negates the determinant
- P7 (Row scaling): Multiplying a row by $c$ multiplies $\det$ by $c$
- P8 (Row addition): Adding a multiple of one row to another leaves $\det$ unchanged
Derivation 3: Multiplicativity of Determinant
We sketch the proof that $\det(AB) = \det(A)\det(B)$. The key insight is that the determinant is the unique alternating multilinear function of the columns (or rows) that equals 1 on the identity matrix.
Fix $B$ and define $\phi(A) = \det(AB)$. Since the columns of $AB$are $A\mathbf{b}_1, \ldots, A\mathbf{b}_n$, and $\det$ is multilinear in columns,$\phi$ is multilinear in the columns of $B$. It is also alternating (if two columns of $B$ are equal, then two columns of $AB$ are equal, so $\det(AB) = 0$). By the uniqueness characterization, $\phi(A) = \phi(I) \cdot \det(B) = \det(A)\det(B)$.
3.4 Cramer's Rule
Theorem: Cramer's Rule
For the system $A\mathbf{x} = \mathbf{b}$ where $A$ is $n \times n$with $\det(A) \neq 0$, the unique solution is:
where $A_j$ is the matrix obtained by replacing column $j$ of $A$with the vector $\mathbf{b}$.
Derivation 4: Proof of Cramer's Rule
We use the adjugate (classical adjoint) matrix. Recall the cofactor matrix$C$ where $C_{ij} = (-1)^{i+j} M_{ij}$. The adjugate is$\text{adj}(A) = C^T$, and it satisfies:
This follows because the $(i,j)$ entry of $A \cdot \text{adj}(A)$ is$\sum_k a_{ik} C_{jk}$, which equals $\det(A)$ when $i = j$(cofactor expansion along row $i$) and 0 when $i \neq j$ (expansion of a matrix with two identical rows).
Therefore $A^{-1} = \frac{1}{\det(A)} \text{adj}(A)$, and the solution$\mathbf{x} = A^{-1}\mathbf{b}$ gives $x_j = \frac{1}{\det(A)} \sum_k C_{jk} b_k$, which is precisely $\frac{\det(A_j)}{\det(A)}$ by cofactor expansion of $A_j$.
Practical Note
While Cramer's rule is elegant, it is computationally expensive ($O(n \cdot n!)$ via cofactor expansion, or $O(n^4)$ if each determinant is computed by elimination). Gaussian elimination ($O(n^3)$) is far more efficient for solving systems. However, Cramer's rule is invaluable for theoretical analysis and for deriving formulas for small systems.
3.5 Invertibility
The Invertible Matrix Theorem
For an $n \times n$ matrix $A$, the following are equivalent:
- 1. $A$ is invertible
- 2. $\det(A) \neq 0$
- 3. $\text{rank}(A) = n$
- 4. $\ker(A) = \{\mathbf{0}\}$
- 5. The columns of $A$ are linearly independent
- 6. The columns of $A$ span $\mathbb{F}^n$
- 7. $A\mathbf{x} = \mathbf{b}$ has a unique solution for every $\mathbf{b}$
- 8. $A^T$ is invertible
- 9. The eigenvalues of $A$ are all nonzero
- 10. $0$ is not a singular value of $A$
Derivation 5: Inverse via Adjugate
The explicit formula for the inverse in terms of cofactors is:
For a $2 \times 2$ matrix, this simplifies to the well-known formula:
3.6 Historical Development
Seki Takakazu and Leibniz (1683β1693)
The determinant was independently discovered by the Japanese mathematician Seki Takakazu (1683) and Gottfried Wilhelm Leibniz (1693). Seki developed the concept while solving systems of equations using what he called "kihou" (literal: technique of simultaneous equations). Leibniz recognized that the sign pattern of determinants encodes permutation parity.
Gabriel Cramer (1750)
Cramer published his rule for solving systems of linear equations in his Introduction to the Analysis of Algebraic Curves. Though the formula bears his name, similar ideas had been explored by Colin Maclaurin as early as 1729.
Augustin-Louis Cauchy (1812)
Cauchy systematized the theory of determinants, proving the multiplicative property$\det(AB) = \det(A)\det(B)$ and establishing much of the modern notation. His 84-page memoir on the subject was a landmark in linear algebra.
James Joseph Sylvester (1850)
Sylvester coined the term "matrix" in 1850 (from the Latin for "womb"), viewing matrices as generating arrays from which determinants could be "born." His collaboration with Cayley established matrix algebra as a distinct branch of mathematics.
3.7 Applications
Volume and Orientation
The absolute value $|\det(A)|$ gives the volume scaling factor of the linear transformation. For vectors $\mathbf{v}_1, \ldots, \mathbf{v}_n$, the volume of the parallelepiped they span is $|\det(\mathbf{v}_1 \mid \cdots \mid \mathbf{v}_n)|$. The sign of the determinant indicates whether the map preserves or reverses orientation.
Cross Product
The cross product in $\mathbb{R}^3$ can be expressed as a determinant:$\mathbf{a} \times \mathbf{b} = \det\begin{pmatrix} \mathbf{i} & \mathbf{j} & \mathbf{k} \\ a_1 & a_2 & a_3 \\ b_1 & b_2 & b_3 \end{pmatrix}$. This extends to the $n$-dimensional setting via the exterior product.
Change of Variables in Integration
The Jacobian determinant appears in the multivariable change of variables formula:$\int_{\phi(U)} f(\mathbf{y})\, d\mathbf{y} = \int_U f(\phi(\mathbf{x})) \left|\det\left(\frac{\partial \phi}{\partial \mathbf{x}}\right)\right| d\mathbf{x}$. This is why the determinant measures volume distortion.
Characteristic Polynomial
The characteristic polynomial $p(\lambda) = \det(A - \lambda I)$ is the key to eigenvalue theory. Its roots are the eigenvalues, and its coefficients encode the trace ($\text{tr}(A) = \sum \lambda_i$) and determinant ($\det(A) = \prod \lambda_i$).
3.8 Computational Exploration
The following simulation implements determinant computation via both the Leibniz formula and cofactor expansion, verifies key determinant properties, demonstrates Cramer's rule, and visualizes how determinants relate to geometric transformations.
Matrices and Determinants: Computation and Visualization
PythonClick Run to execute the Python code
Code will be executed with Python 3 on the server