Matrices
Matrices provide a compact algebraic framework for representing and manipulating systems of linear
equations, geometric transformations, and — at a more advanced level — quantum mechanical states and
data structures. This topic develops the algebra of matrices and their interpretation as linear
transformations of the plane and space.
Board Coverage
Board Paper Notes AQA Paper 1 3D transformations, eigenvalues and eigenvectors Edexcel FP1 2D transformations, 3 × 3 3\times3 3 × 3 matrices, determinants OCR (A) Paper 1 2D transformations, 3 × 3 3\times3 3 × 3 matrices CIE P1 2D and basic 3D transformations, inverses, determinants
1. Matrix Notation and Operations
Definition. An m × n m \times n m × n matrix is a rectangular array of numbers arranged in m m m rows and
n n n columns. We write A = ( a i j ) \mathbf{A} = (a_{ij}) A = ( a ij ) where a i j a_{ij} a ij is the entry in row i i i , column j j j .
The set of all m × n m \times n m × n matrices with real entries is denoted M m × n ( R ) M_{m\times n}(\mathbb{R}) M m × n ( R ) .
A matrix with a single row is a row vector , and a matrix with a single column is a column
vector .
1.1 Matrix Addition and Scalar Multiplication
For matrices A , B ∈ M m × n ( R ) \mathbf{A}, \mathbf{B} \in M_{m \times n}(\mathbb{R}) A , B ∈ M m × n ( R ) and scalar k ∈ R k \in \mathbb{R} k ∈ R :
( A + B ) i j = a i j + b i j ( k A ) i j = k ⋅ a i j \begin{aligned}
(\mathbf{A} + \mathbf{B})_{ij} &= a_{ij} + b_{ij} \\
(k\mathbf{A})_{ij} &= k \cdot a_{ij}
\end{aligned} ( A + B ) ij ( k A ) ij = a ij + b ij = k ⋅ a ij
Matrix addition is commutative (A + B = B + A \mathbf{A} + \mathbf{B} = \mathbf{B} + \mathbf{A} A + B = B + A ) and
associative.
1.2 Matrix Multiplication
Definition. If A ∈ M m × p ( R ) \mathbf{A} \in M_{m \times p}(\mathbb{R}) A ∈ M m × p ( R ) and
B ∈ M p × n ( R ) \mathbf{B} \in M_{p \times n}(\mathbb{R}) B ∈ M p × n ( R ) , the product
A B ∈ M m × n ( R ) \mathbf{AB} \in M_{m \times n}(\mathbb{R}) AB ∈ M m × n ( R ) is defined by:
( A B ) i j = ∑ k = 1 p a i k b k j \boxed{(\mathbf{AB})_{ij} = \sum_{k=1}^{p} a_{ik}\, b_{kj}} ( AB ) ij = k = 1 ∑ p a ik b k j
Matrix multiplication is associative but not commutative in general:
A B ≠ B A \mathbf{AB} \neq \mathbf{BA} AB = BA .
The n × n n \times n n × n identity matrix I n \mathbf{I}_n I n satisfies
A I n = I n A = A \mathbf{A}\mathbf{I}_n = \mathbf{I}_n\mathbf{A} = \mathbf{A} A I n = I n A = A for any
A ∈ M m × n ( R ) \mathbf{A} \in M_{m \times n}(\mathbb{R}) A ∈ M m × n ( R ) .
warning "apply
B \mathbf{B} B first, then
A \mathbf{A} A ." When composing transformations,
the rightmost matrix is applied first. :::
2. Determinants 2.1 The 2 × 2 2 \times 2 2 × 2 Determinant Definition. For A = ( a b c d ) \mathbf{A} = \begin{pmatrix} a & b \\ c & d \end{pmatrix} A = ( a c b d ) :
det ( A ) = a d − b c \boxed{\det(\mathbf{A}) = ad - bc} det ( A ) = a d − b c
2.2 The 3 × 3 3 \times 3 3 × 3 Determinant Definition. For
A = ( a 11 a 12 a 13 a 21 a 22 a 23 a 31 a 32 a 33 ) \mathbf{A} = \begin{pmatrix} a_{11} & a_{12} & a_{13} \\ a_{21} & a_{22} & a_{23} \\ a_{31} & a_{32} & a_{33} \end{pmatrix} A = a 11 a 21 a 31 a 12 a 22 a 32 a 13 a 23 a 33 ,
the determinant is computed by cofactor expansion along any row or column:
det ( A ) = a 11 ∣ a 22 a 23 a 32 a 33 ∣ − a 12 ∣ a 21 a 23 a 31 a 33 ∣ + a 13 ∣ a 21 a 22 a 31 a 32 ∣ \boxed{\det(\mathbf{A}) = a_{11}\begin{vmatrix} a_{22} & a_{23} \\ a_{32} & a_{33} \end{vmatrix} - a_{12}\begin{vmatrix} a_{21} & a_{23} \\ a_{31} & a_{33} \end{vmatrix} + a_{13}\begin{vmatrix} a_{21} & a_{22} \\ a_{31} & a_{32} \end{vmatrix}} det ( A ) = a 11 a 22 a 32 a 23 a 33 − a 12 a 21 a 31 a 23 a 33 + a 13 a 21 a 31 a 22 a 32
The signs alternate + + + , − - − , + + + along the first row (following the checkerboard pattern).
2.3 Properties of Determinants For n × n n \times n n × n matrices A , B \mathbf{A}, \mathbf{B} A , B and scalar k k k :
det ( I n ) = 1 \det(\mathbf{I}_n) = 1 det ( I n ) = 1
det ( A T ) = det ( A ) \det(\mathbf{A}^T) = \det(\mathbf{A}) det ( A T ) = det ( A )
det ( k A ) = k n det ( A ) \det(k\mathbf{A}) = k^n \det(\mathbf{A}) det ( k A ) = k n det ( A )
Swapping two rows (or columns) multiplies the determinant by − 1 -1 − 1
A matrix with a zero row (or column) has det = 0 \det = 0 det = 0
det ( A ) = 0 \det(\mathbf{A}) = 0 det ( A ) = 0 if and only if A \mathbf{A} A is singular (non-invertible)
Proof of det ( A B ) = det ( A ) det ( B ) \det(\mathbf{AB}) = \det(\mathbf{A})\det(\mathbf{B}) det ( AB ) = det ( A ) det ( B ) (for 2 × 2 2\times2 2 × 2 matrices) Let A = ( a b c d ) \mathbf{A} = \begin{pmatrix} a & b \\ c & d \end{pmatrix} A = ( a c b d ) and
B = ( e f g h ) \mathbf{B} = \begin{pmatrix} e & f \\ g & h \end{pmatrix} B = ( e g f h ) .
A B = ( a e + b g a f + b h c e + d g c f + d h ) \mathbf{AB} = \begin{pmatrix} ae + bg & af + bh \\ ce + dg & cf + dh \end{pmatrix} AB = ( a e + b g ce + d g a f + bh c f + d h )
det ( A B ) = ( a e + b g ) ( c f + d h ) − ( a f + b h ) ( c e + d g ) = a c e f + a d e h + b c f g + b d g h − a c e f − a d f g − b c e h − b d g h = a d e h + b c f g − a d f g − b c e h = a d ( e h − f g ) − b c ( e h − f g ) = ( a d − b c ) ( e h − f g ) = det ( A ) ⋅ det ( B ) □ \begin{aligned}
\det(\mathbf{AB}) &= (ae + bg)(cf + dh) - (af + bh)(ce + dg) \\
&= acef + adeh + bcfg + bdgh - acef - adfg - bceh - bdgh \\
&= adeh + bcfg - adfg - bceh \\
&= ad(eh - fg) - bc(eh - fg) \\
&= (ad - bc)(eh - fg) \\
&= \det(\mathbf{A}) \cdot \det(\mathbf{B}) \quad \square
\end{aligned} det ( AB ) = ( a e + b g ) ( c f + d h ) − ( a f + bh ) ( ce + d g ) = a ce f + a d e h + b c f g + b d g h − a ce f − a df g − b ce h − b d g h = a d e h + b c f g − a df g − b ce h = a d ( e h − f g ) − b c ( e h − f g ) = ( a d − b c ) ( e h − f g ) = det ( A ) ⋅ det ( B ) □ Intuition. The determinant measures how a matrix scales area (in 2D) or volume (in 3D).
Composing two transformations multiplies their area/volume scaling factors, which is why
determinants multiply.
3. Inverse Matrices Definition. The inverse of a square matrix A \mathbf{A} A , written A − 1 \mathbf{A}^{-1} A − 1 , is the
unique matrix satisfying:
A A − 1 = A − 1 A = I \mathbf{A}\mathbf{A}^{-1} = \mathbf{A}^{-1}\mathbf{A} = \mathbf{I} A A − 1 = A − 1 A = I
An inverse exists if and only if det ( A ) ≠ 0 \det(\mathbf{A}) \neq 0 det ( A ) = 0 . Such a matrix is called non-singular .
3.1 Inverse of a 2 × 2 2 \times 2 2 × 2 Matrix For A = ( a b c d ) \mathbf{A} = \begin{pmatrix} a & b \\ c & d \end{pmatrix} A = ( a c b d ) with
det ( A ) = a d − b c ≠ 0 \det(\mathbf{A}) = ad - bc \neq 0 det ( A ) = a d − b c = 0 :
A − 1 = 1 a d − b c ( d − b − c a ) \boxed{\mathbf{A}^{-1} = \frac{1}{ad - bc}\begin{pmatrix} d & -b \\ -c & a \end{pmatrix}} A − 1 = a d − b c 1 ( d − c − b a )
3.2 The Adjugate Method (3 × 3 3 \times 3 3 × 3 ) For a 3 × 3 3 \times 3 3 × 3 matrix A \mathbf{A} A :
A − 1 = ◆ L B ◆ 1 ◆ R B ◆◆ L B ◆ det ( A ) ◆ R B ◆ adj ( A ) \boxed{\mathbf{A}^{-1} = \frac◆LB◆1◆RB◆◆LB◆\det(\mathbf{A})◆RB◆\operatorname{adj}(\mathbf{A})} A − 1 = L ◆ B ◆1◆ R B ◆◆ L B ◆ det ( A ) ◆ R B ◆ adj ( A )
where the adjugate (or adjoint ) matrix adj ( A ) \operatorname{adj}(\mathbf{A}) adj ( A ) is the transpose of the
cofactor matrix .
Definition. The cofactor C i j C_{ij} C ij of entry a i j a_{ij} a ij is ( − 1 ) i + j (-1)^{i+j} ( − 1 ) i + j times the determinant of
the submatrix obtained by deleting row i i i and column j j j . The cofactor matrix has entries
C i j C_{ij} C ij , and adj ( A ) = ( C i j ) T \operatorname{adj}(\mathbf{A}) = (C_{ij})^T adj ( A ) = ( C ij ) T .
Proof that the inverse is unique Suppose B \mathbf{B} B and C \mathbf{C} C are both inverses of A \mathbf{A} A . Then:
B = B I = B ( A C ) = ( B A ) C = I C = C □ \mathbf{B} = \mathbf{B}\mathbf{I} = \mathbf{B}(\mathbf{AC}) = (\mathbf{BA})\mathbf{C} = \mathbf{IC} = \mathbf{C} \quad \square B = BI = B ( AC ) = ( BA ) C = IC = C □
To verify your inverse, always check that
A A − 1 = I \mathbf{A}\mathbf{A}^{-1} = \mathbf{I} A A − 1 = I . This
catches sign errors and arithmetic mistakes immediately. :::
Worked Example: 3 × 3 3\times3 3 × 3 inverse Find A − 1 \mathbf{A}^{-1} A − 1 where
A = ( 1 2 0 0 1 3 1 0 1 ) \mathbf{A} = \begin{pmatrix} 1 & 2 & 0 \\ 0 & 1 & 3 \\ 1 & 0 & 1 \end{pmatrix} A = 1 0 1 2 1 0 0 3 1 .
det ( A ) = 1 ∣ 1 3 0 1 ∣ − 2 ∣ 0 3 1 1 ∣ + 0 = 1 ( 1 ) − 2 ( − 3 ) = 7 \det(\mathbf{A}) = 1\begin{vmatrix} 1 & 3 \\ 0 & 1 \end{vmatrix} - 2\begin{vmatrix} 0 & 3 \\ 1 & 1 \end{vmatrix} + 0 = 1(1) - 2(-3) = 7 det ( A ) = 1 1 0 3 1 − 2 0 1 3 1 + 0 = 1 ( 1 ) − 2 ( − 3 ) = 7 .
Cofactors:
C 11 = + ∣ 1 3 0 1 ∣ = 1 , C 12 = − ∣ 0 3 1 1 ∣ = 3 , C 13 = + ∣ 0 1 1 0 ∣ = − 1 C 21 = − ∣ 2 0 0 1 ∣ = − 2 , C 22 = + ∣ 1 0 1 1 ∣ = 1 , C 23 = − ∣ 1 2 1 0 ∣ = 2 C 31 = + ∣ 2 0 1 3 ∣ = 6 , C 32 = − ∣ 1 0 0 3 ∣ = − 3 , C 33 = + ∣ 1 2 0 1 ∣ = 1 \begin{aligned}
C_{11} &= +\begin{vmatrix} 1 & 3 \\ 0 & 1 \end{vmatrix} = 1, \quad C_{12} &= -\begin{vmatrix} 0 & 3 \\ 1 & 1 \end{vmatrix} = 3, \quad C_{13} &= +\begin{vmatrix} 0 & 1 \\ 1 & 0 \end{vmatrix} = -1 \\
C_{21} &= -\begin{vmatrix} 2 & 0 \\ 0 & 1 \end{vmatrix} = -2, \quad C_{22} &= +\begin{vmatrix} 1 & 0 \\ 1 & 1 \end{vmatrix} = 1, \quad C_{23} &= -\begin{vmatrix} 1 & 2 \\ 1 & 0 \end{vmatrix} = 2 \\
C_{31} &= +\begin{vmatrix} 2 & 0 \\ 1 & 3 \end{vmatrix} = 6, \quad C_{32} &= -\begin{vmatrix} 1 & 0 \\ 0 & 3 \end{vmatrix} = -3, \quad C_{33} &= +\begin{vmatrix} 1 & 2 \\ 0 & 1 \end{vmatrix} = 1
\end{aligned} C 11 C 21 C 31 = + 1 0 3 1 = 1 , C 12 = − 2 0 0 1 = − 2 , C 22 = + 2 1 0 3 = 6 , C 32 = − 0 1 3 1 = 3 , C 13 = + 1 1 0 1 = 1 , C 23 = − 1 0 0 3 = − 3 , C 33 = + 0 1 1 0 = − 1 = − 1 1 2 0 = 2 = + 1 0 2 1 = 1 adj ( A ) = ( 1 − 2 6 3 1 − 3 − 1 2 1 ) \operatorname{adj}(\mathbf{A}) = \begin{pmatrix} 1 & -2 & 6 \\ 3 & 1 & -3 \\ -1 & 2 & 1 \end{pmatrix} adj ( A ) = 1 3 − 1 − 2 1 2 6 − 3 1
A − 1 = 1 7 ( 1 − 2 6 3 1 − 3 − 1 2 1 ) \mathbf{A}^{-1} = \frac{1}{7}\begin{pmatrix} 1 & -2 & 6 \\ 3 & 1 & -3 \\ -1 & 2 & 1 \end{pmatrix} A − 1 = 7 1 1 3 − 1 − 2 1 2 6 − 3 1
4. Solving Systems of Linear Equations A system of n n n linear equations in n n n unknowns can be written as
A x = b \mathbf{A}\mathbf{x} = \mathbf{b} Ax = b where A \mathbf{A} A is the coefficient matrix, x \mathbf{x} x is
the column vector of unknowns, and b \mathbf{b} b is the column vector of constants.
If A \mathbf{A} A is non-singular, the unique solution is:
x = A − 1 b \boxed{\mathbf{x} = \mathbf{A}^{-1}\mathbf{b}} x = A − 1 b
4.1 Geometric Interpretation (2D) For a 2 × 2 2 \times 2 2 × 2 system:
det ( A ) ≠ 0 \det(\mathbf{A}) \neq 0 det ( A ) = 0 : the two lines intersect at a unique point.
det ( A ) = 0 \det(\mathbf{A}) = 0 det ( A ) = 0 and the equations are consistent: the lines are coincident (infinitely many
solutions).
det ( A ) = 0 \det(\mathbf{A}) = 0 det ( A ) = 0 and the equations are inconsistent: the lines are parallel (no solutions).
4.2 Cramer's Rule For a system A x = b \mathbf{A}\mathbf{x} = \mathbf{b} Ax = b where det ( A ) ≠ 0 \det(\mathbf{A}) \neq 0 det ( A ) = 0 :
x i = ◆ L B ◆ det ( A i ) ◆ R B ◆◆ L B ◆ det ( A ) ◆ R B ◆ x_i = \frac◆LB◆\det(\mathbf{A}_i)◆RB◆◆LB◆\det(\mathbf{A})◆RB◆ x i = L ◆ B ◆ det ( A i ) ◆ R B ◆◆ L B ◆ det ( A ) ◆ R B ◆
where A i \mathbf{A}_i A i is A \mathbf{A} A with column i i i replaced by b \mathbf{b} b .
Worked Example: Solving a 3 × 3 3\times3 3 × 3 system Solve: { x + 2 y = 4 y + 3 z = 5 x + z = 2 \begin{cases} x + 2y = 4 \\ y + 3z = 5 \\ x + z = 2 \end{cases} ⎩ ⎨ ⎧ x + 2 y = 4 y + 3 z = 5 x + z = 2
In matrix form:
( 1 2 0 0 1 3 1 0 1 ) ( x y z ) = ( 4 5 2 ) \begin{pmatrix} 1 & 2 & 0 \\ 0 & 1 & 3 \\ 1 & 0 & 1 \end{pmatrix}\begin{pmatrix} x \\ y \\ z \end{pmatrix} = \begin{pmatrix} 4 \\ 5 \\ 2 \end{pmatrix} 1 0 1 2 1 0 0 3 1 x y z = 4 5 2 .
Using the inverse from the previous example:
( x y z ) = 1 7 ( 1 − 2 6 3 1 − 3 − 1 2 1 ) ( 4 5 2 ) = 1 7 ( 4 − 10 + 12 12 + 5 − 6 − 4 + 10 + 2 ) = 1 7 ( 6 11 8 ) = ( 6 / 7 11 / 7 8 / 7 ) \begin{pmatrix} x \\ y \\ z \end{pmatrix} = \frac{1}{7}\begin{pmatrix} 1 & -2 & 6 \\ 3 & 1 & -3 \\ -1 & 2 & 1 \end{pmatrix}\begin{pmatrix} 4 \\ 5 \\ 2 \end{pmatrix} = \frac{1}{7}\begin{pmatrix} 4 - 10 + 12 \\ 12 + 5 - 6 \\ -4 + 10 + 2 \end{pmatrix} = \frac{1}{7}\begin{pmatrix} 6 \\ 11 \\ 8 \end{pmatrix} = \begin{pmatrix} 6/7 \\ 11/7 \\ 8/7 \end{pmatrix} x y z = 7 1 1 3 − 1 − 2 1 2 6 − 3 1 4 5 2 = 7 1 4 − 10 + 12 12 + 5 − 6 − 4 + 10 + 2 = 7 1 6 11 8 = 6/7 11/7 8/7
A 2 × 2 2 \times 2 2 × 2 matrix T \mathbf{T} T represents a linear transformation of R 2 \mathbb{R}^2 R 2 : the point
( x , y ) (x, y) ( x , y ) is mapped to ( x ′ , y ′ ) (x', y') ( x ′ , y ′ ) where
( x ′ y ′ ) = T ( x y ) \begin{pmatrix} x' \\ y' \end{pmatrix} = \mathbf{T}\begin{pmatrix} x \\ y \end{pmatrix} ( x ′ y ′ ) = T ( x y ) .
5.1 Reflection Reflection in the x x x -axis: ( 1 0 0 − 1 ) \begin{pmatrix} 1 & 0 \\ 0 & -1 \end{pmatrix} ( 1 0 0 − 1 )
Reflection in the y y y -axis: ( − 1 0 0 1 ) \begin{pmatrix} -1 & 0 \\ 0 & 1 \end{pmatrix} ( − 1 0 0 1 )
Reflection in the line y = x y = x y = x : ( 0 1 1 0 ) \begin{pmatrix} 0 & 1 \\ 1 & 0 \end{pmatrix} ( 0 1 1 0 )
Reflection in the line y = tan θ x y = \tan\theta\, x y = tan θ x :
R = ( cos 2 θ sin 2 θ sin 2 θ − cos 2 θ ) \boxed{\mathbf{R} = \begin{pmatrix} \cos 2\theta & \sin 2\theta \\ \sin 2\theta & -\cos 2\theta \end{pmatrix}} R = ( cos 2 θ sin 2 θ sin 2 θ − cos 2 θ )
Proof of the reflection matrix The reflection of a vector in a line through the origin making angle θ \theta θ with the x x x -axis can
be decomposed: first rotate by − θ -\theta − θ to align the mirror with the x x x -axis, reflect in the
x x x -axis, then rotate back by θ \theta θ .
R = ( cos θ − sin θ sin θ cos θ ) ( 1 0 0 − 1 ) ( cos θ sin θ − sin θ cos θ ) \mathbf{R} = \begin{pmatrix} \cos\theta & -\sin\theta \\ \sin\theta & \cos\theta \end{pmatrix}\begin{pmatrix} 1 & 0 \\ 0 & -1 \end{pmatrix}\begin{pmatrix} \cos\theta & \sin\theta \\ -\sin\theta & \cos\theta \end{pmatrix} R = ( cos θ sin θ − sin θ cos θ ) ( 1 0 0 − 1 ) ( cos θ − sin θ sin θ cos θ )
= ( cos θ sin θ sin θ − cos θ ) ( cos θ sin θ − sin θ cos θ ) = \begin{pmatrix} \cos\theta & \sin\theta \\ \sin\theta & -\cos\theta \end{pmatrix}\begin{pmatrix} \cos\theta & \sin\theta \\ -\sin\theta & \cos\theta \end{pmatrix} = ( cos θ sin θ sin θ − cos θ ) ( cos θ − sin θ sin θ cos θ )
= ( cos 2 θ − sin 2 θ cos θ sin θ + sin θ cos θ sin θ cos θ + cos θ sin θ sin 2 θ − cos 2 θ ) = ( cos 2 θ sin 2 θ sin 2 θ − cos 2 θ ) □ = \begin{pmatrix} \cos^2\theta - \sin^2\theta & \cos\theta\sin\theta + \sin\theta\cos\theta \\ \sin\theta\cos\theta + \cos\theta\sin\theta & \sin^2\theta - \cos^2\theta \end{pmatrix} = \begin{pmatrix} \cos 2\theta & \sin 2\theta \\ \sin 2\theta & -\cos 2\theta \end{pmatrix} \quad \square = ( cos 2 θ − sin 2 θ sin θ cos θ + cos θ sin θ cos θ sin θ + sin θ cos θ sin 2 θ − cos 2 θ ) = ( cos 2 θ sin 2 θ sin 2 θ − cos 2 θ ) □
5.2 Rotation Rotation anticlockwise by angle θ \theta θ about the origin:
R θ = ( cos θ − sin θ sin θ cos θ ) \boxed{\mathbf{R}_\theta = \begin{pmatrix} \cos\theta & -\sin\theta \\ \sin\theta & \cos\theta \end{pmatrix}} R θ = ( cos θ sin θ − sin θ cos θ )
Note: det ( R θ ) = cos 2 θ + sin 2 θ = 1 \det(\mathbf{R}_\theta) = \cos^2\theta + \sin^2\theta = 1 det ( R θ ) = cos 2 θ + sin 2 θ = 1 .
5.3 Enlargement Enlargement by scale factor k k k about the origin: ( k 0 0 k ) \begin{pmatrix} k & 0 \\ 0 & k \end{pmatrix} ( k 0 0 k )
det = k 2 \det = k^2 det = k 2 , confirming the area is scaled by k 2 k^2 k 2 .
5.4 Shear Shear parallel to the x x x -axis by factor k k k : ( 1 k 0 1 ) \begin{pmatrix} 1 & k \\ 0 & 1 \end{pmatrix} ( 1 0 k 1 )
Shear parallel to the y y y -axis by factor k k k : ( 1 0 k 1 ) \begin{pmatrix} 1 & 0 \\ k & 1 \end{pmatrix} ( 1 k 0 1 )
Note: det = 1 \det = 1 det = 1 for shears, so area is preserved.
If transformation A \mathbf{A} A is followed by transformation B \mathbf{B} B , the combined
transformation is represented by B A \mathbf{BA} BA (rightmost applied first).
Order matters. A rotation followed by a reflection generally produces a different result
from a reflection followed by a rotation. The combined matrix is B A \mathbf{BA} BA (not A B \mathbf{AB} AB )
when A \mathbf{A} A is applied first. :::
To find the matrix of a combined transformation, multiply the matrices in reverse order of
application. If the question says "reflect then rotate," compute
R r o t × R r e f \mathbf{R}_{\mathrm{rot}} \times \mathbf{R}_{\mathrm{ref}} R rot × R ref . :::
Worked Example: Combined transformation Find the matrix representing a rotation of 90 ∘ 90^\circ 9 0 ∘ anticlockwise about the origin followed by a
reflection in the line y = x y = x y = x .
Rotation by 90 ∘ 90^\circ 9 0 ∘ : R = ( 0 − 1 1 0 ) \mathbf{R} = \begin{pmatrix} 0 & -1 \\ 1 & 0 \end{pmatrix} R = ( 0 1 − 1 0 )
Reflection in y = x y = x y = x : S = ( 0 1 1 0 ) \mathbf{S} = \begin{pmatrix} 0 & 1 \\ 1 & 0 \end{pmatrix} S = ( 0 1 1 0 )
Combined (reflection applied after rotation):
S R = ( 0 1 1 0 ) ( 0 − 1 1 0 ) = ( 1 0 0 − 1 ) \mathbf{S}\mathbf{R} = \begin{pmatrix} 0 & 1 \\ 1 & 0 \end{pmatrix}\begin{pmatrix} 0 & -1 \\ 1 & 0 \end{pmatrix} = \begin{pmatrix} 1 & 0 \\ 0 & -1 \end{pmatrix} SR = ( 0 1 1 0 ) ( 0 1 − 1 0 ) = ( 1 0 0 − 1 )
This is a reflection in the x x x -axis. det = − 1 \det = -1 det = − 1 , consistent with an orientation-reversing
transformation.
3D transformations are required by AQA and appear on CIE P1. Edexcel and OCR focus primarily
on 2D but may include basic 3 × 3 3\times3 3 × 3 determinant and inverse calculations. :::
6.1 Rotations in 3D Rotation about the x x x -axis by angle θ \theta θ :
R x = ( 1 0 0 0 cos θ − sin θ 0 sin θ cos θ ) \mathbf{R}_x = \begin{pmatrix} 1 & 0 & 0 \\ 0 & \cos\theta & -\sin\theta \\ 0 & \sin\theta & \cos\theta \end{pmatrix} R x = 1 0 0 0 cos θ sin θ 0 − sin θ cos θ
Rotation about the y y y -axis by angle θ \theta θ :
R y = ( cos θ 0 sin θ 0 1 0 − sin θ 0 cos θ ) \mathbf{R}_y = \begin{pmatrix} \cos\theta & 0 & \sin\theta \\ 0 & 1 & 0 \\ -\sin\theta & 0 & \cos\theta \end{pmatrix} R y = cos θ 0 − sin θ 0 1 0 sin θ 0 cos θ
Rotation about the z z z -axis by angle θ \theta θ :
R z = ( cos θ − sin θ 0 sin θ cos θ 0 0 0 1 ) \mathbf{R}_z = \begin{pmatrix} \cos\theta & -\sin\theta & 0 \\ \sin\theta & \cos\theta & 0 \\ 0 & 0 & 1 \end{pmatrix} R z = cos θ sin θ 0 − sin θ cos θ 0 0 0 1
Each has det = 1 \det = 1 det = 1 and represents a rigid motion preserving distances and orientation.
6.2 Reflections in 3D Reflection in the plane x = 0 x = 0 x = 0 (the y z yz y z -plane):
( − 1 0 0 0 1 0 0 0 1 ) \begin{pmatrix} -1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & 1 \end{pmatrix} − 1 0 0 0 1 0 0 0 1
Reflection in the plane y = 0 y = 0 y = 0 (the x z xz x z -plane):
( 1 0 0 0 − 1 0 0 0 1 ) \begin{pmatrix} 1 & 0 & 0 \\ 0 & -1 & 0 \\ 0 & 0 & 1 \end{pmatrix} 1 0 0 0 − 1 0 0 0 1
Reflection in the plane z = 0 z = 0 z = 0 (the x y xy x y -plane):
( 1 0 0 0 1 0 0 0 − 1 ) \begin{pmatrix} 1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & -1 \end{pmatrix} 1 0 0 0 1 0 0 0 − 1
Each has det = − 1 \det = -1 det = − 1 , confirming orientation reversal.
7. Eigenvalues and Eigenvectors info require this topic at A Level. :::
Definition. Let A \mathbf{A} A be an n × n n \times n n × n matrix. A scalar λ \lambda λ is an eigenvalue of
A \mathbf{A} A if there exists a non-zero vector v \mathbf{v} v such that:
A v = λ v \boxed{\mathbf{A}\mathbf{v} = \lambda\mathbf{v}} Av = λ v
The vector v \mathbf{v} v is called an eigenvector corresponding to λ \lambda λ .
7.1 The Characteristic Equation A v = λ v ⟺ ( A − λ I ) v = 0 \mathbf{A}\mathbf{v} = \lambda\mathbf{v} \iff (\mathbf{A} - \lambda\mathbf{I})\mathbf{v} = \mathbf{0} Av = λ v ⟺ ( A − λ I ) v = 0 .
For a non-trivial solution (v ≠ 0 \mathbf{v} \neq \mathbf{0} v = 0 ), we require
det ( A − λ I ) = 0 \det(\mathbf{A} - \lambda\mathbf{I}) = 0 det ( A − λ I ) = 0 .
det ( A − λ I ) = 0 \boxed{\det(\mathbf{A} - \lambda\mathbf{I}) = 0} det ( A − λ I ) = 0
This is the characteristic equation of A \mathbf{A} A . Its roots are the eigenvalues.
7.2 Finding Eigenvectors For each eigenvalue λ \lambda λ , solve ( A − λ I ) v = 0 (\mathbf{A} - \lambda\mathbf{I})\mathbf{v} = \mathbf{0} ( A − λ I ) v = 0 by
row reduction.
7.3 Diagonalisation Definition. A matrix A \mathbf{A} A is diagonalisable if there exists an invertible matrix
P \mathbf{P} P and a diagonal matrix D \mathbf{D} D such that:
A = P D P − 1 \boxed{\mathbf{A} = \mathbf{P}\mathbf{D}\mathbf{P}^{-1}} A = PD P − 1
The columns of P \mathbf{P} P are the eigenvectors of A \mathbf{A} A , and the diagonal entries of
D \mathbf{D} D are the corresponding eigenvalues.
A matrix is diagonalisable if and only if it has n n n linearly independent eigenvectors (always true
for n n n distinct eigenvalues).
Intuition. Diagonalisation changes to a coordinate system where the transformation acts
independently on each axis (stretching by eigenvalues). In this basis, the matrix takes its simplest
possible form.
Worked Example: Eigenvalues and eigenvectors Find the eigenvalues and eigenvectors of
A = ( 4 1 2 3 ) \mathbf{A} = \begin{pmatrix} 4 & 1 \\ 2 & 3 \end{pmatrix} A = ( 4 2 1 3 ) .
Characteristic equation:
det ( 4 − λ 1 2 3 − λ ) = 0 \det\!\begin{pmatrix} 4 - \lambda & 1 \\ 2 & 3 - \lambda \end{pmatrix} = 0 det ( 4 − λ 2 1 3 − λ ) = 0
( 4 − λ ) ( 3 − λ ) − 2 = λ 2 − 7 λ + 10 = ( λ − 5 ) ( λ − 2 ) = 0 (4 - \lambda)(3 - \lambda) - 2 = \lambda^2 - 7\lambda + 10 = (\lambda - 5)(\lambda - 2) = 0 ( 4 − λ ) ( 3 − λ ) − 2 = λ 2 − 7 λ + 10 = ( λ − 5 ) ( λ − 2 ) = 0
Eigenvalues: λ 1 = 5 \lambda_1 = 5 λ 1 = 5 , λ 2 = 2 \lambda_2 = 2 λ 2 = 2 .
For λ 1 = 5 \lambda_1 = 5 λ 1 = 5 :
( − 1 1 2 − 2 ) ( x y ) = ( 0 0 ) ⟹ − x + y = 0 ⟹ y = x \begin{pmatrix} -1 & 1 \\ 2 & -2 \end{pmatrix}\begin{pmatrix} x \\ y \end{pmatrix} = \begin{pmatrix} 0 \\ 0 \end{pmatrix} \implies -x + y = 0 \implies y = x ( − 1 2 1 − 2 ) ( x y ) = ( 0 0 ) ⟹ − x + y = 0 ⟹ y = x .
Eigenvector: v 1 = ( 1 1 ) \mathbf{v}_1 = \begin{pmatrix} 1 \\ 1 \end{pmatrix} v 1 = ( 1 1 ) (or any non-zero scalar
multiple).
For λ 2 = 2 \lambda_2 = 2 λ 2 = 2 :
( 2 1 2 1 ) ( x y ) = ( 0 0 ) ⟹ 2 x + y = 0 ⟹ y = − 2 x \begin{pmatrix} 2 & 1 \\ 2 & 1 \end{pmatrix}\begin{pmatrix} x \\ y \end{pmatrix} = \begin{pmatrix} 0 \\ 0 \end{pmatrix} \implies 2x + y = 0 \implies y = -2x ( 2 2 1 1 ) ( x y ) = ( 0 0 ) ⟹ 2 x + y = 0 ⟹ y = − 2 x .
Eigenvector: v 2 = ( 1 − 2 ) \mathbf{v}_2 = \begin{pmatrix} 1 \\ -2 \end{pmatrix} v 2 = ( 1 − 2 ) .
Diagonalisation: P = ( 1 1 1 − 2 ) \mathbf{P} = \begin{pmatrix} 1 & 1 \\ 1 & -2 \end{pmatrix} P = ( 1 1 1 − 2 ) ,
D = ( 5 0 0 2 ) \mathbf{D} = \begin{pmatrix} 5 & 0 \\ 0 & 2 \end{pmatrix} D = ( 5 0 0 2 ) , giving
A = P D P − 1 \mathbf{A} = \mathbf{P}\mathbf{D}\mathbf{P}^{-1} A = PD P − 1 .
8. Summary of Key Results det ( A B ) = det ( A ) det ( B ) \boxed{\det(\mathbf{AB}) = \det(\mathbf{A})\det(\mathbf{B})} det ( AB ) = det ( A ) det ( B )
A − 1 = ◆ L B ◆ 1 ◆ R B ◆◆ L B ◆ det ( A ) ◆ R B ◆ adj ( A ) w h e n det ( A ) ≠ 0 \boxed{\mathbf{A}^{-1} = \frac◆LB◆1◆RB◆◆LB◆\det(\mathbf{A})◆RB◆\operatorname{adj}(\mathbf{A}) \quad \mathrm{when } \det(\mathbf{A}) \neq 0} A − 1 = L ◆ B ◆1◆ R B ◆◆ L B ◆ det ( A ) ◆ R B ◆ adj ( A ) when det ( A ) = 0
R o t a t i o n b y θ : ( cos θ − sin θ sin θ cos θ ) \boxed{\mathrm{Rotation by } \theta: \begin{pmatrix} \cos\theta & -\sin\theta \\ \sin\theta & \cos\theta \end{pmatrix}} Rotationby θ : ( cos θ sin θ − sin θ cos θ )
R e f l e c t i o n i n y = ( tan θ ) x : ( cos 2 θ sin 2 θ sin 2 θ − cos 2 θ ) \boxed{\mathrm{Reflection in } y = (\tan\theta)x: \begin{pmatrix} \cos 2\theta & \sin 2\theta \\ \sin 2\theta & -\cos 2\theta \end{pmatrix}} Reflectionin y = ( tan θ ) x : ( cos 2 θ sin 2 θ sin 2 θ − cos 2 θ )
det ( A − λ I ) = 0 ⟹ e i g e n v a l u e s o f A \boxed{\det(\mathbf{A} - \lambda\mathbf{I}) = 0 \implies \mathrm{eigenvalues of } \mathbf{A}} det ( A − λ I ) = 0 ⟹ eigenvaluesof A
Problems Problem 1. Given A = ( 3 − 1 2 4 ) \mathbf{A} = \begin{pmatrix} 3 & -1 \\ 2 & 4 \end{pmatrix} A = ( 3 2 − 1 4 ) and
B = ( 1 5 − 2 0 ) \mathbf{B} = \begin{pmatrix} 1 & 5 \\ -2 & 0 \end{pmatrix} B = ( 1 − 2 5 0 ) , compute A B − B A \mathbf{AB} - \mathbf{BA} AB − BA .
Hint Compute both products separately and subtract. They will not be equal.
Answer A B = ( 3 − 1 2 4 ) ( 1 5 − 2 0 ) = ( 5 15 − 6 10 ) \mathbf{AB} = \begin{pmatrix} 3 & -1 \\ 2 & 4 \end{pmatrix}\begin{pmatrix} 1 & 5 \\ -2 & 0 \end{pmatrix} = \begin{pmatrix} 5 & 15 \\ -6 & 10 \end{pmatrix} AB = ( 3 2 − 1 4 ) ( 1 − 2 5 0 ) = ( 5 − 6 15 10 )
B A = ( 1 5 − 2 0 ) ( 3 − 1 2 4 ) = ( 13 19 − 6 2 ) \mathbf{BA} = \begin{pmatrix} 1 & 5 \\ -2 & 0 \end{pmatrix}\begin{pmatrix} 3 & -1 \\ 2 & 4 \end{pmatrix} = \begin{pmatrix} 13 & 19 \\ -6 & 2 \end{pmatrix} BA = ( 1 − 2 5 0 ) ( 3 2 − 1 4 ) = ( 13 − 6 19 2 )
A B − B A = ( 5 15 − 6 10 ) − ( 13 19 − 6 2 ) = ( − 8 − 4 0 8 ) \mathbf{AB} - \mathbf{BA} = \begin{pmatrix} 5 & 15 \\ -6 & 10 \end{pmatrix} - \begin{pmatrix} 13 & 19 \\ -6 & 2 \end{pmatrix} = \begin{pmatrix} -8 & -4 \\ 0 & 8 \end{pmatrix} AB − BA = ( 5 − 6 15 10 ) − ( 13 − 6 19 2 ) = ( − 8 0 − 4 8 )
This confirms A B ≠ B A \mathbf{AB} \neq \mathbf{BA} AB = BA .
Problem 2. Find the determinant and inverse of
A = ( 2 1 3 0 − 1 2 1 0 1 ) \mathbf{A} = \begin{pmatrix} 2 & 1 & 3 \\ 0 & -1 & 2 \\ 1 & 0 & 1 \end{pmatrix} A = 2 0 1 1 − 1 0 3 2 1 .
Hint Expand the determinant along the first row. Then compute cofactors for the adjugate.
Answer det ( A ) = 2 ∣ − 1 2 0 1 ∣ − 1 ∣ 0 2 1 1 ∣ + 3 ∣ 0 − 1 1 0 ∣ \det(\mathbf{A}) = 2\begin{vmatrix} -1 & 2 \\ 0 & 1 \end{vmatrix} - 1\begin{vmatrix} 0 & 2 \\ 1 & 1 \end{vmatrix} + 3\begin{vmatrix} 0 & -1 \\ 1 & 0 \end{vmatrix} det ( A ) = 2 − 1 0 2 1 − 1 0 1 2 1 + 3 0 1 − 1 0
= 2 ( − 1 ) − 1 ( − 2 ) + 3 ( 1 ) = − 2 + 2 + 3 = 3 = 2(-1) - 1(-2) + 3(1) = -2 + 2 + 3 = 3 = 2 ( − 1 ) − 1 ( − 2 ) + 3 ( 1 ) = − 2 + 2 + 3 = 3
Cofactors:
C 11 = + ( − 1 ) = − 1 , C 12 = − ( 0 ⋅ 1 − 2 ⋅ 1 ) = 2 , C 13 = + ( 0 ⋅ 0 − ( − 1 ) ⋅ 1 ) = 1 C 21 = − ( 1 ⋅ 1 − 3 ⋅ 0 ) = − 1 , C 22 = + ( 2 ⋅ 1 − 3 ⋅ 1 ) = − 1 , C 23 = − ( 2 ⋅ 0 − 1 ⋅ 1 ) = 1 C 31 = + ( 2 + 2 ) = 4 , C 32 = − ( 4 − 0 ) = − 4 , C 33 = + ( − 2 − 0 ) = − 2 \begin{aligned}
C_{11} = +(-1) = -1, \quad C_{12} = -(0 \cdot 1 - 2 \cdot 1) = 2, \quad C_{13} = +(0 \cdot 0 - (-1) \cdot 1) = 1 \\
C_{21} = -(1 \cdot 1 - 3 \cdot 0) = -1, \quad C_{22} = +(2 \cdot 1 - 3 \cdot 1) = -1, \quad C_{23} = -(2 \cdot 0 - 1 \cdot 1) = 1 \\
C_{31} = +(2 + 2) = 4, \quad C_{32} = -(4 - 0) = -4, \quad C_{33} = +(-2 - 0) = -2
\end{aligned} C 11 = + ( − 1 ) = − 1 , C 12 = − ( 0 ⋅ 1 − 2 ⋅ 1 ) = 2 , C 13 = + ( 0 ⋅ 0 − ( − 1 ) ⋅ 1 ) = 1 C 21 = − ( 1 ⋅ 1 − 3 ⋅ 0 ) = − 1 , C 22 = + ( 2 ⋅ 1 − 3 ⋅ 1 ) = − 1 , C 23 = − ( 2 ⋅ 0 − 1 ⋅ 1 ) = 1 C 31 = + ( 2 + 2 ) = 4 , C 32 = − ( 4 − 0 ) = − 4 , C 33 = + ( − 2 − 0 ) = − 2 adj ( A ) = ( − 1 − 1 4 2 − 1 − 4 1 1 − 2 ) \operatorname{adj}(\mathbf{A}) = \begin{pmatrix} -1 & -1 & 4 \\ 2 & -1 & -4 \\ 1 & 1 & -2 \end{pmatrix} adj ( A ) = − 1 2 1 − 1 − 1 1 4 − 4 − 2
A − 1 = 1 3 ( − 1 − 1 4 2 − 1 − 4 1 1 − 2 ) \mathbf{A}^{-1} = \frac{1}{3}\begin{pmatrix} -1 & -1 & 4 \\ 2 & -1 & -4 \\ 1 & 1 & -2 \end{pmatrix} A − 1 = 3 1 − 1 2 1 − 1 − 1 1 4 − 4 − 2
Problem 3. Find the matrix representing a reflection in the line y = 3 x y = \sqrt{3}\,x y = 3 x , and verify
that M 2 = I \mathbf{M}^2 = \mathbf{I} M 2 = I .
Hint The line makes angle θ \theta θ with the x x x -axis where tan θ = 3 \tan\theta = \sqrt{3} tan θ = 3 . A reflection applied
twice is the identity.
Answer tan θ = 3 ⟹ θ = ◆ L B ◆ π ◆ R B ◆◆ L B ◆ 3 ◆ R B ◆ \tan\theta = \sqrt{3} \implies \theta = \dfrac◆LB◆\pi◆RB◆◆LB◆3◆RB◆ tan θ = 3 ⟹ θ = L ◆ B ◆ π ◆ R B ◆◆ L B ◆3◆ R B ◆ .
M = ( cos ◆ L B ◆ 2 π ◆ R B ◆◆ L B ◆ 3 ◆ R B ◆ sin ◆ L B ◆ 2 π ◆ R B ◆◆ L B ◆ 3 ◆ R B ◆ sin ◆ L B ◆ 2 π ◆ R B ◆◆ L B ◆ 3 ◆ R B ◆ − cos ◆ L B ◆ 2 π ◆ R B ◆◆ L B ◆ 3 ◆ R B ◆ ) = ( − 1 2 ◆ L B ◆ 3 ◆ R B ◆◆ L B ◆ 2 ◆ R B ◆ ◆ L B ◆ 3 ◆ R B ◆◆ L B ◆ 2 ◆ R B ◆ 1 2 ) \mathbf{M} = \begin{pmatrix} \cos\frac◆LB◆2\pi◆RB◆◆LB◆3◆RB◆ & \sin\frac◆LB◆2\pi◆RB◆◆LB◆3◆RB◆ \\ \sin\frac◆LB◆2\pi◆RB◆◆LB◆3◆RB◆ & -\cos\frac◆LB◆2\pi◆RB◆◆LB◆3◆RB◆ \end{pmatrix} = \begin{pmatrix} -\frac{1}{2} & \frac◆LB◆\sqrt{3}◆RB◆◆LB◆2◆RB◆ \\ \frac◆LB◆\sqrt{3}◆RB◆◆LB◆2◆RB◆ & \frac{1}{2} \end{pmatrix} M = ( cos L ◆ B ◆2 π ◆ R B ◆◆ L B ◆3◆ R B ◆ sin L ◆ B ◆2 π ◆ R B ◆◆ L B ◆3◆ R B ◆ sin L ◆ B ◆2 π ◆ R B ◆◆ L B ◆3◆ R B ◆ − cos L ◆ B ◆2 π ◆ R B ◆◆ L B ◆3◆ R B ◆ ) = ( − 2 1 L ◆ B ◆ 3 ◆ R B ◆◆ L B ◆2◆ R B ◆ L ◆ B ◆ 3 ◆ R B ◆◆ L B ◆2◆ R B ◆ 2 1 )
Verification:
M 2 = ( − 1 2 ◆ L B ◆ 3 ◆ R B ◆◆ L B ◆ 2 ◆ R B ◆ ◆ L B ◆ 3 ◆ R B ◆◆ L B ◆ 2 ◆ R B ◆ 1 2 ) ( − 1 2 ◆ L B ◆ 3 ◆ R B ◆◆ L B ◆ 2 ◆ R B ◆ ◆ L B ◆ 3 ◆ R B ◆◆ L B ◆ 2 ◆ R B ◆ 1 2 ) = ( 1 4 + 3 4 − ◆ L B ◆ 3 ◆ R B ◆◆ L B ◆ 4 ◆ R B ◆ + ◆ L B ◆ 3 ◆ R B ◆◆ L B ◆ 4 ◆ R B ◆ − ◆ L B ◆ 3 ◆ R B ◆◆ L B ◆ 4 ◆ R B ◆ + ◆ L B ◆ 3 ◆ R B ◆◆ L B ◆ 4 ◆ R B ◆ 3 4 + 1 4 ) = ( 1 0 0 1 ) = I □ \mathbf{M}^2 = \begin{pmatrix} -\frac{1}{2} & \frac◆LB◆\sqrt{3}◆RB◆◆LB◆2◆RB◆ \\ \frac◆LB◆\sqrt{3}◆RB◆◆LB◆2◆RB◆ & \frac{1}{2} \end{pmatrix}\begin{pmatrix} -\frac{1}{2} & \frac◆LB◆\sqrt{3}◆RB◆◆LB◆2◆RB◆ \\ \frac◆LB◆\sqrt{3}◆RB◆◆LB◆2◆RB◆ & \frac{1}{2} \end{pmatrix} = \begin{pmatrix} \frac{1}{4} + \frac{3}{4} & -\frac◆LB◆\sqrt{3}◆RB◆◆LB◆4◆RB◆ + \frac◆LB◆\sqrt{3}◆RB◆◆LB◆4◆RB◆ \\ -\frac◆LB◆\sqrt{3}◆RB◆◆LB◆4◆RB◆ + \frac◆LB◆\sqrt{3}◆RB◆◆LB◆4◆RB◆ & \frac{3}{4} + \frac{1}{4} \end{pmatrix} = \begin{pmatrix} 1 & 0 \\ 0 & 1 \end{pmatrix} = \mathbf{I} \quad \square M 2 = ( − 2 1 L ◆ B ◆ 3 ◆ R B ◆◆ L B ◆2◆ R B ◆ L ◆ B ◆ 3 ◆ R B ◆◆ L B ◆2◆ R B ◆ 2 1 ) ( − 2 1 L ◆ B ◆ 3 ◆ R B ◆◆ L B ◆2◆ R B ◆ L ◆ B ◆ 3 ◆ R B ◆◆ L B ◆2◆ R B ◆ 2 1 ) = ( 4 1 + 4 3 − L ◆ B ◆ 3 ◆ R B ◆◆ L B ◆4◆ R B ◆ + L ◆ B ◆ 3 ◆ R B ◆◆ L B ◆4◆ R B ◆ − L ◆ B ◆ 3 ◆ R B ◆◆ L B ◆4◆ R B ◆ + L ◆ B ◆ 3 ◆ R B ◆◆ L B ◆4◆ R B ◆ 4 3 + 4 1 ) = ( 1 0 0 1 ) = I □
Problem 4. The triangle with vertices ( 0 , 0 ) (0, 0) ( 0 , 0 ) , ( 2 , 0 ) (2, 0) ( 2 , 0 ) , ( 0 , 1 ) (0, 1) ( 0 , 1 ) is transformed by the matrix
T = ( 3 1 0 2 ) \mathbf{T} = \begin{pmatrix} 3 & 1 \\ 0 & 2 \end{pmatrix} T = ( 3 0 1 2 ) . Find the coordinates of the vertices of
the image, and verify that the area scales by ∣ det ( T ) ∣ |\det(\mathbf{T})| ∣ det ( T ) ∣ .
Hint Apply T \mathbf{T} T to each vertex. The original triangle has area 1.
Answer ( 0 , 0 ) ↦ ( 0 , 0 ) (0, 0) \mapsto (0, 0) ( 0 , 0 ) ↦ ( 0 , 0 ) , ( 2 , 0 ) ↦ ( 6 , 0 ) (2, 0) \mapsto (6, 0) ( 2 , 0 ) ↦ ( 6 , 0 ) , ( 0 , 1 ) ↦ ( 1 , 2 ) (0, 1) \mapsto (1, 2) ( 0 , 1 ) ↦ ( 1 , 2 ) .
Image vertices: ( 0 , 0 ) (0, 0) ( 0 , 0 ) , ( 6 , 0 ) (6, 0) ( 6 , 0 ) , ( 1 , 2 ) (1, 2) ( 1 , 2 ) .
Original area: 1 2 × 2 × 1 = 1 \dfrac{1}{2} \times 2 \times 1 = 1 2 1 × 2 × 1 = 1 .
Image area using the determinant formula:
1 2 ∣ 6 ⋅ 2 − 0 ⋅ 1 ∣ = 1 2 × 12 = 6 \dfrac{1}{2}\left|6 \cdot 2 - 0 \cdot 1\right| = \dfrac{1}{2} \times 12 = 6 2 1 ∣ 6 ⋅ 2 − 0 ⋅ 1 ∣ = 2 1 × 12 = 6 .
det ( T ) = 6 \det(\mathbf{T}) = 6 det ( T ) = 6 , and ∣ det ( T ) ∣ × o r i g i n a l a r e a = 6 × 1 = 6 |\det(\mathbf{T})| \times \mathrm{original area} = 6 \times 1 = 6 ∣ det ( T ) ∣ × originalarea = 6 × 1 = 6 . ✓
Problem 5. Solve the system of equations using matrices:
{ 2 x + y − z = 3 x − y + 2 z = 1 3 x + 2 y + z = 10 \begin{cases} 2x + y - z = 3 \\ x - y + 2z = 1 \\ 3x + 2y + z = 10 \end{cases} ⎩ ⎨ ⎧ 2 x + y − z = 3 x − y + 2 z = 1 3 x + 2 y + z = 10
Hint Write as A x = b \mathbf{A}\mathbf{x} = \mathbf{b} Ax = b and compute x = A − 1 b \mathbf{x} = \mathbf{A}^{-1}\mathbf{b} x = A − 1 b .
Answer A = ( 2 1 − 1 1 − 1 2 3 2 1 ) \mathbf{A} = \begin{pmatrix} 2 & 1 & -1 \\ 1 & -1 & 2 \\ 3 & 2 & 1 \end{pmatrix} A = 2 1 3 1 − 1 2 − 1 2 1 ,
b = ( 3 1 10 ) \mathbf{b} = \begin{pmatrix} 3 \\ 1 \\ 10 \end{pmatrix} b = 3 1 10 .
det ( A ) = 2 ∣ − 1 2 2 1 ∣ − 1 ∣ 1 2 3 1 ∣ + ( − 1 ) ∣ 1 − 1 3 2 ∣ \det(\mathbf{A}) = 2\begin{vmatrix} -1 & 2 \\ 2 & 1 \end{vmatrix} - 1\begin{vmatrix} 1 & 2 \\ 3 & 1 \end{vmatrix} + (-1)\begin{vmatrix} 1 & -1 \\ 3 & 2 \end{vmatrix} det ( A ) = 2 − 1 2 2 1 − 1 1 3 2 1 + ( − 1 ) 1 3 − 1 2
= 2 ( − 5 ) − 1 ( − 5 ) − 1 ( 5 ) = − 10 + 5 − 5 = − 10 = 2(-5) - 1(-5) - 1(5) = -10 + 5 - 5 = -10 = 2 ( − 5 ) − 1 ( − 5 ) − 1 ( 5 ) = − 10 + 5 − 5 = − 10
Cofactors: C 11 = − 5 C_{11} = -5 C 11 = − 5 , C 12 = 5 C_{12} = 5 C 12 = 5 , C 13 = 5 C_{13} = 5 C 13 = 5 , C 21 = − 3 C_{21} = -3 C 21 = − 3 , C 22 = 5 C_{22} = 5 C 22 = 5 , C 23 = − 1 C_{23} = -1 C 23 = − 1 ,
C 31 = 1 C_{31} = 1 C 31 = 1 , C 32 = − 5 C_{32} = -5 C 32 = − 5 , C 33 = − 3 C_{33} = -3 C 33 = − 3 .
A − 1 = − 1 10 ( − 5 − 3 1 5 5 − 5 5 − 1 − 3 ) \mathbf{A}^{-1} = -\frac{1}{10}\begin{pmatrix} -5 & -3 & 1 \\ 5 & 5 & -5 \\ 5 & -1 & -3 \end{pmatrix} A − 1 = − 10 1 − 5 5 5 − 3 5 − 1 1 − 5 − 3
x = − 1 10 ( − 5 − 3 1 5 5 − 5 5 − 1 − 3 ) ( 3 1 10 ) = − 1 10 ( − 15 − 3 + 10 15 + 5 − 50 15 − 1 − 30 ) = − 1 10 ( − 8 − 30 − 16 ) = ( 4 / 5 3 8 / 5 ) \mathbf{x} = -\frac{1}{10}\begin{pmatrix} -5 & -3 & 1 \\ 5 & 5 & -5 \\ 5 & -1 & -3 \end{pmatrix}\begin{pmatrix} 3 \\ 1 \\ 10 \end{pmatrix} = -\frac{1}{10}\begin{pmatrix} -15 - 3 + 10 \\ 15 + 5 - 50 \\ 15 - 1 - 30 \end{pmatrix} = -\frac{1}{10}\begin{pmatrix} -8 \\ -30 \\ -16 \end{pmatrix} = \begin{pmatrix} 4/5 \\ 3 \\ 8/5 \end{pmatrix} x = − 10 1 − 5 5 5 − 3 5 − 1 1 − 5 − 3 3 1 10 = − 10 1 − 15 − 3 + 10 15 + 5 − 50 15 − 1 − 30 = − 10 1 − 8 − 30 − 16 = 4/5 3 8/5
So x = 4 5 x = \dfrac{4}{5} x = 5 4 , y = 3 y = 3 y = 3 , z = 8 5 z = \dfrac{8}{5} z = 5 8 .
Problem 6. Find the single 2 × 2 2 \times 2 2 × 2 matrix that represents an enlargement by scale factor 2
about the origin followed by a rotation of 90 ∘ 90^\circ 9 0 ∘ anticlockwise. Show that this is equivalent to
a single rotation of 90 ∘ 90^\circ 9 0 ∘ combined with an enlargement by factor 2.
Hint The enlargement matrix is 2 I 2\mathbf{I} 2 I and the rotation is R π / 2 \mathbf{R}_{\pi/2} R π /2 . Since
2 I 2\mathbf{I} 2 I commutes with all matrices, the order doesn't matter.
Answer Enlargement by 2: E = ( 2 0 0 2 ) \mathbf{E} = \begin{pmatrix} 2 & 0 \\ 0 & 2 \end{pmatrix} E = ( 2 0 0 2 ) .
Rotation by 90 ∘ 90^\circ 9 0 ∘ : R = ( 0 − 1 1 0 ) \mathbf{R} = \begin{pmatrix} 0 & -1 \\ 1 & 0 \end{pmatrix} R = ( 0 1 − 1 0 ) .
Enlargement then rotation:
R E = ( 0 − 1 1 0 ) ( 2 0 0 2 ) = ( 0 − 2 2 0 ) \mathbf{R}\mathbf{E} = \begin{pmatrix} 0 & -1 \\ 1 & 0 \end{pmatrix}\begin{pmatrix} 2 & 0 \\ 0 & 2 \end{pmatrix} = \begin{pmatrix} 0 & -2 \\ 2 & 0 \end{pmatrix} RE = ( 0 1 − 1 0 ) ( 2 0 0 2 ) = ( 0 2 − 2 0 ) .
Rotation then enlargement:
E R = ( 2 0 0 2 ) ( 0 − 1 1 0 ) = ( 0 − 2 2 0 ) \mathbf{E}\mathbf{R} = \begin{pmatrix} 2 & 0 \\ 0 & 2 \end{pmatrix}\begin{pmatrix} 0 & -1 \\ 1 & 0 \end{pmatrix} = \begin{pmatrix} 0 & -2 \\ 2 & 0 \end{pmatrix} ER = ( 2 0 0 2 ) ( 0 1 − 1 0 ) = ( 0 2 − 2 0 ) .
Both give the same result:
( 0 − 2 2 0 ) = 2 ( 0 − 1 1 0 ) \begin{pmatrix} 0 & -2 \\ 2 & 0 \end{pmatrix} = 2\begin{pmatrix} 0 & -1 \\ 1 & 0 \end{pmatrix} ( 0 2 − 2 0 ) = 2 ( 0 1 − 1 0 ) .
This is a rotation by 90 ∘ 90^\circ 9 0 ∘ combined with an enlargement by factor 2, and the order is
irrelevant because scalar multiples of the identity commute with all matrices.
Problem 7. AQA only. Find the eigenvalues and corresponding eigenvectors of
A = ( 5 4 1 2 ) \mathbf{A} = \begin{pmatrix} 5 & 4 \\ 1 & 2 \end{pmatrix} A = ( 5 1 4 2 ) , and write down a matrix P \mathbf{P} P
and diagonal matrix D \mathbf{D} D such that A = P D P − 1 \mathbf{A} = \mathbf{P}\mathbf{D}\mathbf{P}^{-1} A = PD P − 1 .
Hint Solve det ( A − λ I ) = 0 \det(\mathbf{A} - \lambda\mathbf{I}) = 0 det ( A − λ I ) = 0 for λ \lambda λ , then find eigenvectors.
Answer Characteristic equation:
det ( 5 − λ 4 1 2 − λ ) = ( 5 − λ ) ( 2 − λ ) − 4 = λ 2 − 7 λ + 6 = ( λ − 1 ) ( λ − 6 ) = 0 \det\begin{pmatrix} 5 - \lambda & 4 \\ 1 & 2 - \lambda \end{pmatrix} = (5 - \lambda)(2 - \lambda) - 4 = \lambda^2 - 7\lambda + 6 = (\lambda - 1)(\lambda - 6) = 0 det ( 5 − λ 1 4 2 − λ ) = ( 5 − λ ) ( 2 − λ ) − 4 = λ 2 − 7 λ + 6 = ( λ − 1 ) ( λ − 6 ) = 0 .
Eigenvalues: λ 1 = 6 \lambda_1 = 6 λ 1 = 6 , λ 2 = 1 \lambda_2 = 1 λ 2 = 1 .
For λ 1 = 6 \lambda_1 = 6 λ 1 = 6 :
( − 1 4 1 − 4 ) ( x y ) = 0 ⟹ − x + 4 y = 0 ⟹ x = 4 y \begin{pmatrix} -1 & 4 \\ 1 & -4 \end{pmatrix}\begin{pmatrix} x \\ y \end{pmatrix} = \mathbf{0} \implies -x + 4y = 0 \implies x = 4y ( − 1 1 4 − 4 ) ( x y ) = 0 ⟹ − x + 4 y = 0 ⟹ x = 4 y .
Eigenvector: v 1 = ( 4 1 ) \mathbf{v}_1 = \begin{pmatrix} 4 \\ 1 \end{pmatrix} v 1 = ( 4 1 ) .
For λ 2 = 1 \lambda_2 = 1 λ 2 = 1 :
( 4 4 1 1 ) ( x y ) = 0 ⟹ x + y = 0 ⟹ x = − y \begin{pmatrix} 4 & 4 \\ 1 & 1 \end{pmatrix}\begin{pmatrix} x \\ y \end{pmatrix} = \mathbf{0} \implies x + y = 0 \implies x = -y ( 4 1 4 1 ) ( x y ) = 0 ⟹ x + y = 0 ⟹ x = − y .
Eigenvector: v 2 = ( 1 − 1 ) \mathbf{v}_2 = \begin{pmatrix} 1 \\ -1 \end{pmatrix} v 2 = ( 1 − 1 ) .
P = ( 4 1 1 − 1 ) , D = ( 6 0 0 1 ) \mathbf{P} = \begin{pmatrix} 4 & 1 \\ 1 & -1 \end{pmatrix}, \quad \mathbf{D} = \begin{pmatrix} 6 & 0 \\ 0 & 1 \end{pmatrix} P = ( 4 1 1 − 1 ) , D = ( 6 0 0 1 )
Problem 8. AQA only. The matrix A = ( 3 − 2 1 0 ) \mathbf{A} = \begin{pmatrix} 3 & -2 \\ 1 & 0 \end{pmatrix} A = ( 3 1 − 2 0 ) has
eigenvalues 1 and 2. Use this to compute A 5 \mathbf{A}^5 A 5 without multiplying matrices five times.
Hint Diagonalise A = P D P − 1 \mathbf{A} = \mathbf{P}\mathbf{D}\mathbf{P}^{-1} A = PD P − 1 , then
A 5 = P D 5 P − 1 \mathbf{A}^5 = \mathbf{P}\mathbf{D}^5\mathbf{P}^{-1} A 5 = P D 5 P − 1 .
Answer For λ = 1 \lambda = 1 λ = 1 :
( 2 − 2 1 − 1 ) v = 0 ⟹ x = y \begin{pmatrix} 2 & -2 \\ 1 & -1 \end{pmatrix}\mathbf{v} = \mathbf{0} \implies x = y ( 2 1 − 2 − 1 ) v = 0 ⟹ x = y . Eigenvector
( 1 1 ) \begin{pmatrix} 1 \\ 1 \end{pmatrix} ( 1 1 ) .
For λ = 2 \lambda = 2 λ = 2 :
( 1 − 2 1 − 2 ) v = 0 ⟹ x = 2 y \begin{pmatrix} 1 & -2 \\ 1 & -2 \end{pmatrix}\mathbf{v} = \mathbf{0} \implies x = 2y ( 1 1 − 2 − 2 ) v = 0 ⟹ x = 2 y . Eigenvector
( 2 1 ) \begin{pmatrix} 2 \\ 1 \end{pmatrix} ( 2 1 ) .
P = ( 1 2 1 1 ) \mathbf{P} = \begin{pmatrix} 1 & 2 \\ 1 & 1 \end{pmatrix} P = ( 1 1 2 1 ) ,
D = ( 1 0 0 2 ) \mathbf{D} = \begin{pmatrix} 1 & 0 \\ 0 & 2 \end{pmatrix} D = ( 1 0 0 2 ) .
det ( P ) = 1 − 2 = − 1 \det(\mathbf{P}) = 1 - 2 = -1 det ( P ) = 1 − 2 = − 1 , so
P − 1 = ( − 1 2 1 − 1 ) \mathbf{P}^{-1} = \begin{pmatrix} -1 & 2 \\ 1 & -1 \end{pmatrix} P − 1 = ( − 1 1 2 − 1 ) .
D 5 = ( 1 0 0 32 ) \mathbf{D}^5 = \begin{pmatrix} 1 & 0 \\ 0 & 32 \end{pmatrix} D 5 = ( 1 0 0 32 ) .
A 5 = P D 5 P − 1 = ( 1 2 1 1 ) ( 1 0 0 32 ) ( − 1 2 1 − 1 ) = ( 1 64 1 32 ) ( − 1 2 1 − 1 ) = ( 63 − 62 31 − 30 ) \mathbf{A}^5 = \mathbf{P}\mathbf{D}^5\mathbf{P}^{-1} = \begin{pmatrix} 1 & 2 \\ 1 & 1 \end{pmatrix}\begin{pmatrix} 1 & 0 \\ 0 & 32 \end{pmatrix}\begin{pmatrix} -1 & 2 \\ 1 & -1 \end{pmatrix} = \begin{pmatrix} 1 & 64 \\ 1 & 32 \end{pmatrix}\begin{pmatrix} -1 & 2 \\ 1 & -1 \end{pmatrix} = \begin{pmatrix} 63 & -62 \\ 31 & -30 \end{pmatrix} A 5 = P D 5 P − 1 = ( 1 1 2 1 ) ( 1 0 0 32 ) ( − 1 1 2 − 1 ) = ( 1 1 64 32 ) ( − 1 1 2 − 1 ) = ( 63 31 − 62 − 30 )
Problem 9. Find the 3 × 3 3 \times 3 3 × 3 matrix representing a rotation of 90 ∘ 90^\circ 9 0 ∘ anticlockwise
about the z z z -axis. Verify that this matrix has determinant 1 and that it maps
( 1 0 0 ) \begin{pmatrix} 1 \\ 0 \\ 0 \end{pmatrix} 1 0 0 to ( 0 1 0 ) \begin{pmatrix} 0 \\ 1 \\ 0 \end{pmatrix} 0 1 0 .
Hint Use the standard formula for R z \mathbf{R}_z R z with θ = π / 2 \theta = \pi/2 θ = π /2 .
Answer R z = ( cos ◆ L B ◆ π ◆ R B ◆◆ L B ◆ 2 ◆ R B ◆ − sin ◆ L B ◆ π ◆ R B ◆◆ L B ◆ 2 ◆ R B ◆ 0 sin ◆ L B ◆ π ◆ R B ◆◆ L B ◆ 2 ◆ R B ◆ cos ◆ L B ◆ π ◆ R B ◆◆ L B ◆ 2 ◆ R B ◆ 0 0 0 1 ) = ( 0 − 1 0 1 0 0 0 0 1 ) \mathbf{R}_z = \begin{pmatrix} \cos\frac◆LB◆\pi◆RB◆◆LB◆2◆RB◆ & -\sin\frac◆LB◆\pi◆RB◆◆LB◆2◆RB◆ & 0 \\ \sin\frac◆LB◆\pi◆RB◆◆LB◆2◆RB◆ & \cos\frac◆LB◆\pi◆RB◆◆LB◆2◆RB◆ & 0 \\ 0 & 0 & 1 \end{pmatrix} = \begin{pmatrix} 0 & -1 & 0 \\ 1 & 0 & 0 \\ 0 & 0 & 1 \end{pmatrix} R z = cos L ◆ B ◆ π ◆ R B ◆◆ L B ◆2◆ R B ◆ sin L ◆ B ◆ π ◆ R B ◆◆ L B ◆2◆ R B ◆ 0 − sin L ◆ B ◆ π ◆ R B ◆◆ L B ◆2◆ R B ◆ cos L ◆ B ◆ π ◆ R B ◆◆ L B ◆2◆ R B ◆ 0 0 0 1 = 0 1 0 − 1 0 0 0 0 1
det ( R z ) = 0 ⋅ ( 0 − 0 ) − ( − 1 ) ( 1 − 0 ) + 0 = 1 \det(\mathbf{R}_z) = 0 \cdot (0 - 0) - (-1)(1 - 0) + 0 = 1 det ( R z ) = 0 ⋅ ( 0 − 0 ) − ( − 1 ) ( 1 − 0 ) + 0 = 1 . ✓
R z ( 1 0 0 ) = ( 0 1 0 ) ✓ \mathbf{R}_z\begin{pmatrix} 1 \\ 0 \\ 0 \end{pmatrix} = \begin{pmatrix} 0 \\ 1 \\ 0 \end{pmatrix} \quad \checkmark R z 1 0 0 = 0 1 0 ✓
The x x x -axis is correctly rotated to the y y y -axis by a 90 ∘ 90^\circ 9 0 ∘ anticlockwise rotation about z z z .
Problem 10. Prove that if A \mathbf{A} A and B \mathbf{B} B are non-singular n × n n \times n n × n matrices,
then ( A B ) − 1 = B − 1 A − 1 (\mathbf{AB})^{-1} = \mathbf{B}^{-1}\mathbf{A}^{-1} ( AB ) − 1 = B − 1 A − 1 .
Hint Show that B − 1 A − 1 \mathbf{B}^{-1}\mathbf{A}^{-1} B − 1 A − 1 satisfies the definition of the inverse of A B \mathbf{AB} AB ,
and invoke uniqueness.
Answer We need to show that ( A B ) ( B − 1 A − 1 ) = I (\mathbf{AB})(\mathbf{B}^{-1}\mathbf{A}^{-1}) = \mathbf{I} ( AB ) ( B − 1 A − 1 ) = I and
( B − 1 A − 1 ) ( A B ) = I (\mathbf{B}^{-1}\mathbf{A}^{-1})(\mathbf{AB}) = \mathbf{I} ( B − 1 A − 1 ) ( AB ) = I .
( A B ) ( B − 1 A − 1 ) = A ( B B − 1 ) A − 1 = A I A − 1 = A A − 1 = I (\mathbf{AB})(\mathbf{B}^{-1}\mathbf{A}^{-1}) = \mathbf{A}(\mathbf{B}\mathbf{B}^{-1})\mathbf{A}^{-1} = \mathbf{A}\mathbf{I}\mathbf{A}^{-1} = \mathbf{A}\mathbf{A}^{-1} = \mathbf{I} ( AB ) ( B − 1 A − 1 ) = A ( B B − 1 ) A − 1 = AI A − 1 = A A − 1 = I
( B − 1 A − 1 ) ( A B ) = B − 1 ( A − 1 A ) B = B − 1 I B = B − 1 B = I (\mathbf{B}^{-1}\mathbf{A}^{-1})(\mathbf{AB}) = \mathbf{B}^{-1}(\mathbf{A}^{-1}\mathbf{A})\mathbf{B} = \mathbf{B}^{-1}\mathbf{I}\mathbf{B} = \mathbf{B}^{-1}\mathbf{B} = \mathbf{I} ( B − 1 A − 1 ) ( AB ) = B − 1 ( A − 1 A ) B = B − 1 IB = B − 1 B = I
Since the inverse is unique, ( A B ) − 1 = B − 1 A − 1 (\mathbf{AB})^{-1} = \mathbf{B}^{-1}\mathbf{A}^{-1} ( AB ) − 1 = B − 1 A − 1 . □ \square □
Intuition. The order reverses, just like putting on and taking off socks and shoes. To undo "A
then B," you must undo B first, then undo A.
8. Advanced Worked Examples
Example 8.1: Diagonalisation of a 3×3 matrix
Problem. Diagonalise A = ( 2 1 0 0 2 0 0 1 3 ) \mathbf{A} = \begin{pmatrix}2&1&0\\0&2&0\\0&1&3\end{pmatrix} A = 2 0 0 1 2 1 0 0 3 .
Solution. Find eigenvalues: det ( A − λ I ) = 0 \det(\mathbf{A}-\lambda\mathbf{I}) = 0 det ( A − λ I ) = 0 .
det ( 2 − λ 1 0 0 2 − λ 0 0 1 3 − λ ) = ( 2 − λ ) 2 ( 3 − λ ) = 0 \det\begin{pmatrix}2-\lambda&1&0\\0&2-\lambda&0\\0&1&3-\lambda\end{pmatrix} = (2-\lambda)^2(3-\lambda) = 0 det 2 − λ 0 0 1 2 − λ 1 0 0 3 − λ = ( 2 − λ ) 2 ( 3 − λ ) = 0
λ 1 = 2 \lambda_1 = 2 λ 1 = 2 (repeated), λ 2 = 3 \lambda_2 = 3 λ 2 = 3 .
For λ = 2 \lambda = 2 λ = 2 : ( A − 2 I ) v = 0 (\mathbf{A}-2\mathbf{I})\mathbf{v} = \mathbf{0} ( A − 2 I ) v = 0 gives
( 0 1 0 0 0 0 0 1 1 ) v = 0 \begin{pmatrix}0&1&0\\0&0&0\\0&1&1\end{pmatrix}\mathbf{v} = \mathbf{0} 0 0 0 1 0 1 0 0 1 v = 0 , so v 2 = 0 v_2 = 0 v 2 = 0 and
v 3 = 0 v_3 = 0 v 3 = 0 , with v 1 v_1 v 1 free. Only one eigenvector: ( 1 , 0 , 0 ) (1,0,0) ( 1 , 0 , 0 ) . Since the geometric multiplicity (1) is
less than the algebraic multiplicity (2), A \mathbf{A} A is not diagonalisable .
Example 8.2: Finding A n \mathbf{A}^n A n using Cayley--Hamilton
Problem. For A = ( 1 2 0 3 ) \mathbf{A} = \begin{pmatrix}1&2\\0&3\end{pmatrix} A = ( 1 0 2 3 ) , find A 5 \mathbf{A}^5 A 5 .
Solution. By Cayley--Hamilton, A 2 − 4 A + 3 I = O \mathbf{A}^2 - 4\mathbf{A} + 3\mathbf{I} = \mathbf{O} A 2 − 4 A + 3 I = O , so
A 2 = 4 A − 3 I \mathbf{A}^2 = 4\mathbf{A} - 3\mathbf{I} A 2 = 4 A − 3 I .
A 3 = A ( 4 A − 3 I ) = 4 ( 4 A − 3 I ) − 3 A = 13 A − 12 I \mathbf{A}^3 = \mathbf{A}(4\mathbf{A}-3\mathbf{I}) = 4(4\mathbf{A}-3\mathbf{I}) - 3\mathbf{A} = 13\mathbf{A} - 12\mathbf{I} A 3 = A ( 4 A − 3 I ) = 4 ( 4 A − 3 I ) − 3 A = 13 A − 12 I .
A 4 = 13 ( 4 A − 3 I ) − 12 A = 40 A − 39 I \mathbf{A}^4 = 13(4\mathbf{A}-3\mathbf{I}) - 12\mathbf{A} = 40\mathbf{A} - 39\mathbf{I} A 4 = 13 ( 4 A − 3 I ) − 12 A = 40 A − 39 I .
A 5 = 40 ( 4 A − 3 I ) − 39 A = 121 A − 120 I \mathbf{A}^5 = 40(4\mathbf{A}-3\mathbf{I}) - 39\mathbf{A} = 121\mathbf{A} - 120\mathbf{I} A 5 = 40 ( 4 A − 3 I ) − 39 A = 121 A − 120 I .
= 121 ( 1 2 0 3 ) − 120 ( 1 0 0 1 ) = ( 121 242 0 363 ) − ( 120 0 0 120 ) = ( 1 242 0 243 ) = 121\begin{pmatrix}1&2\\0&3\end{pmatrix} - 120\begin{pmatrix}1&0\\0&1\end{pmatrix} = \begin{pmatrix}121&242\\0&363\end{pmatrix} - \begin{pmatrix}120&0\\0&120\end{pmatrix} = \boxed{\begin{pmatrix}1&242\\0&243\end{pmatrix}} = 121 ( 1 0 2 3 ) − 120 ( 1 0 0 1 ) = ( 121 0 242 363 ) − ( 120 0 0 120 ) = ( 1 0 242 243 )
Example 8.3: Invariant points and invariant lines
Problem. A = ( 3 1 0 2 ) \mathbf{A} = \begin{pmatrix}3&1\\0&2\end{pmatrix} A = ( 3 0 1 2 ) . Find all invariant points and
invariant lines of the transformation x ↦ A x \mathbf{x} \mapsto \mathbf{Ax} x ↦ Ax .
Solution. Invariant points:
A x = x ⟹ ( A − I ) x = 0 \mathbf{Ax} = \mathbf{x} \implies (\mathbf{A}-\mathbf{I})\mathbf{x} = \mathbf{0} Ax = x ⟹ ( A − I ) x = 0 .
( 2 1 0 1 ) x = 0 ⟹ x 2 = 0 , 2 x 1 = 0 \begin{pmatrix}2&1\\0&1\end{pmatrix}\mathbf{x} = \mathbf{0} \implies x_2 = 0, \; 2x_1 = 0 ( 2 0 1 1 ) x = 0 ⟹ x 2 = 0 , 2 x 1 = 0
Only the origin ( 0 , 0 ) (0,0) ( 0 , 0 ) is an invariant point.
Invariant lines through the origin: These are the eigenspaces. Eigenvalues:
( 3 − λ ) ( 2 − λ ) = 0 (3-\lambda)(2-\lambda) = 0 ( 3 − λ ) ( 2 − λ ) = 0 , so λ = 3 \lambda = 3 λ = 3 and λ = 2 \lambda = 2 λ = 2 .
For λ = 3 \lambda = 3 λ = 3 : ( 0 1 0 − 1 ) v = 0 ⟹ v 2 = 0 \begin{pmatrix}0&1\\0&-1\end{pmatrix}\mathbf{v} = \mathbf{0} \implies v_2 = 0 ( 0 0 1 − 1 ) v = 0 ⟹ v 2 = 0 .
Line: y = 0 y = 0 y = 0 (the x x x -axis).
For λ = 2 \lambda = 2 λ = 2 :
( 1 1 0 0 ) v = 0 ⟹ v 1 + v 2 = 0 \begin{pmatrix}1&1\\0&0\end{pmatrix}\mathbf{v} = \mathbf{0} \implies v_1 + v_2 = 0 ( 1 0 1 0 ) v = 0 ⟹ v 1 + v 2 = 0 . Line:
y = − x y = -x y = − x .
Example 8.4: Determinant as a scaling factor
Problem. The matrix T = ( 2 1 − 1 3 ) \mathbf{T} = \begin{pmatrix}2&1\\-1&3\end{pmatrix} T = ( 2 − 1 1 3 ) represents a
transformation. A triangle has vertices ( 0 , 0 ) (0,0) ( 0 , 0 ) , ( 1 , 0 ) (1,0) ( 1 , 0 ) , ( 0 , 1 ) (0,1) ( 0 , 1 ) . Find the area of its image.
Solution. det ( T ) = 6 − ( − 1 ) = 7 \det(\mathbf{T}) = 6 - (-1) = 7 det ( T ) = 6 − ( − 1 ) = 7 .
Original area = 1 2 = \dfrac{1}{2} = 2 1 . Image area
= ∣ det ( T ) ∣ × original area = 7 × 1 2 = 3.5 = |\det(\mathbf{T})| \times \text{original area} = 7 \times \dfrac{1}{2} = \boxed{3.5} = ∣ det ( T ) ∣ × original area = 7 × 2 1 = 3.5 .
Example 8.5: Commutator and non-commuting matrices
Problem. For A = ( 0 1 0 0 ) \mathbf{A} = \begin{pmatrix}0&1\\0&0\end{pmatrix} A = ( 0 0 1 0 ) and
B = ( 0 0 1 0 ) \mathbf{B} = \begin{pmatrix}0&0\\1&0\end{pmatrix} B = ( 0 1 0 0 ) , compute the commutator
[ A , B ] = A B − B A [\mathbf{A}, \mathbf{B}] = \mathbf{AB} - \mathbf{BA} [ A , B ] = AB − BA .
Solution.
A B = ( 0 1 0 0 ) ( 0 0 1 0 ) = ( 1 0 0 0 ) \mathbf{AB} = \begin{pmatrix}0&1\\0&0\end{pmatrix}\begin{pmatrix}0&0\\1&0\end{pmatrix} = \begin{pmatrix}1&0\\0&0\end{pmatrix} AB = ( 0 0 1 0 ) ( 0 1 0 0 ) = ( 1 0 0 0 ) .
B A = ( 0 0 1 0 ) ( 0 1 0 0 ) = ( 0 0 0 1 ) \mathbf{BA} = \begin{pmatrix}0&0\\1&0\end{pmatrix}\begin{pmatrix}0&1\\0&0\end{pmatrix} = \begin{pmatrix}0&0\\0&1\end{pmatrix} BA = ( 0 1 0 0 ) ( 0 0 1 0 ) = ( 0 0 0 1 ) .
[ A , B ] = ( 1 0 0 − 1 ) \boxed{[\mathbf{A},\mathbf{B}] = \begin{pmatrix}1&0\\0&-1\end{pmatrix}} [ A , B ] = ( 1 0 0 − 1 )
Since [ A , B ] ≠ O [\mathbf{A},\mathbf{B}] \neq \mathbf{O} [ A , B ] = O , A \mathbf{A} A and B \mathbf{B} B do not commute.
Problem. Transformation R R R is a reflection in the line y = x 3 y = x\sqrt{3} y = x 3 . Transformation S S S is a
rotation by 90 ° 90° 90° anticlockwise about the origin. Find the matrix representing R S RS R S .
Solution. The line y = x 3 y = x\sqrt{3} y = x 3 makes angle 60 ° 60° 60° with the x x x -axis.
R = ( cos 120 ° sin 120 ° sin 120 ° − cos 120 ° ) = ( − 1 2 ◆ L B ◆ 3 ◆ R B ◆◆ L B ◆ 2 ◆ R B ◆ ◆ L B ◆ 3 ◆ R B ◆◆ L B ◆ 2 ◆ R B ◆ 1 2 ) R = \begin{pmatrix}\cos 120°&\sin 120°\\\sin 120°&-\cos 120°\end{pmatrix} = \begin{pmatrix}-\frac{1}{2}&\frac◆LB◆\sqrt{3}◆RB◆◆LB◆2◆RB◆\\\frac◆LB◆\sqrt{3}◆RB◆◆LB◆2◆RB◆&\frac{1}{2}\end{pmatrix} R = ( cos 120° sin 120° sin 120° − cos 120° ) = ( − 2 1 L ◆ B ◆ 3 ◆ R B ◆◆ L B ◆2◆ R B ◆ L ◆ B ◆ 3 ◆ R B ◆◆ L B ◆2◆ R B ◆ 2 1 ) .
S = ( 0 − 1 1 0 ) S = \begin{pmatrix}0&-1\\1&0\end{pmatrix} S = ( 0 1 − 1 0 ) .
R S = ( − 1 2 ◆ L B ◆ 3 ◆ R B ◆◆ L B ◆ 2 ◆ R B ◆ ◆ L B ◆ 3 ◆ R B ◆◆ L B ◆ 2 ◆ R B ◆ 1 2 ) ( 0 − 1 1 0 ) = ( ◆ L B ◆ 3 ◆ R B ◆◆ L B ◆ 2 ◆ R B ◆ 1 2 1 2 − ◆ L B ◆ 3 ◆ R B ◆◆ L B ◆ 2 ◆ R B ◆ ) RS = \begin{pmatrix}-\frac{1}{2}&\frac◆LB◆\sqrt{3}◆RB◆◆LB◆2◆RB◆\\\frac◆LB◆\sqrt{3}◆RB◆◆LB◆2◆RB◆&\frac{1}{2}\end{pmatrix}\begin{pmatrix}0&-1\\1&0\end{pmatrix} = \begin{pmatrix}\frac◆LB◆\sqrt{3}◆RB◆◆LB◆2◆RB◆&\frac{1}{2}\\\frac{1}{2}&-\frac◆LB◆\sqrt{3}◆RB◆◆LB◆2◆RB◆\end{pmatrix} R S = ( − 2 1 L ◆ B ◆ 3 ◆ R B ◆◆ L B ◆2◆ R B ◆ L ◆ B ◆ 3 ◆ R B ◆◆ L B ◆2◆ R B ◆ 2 1 ) ( 0 1 − 1 0 ) = ( L ◆ B ◆ 3 ◆ R B ◆◆ L B ◆2◆ R B ◆ 2 1 2 1 − L ◆ B ◆ 3 ◆ R B ◆◆ L B ◆2◆ R B ◆ ) .
det ( R S ) = − 3 4 − 1 4 = − 1 \det(RS) = -\dfrac{3}{4} - \dfrac{1}{4} = -1 det ( R S ) = − 4 3 − 4 1 = − 1 and tr ( R S ) = 0 \text{tr}(RS) = 0 tr ( R S ) = 0 , confirming this is a
reflection.
Example 8.7: Finding the inverse of a 3×3 matrix
Problem. Find A − 1 \mathbf{A}^{-1} A − 1 where
A = ( 1 0 2 0 1 − 1 1 1 0 ) \mathbf{A} = \begin{pmatrix}1&0&2\\0&1&-1\\1&1&0\end{pmatrix} A = 1 0 1 0 1 1 2 − 1 0 .
Solution. det ( A ) = 1 ( 0 + 1 ) − 0 + 2 ( 0 − 1 ) = 1 − 2 = − 1 ≠ 0 \det(\mathbf{A}) = 1(0+1) - 0 + 2(0-1) = 1 - 2 = -1 \neq 0 det ( A ) = 1 ( 0 + 1 ) − 0 + 2 ( 0 − 1 ) = 1 − 2 = − 1 = 0 .
A − 1 = 1 − 1 ( 1 2 − 2 − 1 − 2 1 − 1 − 1 1 ) = ( − 1 − 2 2 1 2 − 1 1 1 − 1 ) \mathbf{A}^{-1} = \dfrac{1}{-1}\begin{pmatrix}1&2&-2\\-1&-2&1\\-1&-1&1\end{pmatrix} = \boxed{\begin{pmatrix}-1&-2&2\\1&2&-1\\1&1&-1\end{pmatrix}} A − 1 = − 1 1 1 − 1 − 1 2 − 2 − 1 − 2 1 1 = − 1 1 1 − 2 2 1 2 − 1 − 1
9. Common Pitfalls
Pitfall Correct Approach Assuming all matrices are diagonalisable Check geometric multiplicity equals algebraic multiplicity for each eigenvalue Forgetting that ( A B ) − 1 = B − 1 A − 1 (\mathbf{AB})^{-1} = \mathbf{B}^{-1}\mathbf{A}^{-1} ( AB ) − 1 = B − 1 A − 1 The order reverses Computing det ( A + B ) = det ( A ) + det ( B ) \det(\mathbf{A}+\mathbf{B}) = \det(\mathbf{A})+\det(\mathbf{B}) det ( A + B ) = det ( A ) + det ( B ) In general, det ( A + B ) ≠ det ( A ) + det ( B ) \det(\mathbf{A}+\mathbf{B}) \neq \det(\mathbf{A})+\det(\mathbf{B}) det ( A + B ) = det ( A ) + det ( B ) Mixing up row and column operations Column operations change the determinant differently from row operations
10. Additional Exam-Style Questions
Question 8
Find the eigenvalues and eigenvectors of A = ( 4 − 1 2 1 ) \mathbf{A} = \begin{pmatrix}4&-1\\2&1\end{pmatrix} A = ( 4 2 − 1 1 ) . Hence
write down P \mathbf{P} P and D \mathbf{D} D such that
P − 1 A P = D \mathbf{P}^{-1}\mathbf{A}\mathbf{P} = \mathbf{D} P − 1 AP = D .
Solution det ( A − λ I ) = ( 4 − λ ) ( 1 − λ ) + 2 = λ 2 − 5 λ + 6 = 0 \det(\mathbf{A}-\lambda\mathbf{I}) = (4-\lambda)(1-\lambda)+2 = \lambda^2-5\lambda+6 = 0 det ( A − λ I ) = ( 4 − λ ) ( 1 − λ ) + 2 = λ 2 − 5 λ + 6 = 0 .
λ = 2 , 3 \lambda = 2, 3 λ = 2 , 3 .
λ = 2 \lambda = 2 λ = 2 :
( 2 − 1 2 − 1 ) v = 0 ⟹ 2 v 1 − v 2 = 0 ⟹ v = ( 1 , 2 ) \begin{pmatrix}2&-1\\2&-1\end{pmatrix}\mathbf{v}=\mathbf{0} \implies 2v_1-v_2=0 \implies \mathbf{v}=(1,2) ( 2 2 − 1 − 1 ) v = 0 ⟹ 2 v 1 − v 2 = 0 ⟹ v = ( 1 , 2 ) .
λ = 3 \lambda = 3 λ = 3 :
( 1 − 1 2 − 2 ) v = 0 ⟹ v 1 = v 2 ⟹ v = ( 1 , 1 ) \begin{pmatrix}1&-1\\2&-2\end{pmatrix}\mathbf{v}=\mathbf{0} \implies v_1=v_2 \implies \mathbf{v}=(1,1) ( 1 2 − 1 − 2 ) v = 0 ⟹ v 1 = v 2 ⟹ v = ( 1 , 1 ) .
P = ( 1 1 2 1 ) \mathbf{P} = \begin{pmatrix}1&1\\2&1\end{pmatrix} P = ( 1 2 1 1 ) ,
D = ( 2 0 0 3 ) \mathbf{D} = \begin{pmatrix}2&0\\0&3\end{pmatrix} D = ( 2 0 0 3 ) .
Question 9
Prove that if λ \lambda λ is an eigenvalue of A \mathbf{A} A with eigenvector v \mathbf{v} v , then
λ 2 \lambda^2 λ 2 is an eigenvalue of A 2 \mathbf{A}^2 A 2 with the same eigenvector.
Solution A v = λ v \mathbf{Av} = \lambda\mathbf{v} Av = λ v .
A 2 v = A ( A v ) = A ( λ v ) = λ ( A v ) = λ ( λ v ) = λ 2 v \mathbf{A}^2\mathbf{v} = \mathbf{A}(\mathbf{Av}) = \mathbf{A}(\lambda\mathbf{v}) = \lambda(\mathbf{Av}) = \lambda(\lambda\mathbf{v}) = \lambda^2\mathbf{v} A 2 v = A ( Av ) = A ( λ v ) = λ ( Av ) = λ ( λ v ) = λ 2 v .
Therefore λ 2 \lambda^2 λ 2 is an eigenvalue of A 2 \mathbf{A}^2 A 2 with eigenvector v \mathbf{v} v .
■ \blacksquare ■
Question 10
The transformation represented by M = ( a b c d ) \mathbf{M} = \begin{pmatrix}a&b\\c&d\end{pmatrix} M = ( a c b d ) maps the unit
square to a parallelogram of area 6. Given a + d = 5 a+d = 5 a + d = 5 and a d − b c = 6 ad-bc = 6 a d − b c = 6 , find the eigenvalues of
M \mathbf{M} M .
Solution The characteristic equation:
λ 2 − ( a + d ) λ + det ( M ) = λ 2 − 5 λ + 6 = 0 \lambda^2 - (a+d)\lambda + \det(\mathbf{M}) = \lambda^2 - 5\lambda + 6 = 0 λ 2 − ( a + d ) λ + det ( M ) = λ 2 − 5 λ + 6 = 0 .
( λ − 2 ) ( λ − 3 ) = 0 (\lambda-2)(\lambda-3) = 0 ( λ − 2 ) ( λ − 3 ) = 0 .
λ = 2 and λ = 3 \boxed{\lambda = 2 \text{ and } \lambda = 3} λ = 2 and λ = 3
:::
11. Connections to Other Topics
11.1 Matrices and complex numbers
Complex eigenvalues lead to rotation-scaling transformations. See
Complex Numbers .
11.2 Matrices and vectors
The cross product can be written as a matrix multiplication. See
Vectors in 3D .
11.3 Matrices and further algebra
Cayley--Hamilton connects matrices to polynomial algebra. See
Further Algebra .
12. Key Results Summary
| Result | Formula/Condition |
| -------------------- | --------------------------------------------------------------------------- | ---------------- | ------------------------------------ |
| Invertibility | A \mathbf{A} A is invertible ⟺ \iff ⟺ det ( A ) ≠ 0 \det(\mathbf{A}) \neq 0 det ( A ) = 0 |
| ( A B ) − 1 (\mathbf{AB})^{-1} ( AB ) − 1 | B − 1 A − 1 \mathbf{B}^{-1}\mathbf{A}^{-1} B − 1 A − 1 |
| ( A B ) T (\mathbf{AB})^T ( AB ) T | B T A T \mathbf{B}^T\mathbf{A}^T B T A T |
| det ( A B ) \det(\mathbf{AB}) det ( AB ) | det ( A ) det ( B ) \det(\mathbf{A})\det(\mathbf{B}) det ( A ) det ( B ) |
| Trace of product | tr ( A B ) = tr ( B A ) \text{tr}(\mathbf{AB}) = \text{tr}(\mathbf{BA}) tr ( AB ) = tr ( BA ) |
| Cayley--Hamilton | A \mathbf{A} A satisfies its own characteristic equation |
| Diagonalisability | All eigenvalues must have geometric multiplicity = = = algebraic multiplicity |
| Area scaling | ∣ det ( T ) ∣ × | \det(\mathbf{T}) | \times ∣ det ( T ) ∣ × original area = = = image area |
13. Further Exam-Style Questions
Question 11
Find the matrix representing a stretch of scale factor 3 parallel to the y y y -axis followed by a
reflection in the x x x -axis.
Solution Stretch: S = ( 1 0 0 3 ) S = \begin{pmatrix}1&0\\0&3\end{pmatrix} S = ( 1 0 0 3 ) . Reflection:
R = ( 1 0 0 − 1 ) R = \begin{pmatrix}1&0\\0&-1\end{pmatrix} R = ( 1 0 0 − 1 ) .
Combined:
R S = ( 1 0 0 − 1 ) ( 1 0 0 3 ) = ( 1 0 0 − 3 ) RS = \begin{pmatrix}1&0\\0&-1\end{pmatrix}\begin{pmatrix}1&0\\0&3\end{pmatrix} = \begin{pmatrix}1&0\\0&-3\end{pmatrix} R S = ( 1 0 0 − 1 ) ( 1 0 0 3 ) = ( 1 0 0 − 3 ) .
( 1 0 0 − 3 ) \boxed{\begin{pmatrix}1&0\\0&-3\end{pmatrix}} ( 1 0 0 − 3 )
Question 12
Prove that det ( A T ) = det ( A ) \det(\mathbf{A}^T) = \det(\mathbf{A}) det ( A T ) = det ( A ) for any square matrix A \mathbf{A} A .
Solution The determinant can be computed by cofactor expansion along any row or column. Expanding
det ( A ) \det(\mathbf{A}) det ( A ) along row i i i and det ( A T ) \det(\mathbf{A}^T) det ( A T ) along column i i i (which is row i i i of
A \mathbf{A} A ) gives the same expression, since the cofactors are the same.
Formally: this follows from the permutation definition of the determinant. ■ \blacksquare ■
14. Advanced Topics
14.1 Eigenvalues and the characteristic polynomial
The characteristic polynomial of A \mathbf{A} A is p ( λ ) = det ( A − λ I ) p(\lambda) = \det(\mathbf{A}-\lambda\mathbf{I}) p ( λ ) = det ( A − λ I ) .
Properties:
The sum of eigenvalues equals the trace: ∑ λ i = t r ( A ) \sum \lambda_i = \mathrm{tr}(\mathbf{A}) ∑ λ i = tr ( A )
The product of eigenvalues equals the determinant: ∏ λ i = det ( A ) \prod \lambda_i = \det(\mathbf{A}) ∏ λ i = det ( A )
A \mathbf{A} A is invertible iff no eigenvalue is zero
Not all matrices are diagonalisable. The Jordan normal form is a generalisation where the diagonal
matrix D \mathbf{D} D may have 1s on the superdiagonal (Jordan blocks).
For example, if A \mathbf{A} A has a repeated eigenvalue λ \lambda λ with only one eigenvector:
P − 1 A P = ( λ 1 0 λ ) \mathbf{P}^{-1}\mathbf{AP} = \begin{pmatrix}\lambda&1\\0&\lambda\end{pmatrix} P − 1 AP = ( λ 0 1 λ )
14.3 Orthogonal diagonalisation
A real symmetric matrix A \mathbf{A} A can always be diagonalised by an orthogonal matrix:
A = Q D Q T \mathbf{A} = \mathbf{Q}\mathbf{D}\mathbf{Q}^T A = QD Q T where Q T = Q − 1 \mathbf{Q}^T = \mathbf{Q}^{-1} Q T = Q − 1 .
The Spectral Theorem states that A = ∑ i = 1 n λ i q i q i T \mathbf{A} = \sum_{i=1}^{n} \lambda_i \mathbf{q}_i\mathbf{q}_i^T A = ∑ i = 1 n λ i q i q i T
where λ i \lambda_i λ i are eigenvalues and q i \mathbf{q}_i q i are orthonormal eigenvectors.
14.4 Matrix norms
The Frobenius norm:
∥ A ∥ F = ◆ L B ◆ ∑ i , j a i j 2 ◆ R B ◆ = ◆ L B ◆ t r ( A T A ) ◆ R B ◆ \|\mathbf{A}\|_F = \sqrt◆LB◆\sum_{i,j} a_{ij}^2◆RB◆ = \sqrt◆LB◆\mathrm{tr}(\mathbf{A}^T\mathbf{A})◆RB◆ ∥ A ∥ F = ◆ L B ◆ ∑ i , j a ij 2 ◆ R B ◆ = ◆ L B ◆ tr ( A T A ) ◆ R B ◆ .
The spectral norm: ∥ A ∥ 2 = σ max \|\mathbf{A}\|_2 = \sigma_{\max} ∥ A ∥ 2 = σ m a x (largest singular value).
15. Further Exam-Style Questions
Question 14
Find the eigenvalues and eigenvectors of
A = ( 1 1 1 1 1 1 1 1 1 ) \mathbf{A} = \begin{pmatrix}1&1&1\\1&1&1\\1&1&1\end{pmatrix} A = 1 1 1 1 1 1 1 1 1 .
Solution det ( A − λ I ) = ( 1 − λ ) 3 − 3 ( 1 − λ ) − 2 + 3 ( 1 − λ ) = ( 1 − λ ) 3 − 3 ( 1 − λ ) + 3 ( 1 − λ ) − 2 \det(\mathbf{A}-\lambda\mathbf{I}) = (1-\lambda)^3 - 3(1-\lambda) - 2 + 3(1-\lambda) = (1-\lambda)^3 - 3(1-\lambda) + 3(1-\lambda) - 2 det ( A − λ I ) = ( 1 − λ ) 3 − 3 ( 1 − λ ) − 2 + 3 ( 1 − λ ) = ( 1 − λ ) 3 − 3 ( 1 − λ ) + 3 ( 1 − λ ) − 2 .
Wait, let me compute directly.
det ( 1 − λ 1 1 1 1 − λ 1 1 1 1 − λ ) \det\begin{pmatrix}1-\lambda&1&1\\1&1-\lambda&1\\1&1&1-\lambda\end{pmatrix} det 1 − λ 1 1 1 1 − λ 1 1 1 1 − λ .
= ( 1 − λ ) [ ( 1 − λ ) 2 − 1 ] − 1 [ ( 1 − λ ) − 1 ] + 1 [ 1 − ( 1 − λ ) ] = (1-\lambda)[(1-\lambda)^2-1] - 1[(1-\lambda)-1] + 1[1-(1-\lambda)] = ( 1 − λ ) [( 1 − λ ) 2 − 1 ] − 1 [( 1 − λ ) − 1 ] + 1 [ 1 − ( 1 − λ )]
= ( 1 − λ ) [ 1 − 2 λ + λ 2 − 1 ] + λ + λ = ( 1 − λ ) ( λ 2 − 2 λ ) + 2 λ = (1-\lambda)[1-2\lambda+\lambda^2-1] + \lambda + \lambda = (1-\lambda)(\lambda^2-2\lambda) + 2\lambda = ( 1 − λ ) [ 1 − 2 λ + λ 2 − 1 ] + λ + λ = ( 1 − λ ) ( λ 2 − 2 λ ) + 2 λ
= ( 1 − λ ) λ ( λ − 2 ) + 2 λ = λ [ ( 1 − λ ) ( λ − 2 ) + 2 ] = λ [ λ − 2 − λ 2 + 2 λ + 2 ] = λ [ − λ 2 + 3 λ ] = λ 2 ( 3 − λ ) = (1-\lambda)\lambda(\lambda-2) + 2\lambda = \lambda[(1-\lambda)(\lambda-2)+2] = \lambda[\lambda-2-\lambda^2+2\lambda+2] = \lambda[-\lambda^2+3\lambda] = \lambda^2(3-\lambda) = ( 1 − λ ) λ ( λ − 2 ) + 2 λ = λ [( 1 − λ ) ( λ − 2 ) + 2 ] = λ [ λ − 2 − λ 2 + 2 λ + 2 ] = λ [ − λ 2 + 3 λ ] = λ 2 ( 3 − λ ) .
Eigenvalues: λ = 0 \lambda = 0 λ = 0 (double) and λ = 3 \lambda = 3 λ = 3 .
λ = 0 \lambda = 0 λ = 0 :
( 1 1 1 1 1 1 1 1 1 ) v = 0 ⟹ v 1 + v 2 + v 3 = 0 \begin{pmatrix}1&1&1\\1&1&1\\1&1&1\end{pmatrix}\mathbf{v}=\mathbf{0} \implies v_1+v_2+v_3=0 1 1 1 1 1 1 1 1 1 v = 0 ⟹ v 1 + v 2 + v 3 = 0 . Two
independent eigenvectors: ( 1 , − 1 , 0 ) (1,-1,0) ( 1 , − 1 , 0 ) and ( 1 , 0 , − 1 ) (1,0,-1) ( 1 , 0 , − 1 ) .
λ = 3 \lambda = 3 λ = 3 :
( − 2 1 1 1 − 2 1 1 1 − 2 ) v = 0 ⟹ v 1 = v 2 = v 3 \begin{pmatrix}-2&1&1\\1&-2&1\\1&1&-2\end{pmatrix}\mathbf{v}=\mathbf{0} \implies v_1=v_2=v_3 − 2 1 1 1 − 2 1 1 1 − 2 v = 0 ⟹ v 1 = v 2 = v 3 .
Eigenvector: ( 1 , 1 , 1 ) (1,1,1) ( 1 , 1 , 1 ) .
Question 15
Prove that similar matrices have the same eigenvalues.
Solution If B = P − 1 A P \mathbf{B} = \mathbf{P}^{-1}\mathbf{AP} B = P − 1 AP , then:
det ( B − λ I ) = det ( P − 1 A P − λ I ) = det ( P − 1 ( A − λ I ) P ) \det(\mathbf{B}-\lambda\mathbf{I}) = \det(\mathbf{P}^{-1}\mathbf{AP}-\lambda\mathbf{I}) = \det(\mathbf{P}^{-1}(\mathbf{A}-\lambda\mathbf{I})\mathbf{P}) det ( B − λ I ) = det ( P − 1 AP − λ I ) = det ( P − 1 ( A − λ I ) P )
= det ( P − 1 ) det ( A − λ I ) det ( P ) = det ( A − λ I ) = \det(\mathbf{P}^{-1})\det(\mathbf{A}-\lambda\mathbf{I})\det(\mathbf{P}) = \det(\mathbf{A}-\lambda\mathbf{I}) = det ( P − 1 ) det ( A − λ I ) det ( P ) = det ( A − λ I ) .
Since the characteristic polynomials are identical, the eigenvalues are the same. ■ \blacksquare ■
16. Further Advanced Topics
16.1 LU decomposition
Any square matrix A \mathbf{A} A can be decomposed as A = L U \mathbf{A} = \mathbf{L}\mathbf{U} A = LU where
L \mathbf{L} L is lower triangular and U \mathbf{U} U is upper triangular. This is used for efficient
numerical solution of systems A x = b \mathbf{Ax} = \mathbf{b} Ax = b .
16.2 The Cayley--Hamilton theorem — applications
Since A \mathbf{A} A satisfies p ( A ) = O p(\mathbf{A}) = \mathbf{O} p ( A ) = O where p p p is the characteristic
polynomial:
A − 1 \mathbf{A}^{-1} A − 1 can be computed from A n \mathbf{A}^n A n terms
A n \mathbf{A}^n A n for large n n n can be reduced using the recurrence
16.3 Singular Value Decomposition (SVD)
Any m × n m \times n m × n matrix A \mathbf{A} A can be written as
A = U Σ V T \mathbf{A} = \mathbf{U}\boldsymbol{\Sigma}\mathbf{V}^T A = U Σ V T where U \mathbf{U} U and V \mathbf{V} V are
orthogonal and Σ \boldsymbol{\Sigma} Σ is diagonal with non-negative singular values.
16.4 Positive definite matrices
A symmetric matrix A \mathbf{A} A is positive definite if x T A x > 0 \mathbf{x}^T\mathbf{A}\mathbf{x} > 0 x T Ax > 0 for
all x ≠ 0 \mathbf{x} \neq \mathbf{0} x = 0 .
Equivalent conditions: all eigenvalues positive, all leading principal minors positive.
17. Further Exam-Style Questions
Question 16
Find det ( A ) \det(\mathbf{A}) det ( A ) where A = ( 1 2 3 4 5 6 7 8 0 ) \mathbf{A} = \begin{pmatrix}1&2&3\\4&5&6\\7&8&0\end{pmatrix} A = 1 4 7 2 5 8 3 6 0 .
Solution det ( A ) = 1 ( 0 − 48 ) − 2 ( 0 − 42 ) + 3 ( 32 − 35 ) = − 48 + 84 − 9 = 27 \det(\mathbf{A}) = 1(0-48) - 2(0-42) + 3(32-35) = -48 + 84 - 9 = \boxed{27} det ( A ) = 1 ( 0 − 48 ) − 2 ( 0 − 42 ) + 3 ( 32 − 35 ) = − 48 + 84 − 9 = 27 .
Question 17
Prove that the trace of a matrix equals the sum of its eigenvalues (with multiplicity).
Solution The characteristic polynomial is
p ( λ ) = ( − 1 ) n [ λ n − ( tr A ) λ n − 1 + ⋯ + ( − 1 ) n det ( A ) ] p(\lambda) = (-1)^n[\lambda^n - (\text{tr}\,\mathbf{A})\lambda^{n-1} + \cdots + (-1)^n\det(\mathbf{A})] p ( λ ) = ( − 1 ) n [ λ n − ( tr A ) λ n − 1 + ⋯ + ( − 1 ) n det ( A )] .
By Vieta's formulae, the coefficient of λ n − 1 \lambda^{n-1} λ n − 1 equals
− ( λ 1 + λ 2 + ⋯ + λ n ) -(\lambda_1 + \lambda_2 + \cdots + \lambda_n) − ( λ 1 + λ 2 + ⋯ + λ n ) .
Therefore tr A = λ 1 + λ 2 + ⋯ + λ n \text{tr}\,\mathbf{A} = \lambda_1 + \lambda_2 + \cdots + \lambda_n tr A = λ 1 + λ 2 + ⋯ + λ n . ■ \blacksquare ■