Linear algebra - Matrices Linear Systems and Determinants
Understand how matrices represent linear maps, how Gaussian elimination solves linear systems, and how determinants determine invertibility and enable Cramer’s rule.
Summary
Read Summary
Flashcards
Save Flashcards
Quiz
Take Quiz
Quick Practice
What is a vector in a finite-dimensional vector space identified with once a basis is chosen?
1 of 13
Summary
Matrices and Linear Systems: A Foundation for Linear Algebra
Introduction
Matrices are one of the most powerful tools in linear algebra. They provide a concrete way to represent abstract linear maps and vector spaces, transforming geometric intuition into computational algorithms. In this guide, we'll explore how matrices encode vector transformations, how we solve systems of linear equations using matrices, and how the determinant tells us fundamental properties of these transformations.
The journey from abstract linear algebra to practical computation relies entirely on the insights in this section.
Part 1: Matrices and Vector Representation
Why Matrices Represent Vectors and Linear Maps
When we have a finite-dimensional vector space, we need a way to work with its vectors concretely. A basis gives us this concrete representation. Once we fix a basis, every vector in the space can be uniquely written as a linear combination of basis vectors. The coordinates of a vector relative to this basis are the coefficients in this linear combination.
These coordinates are naturally organized into a column matrix (or column vector): a matrix with one column and $n$ rows, where $n$ is the dimension of the space.
Example: In $\mathbb{R}^3$ with the standard basis $\{(1,0,0), (0,1,0), (0,0,1)\}$, the vector $(5, -2, 3)$ is represented as the column matrix:
$$\begin{bmatrix} 5 \\ -2 \\ 3 \end{bmatrix}$$
Representing Linear Maps as Matrices
Now here's the key insight: a linear map is completely determined by where it sends the basis vectors. Why? Because any vector is a linear combination of basis vectors, and the linearity of the map tells us exactly how to compute where that vector goes.
To represent a linear map $T: V \to W$ as a matrix:
Take each basis vector $\mathbf{e}i$ of the domain $V$
Compute $T(\mathbf{e}i)$ and express it as coordinates in the basis of $W$
Write these coordinate vectors as the columns of a matrix
Example: Consider the linear map $T: \mathbb{R}^2 \to \mathbb{R}^2$ that rotates vectors counterclockwise by 90°. Using the standard basis:
$T\left(\begin{bmatrix} 1 \\ 0 \end{bmatrix}\right) = \begin{bmatrix} 0 \\ 1 \end{bmatrix}$
$T\left(\begin{bmatrix} 0 \\ 1 \end{bmatrix}\right) = \begin{bmatrix} -1 \\ 0 \end{bmatrix}$
So the matrix representation is:
$$[T] = \begin{bmatrix} 0 & -1 \\ 1 & 0 \end{bmatrix}$$
Matrix-Vector Multiplication as Application of the Linear Map
When we multiply a matrix $A$ (representing a linear map $T$) by a column vector $\mathbf{v}$ (representing a vector $v$), we get:
$$A\mathbf{v} = \text{the coordinate vector of } T(v)$$
This is exactly the formula for applying the linear map. The matrix-vector product isn't just an arbitrary operation—it's the concrete computation of the linear transformation.
Example: Using the rotation matrix above:
$$\begin{bmatrix} 0 & -1 \\ 1 & 0 \end{bmatrix} \begin{bmatrix} 3 \\ 2 \end{bmatrix} = \begin{bmatrix} 0 \cdot 3 + (-1) \cdot 2 \\ 1 \cdot 3 + 0 \cdot 2 \end{bmatrix} = \begin{bmatrix} -2 \\ 3 \end{bmatrix}$$
This correctly rotates $(3, 2)$ to $(-2, 3)$.
Part 2: Matrix Multiplication and Composition
The Deep Connection: Composition of Maps and Matrix Multiplication
Suppose we have two linear maps: $T: U \to V$ and $S: V \to W$. We can compose them to get $S \circ T: U \to W$.
Here's the remarkable fact: if $T$ is represented by matrix $B$ and $S$ is represented by matrix $A$, then $S \circ T$ is represented by the matrix product $AB$.
In other words: matrix multiplication is the computational manifestation of function composition.
This explains why matrix multiplication is defined the way it is. The definition isn't arbitrary—it's the only definition that makes this correspondence work!
$$A(B\mathbf{v}) = (AB)\mathbf{v}$$
Example: Let's compose two transformations in $\mathbb{R}^2$:
First apply rotation by 90°: $B = \begin{bmatrix} 0 & -1 \\ 1 & 0 \end{bmatrix}$
Then scale by 2: $A = \begin{bmatrix} 2 & 0 \\ 0 & 2 \end{bmatrix}$
The composition $A \circ B$ should scale all vectors by 2, then rotate by 90°. Let's compute $AB$:
$$AB = \begin{bmatrix} 2 & 0 \\ 0 & 2 \end{bmatrix} \begin{bmatrix} 0 & -1 \\ 1 & 0 \end{bmatrix} = \begin{bmatrix} 0 & -2 \\ 2 & 0 \end{bmatrix}$$
This represents "rotate by 90°, then scale by 2" (or equivalently, scale by 2 then rotate—these compose to give the same result for these particular maps).
Part 3: Similarity and Change of Basis
The Same Linear Map, Different Bases
Here's a subtle but important point: the same linear map can be represented by different matrices, depending on which basis we choose for the domain and codomain.
Why does this matter? In practice, we might choose different bases for computational convenience. For instance, if we're studying a transformation, there might be special basis vectors called eigenvectors that the transformation just scales. In the eigenvector basis, the matrix representation becomes much simpler.
Similarity: When Matrices Represent the Same Endomorphism
An endomorphism is a linear map from a vector space to itself. When the domain and codomain are the same, changing basis gives us a special relationship.
If $A$ and $B$ are two matrices representing the same endomorphism in different bases, they are similar: there exists an invertible matrix $P$ such that:
$$A = PBP^{-1}$$
The matrix $P$ is the change of basis matrix: its columns are the coordinates of the new basis vectors expressed in the old basis.
Intuition: The formula $A = PBP^{-1}$ says:
$P^{-1}$ converts from the new basis to the old basis
$B$ applies the transformation in the old basis
$P$ converts back from the old basis to the new basis
Computing Changes of Basis: Elementary Row and Column Operations
While we won't need all the details for basic understanding, the key point is that we can systematically find similar matrices using elementary row and column operations:
Swapping rows or columns
Multiplying a row or column by a nonzero scalar
Adding a multiple of one row/column to another
These operations preserve similarity and help us find simpler forms of matrices (like diagonal form, which is much easier to work with).
Part 4: Gaussian Elimination
The Fundamental Algorithm
Gaussian elimination is the algorithm we use to solve linear systems in practice. It works by systematically performing elementary row operations to transform a matrix into a simpler form.
The process has two phases:
Forward Elimination: Transform the matrix into row echelon form, where:
All zero rows are at the bottom
In each non-zero row, the first nonzero entry (the pivot) is to the right of the pivot in the row above
Below each pivot, all entries are zero
Back Substitution: Continue to reduced row echelon form, where additionally:
Each pivot equals 1
Above each pivot, all entries are zero
What Gaussian Elimination Reveals
After reducing to reduced row echelon form, the matrix tells us essential information about the original linear system:
The rank of the matrix equals the number of nonzero rows (the number of independent constraints)
The kernel (the null space—solutions to $A\mathbf{x} = \mathbf{0}$) can be read off from the free variables
Whether an inverse exists: for a square matrix, it's invertible if and only if the reduced form is the identity matrix
Example: Consider the system: $$\begin{align} x + 2y &= 5 \\ 2x + 4y &= 10 \end{align}$$
The coefficient matrix is $\begin{bmatrix} 1 & 2 \\ 2 & 4 \end{bmatrix}$, and the augmented matrix is $\left[\begin{array}{cc|c} 1 & 2 & 5 \\ 2 & 4 & 10 \end{array}\right]$.
Applying elementary row operations (subtract 2× row 1 from row 2): $$\left[\begin{array}{cc|c} 1 & 2 & 5 \\ 0 & 0 & 0 \end{array}\right]$$
The second equation is redundant! We have only one constraint: $x + 2y = 5$. Setting $y = t$ (a free variable), we get $x = 5 - 2t$. The solution set is a line, not a single point.
Part 5: Linear Systems
What is a Linear System?
A linear system is a finite set of linear equations in a finite set of variables. "Linear" means each equation has degree 1 in each variable—no squares, products, or other nonlinearities.
A general linear system with $m$ equations in $n$ unknowns looks like:
$$\begin{align} a{11}x1 + a{12}x2 + \cdots + a{1n}xn &= b1 \\ a{21}x1 + a{22}x2 + \cdots + a{2n}xn &= b2 \\ &\vdots \\ a{m1}x1 + a{m2}x2 + \cdots + a{mn}xn &= bm \end{align}$$
Matrix Formulation: Unifying View
All this complexity simplifies dramatically when we write it as a single matrix equation:
$$A\mathbf{x} = \mathbf{b}$$
where:
$A$ is the $m \times n$ coefficient matrix whose entries are the $a{ij}$ values
$\mathbf{x} = \begin{bmatrix} x1 \\ x2 \\ \vdots \\ xn \end{bmatrix}$ is the column vector of unknowns
$\mathbf{b} = \begin{bmatrix} b1 \\ b2 \\ \vdots \\ bm \end{bmatrix}$ is the right-hand side vector
This formulation connects linear systems to linear maps: solving $A\mathbf{x} = \mathbf{b}$ means finding all vectors $\mathbf{x}$ that map to $\mathbf{b}$ under the linear map represented by $A$.
Example: The system $$\begin{align} 2x - y &= 1 \\ x + 3y &= 4 \end{align}$$
becomes:
$$\begin{bmatrix} 2 & -1 \\ 1 & 3 \end{bmatrix} \begin{bmatrix} x \\ y \end{bmatrix} = \begin{bmatrix} 1 \\ 4 \end{bmatrix}$$
Homogeneous vs. Non-Homogeneous Systems
When $\mathbf{b} = \mathbf{0}$, the system $A\mathbf{x} = \mathbf{0}$ is called homogeneous. This is asking: which vectors get mapped to the zero vector by the linear map represented by $A$? The answer is precisely the kernel (or null space) of $A$.
The solution set of a homogeneous system always includes the zero vector, and it has special algebraic structure: if $\mathbf{x}1$ and $\mathbf{x}2$ are solutions, so is any linear combination $c1\mathbf{x}1 + c2\mathbf{x}2$ (by linearity of $A$).
A non-homogeneous system has $\mathbf{b} \neq \mathbf{0}$. Its solution set (when it exists) is typically not a vector space—it's an affine subspace (a translated version of a subspace). If $\mathbf{x}p$ is one particular solution and $\mathbf{x}h$ is any solution to the homogeneous system, then $\mathbf{x}p + \mathbf{x}h$ is also a solution.
Key principle: The general solution to $A\mathbf{x} = \mathbf{b}$ is: $$\mathbf{x} = \mathbf{x}{\text{particular}} + \mathbf{x}{\text{homogeneous}}$$
where $\mathbf{x}{\text{particular}}$ is any single solution to the non-homogeneous system, and $\mathbf{x}{\text{homogeneous}}$ ranges over all solutions to $A\mathbf{x} = \mathbf{0}$.
Solving Linear Systems
For square invertible matrices: If $A$ is an $n \times n$ matrix and $\det(A) \neq 0$ (we'll define the determinant shortly), then $A$ has an inverse $A^{-1}$. The unique solution is:
$$\mathbf{x} = A^{-1}\mathbf{b}$$
We can check: $A(A^{-1}\mathbf{b}) = (AA^{-1})\mathbf{b} = I\mathbf{b} = \mathbf{b}$ ✓
For any system: Use Gaussian elimination. This method works when the matrix is square, rectangular, singular, or non-singular. It reveals:
Whether a solution exists
Whether the solution is unique
If solutions aren't unique, how to describe them all (using free variables)
Part 6: Endomorphisms and Square Matrices
Endomorphisms: Linear Maps to Themselves
An endomorphism of a vector space $V$ is a linear map $T: V \to V$ from the space to itself. These are special because the domain and codomain are the same space, which creates nice mathematical structure.
With respect to a chosen basis of $V$ of size $n$, an endomorphism is represented by an $n \times n$ square matrix. Square matrices inherit all the special properties of endomorphisms:
They can be composed with themselves: $A \cdot A$, written $A^2$
They can have eigenvalues and eigenvectors
They might be invertible (if $\det(A) \neq 0$)
Where Endomorphisms Appear
Endomorphisms model many important situations:
Geometric transformations: Rotations, reflections, scalings of space
Coordinate changes: Expressing the same geometric configuration in different coordinate systems
Quadratic forms: Expressions like $x^2 + 2xy + 3y^2$ can be encoded as an endomorphism and analyzed using matrix tools
Recurrence relations: The state of a system evolving over time
Part 7: The Determinant
What is the Determinant?
The determinant of a square matrix $A$ is a single number, denoted $\det(A)$ or $|A|$, that encodes crucial information about the linear map represented by $A$.
Geometric meaning: For an endomorphism, the absolute value of the determinant is the factor by which the transformation scales volumes. A positive determinant means the transformation preserves orientation (handedness); a negative determinant means it flips orientation.
If $|\det(A)| > 1$: the transformation expands volumes
If $|\det(A)| < 1$: the transformation shrinks volumes
If $\det(A) = 0$: the transformation collapses space to a lower dimension (it's not invertible)
The Invertibility Criterion
Here's the most important property for solving linear systems:
A square matrix $A$ is invertible if and only if $\det(A) \neq 0$.
This is why the determinant matters: it tells us whether a system $A\mathbf{x} = \mathbf{b}$ has a unique solution (when $\det(A) \neq 0$) or whether something more complicated is happening (when $\det(A) = 0$).
Why it's true: If $\det(A) = 0$, the endomorphism represented by $A$ collapses to a lower-dimensional subspace. This means $A$ is not injective (not one-to-one), so it doesn't have an inverse. Conversely, if $\det(A) \neq 0$, the transformation is full-dimensional and can be reversed.
Cramer's Rule: Solving Systems Using Determinants
For a square system $A\mathbf{x} = \mathbf{b}$ where $\det(A) \neq 0$, we can express the solution explicitly using determinants:
$$xi = \frac{\det(Ai)}{\det(A)}$$
where $Ai$ is the matrix obtained by replacing the $i$-th column of $A$ with the right-hand side vector $\mathbf{b}$.
Example: Solve $\begin{bmatrix} 2 & 1 \\ 1 & 3 \end{bmatrix} \begin{bmatrix} x \\ y \end{bmatrix} = \begin{bmatrix} 5 \\ 6 \end{bmatrix}$
First, compute $\det(A) = 2(3) - 1(1) = 6 - 1 = 5 \neq 0$, so a unique solution exists.
For $x$: replace the first column with $\mathbf{b}$: $A1 = \begin{bmatrix} 5 & 1 \\ 6 & 3 \end{bmatrix}$
$$\det(A1) = 5(3) - 1(6) = 15 - 6 = 9$$ $$x = \frac{9}{5}$$
For $y$: replace the second column with $\mathbf{b}$: $A2 = \begin{bmatrix} 2 & 5 \\ 1 & 6 \end{bmatrix}$
$$\det(A2) = 2(6) - 5(1) = 12 - 5 = 7$$ $$y = \frac{7}{5}$$
Important note: Cramer's rule is elegant and theoretically important, but for large systems, Gaussian elimination is typically faster computationally.
Summary
The topics in this section form a unified framework:
Matrices represent vectors and linear maps concretely
Matrix multiplication encodes composition of linear maps
Gaussian elimination is the practical algorithm for solving systems
Determinants determine invertibility and provide alternative solution methods
Endomorphisms and similarity reveal how the choice of basis affects representation
Master these concepts, and you'll have the tools to solve problems across linear algebra and its applications.
Flashcards
What is a vector in a finite-dimensional vector space identified with once a basis is chosen?
A coordinate column matrix
How are the columns of a matrix representing a linear map determined?
They are the images of the basis vectors of the domain
What operation between linear maps does matrix multiplication correspond to?
Composition
When do two matrices represent the same linear transformation in different bases?
If and only if they are similar
What does Gaussian elimination determine regarding the solutions of a linear system?
Existence of solutions
Uniqueness of solutions
How is a linear system formulated as a matrix equation?
$A\mathbf{x} = \mathbf{b}$ (where $A$ is the coefficient matrix, $\mathbf{x}$ is the variable vector, and $\mathbf{b}$ is the right-hand side vector)
If the coefficient matrix $A$ is invertible, what is the unique solution to the system $A\mathbf{x} = \mathbf{b}$?
$\mathbf{x} = A^{-1}\mathbf{b}$
What defines a homogeneous linear system?
The right-hand side vector is set to zero ($A\mathbf{x} = 0$)
What algebraic structure is formed by the solutions to a homogeneous system $A\mathbf{x} = 0$?
The kernel of the associated linear map
What is the definition of an endomorphism?
A linear map from a vector space to itself
What type of matrix represents an endomorphism in a vector space of dimension $n$?
An $n \times n$ square matrix
What is the criterion for a square matrix to be invertible based on its determinant?
The determinant must be nonzero
How does Cramer's rule solve a linear system?
By using determinants of matrices where columns of the coefficient matrix are replaced by the right-hand side vector
Quiz
Linear algebra - Matrices Linear Systems and Determinants Quiz Question 1: In the matrix equation \(A\mathbf{x} = \mathbf{b}\) that represents a linear system, what does the matrix \(A\) represent?
- The coefficient matrix (correct)
- The solution vector
- The variable vector
- The right‑hand side vector
Linear algebra - Matrices Linear Systems and Determinants Quiz Question 2: An endomorphism on an n‑dimensional vector space, expressed with respect to a basis, is represented by which type of matrix?
- An \(n\times n\) square matrix (correct)
- An \(n\times 1\) column vector
- An \(n\times m\) rectangular matrix
- A scalar
Linear algebra - Matrices Linear Systems and Determinants Quiz Question 3: A square matrix is invertible precisely when which condition on its determinant holds?
- The determinant is nonzero (correct)
- The determinant equals one
- The determinant is positive
- The determinant is negative
Linear algebra - Matrices Linear Systems and Determinants Quiz Question 4: If the coefficient matrix \(A\) is invertible, what is the unique solution to the linear system \(A\mathbf{x}=\mathbf{b}\)?
- \(\mathbf{x}=A^{-1}\mathbf{b}\) (correct)
- \(\mathbf{x}= \mathbf{b}A^{-1}\)
- \(\mathbf{x}=A\mathbf{b}\)
- \(\mathbf{x}=A^{T}\mathbf{b}\)
In the matrix equation \(A\mathbf{x} = \mathbf{b}\) that represents a linear system, what does the matrix \(A\) represent?
1 of 4
Key Concepts
Matrix Operations
Matrix
Matrix multiplication
Gaussian elimination
Determinant
Cramer's rule
Linear Transformations
Linear map
Similarity (linear algebra)
Endomorphism
Linear Systems
Linear system
Homogeneous system
Definitions
Matrix
A rectangular array of numbers that represents a linear transformation between finite‑dimensional vector spaces.
Linear map
A function between vector spaces that preserves vector addition and scalar multiplication.
Matrix multiplication
An operation that composes the linear maps represented by two matrices, yielding a matrix for the combined transformation.
Similarity (linear algebra)
A relation between two square matrices indicating they represent the same linear transformation in different bases, expressed as \(B = P^{-1}AP\).
Gaussian elimination
A systematic method of applying elementary row operations to reduce a matrix to row‑echelon form, used to solve linear systems and determine rank.
Linear system
A collection of linear equations in several variables, often written in matrix form \(A\mathbf{x} = \mathbf{b}\).
Homogeneous system
A linear system whose right‑hand side vector is zero, \(A\mathbf{x}=0\), whose solutions form the kernel of the associated linear map.
Endomorphism
A linear map from a vector space to itself, represented by an \(n\times n\) square matrix when a basis of size \(n\) is chosen.
Determinant
A scalar value computed from a square matrix that indicates invertibility; the matrix is invertible iff its determinant is nonzero.
Cramer's rule
A formula that solves a square linear system by expressing each variable as a ratio of determinants of matrices derived from the coefficient matrix.