Let be a linear transformation. Let be a basis for , and let be a basis for . Let and be the standard basis for and respectively. A transformation matrix of is a matrix such that the following diagram commutes:
where and , i.e., we have the functional equation .
Example. Consider the linear transformation . Let be the standard bases on Let .
Then, the transformation matrix is
Example. Consider the differentiation operator acting on polynomials of degree . Let's find its transformation matrix. Consider the basis for and the basis for . We define the vertical maps by and by , where is a basis for and is a basis for . Therefore, and . We can then write the transformation matrix
Column space, row space, and range
Let be an matrix. We can treat as a linear transformation from to . Let's explore the relationship between a matrix's column and row spaces and its corresponding linear transformation's range and null space.
The column space directly corresponds to the range of . For any , we have for some . Let be the column vectors of . If we write , then we can express as the column form of ,
Therefore, since any element of can be written as a linear combination of the column vectors of , we can conclude that .
The following theorem will summarize the relationship between the column space and row space.
Theorem. Let be any matrix.
, i.e.,
The rank of any matrix equals the maximum number of linearly independent rows of that matrix; that is, the rank of a matrix is the dimension of the subspace generated by the rows of that matrix
The rows and columns of any matrix generate subspaces of the same dimension, numerically equal to the rank of the matrix.
proof. The key insight is that elementary row operations preserve dimension since elementary row matrices are invertible. Let be a triangular system with elementary matrices . This gives us . Since is a triangular system, its nonzero row vectors form a basis for the row space of . When we observe that the columns also form a triangular system, we can conclude that the dimension of the column space equals the number of nonzero row vectors. Thus, .
Remark. Let be an matrix. Since , we can use the rank-nullity theorem to show that a linear transformation induced by is not one-to-one (equivalently, the null space of is nontrivial) when . The argument follows: first, we have
Therefore, implies that the null space of is nontrivial and thus not one-to-one.
How to find a basis for a null space
Finding a basis for the null space of is crucial because it helps determine the solution set of the linear equation when solutions exist. More precisely, we have the following equality:
where is a solution of and is the null space of . This equality can be proved through set inclusion in both directions.
() If is a solution, then . Therefore, .
() An element can be written as for some . We compute .
We use the following example to show the method of finding a basis for a null space.
Example. Let us find all solutions for
This is equivalent to find a basis for the null space of the matrix
Applying Gaussian elimination to the end, we get
(After finding each pivot, we must eliminate all entries above and below it by making them zero)
Now, here's a useful trick to complete our task. We'll label each column with a variable. If the -th column has a pivot, then we can express that variable in terms of other variables. If the -th column has no pivot, then the variable (let's call it ) is free to choose, and we'll add the trivial equation to our system of linear equations.
The last matrix above is the following linear equation system:
Therefore, we have
That is
Determinant
Historically, determinants played a major role in the study of linear algebra. They served as a computational tool to determine whether a linear equation system was singular. Additionally, Cramer's rule used determinants to provide explicit formulas for solving linear equation systems. Today, however, we primarily use determinants for computing eigenvalues.
In this chapter, we will develop an intuitive method to define determinants. We can also provide an axiomatic definition for determinants. From this perspective, we can prove that the only function satisfying these axioms is the one we defined intuitively. We will present the axiomatic definition without developing the related theorems further. Although determinants are no longer central to mathematical research, some of their properties remain essential to know. We will cover these properties in workshops and homework assignments.
Intuitive definition of determinant
The goal is to find an invariant of square matrices that we can use to determine whether a linear equation has a unique solution. In , consider the system
has a unique solution if and only if the vectors and span a nondegenerate parallelogram, i.e., the area of the parallelogram spanned by and is nonzero. We use the following special case to illustrate the idea that the area is .
The area of the parallelogram is . This formula is generally true, and we left the details for the readers (see 3Blue1BrownThe determinant | Chapter 6, Essence of linear algebra).
This motivate us to define
Definition. Let be a matrix. The determinant of is the scalar . We sall denote it by .
Similarly, in , the system
has a unique solution if and only if the volume of the parallelepiped is nonzero. To find this volume, mathematicians expressed the parallelepiped spanned by the three column vectors in terms of . After carefully reorganizing the terms, the volume is given as
You can check the following two expressions are identical:
or
While there are three additional equivalent expressions, all six can be unified in the following definition.
Definition. Let be a matrix. The determinant of is the scalar calculated for any by:
where is the submatrix of obtained by deleting the -th row and -th column. The term is called the cofactor of .
We summarize our discussion in the following table.
Our goal is to extend the definition of determinant to matrices where . This definition must preserve the key property that a matrix has a nonzero determinant if and only if has a unique solution.
Let's examine the patterns we found for and , and see if we can extend them to .