Abstract Linear Algebra 18 | Orthonormal Basis
Based on The Bright Side of Mathematics's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.
An orthonormal basis for a finite-dimensional subspace U is characterized by ⟨B_i,B_j⟩=δ_{ij}, meaning orthogonality for i≠j and unit length for i=j.
Briefing
Orthonormal bases turn the hard parts of orthogonal projection into a fast, almost plug-and-play calculation. In a finite-dimensional subspace U of an inner product space V, choosing basis vectors that are mutually orthogonal and each have length 1 forces all cross-terms in inner products to vanish and makes the Gramian matrix collapse to the identity. That single structural change eliminates matrix inversion and turns the projection problem into a direct formula.
The key condition is written using the Kronecker delta: for basis vectors B_i and B_j in U, the inner product ⟨B_i, B_j⟩ equals δ_{ij}, meaning it is 1 when i=j and 0 when i≠j. Orthogonality is the “nice” part because it enforces right angles between distinct basis vectors, while the unit-length requirement ensures the squared length of each basis vector is exactly 1. Together, these properties define what comes later as an orthonormal system and, when the vectors also span U, an orthonormal basis.
This matters because orthogonal projection onto U typically requires solving a linear system derived from the Gramian matrix. For a general basis of U, the Gramian matrix is built from inner products of basis vectors with each other, and computing the projection involves inverting that matrix (or solving the equivalent system efficiently via row reduction). But when the basis satisfies ⟨B_i, B_j⟩=δ_{ij}, the Gramian matrix becomes the identity: ones on the diagonal and zeros off-diagonal. With that simplification, the coefficients in the projection are immediately determined by the right-hand side inner products.
Concretely, for any vector X in V, the orthogonal projection X_U onto U can be written as a linear combination of the basis vectors. With an orthonormal basis {B_1,…,B_K}, the coefficient in front of B_j is ⟨B_j, X⟩. So the projection takes the form X_U = Σ_{j=1..K} ⟨B_j, X⟩ B_j (often remembered as “basis vector times its inner product with X”). This is not just computational convenience—it also provides a clean conceptual decomposition of X into a component inside U and a perpendicular component outside U.
The transcript then formalizes the terminology. An orthogonal system requires only mutual orthogonality (⟨B_i,B_j⟩=0 for i≠j), while an orthonormal system adds the unit-length condition (⟨B_i,B_i⟩=1). If the vectors also form a basis for U, the system becomes an orthogonal basis or an orthonormal basis, respectively. The standard example is R^3 with the canonical unit vectors under the standard inner product, where the canonical basis is orthonormal and calculations are straightforward.
Finally, the discussion raises the natural next question: can any basis be transformed into an orthonormal one? The answer is yes, via a procedure promised for the next installment, setting up the move from arbitrary bases to orthonormal bases as a practical tool for inner product spaces.
Cornell Notes
Orthonormal bases make orthogonal projection dramatically easier in finite-dimensional inner product spaces. For a subspace U with an orthonormal basis {B_1,…,B_K}, the defining condition is ⟨B_i,B_j⟩=δ_{ij}, so the Gramian matrix built from inner products becomes the identity. That collapse removes the need to invert a matrix when computing the projection of any X∈V onto U. The projection coefficients are simply ⟨B_j,X⟩, giving X_U=∑_{j=1}^K ⟨B_j,X⟩B_j. The transcript also distinguishes orthogonal systems (only mutual orthogonality) from orthonormal systems (orthogonality plus unit length), and notes the standard example in R^3.
What does the Kronecker delta condition ⟨B_i,B_j⟩=δ_{ij} guarantee about the basis vectors?
Why does an orthonormal basis make orthogonal projection easier than a general basis?
How are the coefficients in the projection X_U onto U determined when U has an orthonormal basis?
What’s the difference between an orthogonal system and an orthonormal system?
How do orthogonal bases and orthonormal bases relate to systems?
Review Questions
- In an inner product space, what two inner-product conditions define an orthonormal system?
- Explain why the Gramian matrix becomes the identity matrix when the basis satisfies ⟨B_i,B_j⟩=δ_{ij}.
- Given an orthonormal basis {B_1,…,B_K} for U, write the formula for the orthogonal projection of X onto U.
Key Points
- 1
An orthonormal basis for a finite-dimensional subspace U is characterized by ⟨B_i,B_j⟩=δ_{ij}, meaning orthogonality for i≠j and unit length for i=j.
- 2
Orthogonal projection onto U typically relies on a Gramian matrix built from inner products of the chosen basis vectors.
- 3
When the basis is orthonormal, the Gramian matrix becomes the identity, eliminating the need for matrix inversion in the projection computation.
- 4
For an orthonormal basis {B_1,…,B_K}, the projection of X onto U is X_U=∑_{j=1}^K ⟨B_j,X⟩B_j.
- 5
An orthogonal system requires only ⟨B_i,B_j⟩=0 for i≠j, while an orthonormal system also requires ⟨B_i,B_i⟩=1.
- 6
An orthogonal/orthonormal system becomes an orthogonal/orthonormal basis when the vectors span the subspace U.
- 7
The canonical unit vectors in R^3 under the standard inner product provide a standard example of an orthonormal basis.