Get AI summaries of any video or article — Sign up free
Abstract Linear Algebra 18 | Orthonormal Basis thumbnail

Abstract Linear Algebra 18 | Orthonormal Basis

4 min read

Based on The Bright Side of Mathematics's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

An orthonormal basis for a finite-dimensional subspace U is characterized by ⟨B_i,B_j⟩=δ_{ij}, meaning orthogonality for i≠j and unit length for i=j.

Briefing

Orthonormal bases turn the hard parts of orthogonal projection into a fast, almost plug-and-play calculation. In a finite-dimensional subspace U of an inner product space V, choosing basis vectors that are mutually orthogonal and each have length 1 forces all cross-terms in inner products to vanish and makes the Gramian matrix collapse to the identity. That single structural change eliminates matrix inversion and turns the projection problem into a direct formula.

The key condition is written using the Kronecker delta: for basis vectors B_i and B_j in U, the inner product ⟨B_i, B_j⟩ equals δ_{ij}, meaning it is 1 when i=j and 0 when i≠j. Orthogonality is the “nice” part because it enforces right angles between distinct basis vectors, while the unit-length requirement ensures the squared length of each basis vector is exactly 1. Together, these properties define what comes later as an orthonormal system and, when the vectors also span U, an orthonormal basis.

This matters because orthogonal projection onto U typically requires solving a linear system derived from the Gramian matrix. For a general basis of U, the Gramian matrix is built from inner products of basis vectors with each other, and computing the projection involves inverting that matrix (or solving the equivalent system efficiently via row reduction). But when the basis satisfies ⟨B_i, B_j⟩=δ_{ij}, the Gramian matrix becomes the identity: ones on the diagonal and zeros off-diagonal. With that simplification, the coefficients in the projection are immediately determined by the right-hand side inner products.

Concretely, for any vector X in V, the orthogonal projection X_U onto U can be written as a linear combination of the basis vectors. With an orthonormal basis {B_1,…,B_K}, the coefficient in front of B_j is ⟨B_j, X⟩. So the projection takes the form X_U = Σ_{j=1..K} ⟨B_j, X⟩ B_j (often remembered as “basis vector times its inner product with X”). This is not just computational convenience—it also provides a clean conceptual decomposition of X into a component inside U and a perpendicular component outside U.

The transcript then formalizes the terminology. An orthogonal system requires only mutual orthogonality (⟨B_i,B_j⟩=0 for i≠j), while an orthonormal system adds the unit-length condition (⟨B_i,B_i⟩=1). If the vectors also form a basis for U, the system becomes an orthogonal basis or an orthonormal basis, respectively. The standard example is R^3 with the canonical unit vectors under the standard inner product, where the canonical basis is orthonormal and calculations are straightforward.

Finally, the discussion raises the natural next question: can any basis be transformed into an orthonormal one? The answer is yes, via a procedure promised for the next installment, setting up the move from arbitrary bases to orthonormal bases as a practical tool for inner product spaces.

Cornell Notes

Orthonormal bases make orthogonal projection dramatically easier in finite-dimensional inner product spaces. For a subspace U with an orthonormal basis {B_1,…,B_K}, the defining condition is ⟨B_i,B_j⟩=δ_{ij}, so the Gramian matrix built from inner products becomes the identity. That collapse removes the need to invert a matrix when computing the projection of any X∈V onto U. The projection coefficients are simply ⟨B_j,X⟩, giving X_U=∑_{j=1}^K ⟨B_j,X⟩B_j. The transcript also distinguishes orthogonal systems (only mutual orthogonality) from orthonormal systems (orthogonality plus unit length), and notes the standard example in R^3.

What does the Kronecker delta condition ⟨B_i,B_j⟩=δ_{ij} guarantee about the basis vectors?

It guarantees two things at once: (1) if i≠j, then δ_{ij}=0, so ⟨B_i,B_j⟩=0 and distinct basis vectors are orthogonal; (2) if i=j, then δ_{ii}=1, so ⟨B_i,B_i⟩=1 and each basis vector has length 1 (since ⟨B_i,B_i⟩ is the squared length). Together, these properties define an orthonormal system.

Why does an orthonormal basis make orthogonal projection easier than a general basis?

Orthogonal projection onto U usually requires solving a linear system derived from the Gramian matrix G, which depends on inner products among the chosen basis vectors. With an orthonormal basis, ⟨B_i,B_j⟩=δ_{ij} forces G to be the identity matrix (ones on the diagonal, zeros off-diagonal). Solving the system then becomes immediate: the projection coefficients come directly from the corresponding inner products with X, rather than from inverting G.

How are the coefficients in the projection X_U onto U determined when U has an orthonormal basis?

If {B_1,…,B_K} is an orthonormal basis of U, then the projection is X_U=∑_{j=1}^K ⟨B_j,X⟩B_j. Each coefficient multiplying B_j is exactly the inner product ⟨B_j,X⟩, so the calculation reduces to computing inner products and forming the linear combination.

What’s the difference between an orthogonal system and an orthonormal system?

An orthogonal system requires only mutual orthogonality: ⟨B_i,B_j⟩=0 for i≠j, with no requirement that ⟨B_i,B_i⟩ equals 1. An orthonormal system adds the unit-length condition ⟨B_i,B_i⟩=1 for every i, so each vector has length 1.

How do orthogonal bases and orthonormal bases relate to systems?

The transcript distinguishes “system” from “basis” by spanning. An orthogonal (or orthonormal) system may have fewer vectors than needed to span U. When the vectors also form a basis for U—meaning they span U—then the orthogonal system becomes an orthogonal basis (OB) and the orthonormal system becomes an orthonormal basis (ONB).

Review Questions

  1. In an inner product space, what two inner-product conditions define an orthonormal system?
  2. Explain why the Gramian matrix becomes the identity matrix when the basis satisfies ⟨B_i,B_j⟩=δ_{ij}.
  3. Given an orthonormal basis {B_1,…,B_K} for U, write the formula for the orthogonal projection of X onto U.

Key Points

  1. 1

    An orthonormal basis for a finite-dimensional subspace U is characterized by ⟨B_i,B_j⟩=δ_{ij}, meaning orthogonality for i≠j and unit length for i=j.

  2. 2

    Orthogonal projection onto U typically relies on a Gramian matrix built from inner products of the chosen basis vectors.

  3. 3

    When the basis is orthonormal, the Gramian matrix becomes the identity, eliminating the need for matrix inversion in the projection computation.

  4. 4

    For an orthonormal basis {B_1,…,B_K}, the projection of X onto U is X_U=∑_{j=1}^K ⟨B_j,X⟩B_j.

  5. 5

    An orthogonal system requires only ⟨B_i,B_j⟩=0 for i≠j, while an orthonormal system also requires ⟨B_i,B_i⟩=1.

  6. 6

    An orthogonal/orthonormal system becomes an orthogonal/orthonormal basis when the vectors span the subspace U.

  7. 7

    The canonical unit vectors in R^3 under the standard inner product provide a standard example of an orthonormal basis.

Highlights

Orthonormality forces the Gramian matrix to be the identity, turning projection into a direct inner-product formula.
With an orthonormal basis, the coefficient of B_j in the projection of X is exactly ⟨B_j,X⟩.
Orthogonal systems require right angles only; orthonormal systems add the unit-length constraint.
The projection formula X_U=∑ ⟨B_j,X⟩B_j is presented as a memory-friendly “basis times inner product” rule.
The standard basis of R^3 is used to anchor the abstract definitions before moving to general inner product spaces.

Topics