Skip to main content

Hilbert Spaces and Dirac Notation

Table of Contents

How do We Represent States?

Classical physics relies on a set of assumptions about how states of a system are represented. These include:

  • A quantity like (position) is just a single number.
  • Quantities can never have multiple values simultaneously. A particle cannot be at two places at once, nor can it have two different momenta.
  • The state of a system is completely determined by specifying the values of all such quantities.
  • These quantities are continuous, meaning that there are no gaps between possible values. This is intuitively true because particles cannot just jump instantly from one position to another.
  • The state of a system can be known with arbitrary precision.

These assumptions make it natural to represent the state of a system using a continous function. For example, the position of a particle can be represented by a function that gives the position of the particle at each time .

Quantum mechanics, however, challenges these assumptions:

  • It introduces the concept of superposition, where a particle can exist in multiple states at once. For example, a particle can be in a superposition of being at two different positions simultaneously.
  • It introduces the concept of quantization, where certain quantities can only take on discrete values. For example, the energy of an electron in an atom can only take on certain discrete values, or, if you have read the section on the Stern Gerlach Experiment, the spin of an electron can only be up or down.
  • It introduces the concept of uncertainty, where the state of a system cannot be known with arbitrary precision. This is encapsulated in the various uncertainty principles.

Clearly, we cannot use continuous functions to represent the state of a system in quantum mechanics.

Suppose that we want to write down the state of a particle using its energy (for example, the energy of an electron in an atom). Assume that, at an energy level of exactly , the particle's state is described by some object . We do not currently know what this object is, but we know that it represents the state of the particle at energy level .

Of course, there are multiple outcomes for the energy of the particle, each corresponding to a different state . We can write the state of the particle by combining these states with an unknown operation:

Each state also has its own probability - some states are more likely than others. Hence, it's also appropriate to introduce a scaling factor for each state :

It turns out, the operation is just addition, and introducing scale factors makes the state a linear combination of the states . Since we can add and scale the states , they form a vector space, and is a vector in this space. A vector inside these brackets, , is called a ket.

This is the fundamental idea behind the state vector formalism in quantum mechanics. The state of a system is represented by a vector in a vector space, and the state can be a linear combination of multiple basis states. The reason we do not use the usual notation for vectors (like ) will become clear later.

If the representation of everything about a system with object is not clear, know that this is a common pattern in physics. Later on, we will see that we can do the same in classical mechanics, where we can represent everything about a system with the Lagrangian or Hamiltonian.

The Wavefunction (and a Need for Cauchy Completeness)

In quantum mechanics, we represent physical quantities like position, momentum, and energy using operators. As shown previously, the state of a system is represented by a state vector . One might guess that to find some physical quantity of the system, we can use a matrix to operate on the state vector. This matrix is called an operator.

For example, the position operator acts on the state vector to give the position of the particle. It is denoted as .

The wavefunction is a complex-valued function that represents the state of a system. As shown previously, it is formed by combining different states with scaling factors :

This sum is called a superposition of states.

But of course, some quantities can take on an infinite number of values. A particle can theoretically take on any energy level, so we need to sum over an infinite number of states:

For a continuous quantity, like position, we need to sum over an uncountably infinite number of states. This means that the sum becomes an integral:

It turns out that is the position wavefunction, . For another quantity, like momentum, we would have a different wavefunction , called the momentum wavefunction. Therefore, for any physical quantity , we can define a wavefunction that represents the state of the system in that quantity's space:

The state vectors that are used to represent the state of the system are called basis states. Recall from linear algebra that the dimensionality of a vector space is the number of basis states required to span the space. But if there are an infinite number of basis states, the vector space is infinite-dimensional. This raises a problem. To illustrate, consider a vector space represented by the set of all polynomials. The basis states are the monomials , and polynomials are linear combinations of these basis states. But observe this following infinite polynomial:

This infinite polynomial turns out to be , which is not a polynomial. Hence, an infinite sum of polynomials can give a function that is not a polynomial.

In order for the wavefunction to be a valid representation of the state of the system, we must add one more condition: a limit of a sequence/series of state vectors must converge to a state vector in the space. This is known as Cauchy completeness, and it ensures that the space of state vectors is valid.

Let's summarize the properties of the vector space of state vectors:

  • The state of a system is represented by a state vector .
  • The state vector is a linear combination of basis states .
  • The basis states form an infinite-dimensional vector space.
  • The vector space is Cauchy complete, meaning that the limit of a sequence of state vectors converges to a state vector in the space.

This vector space that we have just described is almost the definition of a Hilbert space . We need to add one more property to make it a Hilbert space: the space has a well-defined inner product.

The Inner Product

Another important concept in quantum mechanics is the inner product. It is a generalization of the dot product in Euclidean space to infinite-dimensional spaces.

Recall that the dot product of two vectors and (in an orthonormal basis) is given by:

note

Outside orthonormal bases, the dot product is given by , where is the metric tensor.

Defining a dot product helps us define angles between vectors, as well as lengths and projections. However, it seems that the product does not have an equally simple geometric interpretation in abstract vector spaces like the space of state vectors.

Suppose and are two state vectors in the space of state vectors. The inner product is denoted as . Just like the dot product, the inner product is a way to multiply two vectors to get a scalar. An inner product requires the following properties, all of which should be familiar from the dot product:

  1. Linearity in the second argument: .

  2. Conjugate symmetry: . There is a reason for the conjugation, and we can show this with a simple example.

    Consider an inner product of (with a magnitude of ) with itself: . Next, scale the vectors in the inner product by to get . By linearity, this is . But the inner product defines the magnitude, and with this scaling, the magnitude is .

    Hence, to ensure that the magnitude is real, we need to conjugate one of the vectors when flipping the order: . The following shows that the magnitude of is real:

  3. Positive definiteness: If , then . This ensures that the inner product is a valid measure of the magnitude of a vector. After all, the magnitude of a vector should not be zero if the vector itself is not the zero vector.

Corollary: Antilinearity in the First Argument

For the linearity of the first argument, we can show that it is antilinear from the linearity of the second argument and conjugate symmetry:

Corollary: Magnitude and Orthogonality

  • The magnitude of a vector is given by , just like the magnitude of a vector in Euclidean space.
  • Two vectors and are orthogonal if .

Applying to Discrete Basis States

Let's apply the inner product to quantum mechanics, where the basis states are discrete. In an orthnormal basis, the inner product of two basis states and is given by:

This is the Kronecker delta, which is if and otherwise.

Recall that the state vector is a linear combination of basis states :

Suppose we want to find the coefficient . This is similar to finding the component of a vector along a direction in Euclidean space. We can find by taking the inner product of with :

Therefore:

This is the projection of onto .

Dirac Delta Function

Previously, we applied the inner product to discrete basis states. But what if the basis states are continuous, like the position basis ?

Recall that the state vector can be written as an integral of basis states :

Suppose we want to find the coefficient when the basis state is (where is a specific value of ). We can find by taking the inner product of with , just like we did with discrete basis states:

The inner product is a function that is everywhere except at , where it is infinite. (This is because the basis states are orthogonal to each other, except when .) This function is the Dirac delta function .

What is the Dirac Delta Function?

The Dirac Delta can be thought of as a continuous analog of the Kronecker delta. A common interpretation of the Dirac delta function is that it is a function that is zero everywhere except at a single point, where it is infinite:

Additionally, it satisfies the following property:

Intuitively, the Dirac delta function is like a infinitely big spike at a single point. We can shift the spike to any point by writing . The integral property is then:

We can also multiply the integrand by a function . The only time when the integrand is non-zero is when , so the integral is just :

However, the big-spike interpretation is not the complete picture. The Dirac delta function can be constructed as the limit of a function. Consider the normalized Gaussian function:

This function is a bell curve centered at with a width of . As , the bell curve becomes narrower and taller, and the area under the curve remains , as required for a probability density function. Thus, the limit of as is the Dirac delta function.

The graph of this distribution is shown below:

Normalized Gaussian Distribution

Next, consider another function:

This function is a sine wave that oscillates faster as decreases. A graph is also shown below:

Sine Distribution

As , the sine wave oscillates faster and faster. It does have a big spike at , but it also oscillates at other points. However, it turns out that the limit of this function as is also the Dirac delta function because it satisfies its integral property. This function is used extensively in the Fourier transform.

Instead, a better definition of the Dirac delta function is the following:

The Dirac delta function is a distribution that satisfies the following property:

Corollary: Inner Product of Wavefunctions

Consider the inner product of two wavefunctions and . Each wavefunction can be written as an integral of basis states :

(We use in the second wavefunction to avoid confusion with .) The inner product of these wavefunctions is then:

Since this is only nonzero when , we can collapse the integral and just replace with :

This is similar to the inner product in the discrete case:

Back to the Inner Product

Going back to trying to find from the inner product , we can write it as:

This is just :

Dual Space and Bra Vectors

Consider a linear map that acts on a vector to give a scalar :

This type of linear map is called a linear functional. For example, consider an Euclidean example: define as a linear functional that acts on a vector to give the -component of . Then, .

Linear functionals can be represented by a matrix, where is the dimension of the vector space. In other words, they are row vectors. These objects have many names - linear functionals, covectors, dual vectors, and row vectors. For a more detailed explanation and some visual intuition, see this Eigenchris video. It details how linear functionals can be visualized by contour lines.

The set of all linear functionals on a vector space forms a vector space itself, called the dual space (denoted as ). Linear functionals appear in quantum mechanics because we need to convert from the vector (state vector) to a scalar (e.g. probability, expectation value, etc.). In quantum mechanics, the dual space is denoted as (called a bra vector), and they exist in the dual Hilbert space . A linear functional acting on a state vector is denoted as .

Notice that inner products, just like linear functionals, map a vector to a scalar. This means that inner products and linear functionals are fundamentally the same. For any linear functional, you can find a corresponding vector that represents it in an inner product. You can watch this 3b1b video for a more visual explanation. This property is known as the Riesz representation theorem:

For any linear functional in a Hilbert space , there exists a unique vector in such that:

To show this in our notation, we literally write the linear functional as an inner product:

This set of notation is called bra-ket notation or Dirac notation. It is very helpful because it allows us to interchange between vectors and linear functionals easily without having to worry about the details.

Resolution of the Identity

The resolution of the identity is a property of the basis states in a Hilbert space. We will prove this in order to show how Dirac notation simplifies calculations.

Consider a set of orthonormal basis states . A quantum state can be written as a linear combination of these basis states:

In an orthonormal basis, can be found by taking the inner product of with (from Equation ):

Dirac notation allows us to split the braket into two parts:

Since is constant regardless of the summation index, we can write it outside the sum:

This is an equation that holds for any state . This must mean that the quantity in the parentheses is the identity operator :

The resolution of the identity states that:

This is a very important property of the basis states in a Hilbert space. It is also known as the completeness relation. In the continuous case, the sum becomes an integral:

Summary and Next Steps

In this note, we have introduced the concept of the state vector in quantum mechanics. The state of a system is represented by a state vector , which is a linear combination of basis states .

Here are the key points to remember:

  • The state of a system is represented by a state vector in a Hilbert space .

  • A Hilbert space is an infinite-dimensional vector space that is Cauchy complete.

  • The state vector is a linear combination of basis states . In a discrete basis, the state vector is a sum over basis states:

    In a continuous basis, the state vector is an integral over basis states:

  • The inner product of two state vectors and is an operation that gives a scalar. It is denoted as and satisfies these properties:

    1. Linearity in the second argument: .
    2. Conjugate symmetry: .
    3. Positive definiteness: if .
  • The Dirac delta function is a distribution that satisfies the property:

    It can be interpreted as a big spike at a single point, but that is not the complete picture.

  • The value of can be found by taking the inner product of with :

  • Linear functionals in a Hilbert space are represented by bra vectors . They exist in the dual Hilbert space . The inner product of a bra vector and a state vector is denoted as . This forms the basis of Dirac notation.

  • The resolution of the identity/completeness relation states that the identity operator can be written as:

    In the continuous case, the sum becomes an integral:

This was a lot of information and it is completely normal for the reader to feel overwhelmed. While I only used one page to explain these concepts, it took physicists decades to develop them, not mentioning the centuries of mathematical development that preceded them. Furthermore, quantum mechanics itself is a very complex theory that opposes human intuition.

In the next note, we will move on to how operators act on state vectors and how they relate to observable quantities.

References

  • Quantum Sense, "Maths of Quantum Mechanics", a Youtube Playlist.
  • J.J. Sakurai, "Modern Quantum Mechanics", section 1.4.