**Required math: calculus, vectors**

**Required physics: **SchrÃ¶dinger equation

References: Griffiths, David J. (2005), Introduction to Quantum Mechanics, 2nd Edition; Pearson Education – Chapter 3, PostÂ 21.

We’ve been writing vector inner products using the Dirac bra-ket notation, so the inner product of two vectors is . Dirac’s idea was to break this notation into two pieces, the ‘bra’ and the ‘ket’. The meaning of the ket part is fairly obvious: it’s just the original vector. But what exactly is the ‘bra’ part? Essentially, it’s a linear operator whose operand is a vector and output is a complex number (scalar). If the vector space is discrete (containing any number of dimensions, finite or infinite), then applying a bra to a ket results in the ordinary scalar product (the ‘dot product’ familiar from linear algebra). If the vector space is continuous, as with position or momentum, then applying a bra to a ket results in an integral over the relevant domain.

It’s worth pointing out that some authors such as Griffiths call the bra a linear function of vectors rather than an operator, preferring to reserve the term ‘operator’ for something which operates on a vector and returns another vector. I don’t see any particular value in such a fine distinction, and since the bra certainly does ‘operate’ on a vector (even though it produces a scalar as the result), the term ‘operator’ seems appropriate.

Although the bra has no physical meaning on its own, it can still simplify the notation for some other operators. One such example is the *projection operator*. If you can remember your linear algebra, you might recall that, given two vectors and , you can find the perpendicular projection of on from the formula

If is normalized (that is, it’s a unit vector), then this formula reduces to

This amounts to taking the inner (dot) product of with and multiplying by the vector . That is, we take a bra of and have it operate on a ket of , then multiply the result into the ket of .

In bra-ket notation, we can define the projection operator as

where is a normalized vector. Applying this to any other vector gives the projection of along :

We’ll have a look at a few properties of the projection operator.

First, the projection operator is *idempotent*, which means that . The consequence of this is that it doesn’t matter how many times you apply a given projection operator; it will have the same result as applying it just once. This makes sense from a geometric viewpoint, since once you’ve projected a vector onto another vector, projecting the projection just gives you the same projection back again.

The proof of the idempotent property is quite simple: , since .

Since it’s an operator that returns a vector, we can find its eigenvalues. Using the idempotent property we get

where is an eigenvalue of . Thus the only two eigenvalues possible are 0 and 1.

For an eigenvalue of 1, the corresponding eigenvector must satisfy . Thus the eigenvector for eigenvalue 1 is for some constant . Thus any vector parallel to is an eigenvector.

For an eigenvalue of 0, we have , so the eigenvector is any vector orthogonal to .