In physics, an operator is a function over a space of physical states onto another space of physical states. The simplest example of the utility of operators is the study of symmetry (which makes the concept of a group useful in this context). Because of this, they are useful tools in classical mechanics. Operators are even more important in quantum mechanics, where they form an intrinsic part of the formulation of the theory.
In classical mechanics, the movement of a particle (or system of particles) is completely determined by the Lagrangian [math]\displaystyle{ L(q, \dot{q}, t) }[/math] or equivalently the Hamiltonian [math]\displaystyle{ H(q, p, t) }[/math], a function of the generalized coordinates q, generalized velocities [math]\displaystyle{ \dot{q} = \mathrm{d} q / \mathrm{d} t }[/math] and its conjugate momenta:
If either L or H is independent of a generalized coordinate q, meaning the L and H do not change when q is changed, which in turn means the dynamics of the particle are still the same even when q changes, the corresponding momenta conjugate to those coordinates will be conserved (this is part of Noether's theorem, and the invariance of motion with respect to the coordinate q is a symmetry). Operators in classical mechanics are related to these symmetries.
More technically, when H is invariant under the action of a certain group of transformations G:
The elements of G are physical operators, which map physical states among themselves.
Transformation | Operator | Position | Momentum |
---|---|---|---|
Translational symmetry | [math]\displaystyle{ X(\mathbf{a}) }[/math] | [math]\displaystyle{ \mathbf{r}\rightarrow \mathbf{r} + \mathbf{a} }[/math] | [math]\displaystyle{ \mathbf{p}\rightarrow \mathbf{p} }[/math] |
Time translation symmetry | [math]\displaystyle{ U(t_0) }[/math] | [math]\displaystyle{ \mathbf{r}(t)\rightarrow \mathbf{r}(t+t_0) }[/math] | [math]\displaystyle{ \mathbf{p}(t)\rightarrow \mathbf{p}(t+t_0) }[/math] |
Rotational invariance | [math]\displaystyle{ R(\mathbf{\hat{n}},\theta) }[/math] | [math]\displaystyle{ \mathbf{r}\rightarrow R(\mathbf{\hat{n}},\theta)\mathbf{r} }[/math] | [math]\displaystyle{ \mathbf{p}\rightarrow R(\mathbf{\hat{n}},\theta)\mathbf{p} }[/math] |
Galilean transformations | [math]\displaystyle{ G(\mathbf{v}) }[/math] | [math]\displaystyle{ \mathbf{r}\rightarrow \mathbf{r} + \mathbf{v}t }[/math] | [math]\displaystyle{ \mathbf{p}\rightarrow \mathbf{p} + m\mathbf{v} }[/math] |
Parity | [math]\displaystyle{ P }[/math] | [math]\displaystyle{ \mathbf{r}\rightarrow -\mathbf{r} }[/math] | [math]\displaystyle{ \mathbf{p}\rightarrow -\mathbf{p} }[/math] |
T-symmetry | [math]\displaystyle{ T }[/math] | [math]\displaystyle{ \mathbf{r}\rightarrow \mathbf{r}(-t) }[/math] | [math]\displaystyle{ \mathbf{p}\rightarrow -\mathbf{p}(-t) }[/math] |
where [math]\displaystyle{ R(\hat{\boldsymbol{n}}, \theta) }[/math] is the rotation matrix about an axis defined by the unit vector [math]\displaystyle{ \hat{\boldsymbol{n}} }[/math] and angle θ.
If the transformation is infinitesimal, the operator action should be of the form
where [math]\displaystyle{ I }[/math] is the identity operator, [math]\displaystyle{ \epsilon }[/math] is a parameter with a small value, and [math]\displaystyle{ A }[/math] will depend on the transformation at hand, and is called a generator of the group. Again, as a simple example, we will derive the generator of the space translations on 1D functions.
As it was stated, [math]\displaystyle{ T_a f(x)=f(x-a) }[/math]. If [math]\displaystyle{ a=\epsilon }[/math] is infinitesimal, then we may write
This formula may be rewritten as
where [math]\displaystyle{ D }[/math] is the generator of the translation group, which in this case happens to be the derivative operator. Thus, it is said that the generator of translations is the derivative.
The whole group may be recovered, under normal circumstances, from the generators, via the exponential map. In the case of the translations the idea works like this.
The translation for a finite value of [math]\displaystyle{ a }[/math] may be obtained by repeated application of the infinitesimal translation:
with the [math]\displaystyle{ \cdots }[/math] standing for the application [math]\displaystyle{ N }[/math] times. If [math]\displaystyle{ N }[/math] is large, each of the factors may be considered to be infinitesimal:
But this limit may be rewritten as an exponential:
To be convinced of the validity of this formal expression, we may expand the exponential in a power series:
The right-hand side may be rewritten as
which is just the Taylor expansion of [math]\displaystyle{ f(x-a) }[/math], which was our original value for [math]\displaystyle{ T_a f(x) }[/math].
The mathematical properties of physical operators are a topic of great importance in itself. For further information, see C*-algebra and Gelfand–Naimark theorem.
The mathematical formulation of quantum mechanics (QM) is built upon the concept of an operator.
Physical pure states in quantum mechanics are represented as unit-norm vectors (probabilities are normalized to one) in a special complex Hilbert space. Time evolution in this vector space is given by the application of the evolution operator.
Any observable, i.e., any quantity which can be measured in a physical experiment, should be associated with a self-adjoint linear operator. The operators must yield real eigenvalues, since they are values which may come up as the result of the experiment. Mathematically this means the operators must be Hermitian.[1] The probability of each eigenvalue is related to the projection of the physical state on the subspace related to that eigenvalue. See below for mathematical details about Hermitian operators.
In the wave mechanics formulation of QM, the wavefunction varies with space and time, or equivalently momentum and time (see position and momentum space for details), so observables are differential operators.
In the matrix mechanics formulation, the norm of the physical state should stay fixed, so the evolution operator should be unitary, and the operators can be represented as matrices. Any other symmetry, mapping a physical state into another, should keep this restriction.
The wavefunction must be square-integrable (see Lp spaces), meaning:
and normalizable, so that:
Two cases of eigenstates (and eigenvalues) are:
Let ψ be the wavefunction for a quantum system, and [math]\displaystyle{ \hat{A} }[/math] be any linear operator for some observable A (such as position, momentum, energy, angular momentum etc.). If ψ is an eigenfunction of the operator [math]\displaystyle{ \hat{A} }[/math], then
where a is the eigenvalue of the operator, corresponding to the measured value of the observable, i.e. observable A has a measured value a.
If ψ is an eigenfunction of a given operator [math]\displaystyle{ \hat{A} }[/math], then a definite quantity (the eigenvalue a) will be observed if a measurement of the observable A is made on the state ψ. Conversely, if ψ is not an eigenfunction of [math]\displaystyle{ \hat{A} }[/math], then it has no eigenvalue for [math]\displaystyle{ \hat{A} }[/math], and the observable does not have a single definite value in that case. Instead, measurements of the observable A will yield each eigenvalue with a certain probability (related to the decomposition of ψ relative to the orthonormal eigenbasis of [math]\displaystyle{ \hat{A} }[/math]).
In bra–ket notation the above can be written;
that are equal if [math]\displaystyle{ \left| \psi \right\rangle }[/math] is an eigenvector, or eigenket of the observable A.
Due to linearity, vectors can be defined in any number of dimensions, as each component of the vector acts on the function separately. One mathematical example is the del operator, which is itself a vector (useful in momentum-related quantum operators, in the table below).
An operator in n-dimensional space can be written:
where ej are basis vectors corresponding to each component operator Aj. Each component will yield a corresponding eigenvalue [math]\displaystyle{ a_j }[/math]. Acting this on the wave function ψ:
in which we have used [math]\displaystyle{ \hat{A}_j \psi = a_j \psi . }[/math]
In bra–ket notation:
If two observables A and B have linear operators [math]\displaystyle{ \hat{A} }[/math] and [math]\displaystyle{ \hat{B} }[/math], the commutator is defined by,
The commutator is itself a (composite) operator. Acting the commutator on ψ gives:
If ψ is an eigenfunction with eigenvalues a and b for observables A and B respectively, and if the operators commute:
then the observables A and B can be measured simultaneously with infinite precision, i.e., uncertainties [math]\displaystyle{ \Delta A = 0 }[/math], [math]\displaystyle{ \Delta B = 0 }[/math] simultaneously. ψ is then said to be the simultaneous eigenfunction of A and B. To illustrate this:
It shows that measurement of A and B does not cause any shift of state, i.e., initial and final states are same (no disturbance due to measurement). Suppose we measure A to get value a. We then measure B to get the value b. We measure A again. We still get the same value a. Clearly the state (ψ) of the system is not destroyed and so we are able to measure A and B simultaneously with infinite precision.
If the operators do not commute:
they cannot be prepared simultaneously to arbitrary precision, and there is an uncertainty relation between the observables
even if ψ is an eigenfunction the above relation holds. Notable pairs are position-and-momentum and energy-and-time uncertainty relations, and the angular momenta (spin, orbital and total) about any two orthogonal axes (such as Lx and Ly, or sy and sz, etc.).[2]
The expectation value (equivalently the average or mean value) is the average measurement of an observable, for particle in region R. The expectation value [math]\displaystyle{ \left\langle \hat{A} \right\rangle }[/math] of the operator [math]\displaystyle{ \hat{A} }[/math] is calculated from:[3]
This can be generalized to any function F of an operator:
An example of F is the 2-fold action of A on ψ, i.e. squaring an operator or doing it twice:
The definition of a Hermitian operator is:[1]
Following from this, in bra–ket notation:
Important properties of Hermitian operators include:
An operator can be written in matrix form to map one basis vector to another. Since the operators are linear, the matrix is a linear transformation (aka transition matrix) between bases. Each basis element [math]\displaystyle{ \phi_j }[/math] can be connected to another,[3] by the expression:
which is a matrix element:
A further property of a Hermitian operator is that eigenfunctions corresponding to different eigenvalues are orthogonal.[1] In matrix form, operators allow real eigenvalues to be found, corresponding to measurements. Orthogonality allows a suitable basis set of vectors to represent the state of the quantum system. The eigenvalues of the operator are also evaluated in the same way as for the square matrix, by solving the characteristic polynomial:
where I is the n × n identity matrix, as an operator it corresponds to the identity operator. For a discrete basis:
while for a continuous basis:
A non-singular operator [math]\displaystyle{ \hat{A} }[/math] has an inverse [math]\displaystyle{ \hat{A}^{-1} }[/math] defined by:
If an operator has no inverse, it is a singular operator. In a finite-dimensional space, an operator is non-singular if and only if its determinant is nonzero:
and hence the determinant is zero for a singular operator.
The operators used in quantum mechanics are collected in the table below (see for example[1][4]). The bold-face vectors with circumflexes are not unit vectors, they are 3-vector operators; all three spatial components taken together.
Operator (common name/s) | Cartesian component | General definition | SI unit | Dimension |
---|---|---|---|---|
Position | [math]\displaystyle{ \begin{align} \hat{x} &= x, & \hat{y} &= y, & \hat{z} &= z \end{align} }[/math] | [math]\displaystyle{ \mathbf{\hat{r}} = \mathbf{r} \,\! }[/math] | m | [L] |
Momentum | General
[math]\displaystyle{ \begin{align} \hat{p}_x & = -i \hbar \frac{\partial}{\partial x}, & \hat{p}_y & = -i \hbar \frac{\partial}{\partial y}, & \hat{p}_z & = -i \hbar \frac{\partial}{\partial z} \end{align} }[/math] |
General
[math]\displaystyle{ \mathbf{\hat{p}} = -i \hbar \nabla \,\! }[/math] |
J s m−1 = N s | [M] [L] [T]−1 |
Electromagnetic field
[math]\displaystyle{ \begin{align} \hat{p}_x = -i \hbar \frac{\partial}{\partial x} - qA_x \\ \hat{p}_y = -i \hbar \frac{\partial}{\partial y} - qA_y \\ \hat{p}_z = -i \hbar \frac{\partial}{\partial z} - qA_z \end{align} }[/math] |
Electromagnetic field (uses kinetic momentum; A, vector potential)
[math]\displaystyle{ \begin{align} \mathbf{\hat{p}} & = \mathbf{\hat{P}} - q\mathbf{A} \\ & = -i \hbar \nabla - q\mathbf{A} \\ \end{align}\,\! }[/math] |
J s m−1 = N s | [M] [L] [T]−1 | |
Kinetic energy | Translation
[math]\displaystyle{ \begin{align} \hat{T}_x & = -\frac{\hbar^2}{2m}\frac{\partial^2}{\partial x^2} \\[2pt] \hat{T}_y & = -\frac{\hbar^2}{2m}\frac{\partial^2}{\partial y^2} \\[2pt] \hat{T}_z & = -\frac{\hbar^2}{2m}\frac{\partial^2}{\partial z^2} \\ \end{align} }[/math] |
[math]\displaystyle{ \begin{align} \hat{T} & = \frac{1}{2m}\mathbf{\hat{p}}\cdot\mathbf{\hat{p}} \\ & = \frac{1}{2m}(-i \hbar \nabla)\cdot(-i \hbar \nabla) \\ & = \frac{-\hbar^2 }{2m}\nabla^2 \end{align}\,\! }[/math] |
J | [M] [L]2 [T]−2 |
Electromagnetic field
[math]\displaystyle{ \begin{align} \hat{T}_x & = \frac{1}{2m}\left(-i \hbar \frac{\partial}{\partial x} - q A_x \right)^2 \\ \hat{T}_y & = \frac{1}{2m}\left(-i \hbar \frac{\partial}{\partial y} - q A_y \right)^2 \\ \hat{T}_z & = \frac{1}{2m}\left(-i \hbar \frac{\partial}{\partial z} - q A_z \right)^2 \end{align}\,\! }[/math] |
Electromagnetic field (A, vector potential)
[math]\displaystyle{ \begin{align} \hat{T} & = \frac{1}{2m}\mathbf{\hat{p}}\cdot\mathbf{\hat{p}} \\ & = \frac{1}{2m}(-i \hbar \nabla - q\mathbf{A})\cdot(-i \hbar \nabla - q\mathbf{A}) \\ & = \frac{1}{2m}(-i \hbar \nabla - q\mathbf{A})^2 \end{align}\,\! }[/math] |
J | [M] [L]2 [T]−2 | |
Rotation (I, moment of inertia)
[math]\displaystyle{ \begin{align} \hat{T}_{xx} & = \frac{\hat{J}_x^2}{2I_{xx}} \\ \hat{T}_{yy} & = \frac{\hat{J}_y^2}{2I_{yy}} \\ \hat{T}_{zz} & = \frac{\hat{J}_z^2}{2I_{zz}} \\ \end{align}\,\! }[/math] |
Rotation
[math]\displaystyle{ \hat{T} = \frac{\mathbf{\hat{J}}\cdot\mathbf{\hat{J}}}{2I} \,\! }[/math][citation needed] |
J | [M] [L]2 [T]−2 | |
Potential energy | N/A | [math]\displaystyle{ \hat{V} = V\left( \mathbf{r}, t \right) = V \,\! }[/math] | J | [M] [L]2 [T]−2 |
Total energy | N/A | Time-dependent potential: [math]\displaystyle{ \hat{E} = i \hbar \frac{\partial}{\partial t} \,\! }[/math] Time-independent: |
J | [M] [L]2 [T]−2 |
Hamiltonian | [math]\displaystyle{ \begin{align} \hat{H} & = \hat{T} + \hat{V} \\ & = \frac{1}{2m}\mathbf{\hat{p}}\cdot\mathbf{\hat{p}} + V \\ & = \frac{1}{2m}\hat{p}^2 + V \\ \end{align} \,\! }[/math] | J | [M] [L]2 [T]−2 | |
Angular momentum operator | [math]\displaystyle{ \begin{align} \hat{L}_x & = -i\hbar \left(y {\partial \over \partial z} - z {\partial \over \partial y}\right) \\ \hat{L}_y & = -i\hbar \left(z {\partial \over \partial x} - x {\partial \over \partial z}\right) \\ \hat{L}_z & = -i\hbar \left(x {\partial \over \partial y} - y {\partial \over \partial x}\right) \end{align} }[/math] | [math]\displaystyle{ \mathbf{\hat{L}} = \mathbf{r} \times -i\hbar \nabla }[/math] | J s = N s m | [M] [L]2 [T]−1 |
Spin angular momentum | [math]\displaystyle{ \begin{align}
\hat{S}_x &= {\hbar \over 2} \sigma_x &
\hat{S}_y &= {\hbar \over 2} \sigma_y &
\hat{S}_z &= {\hbar \over 2} \sigma_z
\end{align} }[/math]
where [math]\displaystyle{ \begin{align} \sigma_x &= \begin{pmatrix} 0 & 1 \\ 1 & 0 \end{pmatrix} \\ \sigma_y &= \begin{pmatrix} 0 & -i \\ i & 0 \end{pmatrix} \\ \sigma_z &= \begin{pmatrix} 1 & 0 \\ 0 & -1 \end{pmatrix} \end{align} }[/math] are the Pauli matrices for spin-½ particles. |
[math]\displaystyle{ \mathbf{\hat{S}} = {\hbar \over 2} \boldsymbol{\sigma} \,\! }[/math]
where σ is the vector whose components are the Pauli matrices. |
J s = N s m | [M] [L]2 [T]−1 |
Total angular momentum | [math]\displaystyle{ \begin{align} \hat{J}_x & = \hat{L}_x + \hat{S}_x \\ \hat{J}_y & = \hat{L}_y + \hat{S}_y \\ \hat{J}_z & = \hat{L}_z + \hat{S}_z \end{align} }[/math] | [math]\displaystyle{ \begin{align} \mathbf{\hat{J}} & = \mathbf{\hat{L}} + \mathbf{\hat{S}} \\ & = -i\hbar \mathbf{r}\times\nabla + \frac{\hbar}{2}\boldsymbol{\sigma} \end{align} }[/math] | J s = N s m | [M] [L]2 [T]−1 |
Transition dipole moment (electric) | [math]\displaystyle{ \begin{align} \hat{d}_x & = q\hat{x}, & \hat{d}_y & = q\hat{y}, & \hat{d}_z & = q\hat{z} \end{align} }[/math] | [math]\displaystyle{ \mathbf{\hat{d}} = q \mathbf{\hat{r}} }[/math] | C m | [I] [T] [L] |
The procedure for extracting information from a wave function is as follows. Consider the momentum p of a particle as an example. The momentum operator in position basis in one dimension is:
Letting this act on ψ we obtain:
if ψ is an eigenfunction of [math]\displaystyle{ \hat{p} }[/math], then the momentum eigenvalue p is the value of the particle's momentum, found by:
For three dimensions the momentum operator uses the nabla operator to become:
In Cartesian coordinates (using the standard Cartesian basis vectors ex, ey, ez) this can be written;
that is:
The process of finding eigenvalues is the same. Since this is a vector and operator equation, if ψ is an eigenfunction, then each component of the momentum operator will have an eigenvalue corresponding to that component of momentum. Acting [math]\displaystyle{ \mathbf{\hat{p}} }[/math] on ψ obtains:
de:Operator (Mathematik)#Operatoren der Physik