Hermitian Operators Eigenvectors of a Hermitian operator
[Pages:7]Hermitian Operators
? Definition: an operator is said to be Hermitian if it satisfies: A=A
? Alternatively called `self adjoint' ? In QM we will see that all observable properties
must be represented by Hermitian operators
? Theorem: all eigenvalues of a Hermitian operator are real
? Proof: ? Start from Eigenvalue Eq.: A am = am am
? Take the H.c. (of both sides): am A = am! am
? Use A=A:
am A = am! am
? Combine to give:
am A am = am! am am = am am am
? Since !am |am" # 0 it follows that
am! = am
Eigenvectors of a Hermitian operator
? Note: all eigenvectors are defined only up to a multiplicative c-number constant
( ) ( ) A am = am am ! A c am = am c am
? Thus we can choose the normalization !am|am"=1
? THEOREM: all eigenvectors corresponding to distinct eigenvalues are orthogonal
? Proof: ? Start from eigenvalue equation: A am = am am
? Take H.c. with m $ n:
an A = an an
? Combine to give:
an A am = an an am = am an am
? This can be written as: (an ! am ) an am = 0
? So either am = an in which case they are not distinct, or !am|an"=0, which means the eigenvectors are orthogonal
Completeness of Eigenvectors of a Hermitian operator
? THEOREM: If an operator in an M-dimensional Hilbert space has M distinct eigenvalues (i.e. no degeneracy), then its eigenvectors form a `complete set' of unit vectors (i.e a complete `basis')
? Proof: M orthonormal vectors must span an M-dimensional space.
? Thus we can use them to form a representation of the identity operator:
Degeneracy
? Definition: If there are at least two linearly independent eigenvectors associated with the same eigenvalue, then the eigenvalue is degenerate.
? The `degree of degeneracy' of an eigenvalue is the number of linearly independent eigenvectors that are associated with it
? Let dm be the degeneracy of the mth eigenvalue ? Then dm is the dimension of the degenerate
subspace
? Example: The d=2 case
? Let's refer to the two linearly independent eigenvectors |%n" and |&n"
? There is some operator W such that for some n we have: W |%n"= %n|%n" and W | &n"= &n| &n"
? Also we choose to normalize these states: !%n|%n"=1 and ! &n| &n"=1
? Linear independence means !%n |&n" # 1.
? If they are not orthogonal (!%n |&n" # 0), we can always use Gram-Schmidt Orthogonalization to get an orthonormal set
Gram-Schmidt Orthogonalization
? Procedure:
? Let
!n ,1 " !n
? A second orthogonal vector is then
$n,2 #
!n " $n !n " $n
$n !n $n !n
? Proof:
$n ,1 $n ,2
#
$n
!n !n
" $n $n " $n $n
$n !n !n
? but
!n !n = 1
? Therefore
!n ,1 !n ,2 = 0
? Can be continued for higher degree of degeneracy
? Analogy in 3-d:
rr = erxrx + ery ry + erz rz
( ) rr = erx (erx " rr) + ery ery " rr + erz (erz " rr)
rr " erx (erx # rr) $ erx
r " ex ex r # ex
? Result: From M linearly independent degenerate
!eigenvectors we can always form M orthonormal
unit vectors which span the M-dimensional
!
d?egeIfntheirsaitsedsounbes,!pthaecnet.he eigenvectors of a Hermitian
operator form a complete basis even with degeneracy
present
Phy851/Lecture 4: Basis sets and representations
? A `basis' is a set of orthogonal unit vectors in Hilbert space
? analogous to choosing a coordinate system in 3D space
? A basis is a complete set of unit vectors that spans the state space
? Basis sets come in two flavors: `discrete' and `continuous'
? A discrete basis is what we have been considering so far. The unit vectors can be labeled by integers, e.g. {|1", |2",..., |M"}, where M can be either finite or infinite
? The number of basis vectors is either finite or `countable infinity'.
? A continuous basis is a generalization whereby the unit vectors are labeled by real numbers, e.g. {|x"}; xmin< x < xmax, where the upper and lower bounds can be either finite or infinite
? The number of basis vectors is `uncountable infinity'.
Properties of basis vectors
property
discrete
continuous
orthogonality normalization
j k = ! jk j j =1
x x! = # (x " x!) x x =!
state expansion component/ wavefunction projector
operator expansion Matrix element
" = ! j cj j cj " j !
1=! j j j
! A = j Ajk k jk Ajk ! j A k
" = ! dx x " (x)
! (x)" x !
1 = ! dx x x A = ! dxdx" x A(x, x") x"
A(x, x!)" x A x!
12 = 1
[ ]2
" dx x x = " dx dx# x x x# x# = " dx dx# x $(x % x#) x# = " dx x x
!
Example 1
? Consider the relation: ! ' = A!
? To know |' _ " or |'" you must know its components in some basis
? Here we will go from the abstract form to the specific relation between components
Abstract equation:
Project onto a single unit vector:
##" "==AA## j j##" "== j jAA##
Insert the projector:
j j##" "==!k! j jAAkk kk## k
Translate to vector notation:
!! c"cj "=j =k AAjkcjkkck k
Same procedure for continuous basis:
#" = A# x #" = x A#
x #" = $ dx" x A x" x" # #"(x) = $ dx"A(x, x")#(x")
!
Example 2: Combining different basis sets in a single expression
? Let's assume we know the components of |(" in the basis {|1",|2",|3",... }
? cj)!j|("
? Let's suppose that we only know the wavefunction
of |'" in the continuous basis {|x"} ? '(x) )!x|'"
? In addition, we only know the matrix elements of A
in the alternate continuous basis {|k"} ? A(k,k') )!k|A|k'" ? How would we compute the matrix element !(|A|'"?
" A! = " A!
= ! # j j A" j
= # $ dx " j j A x x ! j
= $ % dxdk dk# " j j k k A k# k# x x ! j
=
$
%
dxdk
dk
"
c
# j
jk
A(k, k") k" x ! (x)
j
? We see that in order to compute this number, we need the inner-products !j|k" and !k|'"
? These are the transformation coefficients to go from one basis to another
Change of Basis
? Let the sets {|1",|2",|3",...} and {|u1",|u2",|u3",...} be two different orthonormal basis sets
? Suppose we know the components of |'" in the basis {|1",|2",|3",...}, this means we know the elements {cj}:
? How do we find the components {Cj} of |'" in the alternate basis {|u1",|u2",|u3",...}
? This is easily handled with Dirac notation:
" A! = % " j j A! j
= % & dx " j j A x x ! j
= % & dxdk dk# " j j k k A k# k# x x ! j
=
%
&
dx
dk
dk
#
c
$ j
jk
A(k, k#) k# x ! (x)
j
? The change of basis is accomplished by multiplying
the original column vector by a transformation
matrix U.
The Transformation matrix
? The transformation matrix looks like this
$& u1 1 u1 2 u1 3 L!#
U
=
$ $
$$%
u2 1 u3 1
M
u2 2 u3 2
M
u2 3 L!
u3 3 M
O L!!!"
? The columns of U are the components of the old unit vectors in the new basis
? If we specify at least one basis set in physical terms, then we can define other basis sets by specifying the elements of the transformation matrix
Example: 2-D rotation
? Let's do a familiar problem using the new notation
? Consider a clockwise rotation of 2-dimensional Cartesian coordinates:
Continued
Insert projector onto `known'
basis
Summary
? Basis sets can be continuous or discrete
? The important equations are:
1=! j j j j k = ! jk
1 = ! dx x x
x x! = # (x " x!)
? Change of basis is simple with Dirac notation:
1. Write unknown quantity 2. Insert projector onto known basis 3. Evaluate the transformation matrix elements 4. Perform the required summations
Cj = uj "
=! uj k k " k
! = u j k ck j
................
................
In order to avoid copyright disputes, this page is only a partial summary.
To fulfill the demand for quickly locating and searching documents.
It is intelligent file search solution for home and business.
Related download
- vectors matrices and coordinate transformations
- vectors and matrices a
- vector norms usm
- chapter 3 vectors
- hermitian operators eigenvectors of a hermitian operator
- mechanics i j notation ucl
- openstax university physics volume i unit 1 mechanics
- transpose dot product
- directional derivatives and the gradient vector
- a review of vectors and tensors
Related searches
- what is a theme of a story
- a memoir of a family and culture in cris
- derivative of a function at a point
- calculate a percent of a total
- find a free address of a person
- a theme of a story
- responsibilities of a trustee of a trust
- a example of a letter
- definition of a derivative with a fraction
- perimeter of a triangle on a graph
- example of a profile on a person
- finding a percent of a number worksheet