Professional Documents
Culture Documents
Chem532 Notes
Chem532 Notes
Structure Theory
Francesco A. Evangelista
Emory University
2 Hückel Theory 13
2.1 Foundation of the Hückel MO Method . . . . . . . . . . . . . . . . . . . . 13
2.2 Determination of the MO coefficients . . . . . . . . . . . . . . . . . . . . . 14
2.3 The Hückel Method in Practice . . . . . . . . . . . . . . . . . . . . . . . . 16
2.4 Heteroatoms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
2.5 Pairing Theorem for Alternant Hydrocarbons (AHs) . . . . . . . . . . . . . 18
2.6 Symmetry of MOs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
2.7 MO electron density and bond order, total electron density and total bond order 22
2.8 Orbital Interactions and Symmetry . . . . . . . . . . . . . . . . . . . . . . . 26
2.9 Two-center interactions . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
2.10 Cyclization Reaction Between Two Molecules . . . . . . . . . . . . . . . . . 28
2.11 Orbital mixing rules . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29
2.12 Examples and Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
2.12.1 Molecular orbitals of H3 . . . . . . . . . . . . . . . . . . . . . . . . 31
3
4 CONTENTS
5 Hartree–Fock theory 57
5.1 The Hartree–Fock wave function . . . . . . . . . . . . . . . . . . . . . . . 57
5.1.1 Minimization of the Hartree–Fock energy functional . . . . . . . . . 57
5.1.2 Hartree–Fock equations in the canonical form . . . . . . . . . . . . 60
5.1.3 Meaning of the Hartree–Fock orbital energies. Koopman’s theorem . 60
5.1.4 Generalized, unrestricted, and restricted Hartree–Fock theory . . . . 62
5.1.5 Closed-shell restricted Hartree–Fock theory . . . . . . . . . . . . . 62
5.1.6 The Roothan equations . . . . . . . . . . . . . . . . . . . . . . . . 63
5.1.7 Outline of the Hartree–Fock–Roothan procedure . . . . . . . . . . . 65
5
6 CHAPTER 1. THE MOLECULAR HAMILTONIAN
where
h̄2 X 2
N
T̂e = − ∇i , (1.6)
2me i
h̄2 X 1
M
T̂N = − ∇2 , (1.7)
2 MA A
A
X
N X
N
1 e2
V̂ee = + , (1.8)
i j>i
4πϵ0 rij
X
N X
M
1 e2 ZA
V̂eN = − , (1.9)
i
4πϵ0 riA
A
X
M X
M
1 e2 ZA ZB
V̂NN = + , (1.10)
4πϵ0 rAB
A B>A
and the quantity rij is the distance between electrons labeled with indices i and j
q
rij = |⃗ri − ⃗rj | = (xi − xj )2 + (yi − yj )2 + (zi − zj )2 , (1.11)
1X 2 1X 1 X
N X
1 X X ZA X X ZA ZB
N M N N M M M
Ĥ = − ∇i − ∇2A + − + , (1.12)
2 i 2 MA r
i j>i ij i
riA rAB
A A A B>A
Problem 1.2.1 What does the spectrum of Ĥ look like for a diatomic molecule?
Note that the nuclear coordinates R enter in the electronic wave functions as a parameter,
hence we use the symbol “;” to separate it from the electron coordinates r. The nuclear wave
1.3. THE BORN–OPPENHEIMER APPROXIMATION 7
function is instead independent of the electron coordinates; however, this does not mean that
it will be independent from the electronic wave function (see later).
If we plug this trial state in the Schrödinger equation, we get the following equation
(T̂e + T̂N + V̂ee + V̂eN + V̂NN )Φel (r; R)χnuc (R) = EΦel (r; R)χnuc (R) (1.15)
These five operators act on the wave function in different ways. The kinetic energy operators
take derivates with respect to coordinates, while the potential operators just multiply the wave
function by a number. The kinetic energy operator for the electrons acts only on the electronic
coordinates and so we can write
T̂e Φel (r; R)χnuc (R) = χnuc (R)T̂e Φel (r; R). (1.16)
The nuclear kinetic energy operator complicates things a bit, because when we apply it to the
wave function we get three terms
1X 1
M
T̂N Φel (r; R)χnuc (R) = − ∇2 [Φel (r; R)χnuc (R)]
2 MA A
A
1 X
M
1
=− χnuc (R)∇2A Φel (r; R)
2 MA
A
(1.17)
X
M
1
− ∇A χnuc (R) · ∇A Φel (r; R)
MA
A
1X 1
M
− Φel (r; R)∇2A χnuc (R).
2 MA
A
The first and second terms on the right hand side of this equation couple the electronic and
nuclear degrees of freedom. If we neglect them and keep only the last term, we can rewrite
the Schrödinger equation as
Φel (r; R)T̂N χnuc (R) + χnuc (R)(T̂e + V̂ee + V̂eN + V̂NN )Φel (r; R) = EΦel (r; R)χnuc (R)
(1.18)
To get this expression we collected all the wave functions not affected by operators to the left
side of each term. Next we collect all the terms in the following way
χnuc (R)(T̂e + V̂ee + V̂eN + V̂NN )Φel (r; R) = Φel (r; R)(E − T̂N )χnuc (R), (1.19)
and finally divide each side by Φel (r; R)χnuc (R) to get
1 1
(T̂e + V̂ee + V̂eN + V̂NN )Φel (r; R) = (E − T̂N )χnuc (R) (1.20)
Φel (r; R) χnuc (R)
The left and right sides of this equation must be equal for any values of r and R, but since the
right side does not depend on r, both sides must depend only the variable R. Let us call this
quantity V (R). Then Eq. (1.20) corresponds to two set of conditions
(T̂e + V̂ee + V̂eN + V̂NN )Φel (r; R) = V (R)Φel (r; R), (1.21)
and
[T̂N + V (R)]χnuc (R) = Eχnuc (R). (1.22)
Eqs. (1.21) and (1.22) are coupled equations. We can find their solution by first solving
the Schrödinger equation for the electrons assuming that the nuclei are fixed in space (R is a
constant). Eq. (1.21) is equivalent to the following electronic Schrödinger equation
Note that there are multiple solutions of the electronic Schrödinger equation, so different pairs
of eigenvalues/eigenfunctions are labeled by the state index k, which can take the values 0, 1,
…. The state with k = 0 is called the ground electronic state, while states with k > 0 are
excited electronic states. The function Vk (R) is commonly known as the potential energy
surface (abbreviates as PES). Note that Vk (R) gives us the potential energy at a given molec-
ular geometry R and therefore it describes a set of surfaces in 3M + 1 dimensions (nuclear
coordinates plus energy).
Note that since for a fixed geometry V̂NN is a constant (it depends only on the relative
positions of the nuclei), we can in practice solve the following eigenvalue problem
which has the same eigenfunctions [Φel,k (r; R)] but different eigenvalues [Eel,k (R)]. The
quantity Eel,k (R) is also called the electronic energy and it is related to Vk (R) via the equa-
tion
X
M X M
ZA ZB
Vk (R) = Eel,k (R) + . (1.26)
rAB
A B>A
The concept of the potential energy surface is at the basis of many concepts used to describe
molecules like, the molecular geometry, transition states, etc. The equilibrium geometry of a
molecule in a given electronic state k, Req , is defined as the minimum of Vk (R), that is, the
gradient of Vk (R) at this geometry is zero
∂Vk (R)
= 0, (1.27)
∂R R=Req
and the curvature of Vk (R) at Req is positive (the eigenvalues of the second derivative matrix
or Hessian are all positive). Note that for a given electronic state the equilibrium geometry
may be unique or there might be several minima with the same energy. In certain cases, the
ground state might not be a bound molecular state (e.g., for dications like Li2 2+ the energy
minimum corresponds to the molecule dissociated into two Li+ atoms).
2.0
1.5
1.0
Nuclear Repulsion Energy
Energy/mHartree
0.5
0.0
Dissociation Energy (H ⋅ + H + )
−0.5
Total Energy
−1.0
Electronic Energy
−1.5
−2.0
0 1 2 3 4 5 6
r(H-H)/a.u.
Although there are 3M nuclear degrees of freedom (DOF), in absence of an external po-
tential translations and rotations of a molecule do not change the value of Vk (R). Therefore,
the number of internal degrees of freedom that change the value of Vk (R) is less than 3M .
For a non-linear molecule we have that
3M − |{z}
internal DOF = |{z} 3 − |{z}
3 = 3M − 6, (1.28)
total DOF rotations translations
3M − |{z}
internal DOF = |{z} 2 − |{z}
3 = 3M − 5. (1.29)
total DOF rotations translations
From this equation we deduce that for E1 and E2 to be degenerate, the following conditions
must be satisfied simultaneously:
that is, the states that form our basis must be degenerate and there should be no coupling.
In general, these two conditions are independent. So suppose we find a point R∗ for which
Eqs.(1.32)-(1.33) are satisfied, then we can ask what is the dimensionality of the space in which
these two states are degenerate? If we consider a system with M atoms and assume that |V |
4
In other words, |V | is only acci- is not zero for all molecular geometries,4 then the subset of molecular geometries for which
dentally zero. two potential energy surfaces intersect has dimension 3M − 6 − 2 = 3M − 8 for non-linear
molecules and 3M − 5 − 2 = 3M − 7 for linear molecules. Here we subtract 2 from the
number of degrees of freedom to account for the two constraints that must be satisfied for the
levels to be degenerate. This result tells us when a crossing can happen, and if a crossing can
happen, it tells us what is the dimension of the space of degenerate electronic energies.
In this approximation the nuclei experience a potential generated by the electrons plus the
nuclear-nuclear repulsion:
Z
Vk (R) = dr Φ∗el,k (r; R)Ĥel (R)Φel,k (r; R) + VNN (R). (1.36)
This expression is an equivalent way to rewrite Vk (R), and it emphasizes the fact that the po-
tential energy surface is the potential experienced by the nuclei after averaging the interaction
with the electrons in the state Φel,k (r; R).
Note that the solutions of the nuclear Schrödinger equation are vibrational levels of the
electronic state k. Since each electronic state has a different shape, the vibrational levels for
two electronic states are, in general, different. In this picture, the electrons react instanta-
neously to the motion of the nuclei, so the nuclei always see the electrons in a fully relaxed
state (adiabatic approximation). The BO approximation is thus valid when nuclear and elec-
tronic motion occur on different time (energy) scales. In practice, the BO approximation is
1.5. THE NUCLEAR SCHRÖDINGER EQUATION 11
accurate when electronic states are well separated from one another, and, conversely, it fails
when two or more electronic states become near-degenerate. The first corrections to the BO
3/2
approximation enter at order λ6 = m M
e
, which for the hydrogen atom are of the order
of 1.3 · 10−5 Eh .
■ Advanced topic
In deriving the nuclear Schrödinger equation we have neglected two terms that arise from the kinetic
energy operator for the nuclei. So what happens if we introduce these terms back but still assume the
BO factorized wave function for the electrons + nuclei? Let us go back to the Schrödinger equation for
the electrons and nuclei where we plugged in the factorized form of the wave function [see Eq. (1.15)]
and left project with the state Φ∗el (r; R) and integrate over r (we momentarily omit the electronic state
index k in the next few equations). This gives us
⟨Φel (r; R)| (T̂e + T̂N + V̂ee + V̂eN + V̂NN ) |Φel (r; R)⟩ χnuc (R) = Eχnuc (R). (1.37)
Using the fact that (T̂e + V̂eN + V̂ee + V̂NN )Φel (r; R) = V (R)Φel (r; R), we can simplify this equation
to
⟨Φel (r; R)| (T̂N + V (R)) |Φel (r; R)⟩ χnuc (R) = Eχnuc (R). (1.38)
The left-hand-side of this equation can be written as the sum of three terms:
1 X
M
1
− χnuc (R) ⟨Φel (r; R)|∇2A Φel (r; R)⟩ .
2 A
MA | {z }
Second-order derivative coupling
The first term appears in the nuclear Schrödinger equation that we have seen before, but the terms
in the second and third lines are new. These are the so-called diagonal corrections to the BO
approximation since they involve derivative couplings between the same electronic state. These
terms can be included in the nuclear Schrödinger equation [Eq. (1.38)] to estimate corrections to
the BO approximation. For heavy nuclei these corrections are somewhat small and so are typically
neglected. However, diagonal corrections are sometime included in highly-accurate computations.
To go beyond diagonal BO corrections one has to consider solutions to the molecular Hamilto-
nian that mix various electronic states. Since at each value of R the eigenfunctions of the electronic
Hamiltonian form a complete basis for the electrons, we can expand the full wave function (electrons
plus nuclei) as
X
∞
Ψ(r, R) = Φel,k (r; R)χnuc,k (R). (1.40)
k=0
When this equation is substituted in the Schrödinger equation it leads to a set of coupled Schrödinger
equations that determine the nuclear wave functions χnuc,k (R). Note, however, that these nuclear
wave functions are different from the ones we have encountered before.
Eq. (1.40) is the most general solution that we can write for a wave function of electrons and
nuclei, and it is required when the assumptions made in the Born–Oppenheimer approximation are
not valid. A large number of interesting and important phenomena are associated with the break
down of the adiabatic assumption and require a significantly more refined treatment than the one
discussed here. ■
We will come back to the nuclear Schrödinger equation after learning how to compute poten-
tial energy surfaces.
12 CHAPTER 1. THE MOLECULAR HAMILTONIAN
Study Questions
1. What are atomic units?
2. What is the unit of energy in atomic units?
3. What are the five terms in the molecular Hamiltonian?
4. What how many dimensions does the wave function for the molecular Hamiltonian
have?
5. What is the basis for the Born–Oppenheimer approximation?
6. What terms in the expression T̂N Φel (r; R) are neglected when deriving the Born–Oppenheimer
approximation?
7. What is the electronic Schrödinger equation?
8. What is the electronic Hamiltonian and what operators enter it?
9. What is the difference between the potential energy surface and the electronic energy?
10. Why do Eqs. (1.23) and (1.25) share the same eigenfunction?
11. How many internal degrees of freedom are there for a linear and a nonlinear molecule?
12. Can two electronic state of the same symmetry cross in a diatomic molecule? What
happens if the have the same symmetry?
13. What is the nuclear Schrödinger equation?
14. Why do the solutions of the nuclear Schrödinger equation depend on the electronic
state?
15. What are the diagonal corrections to the BO approximation?
CHAPTER 2
Hückel Theory
In the Hückel method electrons occupy molecular orbitals (MOs), ψi (r), that are written
as a linear combination of atomic orbitals (LCAO). The i-th molecular orbital ψi (r) is written
as
XN
ψi (r) = χµ (r)Cµi , (2.1)
µ
where:
• N is the number of atomic orbitals (AOs) = number of molecular orbitals (MOs) =
number of atoms that share pz orbitals
• the functions χµ (r) : R3 → R are atomic pz orbitals (also called basis functions)
13
14 CHAPTER 2. HÜCKEL THEORY
• r = (x, y, z) is the electron coordinate (we will neglect spin for now)
• the matrix C with elements (C)µi = Cµi gives the coefficient of the µ-th AO in the
i-th MO. In other words, read the coefficient of ψi from the i-th column of the matrix
C.
■Example 2.1.3
The MOs of ethylene are written as ψi (r) = χ1 (r)C1i + χ2 (r)C2i . ■
Hückel theory postulates that the MOs satisfy the following Schrödinger equation:
where
and V̂ (1) is an “effective” potential energy for an electron. In Hückel we do not assume that
V̂ (1) is known. All the integrals that enter in the theory are parameters that are adjusted to
match experiment.
It is convenient to introduce the following integrals in the AO basis:
Z
hµµ = ⟨χµ | ĥ |χµ ⟩ = dr χ∗µ ĥ(r)χµ (r) (Coulomb integral of AO µ) (2.5)
Z
hµν = ⟨χµ | ĥ |χν ⟩ = dr χ∗µ (r)ĥχν (r) (Resonance integral between AOs µ and ν)
(2.6)
Z
Sµν = ⟨χµ |χν ⟩ = dr χ∗µ (r)χν (r)dr (Overlap integral between AOs µ and ν) (2.7)
X
N X
N
ĥ(1) |χµ ⟩ Cµi = ϵi |χµ ⟩ Cµi , (2.8)
µ µ
Multiply by ⟨χν | from the left. The left hand side becomes:
X
N X
N X
N
⟨χν | ĥ(1) |χµ ⟩ Cµi = ⟨χν | ĥ(1) |χµ ⟩ Cµi = hνµ Cµi (2.9)
µ µ µ
2.2. DETERMINATION OF THE MO COEFFICIENTS 15
X
N X
N X
N
⟨χν | ϵi |χµ ⟩ Cµi = ⟨χν |χµ ⟩ Cµi ϵi = Sνµ Cµi ϵi (2.10)
µ µ µ
Therefore we have:
X
N X
N
hνµ Cµi = Sνµ Cµi ϵi , (2.11)
µ µ
h11 h12 ... h1N C1i S11 S12 ... S1N C1i
h21 h22 ... h2N C2i S21 S22 ... S2N
C2i
.. .. .. .. .. = .. .. .. .. .. ϵi ,
. . . . . . . . . .
hN 1 hN 2 ... hN N CN i SN 1 SN 2 ... SN N CN i
(2.12)
or more compactly:
Hci = Sci ϵi , (2.13)
where ci is the column matrix
C1i
C2i
ci = . . (2.14)
..
CN i
If we combine all the columns together into the matrix C = (c1 , c2 , . . . , cN ), we can write
Eq. (2.13) as
HC = SCϵ, (2.15)
where the matrix ϵ contains all the orbital energies in the diagonal elements
ϵ1 0 . . . 0
0 ϵ2 . . . 0
ϵ=. .. . . .. (2.16)
.. . . .
0 0 ... ϵN
You only need the “connectivity” of a molecule to write down the Hückel Hamiltonian.
1. butadiene
2. naphthalene
3. azulene
In the Hückel method, the energy is given by the sum of the orbital energies (ϵi ) of all the
occupied orbitals X
E= ϵ i ni , (2.24)
i
Note that some times the eigenvalues are also expressed in the following form
ϵi = α + λi β, (2.26)
2.4 Heteroatoms
For hydrocarbons containing heteroatoms we introduce atom-specific and bond-specific ma-
trix elements.
18 CHAPTER 2. HÜCKEL THEORY
Table 2.1: Recommended values of the parameters for heteroatoms in the Hückel method
■ Example 2.5.2 | Pairing theorem for the allyl and benzyl radicals (odd alternant).
In the allyl radical the second MO (ψ2 , ϵ2 = α) is paired to itself. In the benzyl radical the fourth
MO (ψ4 , ϵ4 = α) is paired to itself. ■
Another consequence of this theorem is that the coefficients of paired orbitals are also
paired (MOs ψi and ψN −i+1 are paired). This property is best expressed by labeling alternat-
ing carbon atoms with a star (*) and the relationship
Cµ,N −i+1 = Cµi when µ = starred atom (2.31)
= −Cµi when µ = unstarred atom. (2.32)
20 CHAPTER 2. HÜCKEL THEORY
Energies
1 2(HOMO) 3(LUMO) 4
1.618034 .618034 -.618034 -1.618034 MO energies
MO coefficients
0 .371748 -.601501 .601501 -.371748 <- AO1
1 .601501 -.371748 -.371748 .601501 <- AO2
2 .601501 .371748 -.371748 -.601501 <- AO3
3 .371748 .601501 .601501 .371748 <- AO4
MO1 MO2 MO3 MO4
Note that the energies and coefficients of MOs 1 and 4 and MOs 2 and 3 are paired. Do not forget
that the sign of MO as a whole is arbitrary. i.e. ψi or −ψi are both OK. ■
Energies
1 2(SOMO) 3(LUMO)
1.414214 .000000 -1.414214 MO energies
MO coefficients
0 -.500000 -.707107 .500000 <- AO1
1 -.707107 .000000 -.707107 <- AO2
2 -.500000 .707107 .500000 <- AO3
MO1 MO2 MO3
Note that the energies and coefficients of MOs 1 and 3 are paired. ■
2.6. SYMMETRY OF MOS 21
When MOs are degenerate (i.e., two MOs have same energy), their MO coefficients can
not be uniquely determined. Any linear combination of degenerate MOs is also acceptable
MO of the same energy. Appropriate transformation among degenerate MOs is needed to
show pairing.
In this example, the MOs of the benzene molecule were transformed according to
and chosen either to be such that |C12 | is maximized, which in this case is also equivalent to C13 = 0.
Energies
22 CHAPTER 2. HÜCKEL THEORY
1 2 3 4 5 6
2.000000 1.000000 1.000000 -1.000000 -1.000000 -2.000000
MO coefficients
New Energies
1 2 3 4 5 6
2.000000 1.000000 1.000000 -1.000000 -1.000000 -2.000000
New MO coefficients
1 .408248 .577350 .000000 .577350 .000000 -.408248
2 .408248 .288675 -.500000 -.288675 .500000 .408248
3 .408248 -.288675 -.500000 -.288675 -.500000 -.408248
4 .408248 -.577350 .000000 .577350 .000000 .408248
5 .408248 -.288675 .500000 -.288675 .500000 -.408248
6 .408248 .288675 .500000 -.288675 -.500000 .408248
1 2 3 4 5 6
Shown above are two sets of orbitals for benzene. Note that orbitals 2 and 3 and 4 and 5 are de-
generate. Therefore, we are allowed to separately mix them without changing the orbital energies.
■
7 7
To simplify
∑ ∑the notation
∑ we will We can interpret the last two terms in the following way
write µ ν as µ̸=ν and
ν̸=µ
omit the superscript N . • qµi = |Cµi |2 is the probability of finding an electron in MO ψi on the atomic orbital
χµ . Therefore, we call this the electron density on AO χµ due to the MO ψi .
∗
• piµν = Cµi Cνi may be interpreted as the bond order for bond µ-ν in MO ψi (recall
that the Hückel method assumes Sµν = 0 when µ ̸= ν).
2.7. MO ELECTRON DENSITY AND BOND ORDER, TOTAL ELECTRON DENSITY AND TOTAL BOND ORDER23
■ Example 2.7.1 | Electron density for the first and second MOs of the allyl radical.
For ψ1
For ψ2
X
MO X
MO
qµ = ni |Cµi |2 = ni qµi ni = occupation (2, 1, or 0) (2.35)
i i
X
MO X
MO
∗
pµν = ni Cµi Cνi = ni piµν (2.37)
i i
Note that the orbital energy may be rewritten using the orbital density (qµi = |Cµi |2 ) and
∗
bond order (piµν = Cµi Cνi )
X X
ϵi = ⟨ψi |ĥ|ψi ⟩ = qµi αµ + 2 piµν βµν . (2.38)
µ µ<ν
and the total energy can be also expressed using the total density and bond order as
X
occ X X
E= ni ϵ i = qµ α µ + 2 pµν βµν . (2.39)
i µ µ<ν
24 CHAPTER 2. HÜCKEL THEORY
p12 = 2C11 C21 + C12 C22 = 2 × 0.5 × (−0.7071) + 0.7071 × 0 = 0.7071. (2.40)
Note that for alternant hydrocarbons the pairing theorem always gives qµ = 1. ■
Note that the 1-2 and 3-4 bonds (p12 = p34 = 0.894427) are stronger than the 2-3 bond
(p23 = 0.447214). Also, ignore the negative values for non-nearneighbor atoms. ■
Input Matrix
1 2 3 4 5 6
1 .5000
2 1.1000 .0000
3 .0000 1.0000 .0000
4 .0000 .0000 1.0000 .0000
5 .0000 .0000 .0000 1.0000 .0000
6 1.1000 .0000 .0000 .0000 1.0000 .0000
Energies
1 2 3 4 5 6
2.199322 1.206641 1.000000 -.916933 -1.000000 -1.989029
MO coefficients
1 .558416 .525891 .000000 .528265 .000000 -.364069
2 .431331 .168916 .500000 -.340234 .500000 .411900
3 .334378 -.374659 .500000 -.269119 -.500000 -.418804
4 .304074 -.620995 .000000 .586999 .000000 .421114
5 .334378 -.374659 -.500000 -.269119 .500000 -.418804
2.7. MO ELECTRON DENSITY AND BOND ORDER, TOTAL ELECTRON DENSITY AND TOTAL BOND ORDER25
Total charges:
N1: 1 - 1.177= -0.177
C2: 1 - 0.929= +0.071
C3: 1 - 1.004= -0.104
C4: 1 - 0.956= +0.044
Input Matrix
1 2 3 4 5
1 .8000
2 .9000 .0000
3 .0000 1.0000 .0000
4 .0000 .0000 1.0000 .0000
5 .9000 .0000 .0000 1.0000 .0000
Energies
1 2 3 4 5
2.120048 .920296 .618034 -1.240344 -1.618034
MO coefficients
1 -.583903 -.643479 .000000 .494968 .000000
2 -.428211 -.043004 -.601501 -.561058 -.371748
3 -.382315 .539554 -.371748 .250434 .601501
4 -.382315 .539554 .371748 .250434 -.601501
5 -.428211 -.043004 .601501 -.561058 .371748
Total charges:
N1: 2 - 1.510= +0.490
26 CHAPTER 2. HÜCKEL THEORY
γβ
µ µ
R R
Each molecular orbital ψi from Hückel theory gets stabilized by an energy amount given
by
(2) | ⟨ψi |ĥ′ |χR ⟩ |2
∆Ei = . (2.42)
ϵi − ϵR
Note that there is no summation here because there is only one perturbing state (the empty
orbital of the reagent). Here is a molecular orbital scheme that shows how the interaction of
the reagent empty orbital stabilizes the orbitals in the transition state
(2)
The numerator of the expression for ∆Ei may be easily computed as
X
⟨ψi |ĥ′ |χR ⟩ = Cµi ⟨χν |ĥ′ |χR ⟩ = Cµi ⟨χµ |ĥ′ |χR ⟩ = γβCµi . (2.43)
ν
PN
because ψi = µ χµ Cµi and only the matrix element ⟨χµ |ĥ′ |χR ⟩ is nonzero. The denom-
inator is instead given by
ϵi − ϵR = α + λi β − α = λi β, (2.44)
2.9. TWO-CENTER INTERACTIONS 27
where we assumed that the electrophile is a carbon atom (ϵR = α). Putting everything to-
gether we get
2
(2) (γβCµi )2 Cµi
∆Ei = = γ 2 β. (2.45)
λi β λi
If we sum this contribution over all the doubly occupied orbitals we get a total energy stabi-
lization equal to
X
occ
(2)
Xocc 2
Cµi E 2
2 ∆Ei = 2 γ 2 β = SR γ β. (2.46)
i i
λ i
E
This equation partitions the interaction energy into a molecule specific part SR , called the
2
superdelocalizability and γ β, a term that depends on the strength of the interaction.
The largest contribution to SR E
comes form the MO with the smallest |λi |, which is called
highest occupied molecular orbital (HOMO) and is given by the single term
2
Cµ,HOMO
2 (2.47)
λHOMO
If we ignore the denominator we can consider a simplified reactivity index, the frontier electron
density, given by
fµEl = 2Cµ,HOMO
2
. (2.48)
The most reactive atom towards an electrophile is the one with the largest value of fµEl , among
all possible values of µ. This means that the atom with the largest absolute coefficient in the
HOMO is the most reactive. One can repeat the same analysis for nucleophilic reactions, in
which case the reactivity can be connected to a corresponding frontier electron density
fµNuc = 2Cµ,LUMO
2
, (2.49)
where LUMO indicates the lowest unoccupied MO. These two indices allow to identify the
atoms (labeled with µ) with the highest reactivity towards electrophiles and nucleophiles by
just looking at the HOMO and LUMO of a molecule. For radical reactions, on considers in-
stead the average of fµEl and fRNuc
1 El
fµRad = (f + fµNuc ). (2.50)
2 µ
perturbation
γβ
µ µ′
A B
ν ν′
γβ
Assuming again that each pair of orbitals interacts with a term equal to γβ, we can write
Each orbital ψiA on A will be stabilized by the interaction with the unoccupied orbitals on
B
A,(2)
XB
virtual
| ⟨ψiA |ĥ′ |ψkB ⟩ |2
∆Ei = . (2.52)
ϵAi − ϵBk
k̸=i
Similarly, an orbital ψiB on B will be stabilized by the interaction with the unoccupied orbitals
on A by an amount
B,(2)
X B | ⟨ψ B |ĥ′ |ψ A ⟩ |2
virtual
i k
∆Ei = . (2.53)
ϵBi − ϵAk
k̸=i
Now let us suppose that the most important contribution comes from the interaction of the
HOMO on A with the LUMO on B. The numerator for this contribution is given by
A,(2) | ⟨ψHOMO
A
|ĥ′ |ψLUMO
B
⟩ |2
∆EHOMO = . (2.54)
ϵHOMO − ϵLUMO
A B
If we plug in the definition of these two orbitals we get for the numerator
X
⟨ψHOMO
A
|ĥ′ |ψ BLUMO ⟩ = ⟨χρ |ĥ′ |χσ ⟩ Cρ,HOMO
A B
Cσ,LUMO
ρσ
We can now use this result to analyze the cyclization reaction between two molecules.
For butadiene (A) the HOMO is antisymmetric, that is, it takes values with opposite signs
on the two interacting atoms
A
Cµ,HOMO = −Cν,HOMO
A
(2.56)
and this is also true for the LUMO of ethylene (B)
This interaction leads to a lowering in energy since ϵAHOMO − ϵBLUMO < 0, therefore, we say
that this interaction is allowed. In general, the interaction of orbitals with the same symmetry
is allowed. For example, the interaction of the butadiene LUMO with the ethylene HOMO
(the opposite of what we considered above) is a case where both orbitals are symmetric since
A A
Cµ,LUMO = Cν,LUMO , (2.59)
and
CµB′ ,HOMO = CνB′ ,HOMO . (2.60)
For this case the quantity δ is also nonzero
A
δ = 2Cµ,LUMO CµB′ ,HOMO (2.61)
The interaction of two ethylene molecules is an example where the HOMO and LUMO
have different symmetry
• Rule 2. When a pair of orbitals interact, they form a bonding (in-phase) orbital and an
anti-bonding (out-of-phase) orbital. The bonding orbital is lower in energy than the
antibonding orbital.
• Rule 3. When the energy difference between a pair of orbitals before mixing is small
and/or the overlap between a pair of orbitals is large, the energy separation becomes
large after mixing.
• Rule 4. When a pair of orbitals with different energies interact, the newly formed MO
has large contribution of the original orbital whose energy is closest to the newly formed
MO.
• Rule 5. The same number of new MOs are created from the original orbitals before
mixing (N MOs from N AOs).
• Rule 6. Suppose another orbital χc affects the two orbitals ϕ1 and ϕ2 , which were
formed mainly by interacting two orbitals χa and χb . The perturbation caused by the
third orbital χc causes a small change of the orbital shape of ϕ1 and ϕ2 . When the
energy of χc is higher than the energies of ϕ1 and ϕ2 , the mixing of χc with ϕ1 and ϕ2
is in-phase. When the energy of χc is lower than the energies of ϕ1 and ϕ2 , the mixing
2.12. EXAMPLES AND PROBLEMS 31
of χc into ϕ1 and ϕ2 is out-of-phase. When three orbitals are interacting, the newly
formed three MO’s can be expressed with a linear combination of the original three
orbitals. The mixing coefficients depend on the energies of the original three orbitals.
In the case of triangular H3 , let us assume that atomic hydrogen approaches on the bisecting
axis of the hydrogen molecule. Note that χa interacts only with χb since the symmetry of χc
is such that no mixing happens. The newly formed molecular orbital ψ1 is the bonding com-
bination between χa and χb , and the molecular orbital ψ2 is the corresponding anti-bonding
combination. ψ3 is the original χc .
Linear H3 (H · · · H–H)
In the linear form of H3 , an hydrogen atom approaches the molecular axis of the hydrogen
molecule. The situation becomes different from the triangular H3 case, because all orbitals χa ,
χb , and χc interact with each other. Comparing these two cases, we predict that the hydrogen
exchange reaction H2 +H prefers the linear transition-state structure since orbital interactions
lead to an overall stabilization of the energy.
32 CHAPTER 2. HÜCKEL THEORY
Problem 2.12.1 — Molecular orbitals of H4 . Construct the molecular orbitals for square
H4 and linear H4 species from the two sets of molecular orbitals of the H2 molecule. Draw
the orbital energy diagrams and discuss which structure is more stable.
Problem 2.12.2 — π MOs of benzene. Create the six π-molecular orbitals of benzene by
combining the two sets of π-orbitals of the allyl radical. The picture shows only one side
of π-plane (top view).
CHAPTER 3
3.1 Spin
So far we have neglected to account for spin. To introduce spin, we consider a set of spin
operators ŝx , ŝy , ŝz , analogous to angular momentum operators.8 The spin operators satisfy 8 We use lower case s for spin oper-
the following commutation relationship ators that act only on one particle.
This notation will help us distin-
guish from the spin operators for a
[ŝj , ŝk ] = ŝj ŝk − ŝk ŝj = ih̄εjkl ŝl with j, k, l ∈ {x, y, z}, (3.1) collection of electrons.
we can pick one component, say the z axis and introduce a basis of spin functions |s, ms ⟩ that
are simultaneous eigenfunctions of ŝ2 and ŝz . Note that for a given value of s, ms can take
values s, s − 1, s − 2, . . . , −s. The spin eigenfunctions satisfy:
In the case of electrons we have that s = 21 , and so we label the two spin functions as α
and β 9 9
Other notations are common. For
example |0⟩ and |1⟩ is commonly
1 1 used in quantum computing and
|α⟩ = | , ⟩ , (3.8) |↑⟩ and |↓⟩ in physics.
2 2
1 1
|β⟩ = | , − ⟩ . (3.9)
2 2
33
34 CHAPTER 3. MANY-ELECTRON WAVE FUNCTIONS
When talking about spin it is also convenient to introduce the raising (ŝ+ ) and lowering
(ŝ− ) spin operators. These are defined as:
ŝ+ = ŝx + iŝy (3.15)
ŝ− = ŝx − iŝy (3.16)
1
In the case of s = 2 particles, the action of ŝ+ and ŝ− on the spin eigenfunctions is
ŝ+ |α⟩ = 0, (3.17)
1 1
ŝ+ |β⟩ = |α⟩ (ms : − → ), (3.18)
2 2
1 1
ŝ− |α⟩ = |β⟩ (ms : → − ), (3.19)
2 2
ŝ− |β⟩ = 0. (3.20)
These operator raise or lower the value of ms or return zero if an eigenfunction with higher
(lower) value of ms does not exist. The ŝ2 operator may be represented using the raising and
lowering operators as
ŝ2 = ŝ+ ŝ− + ŝ2z − ŝz . (3.21)
It is also convenient to introduce many-electron generalizations of the spin operators. For
a system of N electrons, we can generalize the six spin operators that we introduced above by
taking the sum of operators acting on each particle
X
N X
N X
N
Ŝx = ŝx (i), Ŝy = ŝy (i), Ŝz = ŝz (i),
i i i
X
N X
N X
N
Ŝ 2 = ŝ2 (i), Ŝ+ = ŝ+ (i), Ŝ− = ŝ− (i),
i i i
where, for example, ŝ2 (i) means that an operator ŝ2 acts only on electron i. These operators
satisfy the same commutation relationships of their one-electron versions (the lowercase op-
erators). This implies that for N electron systems we can still find eigenstates of both Ŝ 2 and
Ŝz . Then the corresponding spin eigenfunctions will be of the form |S, MS ⟩, and these will
satisfy
Ŝ 2 |S, MS ⟩ = S(S + 1) |S, MS ⟩ , (3.22)
Ŝz |S, MS ⟩ = MS |S, MS ⟩ , MS = −S, −S + 1, . . . , S − 1, S. (3.23)
When describing a spin state with quantum number S, it is more common to refer to its
multiplicity, which is given by
3.2. SPIN ORBITALS 35
multiplicity = 2S + 1 (3.24)
States with multiplicity = 1 (S = 0) are called singlets, multiplicity = 2 (S = 1/2) are called
doublets, and so on.
where σ(ω) ∈ {α(ω), β(ω)} indicates a generic spin function. Note that if the spatial orbitals
are orthonormal, then the spin orbital basis is also orthonormal in the following sense
Z
∗
⟨ψi,σ |ψj,σ′ ⟩ = d⃗r dω ψi,σ (⃗r, ω)ψj,σ′ (⃗r, ω)
Z Z
(3.27)
= d⃗r ϕi (⃗r)ϕj (⃗r) dω σ ∗ (ω)σ ′ (ω)
∗
=δij δσσ′ .
Note that even if the spatial part of a spin orbital is the same (ψi = ψj ) the overlap integral
can still be zero if the spin functions σ and σ ′ are different.
To avoid having to write the two subscripts i and σ every time we indicated a spin orbital,
we will often collect the indices (i, σ) into one index k and specify an ordering of the spin
orbitals. For example, for basis of K spatial orbitals we can define 2K spin orbitals as:
Particles whose wave functions satisfies this condition are called fermions.
36 CHAPTER 3. MANY-ELECTRON WAVE FUNCTIONS
notice, however, that the expansions coefficients ai (x2 ) will depend on the value of x2 , in other
words, ai is a function of x2 . We can now expand ai (x2 ) using the same spin orbital basis:
X
∞
ai (x2 ) = aij ψj (x2 ), (3.34)
j
and introduced the quantity aij , which carries both the indices for the x1 and x2 expansions. Com-
bining Eqs. (3.33)–(3.34) we obtain:
X
∞
Φ(x1 , x2 ) = aij ψi (x1 )ψj (x2 ). (3.35)
ij
To find out the implication of this condition onto the coefficients aij , first perform the index inter-
change i ↔ j on the left hand side:
X
∞ X
∞
aij ψi (x1 )ψj (x2 ) = − aji ψj (x2 )ψi (x1 ), (3.37)
ij ij
and the collect terms that multiply the factor ψi (x1 )ψj (x2 ) to obtain the condition:
X
∞
(aij + aji )ψi (x1 )ψj (x2 ) = 0. (3.38)
ij
Equation (3.38) must hold for any value of ψi (x1 )ψj (x2 ), and this is possible only if
As a consequence, the diagonal elements of aij , aii , are zero. To see this set j = i. Then aii + aii =
2aii = 0, so that aii = 0. Eq. (3.39) shows that the when we expand an antisymmetric wave
function using a single-particle basis, the antisymmetry condition is reflected in the properties of the
coefficients aij . The quantity aij is a skew-symmetric or antisymmetric matrix. For N electrons,
the expansion coefficient will be an antisymmetric tensor with N indices, that is ai1 i2 i3 ···iN .
3.3. N -ELECTRON WAVE FUNCTIONS 37
Using Eq. (3.39) we may write the two-electron wave function as:
X
∞ X
∞
Φ(x1 , x2 ) = aij ψi (x1 )ψj (x2 ) + aij ψi (x1 )ψj (x2 )
i<j i>j
X
∞ X
∞
= aij ψi (x1 )ψj (x2 ) + aji ψj (x1 )ψi (x2 )
i<j j>i
(3.41)
X
∞ X
∞
= aij ψi (x1 )ψj (x2 ) − aij ψi (x2 )ψj (x1 )
i<j i<j
X
∞
= aij [ψi (x1 )ψj (x2 ) − ψi (x2 )ψj (x1 )].
i<j
The quantity ψi (x1 )ψj (x2 ) − ψi (x2 )ψj (x1 ) can also be expressed as the determinant of a matrix
since:
ψi (x1 ) ψi (x2 ) ψi (x1 ) ψi (x2 )
det ≡ = ψi (x1 )ψj (x2 ) − ψi (x2 )ψj (x1 ). (3.42)
ψj (x1 ) ψj (x2 ) ψj (x1 ) ψj (x2 )
X
∞
ψi (x1 ) ψi (x2 )
Φ(x1 , x2 ) = aij . (3.43)
ψj (x1 ) ψj (x2 )
i<j
This expansion is called a full configuration interaction (FCI) wave function. The FCI
wave function can provide the exact solution to the electronic Schrödinger equation. In prac-
tice, we always work with a finite spin orbital basis. Then we say that the FCI wave function
provides the exact solution of the Schrödinger equation within a finite basis. For a system
with N electrons and a spin orbital basis of dimension 2K, the number of determinants in the
FCI expansion is given by the number of ways to choose a subset of N electrons, disregarding
their order, from a set of 2K spin orbitals:
2K
NFCI = . (3.47)
N
38 CHAPTER 3. MANY-ELECTRON WAVE FUNCTIONS
NFCI grows very quickly with N and K and so the FCI wave function is impractical for more
than 16–18 electrons.
■ Example 3.3.2 | Hartree products vs. Slater determinants.
Let us compare a wave function that is just a product of spin orbitals with a Slater determinant for a
two-electrons system. A Hartree product is simply the product of orbitals:
The modulus square of ΦHP (x1 , x2 ) shows that the probability distribution of two electrons is just
the product of individual orbital probabilities:
|ΦHP (x1 , x2 )|2 = |ψi (x1 )|2 |ψj (x2 )|2 = Pi (x1 )Pj (x2 ). (3.49)
The first two terms are symmetrized probability distributions for individual orbitals [Pi (x1 )Pj (x2 )+
Pj (x1 )Pi (x2 )], while the third and fourth terms couple the position of electron 1 and 2. If we inte-
grate |ΦSD (x1 , x2 )|2 over the spin variables ω1 and ω2 we can derive a probability distribution that
depends only on the positions of the electrons:
Z
P (r1 , r2 ) = dω1 dω2 |ΦSD (x1 , x2 )|2 . (3.52)
We now distinguish two cases. If electrons have different spin then the probability distribution is
symmetric but not correlated:
1
P (r1 , r2 ) = [Pi (r1 )Pj (r2 ) + Pj (r1 )Pi (r2 )]. (3.53)
2
If instead the two electrons have different spin, then the cross terms [ψi∗ (r1 )ψj (r1 )ψj∗ (r2 )ψi (r2 )]
survive and introduce correlation in the probability distribution
1
P (r1 , r2 ) = [Pi (r1 )Pj (r2 ) + Pj (r1 )Pi (r2 )
2
− ϕ∗i (r1 )ϕj (r1 )ϕ∗j (r2 )ϕi (r2 ) (3.54)
− ϕ∗j (r1 )ϕi (r1 )ϕj (r2 )ϕ∗i (r2 )].
We say that a Slater determinant accounts for Fermi correlation, a consequence of the antisymme-
try of fermionic wave functions. This form of correlation prevents two electrons with same spin to
be found at the same point in space. ■
By combining the spatial orbitals with spin functions we obtain four spin orbitals:
From these spin orbitals we can form Slater determinants. We cannot pick any pair of
spin orbitals. For example, the determinant:
|ψg ψg ⟩ = 0, (3.62)
is zero because we have placed two electrons in the same spin orbital.13 We may form six 13 Just to avoid confusion: a spatial
unique determinants: orbital may accommodate two
electrons (with opposite spin) a
spin orbitals can accommodate
|Φ1 ⟩ = |ψg ψḡ ⟩ = (3.63) only one one electron.
Symmetry greatly simplifies this problem.14 The determinants |Φ1 ⟩ and |Φ4 ⟩ are gerade since 14
Recall that when an operator Ô
the product ψg ψḡ has symmetry equal to g × g = g.15 Determinants |Φ2 ⟩, |Φ3 ⟩, |Φ5 ⟩, |Φ6 ⟩ commutes with the Hamiltonian,
[Ĥ, Ô] = 0, then it is possible to
instead are ungerade. Spin (which is another symmetry) also helps to classify the determinants. find a set of states that simultane-
As shown in the table below, determinants 1–4 have MS = 0, while 5 and 6 have MS = +1 ously diagonalize Ĥ and Ô. For a
and −1, respectively. set of symmetry operations this re-
sult implies that the eigensolutions
of Ĥ may be classified according to
Determinant Occupied spin orbitals Symmetry MS the irreducible representations.
Φ1 ψg ψḡ g 0 15
Recall that for molecules that
Φ2 ψg ψū u 0 possess a center of inversion the
Φ3 ψu ψḡ u 0 (3.69) following rules apply: g × g = g, g
× u = u × g = u, u × u = g.
Φ4 ψu ψū g 0
Φ5 ψg ψu u 1
Φ6 ψḡ ψū u −1
40 CHAPTER 3. MANY-ELECTRON WAVE FUNCTIONS
Why is this classification helpful? Because determinants with different symmetry and spin
will not contribute to the same eigenfunction. We can immediately notice that there are only
two determinants (Φ1 and Φ4 ) of gerade symmetry, and both have MS = 0. Out of these two
determinants one can form eigenfunctions of the form
which will continue to have gerade symmetry and MS = 0. What is spin state (singlet, double,
triplets, etc.) corresponding to these solutions? Remember that for a system with total spin
quantum number S the allowed values of MS are −S, −S + 1, . . . , S − 1, S, in total 2S + 1.
The only value of S that is compatible with the states that we found (g, MS = 0) is S = 0,
since then MS = 0. Therefore, the states of the form C1 |Φ1 ⟩ + C4 |Φ4 ⟩ are singlet states.
From the ungerade determinants with MS = 0 we can form wave functions that look
like:
C2 |Φ2 ⟩ + C3 |Φ3 ⟩ , (3.71)
plus we have two ungerade states with MS = ±1 (|Φ5 ⟩, |Φ6 ⟩) that will not mix with any
other determinants because their symmetry is distinct from that of all other determinants.
and
|χB1s αχB1s β⟩ ≡ (3.74)
since both electrons occupy the atomic orbitals on atom A or B. The remaining contributions
describe covalent bonds in which each atomic orbital is occupied with one electrons. These
are:
|χA1s αχB1s β⟩ ≡ (3.75)
and
|χB1s αχA1s β⟩ ≡ (3.76)
All these four determinants contribute equally, so the determinant Φ1 represents a state that
has mixture of 50% covalent and 50% ionic electronic configurations. The determinant Φ4 is
also a 50/50 combination of covalent and ionic terms, but the sign of the covalent contribu-
tions is different:
Any linear combination of Φ1 and Φ4 may be used to represent covalent or ionic states because
by carefully choosing C1 and C4 it is possible to cancel the ionic or covalent contributions.
3.5. ELECTRONIC STATES OF H2 41
Where in the last step we used the antisymmetry property of Slater determinants and swapped
the two spin orbitals
|χA1s αχB1s α⟩ = − |χB1s αχA1s α⟩ . (3.79)
We also eliminated the Pauli-principle violating terms like this:
In other words, this determinant includes a form of correlation that is a consequence of the
antisymmetry of the wave function of electrons, the so-called Fermi correlation. The state
Φ5 corresponds to having each atomic orbital occupied by one alpha electron.
Finally, we will consider the spin of the determinants. Apply the operator Ŝ 2 to determi-
nant Φ5 :
This result shows that Φ5 is an eigenfunction of Ŝ 2 with eigenvalue equal to 2. In other words,
S(S + 1) = 2, which implies S = 1. Therefore, Φ5 is the component of a triplet state with
MS = +1. Since triplets are triply degenerate, there must be two other components of the
triplet. The MS = −1 component is the determinant Φ6 .
What about the MS = 0 component? It turns out that this component is contained in the
determinants Φ2 and Φ3 . Let us compute Ŝ 2 |Φ2 ⟩. We will need the quantity:
Ŝ+ Ŝ− |Φ2 ⟩ = Ŝ+ Ŝ− |ψg ψū ⟩ = Ŝ+ |ψḡ ψū ⟩ = |ψg ψū ⟩ + |ψḡ ψu ⟩, (3.85)
and
1 1
Ŝz |ψg ψū ⟩ = − |ψg ψū ⟩ = 0. (3.86)
2 2
From these results we find that:
Ŝ 2 |Φ2 ⟩ = |ψg ψū ⟩ + |ψḡ ψu ⟩ = |ψg ψū ⟩ − |ψu ψḡ ⟩ = |Φ2 ⟩ − |Φ3 ⟩ . (3.87)
This means that Φ2 is not an eigenfunction of Ŝ 2 . The fact that Ŝ 2 |Φ2 ⟩ gives back both Φ2
and Φ3 is a hint that a linear combination of these two determinants is an eigenfunction of
spin. The following result:
Ŝ 2 |Φ3 ⟩ = |Φ3 ⟩ − |Φ2 ⟩ . (3.88)
42 CHAPTER 3. MANY-ELECTRON WAVE FUNCTIONS
Note that these are not the Cartesian orbitals 2px , 2py , and 2pz , but they are related to them via
From these three orbitals we can form six spin orbitals, and construct a total of 15 determinants:
Number Determinant ML MS
1 |(2p−1 α)(2p0 α)⟩ −1 1
2 |(2p−1 α)(2p1 α)⟩ 0 1
3 |(2p0 α)(2p1 α)⟩ 1 1
4 |(2p−1 α)(2p−1 β)⟩ −2 0
5 |(2p−1 α)(2p0 β)⟩ −1 0
6 |(2p−1 α)(2p1 β)⟩ 0 0
7 |(2p0 α)(2p−1 β)⟩ −1 0
(3.94)
8 |(2p0 α)(2p0 β)⟩ 0 0
9 |(2p0 α)(2p1 β)⟩ 1 0
10 |(2p1 α)(2p−1 β)⟩ 0 0
11 |(2p1 α)(2p0 β)⟩ 1 0
12 |(2p1 α)(2p1 β)⟩ 2 0
13 |(2p−1 β)(2p0 β)⟩ −1 −1
14 |(2p−1 β)(2p1 β)⟩ 0 −1
15 |(2p0 β)(2p1 β)⟩ 1 −1
We can now represent these determinants in a grid of ML and MS values like indicated below
ML \MS −1 0 1
−2 •
−1 • •• •
(3.95)
0 • ••• •
1 • •• •
2 •
To identify the spin and symmetry of the state that originate from the 15 determinants we look for
a state with the highest possible value of L or S. Consider determinant 12, with ML = 2 and
MS = 0. This determinant arises from a state with L = 2 and S = 0. Why? Because there are
no other values of L and S that could generate a set of determinants consistent with the ones we
found. For example, if this determinant belonged to a state with L = 2 and S = 1, we would have
expected to see determinants withML = 2 and MS = −1, 0, 1 (which is not the case). The state
L = 2 and S = 0 contains five levels with ML = −2, −1, 0, 1, 2, which we can identify with the
five determinants in the middle column. These are indicated in red in the left figure below. Once
these are removed there are only 10 states left, as shown in the right figure below
ML \MS −1 0 1 ML \MS −1 0 1
−2 • −2
−1 • •• • −1 • • •
→
0 • ••• • 0 • •• •
1 • •• • 1 • • •
2 • 2
Of the states left, we then move to consider the one with ML = 1 and MS = 1. This state must
originate from a L = 1 and S = 1 state, to which correspond all the possible combinations of
ML = −1, 0, 1 and MS = −1, 0, 1. These nine combinations are indicated in the left figure below
in red.
ML \MS −1 0 1 ML \MS −1 0 1
−2 −2
−1 • • • −1
→
0 • •• • 0 •
1 • • • 1
2 2
After eliminating the nine states, we are left with one state with ML = 0 and MS = 0. This state
corresponds to the total quantum numbers L = 0 and S = 0. To summarize, we found three
groups of states: 1) a five-fold degenerate state with L = 2, S = 0, 2) a nine-fold degenerate state
with L = 1, S = 1, and a nondegerate state with L = 0, S = 0. Atomic states are given labels
44 CHAPTER 3. MANY-ELECTRON WAVE FUNCTIONS
that depend on the value of L and S. The spin state is specified by its multiplicity (2S + 1) and is
written as a superscript. This is followed by a capital letter that corresponds to the value of L. These
correspond to the symbols used to label atomic orbitals: S (L = 0), P (L = 1), D (L = 2), F (L = 3),
G (L = 4), …. Using this convention the three states that we found are labeled as follows
These orbitals have been labeled “σ” because they are linear combinations of atomic orbitals with
ml = 0. The 1σg orbital has no nodes, while 1σu has one node, and so of the two it is highest in
energy. Similarly, the 2s orbitals yield the molecular orbitals
2σg ∝ 2s + 2s
(3.97)
2σu ∝ 2s − 2s
The combination of the two 2p0 orbitals also gives sigma orbitals
In this case, the 3σu is lower in energy since it has one less node than the 3σg orbital. The combina-
tion of 2p±1 atomic orbitals gives two sets of π orbitals (named so because |ml | = 1).
1σg < 1σu < 2σg < 2σu < 3σu < 1πu < 1πg < 3σg (3.100)
All the molecular orbital shells are filled except for the 1πg , so we can focus on studying the angular
momentum/spin states that originate from the (1πg )2 configuration.
3.5. ELECTRONIC STATES OF H2 45
From these three orbitals we can form eight spin orbitals, and construct a total of 6 determinants,
which we classify according to the values of ML , MS , and their symmetry
If we organize these determinants in a table of ML and MS values we can start to identify one state
with |ML | = 2 and MS = 0.
ML \MS −1 0 1 ML \MS −1 0 1
−2 • −2
−1 −1
→
0 • •• • 0 • •• •
1 1
2 • 2
After we eliminate the first state, we are left with four determinants. This case is similar to that of the
H2 molecule, where we found a triplet state and a singlet. In this case the triplet state has |ML | = 0
and S = 1, and the remaining last state is a singlet with |ML | = 0 and S = 0.
To label the final states, we use the |ML | and S values like in the case of atomic state; however,
we use the symbols: Σ (|ML | = 0), Π (|ML | = 1), ∆ (|ML | = 2), Φ (|ML | = 3), etc. Using this
convention, the three states that we found are labeled as follows
Note that in labeling these states, we have also indicated how the sigma states behave when we apply
a reflection through a plane containing the internuclear axis. This symmetry is considered only for
Σ states since states with higher values of ML always have both values. To determine the “+/−”
symmetry of the sigma states we first consider how the orbitals behave upon reflection of a plane
perpendicular to a plane orthogonal to the x axis. We chose this plane because it only flips the y
component of the π orbitals. One can show that the corresponding symmetry operator (σ̂v ) switches
the two 1πg orbitals
and similarly σ̂v (1πg−1 ) = 1πg1 . To understand how the 3 Σg and 1 Σg states behave under this
symmetry let us consider the the MS = 1 component of the triplet, given by the determinant
|(1πg−1 α)(1πg1 α)⟩. When we apply the σ̂v operation to this determinant, we distribute this oper-
ator to all of its orbitals
σ̂v |(1πg−1 α)(1πg1 α)⟩ = |σ̂v (1πg−1 α)σ̂v (1πg1 α)⟩ = |(1πg1 α)(1πg−1 α)⟩ = − |(1πg−1 α)(1πg1 α)⟩
(3.104)
Here we get a minus sign because after the σ̂v operator switches the two spin orbitals, we swap the
two columns of the Slater determinant to get back to the original ordering. Since the triplet state has
changed sign when we applied the σ̂v operator, the state is “−” and the full symbol for the triplet state
is 3 Σ−
g . In the case of the singlet state,
1
√ |(1πg−1 α)(1πg1 β)⟩ + |(1πg1 α)(1πg−1 β)⟩ (3.105)
2
46 CHAPTER 3. MANY-ELECTRON WAVE FUNCTIONS
we find that the state does not change when we apply the σ̂v operator,
1
σ̂v √ |(1πg−1 α)(1πg1 β)⟩ + |(1πg1 α)(1πg−1 β)⟩
2
1
= √ |(1πg1 α)(1πg−1 β)⟩ + |(1πg−1 α)(1πg1 β)⟩ (3.106)
2
1
= √ |(1πg−1 α)(1πg1 β)⟩ + |(1πg1 α)(1πg−1 β)⟩
2
Study Questions
1. What is spin?
2. What is the commutation relationship for the three components of the spin operator?
3. What are the allowed s and ms values for one electron?
4. What are spin raising and lowering operators?
5. How do spin operator generalize to N particle systems?
6. What is the multiplicity of a spin state?
7. What is a spin orbital?
8. What does it mean to expand a function in a complete basis set?
9. What is special about wave functions for fermions?
10. What is a Slater determinant?
11. What is the full configuration interaction wave function?
12. How do Hartree products and Slater determinant differ?
13. What is Fermi correlation?
14. How many orbitals are there for H2 in a minimal basis set?
15. What are the electronic states of H2 in a minimal basis set?
16. Can you find the electronic states of the p1 and p3 configurations?
17. Can you find the electronic states of the d2 configuration?
CHAPTER 4
To derive the FCI equations we first have to establish the properties of the basis of Slater
determinant. Recall that a determinant may be written as
ψi1 (x1 ) ψi1 (x2 ) ··· ψi1 (xN )
1 ψi2 (x1 ) ψi2 (x2 ) ··· ψi2 (xN )
Φi1 i2 ···iN (x1 , x2 , . . . , xN ) = √ .. .. .. ..
N ! . . . .
ψi (x1 ) ψiN (x2 ) ··· ψi (xN ) (4.2)
N N
1 X
N!
=√ (−1)pπ Pπ [ψi1 (x1 )ψi2 (x2 ) · · · ψiN (xN )] ,
N! π
where π labels a permutation, (−1)pπ is the parity of the permutation π, and Pπ is a permu-
tation operator that acts on the electron coordinates.18 18
For example, for two indices
π = [(1, 2), (2, 1)] and pπ =
The overlap integral ⟨ΦI |ΦI ⟩ for a generic determinant ΦI may be written as [0, 1].
Z
1 XX
N! N!
⟨ΦI |ΦI ⟩ = (−1)pπ +pπ′ dx1 dx2 · · · dxN
N! π (4.3)
π′
∗
Pπ ψi1 (x1 ) · · · ψi∗N (xN ) Pπ [ψi1 (x1 ) · · · ψiN (xN )] .
Since the spin-orbital basis is assumed to be orthonormal, the integrals for each separate vari-
able must contain the square of the same spin orbital to get a non-zero result. This happens
when the two permutations coincide, π = π ′ , in which case the integral simplifies to:
Z
dx1 dx2 · · · dxN Pπ |ψi1 (x1 )|2 |ψi2 (x2 )|2 · · · |ψiN (xN )|2 = 1. (4.4)
47
48 CHAPTER 4. THE FULL CI EQUATIONS AND SLATER RULES
1 X
N!
N!
⟨ΦI |ΦI ⟩ = 1= = 1. (4.5)
N! π N!
To do so, plug in the definition of Slater determinants in ⟨ΦI |ΦJ ⟩, break down this integral
as a product of one-electron integrals, and then show that each term contains one integral
that is always zero.
The orthonormality condition for the Slater determinants may be summarized as:
We can now proceed to derive the FCI equations. Plugging in the FCI wave function in
the Schrödinger equation we obtain
X X
Ĥ cI |ΦI ⟩ = E cI |ΦI ⟩ . (4.8)
I I
Projecting onto the left with ⟨ΦJ | and rearranging terms we get
X X
⟨ΦJ |Ĥ|ΦI ⟩ cI = E ⟨ΦJ |ΦI ⟩ cI , (4.9)
| {z }
I I
δIJ
with the matrix elements HIJ defined as HIJ = ⟨ΦJ |Ĥ|ΦI ⟩. This is an eigenvalue problem
that can be written compactly as
HC = CE, (4.11)
where H is a matrix of dimension NFCI ×NFCI , where NFCI is the number of FCI determinants.
To solve this equation we have to derive equations for the matrix elements of the Hamilto-
nian operator HIJ = ⟨ΦJ |Ĥ|ΦI ⟩. The set of rules that determine the value of these matrix
elements are called Slater rules.
In this section we look at how can we compute matrix elements of the Hamiltonian ⟨ΦI | Ĥ |ΦJ ⟩,
where ΦI and ΦJ are two determinants built out of the same orthonormal spin orbital basis.
19
More general rules exist for 19 For convenience we will write the electronic Hamiltonians as
nonorthonormal spin orbitals, and
when ΦI and ΦJ use different set
of spin orbitals. Ĥ = VNN + Ĥ1 + Ĥ2 , (4.12)
4.2. SLATER RULES 49
where
X
M X
M
ZA ZB
VNN = (4.13)
rAB
A B>A
1 X 2 X X ZA X
N N M
Ĥ1 = − ∇i − = ĥ(i) (4.14)
2 i i
riA i A
X
N XN
1 X
N XN
Ĥ2 = = v̂(i, j). (4.15)
r
i j>i ij i j>i
In this case the only permutations that give nonzero contributions are those for which π = π ′ .
Otherwise, there would be one or more mismatched spin orbitals. For example, consider the
case of a permutation π ′ that swaps coordinates x1 ↔ x2 while π is the identity. The term
with the operator ĥ(x1 )
Z
dx1 dx2 · · · ψi∗1 (x2 )ψi∗2 (x1 ) · · · ĥ(x1 )ψi1 (x1 )ψi2 (x2 ) · · ·
Z Z
(4.18)
= dx1 ψi∗2 (x1 )ĥ(x1 )ψi1 (x1 ) × dx2 ψi∗1 (x2 )ψi2 (x2 ) × · · · = 0.
| {z }
=0
We also get zero if the permutation mixes two variables not involved with the operator ĥ as
in the case of the permutation π ′ ≡ (x2 ↔ x3 )
Z
dx1 dx2 · · · ψi∗1 (x1 )ψi∗3 (x2 ) · · · ĥ(x1 )ψi1 (x1 )ψi2 (x2 ) · · ·
Z Z
(4.19)
= dx1 ψi∗1 (x1 )ĥ(x1 )ψi1 (x1 ) × dx2 ψi∗3 (x2 )ψi2 (x2 ) × · · · = 0.
| {z }
=0
P P
Since for any permutation Pπ j ĥ(xj ) = j ĥ(xj ), we may combine the permutation
operators into a single term
N! Z
1 X
⟨ΦI | Ĥ1 |ΦI ⟩ = dx1 dx2 · · · dxN
N! π
(4.20)
X
Pπ ψi∗1 (x1 ) · · · ψi∗N (xN ) ĥ(xj ) ψi1 (x1 ) · · · ψiN (xN ) .
j
50 CHAPTER 4. THE FULL CI EQUATIONS AND SLATER RULES
1 XX X
N ! occ occ
⟨ΦI | Ĥ1 |ΦI ⟩ = ⟨ψi | ĥ |ψi ⟩ = ⟨i| ĥ |i⟩ . (4.22)
N! π i i
In the last term we introduced the abbreviated notation ⟨i| ĥ |i⟩ ≡ ⟨ψi | ĥ |ψi ⟩.
20
This is simple to see: ⟨ii∥ii⟩ = since the term that corresponds to i = j is null20 and ⟨ji||ji⟩ = ⟨ij||ij⟩.
⟨ii|ii⟩ − ⟨ii|ii⟩ = 0.
X
occ
1X
occ
⟨ΦI | Ĥ |ΦI ⟩ =VNN + ⟨i| ĥ |i⟩ + ⟨ij||ij⟩ ,
i
2 ij
(4.27)
X
occ X
occ
=VNN + ⟨i| ĥ |i⟩ + ⟨ij||ij⟩ ,
i i<j
where each sum is over the set of occupied molecular orbitals in ΦI [also denoted as occ(ΦI )].
|ψ1 ⟩ . (4.28)
where we have eliminated two-electron integrals that are zero due to a mismatch in the spin function.
■
To simplify this notation, sometimes we will use the Coulomb integral (Jij )
Z Z
|ψi (x1 )|2 |ψj (x2 )|2 |ϕi (r1 )|2 |ϕj (r2 )|2
Jij = ⟨ij|ij⟩ = dx1 dx2 = dx1 dx2 , (4.32)
|r1 − r2 | |r1 − r2 |
which is just Coulomb’s law for a two charges with density distribution ρi (r) = |ϕi (r)|2
and ρj (r) = |ϕj (r)|2 . We can also define the exchange integral (Kij ) as
Z
ψi∗ (x1 )ψj∗ (x1 )ψj (x2 )ψi (x2 )
Kij = ⟨ij|ji⟩ = dx1 dx2 . (4.33)
|r1 − r2 |
X
occ(ΦI )
1 X
occ(ΦI )
⟨ΦI | Ĥ |ΦI ⟩ = VN N + ⟨i| ĥ |i⟩ + (Jij − Kij ). (4.34)
i
2 ij
where the spin orbital ψi is replaced by ψa . In this case we say that the two determinant have
been brought to have maximum coincidence in the order of the spin orbitals. The matrix
elements between |ΦI ⟩ and |ΦJ ⟩ is given by
X
occ(ΦI )
⟨ΦI | Ĥ |ΦJ ⟩ = ⟨i| ĥ |a⟩ + ⟨ik||ak⟩ . (4.37)
k
This expression contains contributions from both the one- and two-electron operator.
52 CHAPTER 4. THE FULL CI EQUATIONS AND SLATER RULES
The matrix elements between |ΦI ⟩ and |ΦJ ⟩ is given by a single element of the two-electron
integrals
Equations (4.27), (4.37), and (4.40) are called Slater’s rules and account for all nonzero matrix
elements of a two-body Hamiltonian. When two determinants differ by more than two spin
orbitals then a matrix element is automatically zero.
Determinants that differ from Φ0 by one spin orbital, for example, by replacing ψN →
ψN +1
|ψ1 ψ2 · · · ψN −1 ψN +1 ⟩ , (4.44)
are called singly excited. We can similarly define determinants with two, three, etc. spin
orbital replaced as doubly, triply, etc. excited determinants. In physics, especially in the field
of condensed matter, occupied orbitals are also called holes while the unoccupied orbitals are
called particles. This language is consistent with the view that an excited determinant can
be seen as combination of creating holes (by removing electrons from occupied orbitals) and
particles (by adding electrons to the unoccupied orbitals). Therefore, excited determinants are
21
In band theory another conven- also referred to as particle-hole excitations.21
tion is used. Occupied orbitals are This classification of determinants allows us to rearrange the FCI wave function as
referred to as valence bands while
unoccupied ones as conduction X X X
bands. |ΨFCI ⟩ = c0 |Φ0 ⟩ + cS |ΦS ⟩ + cD |ΦD ⟩ + cT |ΦT ⟩ + . . . , (4.45)
S D T
where summations over the indices S, D, and T , indicate singly, doubly, and triply excited
determinants. For an N -electron system this series truncates when we reach at most N -tuply
excited determinants.
Often, the FCI wave function may be expressed assuming intermediate normalization.
This means that we scale the coefficients so that c0 = 1
X X X
|ΨFCI ⟩ = |Φ0 ⟩ + cS |ΦS ⟩ + cD |ΦD ⟩ + cT |ΦT ⟩ + . . . . (4.46)
S D T
4.4. STRUCTURE OF THE FCI WAVE FUNCTION AND HAMILTONIAN 53
Note that assuming this normalization of the wave function we have that
Writing this expression as E = ⟨Φ0 | Ĥ |ΨFCI ⟩ and plugging in the definition of ΨFCI in inter-
mediate normalization [Eq. (4.46)] we can express the energy in terms of only FCI coefficients
for single and double excitations
X X
E = ⟨Φ0 | Ĥ |ΨFCI ⟩ = ⟨Φ0 | Ĥ |Φ0 ⟩ + cS ⟨Φ0 | Ĥ |ΦS ⟩ + cD ⟨Φ0 | Ĥ |ΦD ⟩ . (4.50)
S D
Terms like ⟨Φ0 | Ĥ |ΦT ⟩ that arise in this expansion are zero due to Slater rules (ΦT and Φ0
differ by more than two spin orbitals!).
Applying a similar logic, we can map out the general structure of the FCI Hamiltonian
matrix. In this case several blocks will be zero due to difference in more than two spin orbitals.
For example the terms that couple the singly (ΦS ) and quadruply (ΦQ ) excited determinants
because ΦS and ΦQ differ by more than two spin orbitals. The general structure of the Hamil-
tonian matrix is indicated below
0 S D T Q
0
S
D
H= (4.52)
T
As you can see, only the diagonal blocks and determinant connected by two or less substitu-
tions are nonzero. This means that the FCI matrix is very sparse. To be more precise, for
each row of the matrix, there will be as many elements as the number of singly and doubly
substituted determinants, which is of the order of (neglecting spin and spatial symmetry)
1
N (2K − N ) + N 2 (2K − N )2 . (4.53)
4
Since the number of rows contains of the order of 2K
N elements, we can see that the number
of nonzero elements divided by the number of matrix elements is
N (2K − N ) + 14 N 2 (2K − N )2
fraction of nonzero elements = 2K
. (4.54)
N
This quantity becomes smaller as N and K increase showing that H is mostly empty.
54 CHAPTER 4. THE FULL CI EQUATIONS AND SLATER RULES
c = c̃ + δc. (4.55)
c̃T Hc̃
Ẽ = (4.56)
c̃T c̃
and write E = Ẽ + δE, where δE is the energy error with respect to the eigenvalue E.
Substitute these quantity in the Schrödinger equation (Hc = Ec) gives
The inverse (H − Ẽ)−1 can be simplified by retaining only the diagonal part of H (Hd ) to get
This equation shows that if we want to improve the trial eigenvector, we just need to evaluate
an expression in which only the product, Hc̃, appears. This quantity is usually called the sigma
vector and it is indicated with σ = Hc̃. Approaches based on direct diagonalization methods
have extended the domain of FCI from several thousand determinants to several billions.
■ Advanced topic | Deriving the FCI equations using the Lagrangian approach.
In this section we will derive the FCI equations using a different approach based on the formalism of
Lagrange multipliers. In this approach we seek to minimize the Rayleigh–Ritz functional
⟨Ψ| Ĥ |Ψ⟩
ERR [Ψ] = . (4.61)
⟨Ψ|Ψ⟩
Note that the denominator ⟨Ψ|Ψ⟩ is necessary to guarantee that the functional is valid irrespective
of the way we normalize the wave function Ψ. If the smallest eigenvalue of the Hamiltonian (E0 )
is bound, that is, E0 > −∞, then minimization of ERR [Ψ] with respect to Ψ will give the lowest
eigenvalue and eigenstate (Ψ0 ) of the Hamiltonian
∂ERR [Ψ]
= 0. (4.63)
∂cI
This approach is a bit clumsy because the expression that we want to minimize contains the quantity
⟨Ψ|Ψ⟩ in the denominator.
The Lagrangian approach offers a more direct way to find the FCI equations. To begin with, we
write a Lagrangian, that is a function that contains the quantity that we want to minimize and the
constraints that we need to satisfy.
4.5. ITERATIVE SOLUTION OF THE FCI EQUATIONS 55
The FCI Lagrangian is given by the energy minus the normalization constraint
∂L[Ψ] X
∗
= HIJ cJ − λcI , (4.65)
∂cI J
from which we conclude that the FCI vector c must satisfy the eigenvalue equation
X
HIJ cJ = λcI . (4.66)
J
The stationary point of L(x, y, λ) with respect to the variables x, y, and λ corresponds to:
∂L(x, y, λ)
= 2x − λ = 0 (4.68)
∂x
∂L(x, y, λ)
= 4y − λ = 0 (4.69)
∂y
∂L(x, y, λ)
= g(x, y) = 0. (4.70)
∂λ
From Eqs. (5.6) and (5.7) we obtain:
x = 2y, (4.71)
which combined with Eq. (5.8) gives:
2
g(2y, y) = 2y + y − 2 = 0 ⇒ y= , (4.72)
3
and x = 43 . ■
Study Questions
1. What is the value of ⟨ΦI |ΦJ ⟩ for two arbitrary Slater determinants ΦI and ΦJ ?
2. What is the equation that determines the FCI energy and eigenvector?
3. What are Slater rules?
4. What is a one-electron integral?
5. What is a two-electron integral?
6. What is the expectation value of the Hamiltonian for a Slater determinant?
7. What are Slater rules for off-diagonal terms?
8. What is a reference determinant?
9. What is an excited determinant?
10. How do we write the FCI wave function using singly, doubly, etc. excited determinants?
56 CHAPTER 4. THE FULL CI EQUATIONS AND SLATER RULES
Hartree–Fock theory
X
occ
1X
occ
E0 [{ψi }] = ⟨i| ĥ |i⟩ + ⟨ij||ij⟩ . (5.4)
i
2 ij
57
58 CHAPTER 5. HARTREE–FOCK THEORY
The stationary point of L(x, y, λ) with respect to the variables x, y, and λ corresponds to:
∂L(x, y, λ)
= 2x − λ = 0 (5.6)
∂x
∂L(x, y, λ)
= 4y − λ = 0 (5.7)
∂y
∂L(x, y, λ)
= g(x, y) = 0. (5.8)
∂λ
From Eqs. (5.6) and (5.7) we obtain:
x = 2y, (5.9)
which combined with Eq. (5.8) gives:
2
g(2y, y) = 2y + y − 2 = 0 ⇒ y= , (5.10)
3
and x = 43 . ■
22
In writing the Lagrange multi- where λji is a matrix of Lagrange multipliers.22 Note that λji must be a Hermitian matrix23 ,
plier we chose for convenience to which is a consequence of requiring the Lagrangian to be a real function.24 To find the mini-
put the index “q” before “p”. This
notation will be convenient later in mum of L[{ψi }] we require that the Lagrangian is stationary with respect to small variations
the derivation. of the spin orbitals {δψi }. This mean that for any arbitrary variation of the spin orbitals {δψi }
23
That is λji = λ∗ji . the Lagrangian must satisfy:
24
A complex Lagrangian cannot be
minimized because there is no such
L[{ψi + δψi }] = L[{ψi }], (5.13)
a thing as the minimum of a com-
plex function. up to first order in {δψi }. Let us first evaluate the variation of the energy:
X
occ
E0 [{ψi + δψi }] = ⟨ψi + δψi | ĥ |ψi + δψi ⟩
i
1X
occ
+ ⟨(ψi + δψi )(ψj + δψj )||(ψi + δψi )(j + δψj )⟩
2 ij
X
occ X
occ
=E0 [{ψi }] + ⟨ψi | ĥ |δψi ⟩ + ⟨δψi | ĥ |ψi ⟩
i i
1X 1X
occ occ
+ ⟨δψi ψj ||ψi ψj ⟩ + ⟨ψi δψj ||ψi ψj ⟩ (5.14)
2 ij 2 ij
1X 1X
occ occ
+ ⟨ψi ψj ||δψi ψj ⟩ + ⟨ψi ψj ||ψi δψj ⟩
2 ij 2 ij
+ (higher order)
X
occ X
occ
=E0 [{ψi }] + ⟨δψi | ĥ |ψi ⟩ + ⟨δψi ψj ||ψi ψj ⟩
i j
Pocc
The variation of the constraint − ij λji (⟨ψp |ψq ⟩ − δpq ) is equal to:
X
occ X
occ
− λji (⟨ψi + δψi |ψj + δψj ⟩ − δij ) = − λji (⟨ψi |ψj ⟩ − δij )
ij ij
X
occ X
occ
(5.15)
− λji ⟨δψi |ψj ⟩ − λji ⟨ψi |δψj ⟩
ij ij
+ (higher order).
Let us identify the change in the Lagrangian up to linear terms, that is:
δL = L[{ψi + δψi }] − L[{ψi }] (5.16)
X
occ X
occ X
occ
δL = ⟨δψi | ĥ |ψi ⟩ + ⟨δψi ψj ||ψi ψj ⟩ − λji ⟨δψp |ψq ⟩ + h.c. (5.17)
i j ij
or
X
occ
fˆ(x1 )ψi (x1 ) = λji ψj (x1 ). (5.23)
j
Eq. (5.23) is called the Hartree–Fock equation in noncanonical form and the operator
Pocc
fˆ(x1 ) = ĥ(x1 ) + j Jˆj (x1 ) − K̂j (x1 ) is called the Fock operator. Note, that this is
not an eigenvalue problem because on the r.h.s. we have a sum of functions. Essentially this
equations says that when we apply the Fock operator to any occupied orbital, we get back a
linear combination of occupied orbitals. This implies that if we apply the Fock operator to a
unoccupied (virtual) orbital, then we get back a linear combination of virtual orbitals.
60 CHAPTER 5. HARTREE–FOCK THEORY
25 −1
U = U† where the matrix U is unitary.25 It is possible to prove that a determinant built from the spin
orbitals ψi′ is related to the determinant in the original basis by a complex phase factor:
From this it can be seen that any property that is given by expectation value is invariant:
It is possible to prove that the Coulomb and exchange operators are invariant:
X X
Jˆj′ (x1 ) = Jˆj (x1 ), (5.27)
j j
X X
K̂j′ (x1 ) = K̂j (x1 ) (5.28)
j j
and consequently the Fock operator is invariant with respect to this orbital rotation, that is,
fˆ′ = fˆ. From Eq. (5.23) it is easy to show that:
λ′ = U† λU. (5.31)
This result tells us that the Lagrange multipliers in the transformed basis are a unitary trans-
formation of the original Lagrange multipliers. If we choose the unitary transformation in
such a way that λ′ is diagonal, then we can rewrite the Hartree–Fock equation as:
Eq. (5.32) is the Hartree–Fock equation in canonical form. The orbitals that satisfy this
equation are called canonical Hartree–Fock MOs. In writing this equation we may assume
that we are using the set of canonical orbitals so we can drop the prime and replace λ′ii with
ϵi :
fˆ(x1 )ψi (x1 ) = ϵi ψi (x1 ). (5.33)
The quantity ϵi is called the canonical orbital energy.
It is important to note that the HF energy is not the sum of orbital energies.
Now consider a Slater determinant built from the Hartree–Fock orbitals in which one of
the spin orbitals, in this case the last one, has been removed:
−1
|ΦN
0 ⟩ = |ψ1 , ψ2 , . . . , ψN −1 ⟩ . (5.36)
−1
The energy of ΦN
0 is:
X −1 N −1
1 X
N
E0N −1 [{ψi }] = ⟨i| ĥ |i⟩ + ⟨ij||ij⟩ . (5.37)
i
2 ij
+ ⟨N N ||N N ⟩ (5.38)
X
N −1
= ⟨N | ĥ |N ⟩ + ⟨iN ||iN ⟩ = ϵN .
i
This result shows that the energy of the last occupied orbital is the energy necessary to remove
one electron (ionize) from the system. This result may be generalized to any occupied orbital.
In general we find that for a general occupied spin orbital (ψi ):
By a similar procedure it is possible to show that for a generic unoccupied orbital (ψa ):
These results are called Koopman’s theorem. Note that orbitals are assumed to be frozen,
which means that we are ignoring relaxation effects to the addition or removal of electrons.
Koopman’s theorem also ignores electron correlation, so the IP and EA will deviate from ex-
periment.
We can also investigate the energy of an electronically excited determinant. Consider for
example the orbital substitution ψi → ψa to give the determinant Φai . The energy of this
excited determinant is:
X
N
⟨Φai | Ĥ |Φai ⟩ = ⟨j| ĥ |j⟩ + ⟨a| ĥ |a⟩ − ⟨i| ĥ |i⟩
j
1X X X
N N N (5.41)
+ ⟨ij||ij⟩ + ⟨aj||aj⟩ − ⟨ij||ij⟩
2 ij j j
− ⟨ia||ia⟩ ,
62 CHAPTER 5. HARTREE–FOCK THEORY
This expression shows that the excitation energy is the difference between orbital energies mi-
nus a term that represents the Coulomb/exchange interaction of the hole created by removing
an electron from ψi and the electron in ψa .
In GHF electrons is also called a non-collinear HF because electron spins are not constrained
to be eigenfunctions of Ŝz .
Unrestricted Hartree–Fock (UHF) restricts spin orbitals to be alpha or beta. If we arrange
spin orbitals in such a way that alpha and beta spin alternate we can write the unrestricted spin
orbitals as:
UHF
ψ2p−1 (r, ω) =ϕ2p−1 (r)α(ω),
UHF
(5.44)
ψ2p (r, ω) =ϕ2p (r)β(ω),
with p = 1, . . . , K, where K is the size of the spatial orbital basis. We are free to arrange
these functions and later we will use an indexing that puts alpha spin orbitals before beta spin
orbitals: (
α(ω) if p < K
ψpUHF (r, ω) = ϕp (r) , (5.45)
β(ω) if p ≥ K
with p = 1, . . . , 2K.
Restricted Hartree–Fock (RHF) theory constrains spin orbitals to have alpha or beta spin.
Spin orbitals come in pairs with identical spatial components:
RHF
ψ2p−1 (r, ω) =ϕp (r)α(ω),
RHF
(5.46)
ψ2p (r, ω) =ϕp (r)β(ω),
with p = 1, . . . , K. If the spin orbitals are arranged so that the alpha spin orbitals come before
the beta spin orbitals then:
(
RHF ϕp (r)α(ω) if p < K
ψp (r, ω) = , (5.47)
ϕp−K (r)β(ω) if p ≥ K
with p = 1, . . . , 2K.
|ΦRHF
0 ⟩ = |ϕ1 α, . . . , ϕN /2 α, ϕ1 β, . . . , ϕN /2 β⟩ . (5.48)
5.1. THE HARTREE–FOCK WAVE FUNCTION 63
Expressions for the energy and Fock matrix in terms of spatial orbitals can be derived by ex-
pressing a sum over spin orbitals in terms of sums over spatial orbitals:
X
N X
N /2
X
N /2
ψi = ϕi α + ϕi β. (5.49)
i i i
In this case the energy expression then takes the simplified form:
X
N /2
X
N /2
E0 = ⟨i| ĥ |i⟩ + ⟨ī| ĥ |ī⟩
i=1 i=1
(5.50)
1X 1X 1X 1X
N /2 N /2 N /2 N /2
⟨ij||ij⟩ + ⟨ij̄||ij̄⟩ + ⟨īj||īj⟩ + ⟨īj̄||īj̄⟩ .
2 ij 2 2 2
ij̄ īj īj̄
X
N /2
X
N /2
E0 = 2 ⟨i| ĥ |i⟩ + [2 ⟨ij|ij⟩ − ⟨ij|ji⟩]. (5.54)
i=1 ij
X
N /2
⟨i| fˆ |j⟩ = ⟨i| ĥ |j⟩ + [2 ⟨ik|jk⟩ − ⟨ik|kj⟩]. (5.55)
k
In general, the computational basis may not be orthonormal. Therefore, we write ⟨χν |χµ ⟩ =
Sνµ , where Sνµ is the overlap matrix. If we denote the matrix elements of the Fock operator
as Fνµ = ⟨χν | fˆ |χµ ⟩, we may write:
X X
Fνµ Cµi = Sνµ Cµi ϵi , (5.59)
µ µ
64 CHAPTER 5. HARTREE–FOCK THEORY
or in matrix notation
FC = SCϵ, (5.60)
where ϵ is a diagonal matrix. This is Roothan’s form of the Hartree–Fock equations. Note
that due to the presence of S, this is a generalized eigenvalue problem.
The matrix F is the Fock operator expressed in the computational basis and is given by:
X
N /2
(F)µν = fµν = ⟨χµ | ĥ |χν ⟩ + [2 ⟨χµ ϕk |χν ϕk ⟩ − ⟨χµ ϕk |ϕk χν ⟩]
k
(5.61)
X
N /2
∗
= hµν + [2 ⟨χµ χρ |χν χσ ⟩ − ⟨χµ χρ |χσ χν ⟩]Cρk Cσk .
k
X
N /2
∗
Pρσ = Cρk Cσk , (5.62)
k
Since the Fock matrix depends on P and ultimately on C, the Hartree–Fock equations are
nonlinear and require an interative solution:
How can we reduce this generalized eigenvalue problem to a normal eigenvalue problem?
Consider a transformation of the computational basis:
X
χµ → χ′µ = χν Xνµ . (5.65)
ν
The molecular orbital represented in the {χµ } and {χ′µ } basis are related via:
X X X
ϕi = χµ Cµi = χ′µ Cµi
′
= ′
χν Xνµ Cµi , (5.66)
µ µ µν
X† SX = 1, (5.68)
then we can take Roothan’s equation, multiply on the left with X† , and substitute C = XC′ to
get:
†
X
| {zFX} C′ = X † ′
| {zSX} C ϵ
F′ 1 (5.69)
F′ C′ = C′ ϵ.
those of the generalized eigenvalue problem. To solve Roothan’s equation in the orthonormal
basis we first compute F′ = X† FX, then solve the eigenvalue equation, F′ C′ = C′ ϵ, and then
obtain the eigenvector from C = XC′ .
How can we pick a transformation X such that X† SX = 1? It is easy to see that there are
several ways to choose X. In the symmetric orthogonalization procedure X is choosen to be
the (matrix) square root of S:
X = S− 2 .
1
(5.70)
This form of X fits the role since:
S− 2 SS− 2 = S− 2 S 2 = 1.
1 1 1 1
(5.71)
− 12
The matrix S may be built by first diagonalizing S with the unitary matrix U:
SU = Us, (5.72)
X = Us− 2 .
1
(5.74)
s− 2 U† SUs− 2 = s− 2 ss− 2 = 1.
1 1 1 1
(5.75)
SU = Us. (5.78)
X = Us− 2 .
1
(5.79)
P(0) = 0. (5.80)
66 CHAPTER 5. HARTREE–FOCK THEORY
10. Update of the density matrix and energy. Using the updated MO coefficient (C) we
compute an updated density matrix
X
N
(k+1) ∗
Pµν = Cµi Cνi . (5.86)
i
Using the updated density matrix we recompute the energy. The change in energy:
∆E (k+1) = E (k+1) − E (k) , (5.87)
and the change in the density matrix:
∆P(k+1) = P(k+1) − P(k) , (5.88)
are used to determine if the Hartree–Fock equation have converged. If |∆E (k+1) | ≤
ηE and ∥∆P(k+1) ∥ ≤ ηP , where ηE and ηP are user-provided convergence thresholds,
then the HF procedure is terminated. Otherwise, we go back to step 5 and build the
matrix G using the updated density matrix. This procedure is iterated until convergence.
Study Questions
1. What are the basic assumptions of the Hartree–Fock method?
2. What mathematical condition is imposed on the wave function in the Hartree–Fock
method?
3. How are the Fock, Coulomb, and exchange operators defined?
4. What is the energy expression of the Hartree–Fock wave function?
5. What is the Hartree–Fock Lagrangian?
6. What does the Hartree–Fock equation in canonical form look like?
7. What is the meaning of the Hartree–Fock orbital energies?
8. What approximations are made in Koopman’s theorem?
9. What is the difference between generalized, unrestricted, and restricted HF?
10. What are the Roothan equations?
11. What is the mathematical expression of the Roothan equations?
12. What is the overlap matrix?
CHAPTER 6
The electron density is multiplied by N so that the integral over all space is the total number
of electrons Z
dx1 ρ(x1 ) = N. (6.2)
In addition to the electron density, we can define the one-electron density matrix γ(x′1 ; x1 ):27 27 Note that the one-electron den-
sity matrix depends on two vari-
Z ables: x1 and x′1 .
′ ′ ∗
γ(x1 ; x1 ) = N dx2 · · · dxN Ψ(x1 , x2 , . . . , xN ) Ψ(x1 , x2 , . . . , xN ). (6.3)
The electron density can be obtainef from the one-electron density matrix by setting x′1 equal
to x1
ρ(x1 ) = γ(x1 ; x1 ), (6.4)
that is, the density is the diagonal component of the density matrix.
Given a one-electron operator Ô1 , defined as:
X
N
Ô1 (x) = ô(xi ), (6.5)
i
the expectation value of Ô1 may be obtained from the one-particle density matrix. To see this
let us compute the expectation value:
N Z
X
⟨Ψ| Ô1 |Ψ⟩ = dx1 · · · dxN Ψ(x1 , x2 , . . . , xN )∗ ô(xi )Ψ(x1 , x2 , . . . , xN ). (6.6)
i
Consider one element of this sum with the operator ô(xi ). It is easy to show that it may be
written in terms of the one-electron density matrix by relabeling the coordinates x1 ↔ xi ,
67
68CHAPTER 6. MOLECULAR PROPERTIES AND CHARACTERIZATION OF STATIONARY POINTS
then using the anti-symmetry of the wave function, and lastly, by introducing the auxiliary
variable x′1
Z
dx1 · · · dxN Ψ∗ (x1 , . . . , xi , . . .)ô(xi )Ψ(x1 , . . . , xi , . . .)
Z
= dx1 · · · dxN Ψ∗ (xi , . . . , x1 , . . .)ô(x1 )Ψ(xi , . . . , x1 , . . .)
Z
= dx1 · · · dxN Ψ∗ (x1 , . . . , xi , . . .)ô(x1 )Ψ(x1 , . . . , xi , . . .) (6.7)
Z Z
= dx1 ô(x1 ) dx2 · · · dxN Ψ∗ (x′1 , . . . , xi , . . .)∗ Ψ(x1 , . . . , xi , . . .)
x′1 =x1
Z
1
= dx1 [ô(x1 )γ(x′1 ; x1 )]x′ =x1 .
N 1
Here we take advantage of the fact that we can relabel electron indices, and that if we simulta-
neously permute indices in the Ψ∗ and Ψ functions the overall permutation factor is one. In
considering the action of ô(x1 ) onto the product Ψ∗ Ψ we have to be careful to consider that
ô(x1 ) only acts on the term on the right Ψ. This can be easily accounted for if we introduce
an auxiliary variable x′1 which, after the application of ô(x1 ) to Ψ∗ Ψ, we set equal to x1 . This
result allows us to write:
Z
⟨Ψ| Ô1 |Ψ⟩ = dx1 [ô(x1 )γ(x′1 ; x1 )]x′ =x1 . (6.8)
1
For a single Slater determinant it is possible to show that the density matrix is given by:
X
N
γ(x′1 ; x1 ) = ψi∗ (x′1 )ψi (x1 ), (6.9)
i
Note that this result is consistent with the energy expression for a Slater determinant. If we
substitute ô(x1 ) for the one-electron Hamiltonian ĥ we recover the first contribution to the
energy of a determinant.
The expectation value of the dipole moment operator for a single Slater determinant |Φ⟩ = |ψ1 · · · ψN ⟩
6.2. ATOMIC CHARGES AND BOND ORDER 69
is
X
N X
M
⟨Φ| µ̂ |Φ⟩ = −e ⟨Φ| ⃗ri |Φ⟩ + e ⃗ A ⟨Φ|Φ⟩
ZA R
i A
(6.12)
X
N X
M
= −e ⟨ψi | ⃗r |ψi ⟩ + e ⃗ A.
ZA R
i A
Let us expand the orbitals in the computational basis {χµ } and integrate over the spin variable
ω. Then the electronic contribution reads:
X
N X
N X
K
∗
⟨ψi | ⃗r |ψi ⟩ = ⟨χµ | ⃗r |χν ⟩ Cµi Cνi
i i µν
X
K X
N
∗
= ⟨χµ | ⃗r |χν ⟩ Cµi Cνi
µν i
| {z }
Pµν (6.13)
X
K
= ⟨χµ | ⃗r |χν ⟩ Pµν
µν
X
K
= (⃗r P∗ )µµ = Tr(⃗r P∗ ),
µ
where r is the matrix representation of the position operator in the computational basis:
Notice that Pµν is the same density matrix that we introduced in the Hartree–Fock method. ■
Another important quantity that can be extracted from the wave function is the two-
electron density (ρ2 ), defined as
Z
N (N − 1)
ρ2 (x1 , x2 ) = dx3 · · · dxN Ψ(x1 , x2 , . . . , xN )∗ Ψ(x1 , x2 , . . . , xN ). (6.15)
2
This quantity gives the probability of finding a pair of electrons in positions x1 and x2 , and
therefore contains information about the correlated motion of electrons. We will see this
quantity again when we will study density functional theory. An analogous two-body density
matrix can also be defined
Z
′ ′ N (N − 1)
γ2 (x1 , x2 ; x1 , x2 ) = dx3 · · · dxN Ψ(x′1 , x′2 , . . . , xN )∗ Ψ(x1 , x2 , . . . , xN ),
2
(6.16)
which reduces to the two-electron density when the primed coordinates are equal to the un-
primed ones
ρ2 (x1 , x2 ) = γ2 (x1 , x2 ; x1 , x2 ). (6.17)
X
N
ϕi (r) = χµ (r)Cµi . (6.18)
µ
X
N X
occ X
N X
N (6.19)
∗
= Sµν Cµi Cµi = Sµν Dνµ = (SD)µµ
µν i µν µ
| {z }
Dνµ
29 29
To simplify
∑ ∑ the notation
∑ we will
write µ ν as µ̸=ν and Using these quantities we define:
ν̸=µ
omit the superscript N .
• Total density on atom µ (qµ ):
X
MO X
MO
qµ = ni |Cµi |2 = ni qµi ni = occupation (2, 1, or 0) (6.20)
i i
From this expression we get an estimate of the gradient with an error that in now proportional
to order h2
f (x0 + h) − f (x0 − h)
f ′ (x0 ) = + O(h2 ). (6.29)
2h
Analogous formulas can be obtained to determine the energy with higher accuracy. For ex-
ample, the following equation provides the gradient approximate to order h4
g0 + h0 (x − x0 ) = 0, (6.34)
or:
g0
x = x0 − . (6.35)
h0
Thus we can find an approximation to the stationary point using the gradient and the Hessian of the
energy. If we apply this equation we are unlikely to land right away onto the stationary point, and an
iterative procedure is usually required. ■
72CHAPTER 6. MOLECULAR PROPERTIES AND CHARACTERIZATION OF STATIONARY POINTS
For a general molecule, we can derive the Newton–Raphson method by expanding the
energy V (R) as a Taylor series
X
∂V
3M
V (R) =V (R0 ) + ∆Rα
α
∂Rα R=R0
1 X ∂ 2 V
3M
+ ∆Rα ∆Rβ + . . . (6.36)
2! ∂Rα ∂Rβ R=R0
αβ
1
=V (R0 ) + gT ∆R + ∆RT F∆R + . . . ,
2
where we introduced the displacement vector (∆R)α = ∆Rα = Rα − R0,α , the gradient
vector
∂V
gα = , (6.37)
∂Rα
and the Hessian matrix (force constant):
∂2V
Fαβ = . (6.38)
∂Rα Rβ
∆R = −F−1 g, (6.40)
or
R = R0 − F−1 g. (6.41)
This procedure for optimizing the geometry is called the Newton–Raphson method. It re-
quires both the energy gradient (g) and the Hessian (F). In a typical quantum chemistry code
the Hessian might not be computed exactly. Instead, one may start from a guess and update
the Hessian with information from the gradient as one explores the potential energy surface
during an optimization.
1 X
3M
1
V (R) ≈ V (R0 ) + ∆RT F∆R = V (R0 ) + ∆Rα Fαβ ∆Rβ , (6.44)
2 2!
αβ
6.5. NORMAL COORDINATE ANALYSIS AND HARMONIC FREQUENCIES 73
1 X 1 ∂2
3M
T̂N = − 2
. (6.45)
2 α Mα ∂Rα
Unfortunately, the potential couples different atomic Cartesian coordinates together, which
prevents a direct solution of the Schrödinger equation. To solve this problem we first rewrite
the kinetic energy operators in terms of Cartesian displacements. Using the chain rule we can
show that taking the derivative with respect to a coordinate Rα is equivalent to taking the
derivative with respect to its corresponding displacement ∆Rα
∂ X ∂∆Rβ ∂ X ∂ ∂
= = δαβ = . (6.46)
∂Rα ∂Rα ∂∆Rβ ∂∆Rβ ∂∆Rα
β β
1X 1 1X
3M 3M
∂2
Ĥnuc = − 2
+ ∆Rα Fαβ ∆Rβ . (6.47)
2 α Mα ∂∆Rα 2
αβ
Next, we want to hide the mass dependence. To this end, introduce mass-weighted coordi-
nates (qα ), defined as p
qα = Mα ∆Rα . (6.48)
With this transformation the derivative with respect to ∆Rα may be written as
∂ X ∂qβ ∂ Xp ∂ p ∂
= = Mβ δαβ = Mα . (6.49)
∂∆Rα ∂∆Rα ∂qβ ∂qβ ∂qα
β β
1 X ∂2 1X
3M 3M
Ĥnuc = − + qα F̃αβ qβ , (6.50)
2 α ∂qα2 2
αβ
Fαβ
F̃αβ = p . (6.51)
M α Mβ
The last step consists in diagonalizing the mass-weighted Hessian. Consider the orthogo-
nal matrix L that diagonalizes F̃:
F̃L = Lλ. (6.54)
If we express the mass-weighted Hessian as F̃ = LλLT , we can rewrite the potential energy
term as:
1X 1X X X
3M 3M
qα Lαγ λγ Lβγ qβ = ( qα Lαγ )λγ ( qβ Lβγ ), (6.55)
2 2 γ α
αβγ β
74CHAPTER 6. MOLECULAR PROPERTIES AND CHARACTERIZATION OF STATIONARY POINTS
which suggests introducing a new set of variables, the normal modes (Qγ ):
X
Qγ = qα Lαγ . (6.56)
α
Since L is an orthogonal transformation, the structure of the kinetic energy operator does not
change when transforming the coordinate systems. Hence, we can finally write the Hamilto-
nian as
3M
1 X ∂2 1X X
3M 3M
1 ∂2 1
Ĥnuc =− + 2
λα Q α = − 2
+ λα Q α . (6.57)
2 α ∂Q2α 2 α α
2 ∂Q2α 2
Note that in the normal coordinate system the potential is a quadratic function:
X
3M
V (Q) = λα Q2α . (6.58)
α
Therefore, if all λα are positive (λα > 0), then the stationary point is a minimum. If one or
more λα < 0 we have a saddle point. If all λα < 0 then we are at a maximum. Thus, a har-
monic vibrational analysis can be used after geometry optimization to characterize the nature
of a stationary point. Note that for saddle points and maxima the concept of vibrational fre-
quency breaks down because we have to take the square root of a negative number. In this case
we talk about imaginary frequencies. Minima are said to have no imaginary frequencies.
In the normal mode coordinates, the Hamiltonian is a sum of terms of the form
1 ∂2 1
Ĥα = − + λα Q2α , (6.59)
2 ∂Q2α 2
which is the Hamiltonian for a harmonic oscillator with unity mass and force constant k = λα .
The eigenvalues for this Hamiltonian are given by
p 1
Eα(vα ) = λα + vα , (6.60)
2
where vα = 0, 1, . . . is the vibrational quantum number for the normal mode Qα . The total
nuclear energy is given by:
X 1 Xp Xp
Ev = Eα(vα ) = λα + λα v α . (6.61)
α
2 α α
The first term in this expansion is the zero-point vibrational energy (ZPVE) in the harmonic
approximation. This is the energy of the nuclei in the ground vibrational state.
The normal modes may be related to the Cartesian displacements by combining all the
coordinate transformations. Starting from the normal modes one finds that:
You are very familiar with the concept of a function. In its simplest form, a function f (x) is
a rule to map a number x to another number f (x). The hydrogen atom wave functions are a
function that maps points in R3 to a complex number.
Functionals generalize the concept of function. In a functional the input is a function, say
f (x), and the output is a number. For example, consider the following functional F :
Z b
F [ρ(x)] = dx|ρ(x)|2 , (7.1)
a
takes a generic function ρ(x) and returns the integral of its modulus square integrated in the
range [a, b]. We will often encounter functionals of the form:
Z b
F [ρ(x)] = dx f (ρ(x), ρ′ (x)), (7.2)
a
Z
F [ρ + ϵϕ] − F [ρ] δF (x)
lim = dx ϕ(x). (7.3)
ϵ→0 ϵ δρ(x)
In other words, for a small variation in the function ρ(x), δρ(x), the difference in the value
of the functional is linear in the functional derivative:
Z
δF [ρ]
δF [ρ] = F [ρ + δρ] − F [ρ] = dx δρ(x) + terms of order [δρ(x)]2 and higher.
δρ(x)
(7.4)
The central idea of DFT is to replace the complex wave function for a N electron system,
Ψ(x1 , . . . , xN ) with the electron density ρ(r). This would allow to reduce the problem of
optimizing a function of 4N variables to the simpler case of a function of 3 dimensions.
75
76 CHAPTER 7. DENSITY FUNCTIONAL THEORY
It is easy to see that one can write the energy of a system of N electrons using only the
one- and two-particle density matrices. In the case of the kinetic energy operator:
N Z
X
∗ 1 2
⟨Ψ| T̂ |Ψ⟩ = dx1 · · · dxN Ψ(x1 , . . . , xN ) − ∇i Ψ(x1 , . . . , xN )
i
2
Z
N
=− dx1 · · · dxN Ψ(x1 , . . . , xN )∗ ∇21 Ψ(x1 , . . . , xN )
2
Z Z
N
=− dx1 ∇21 dx2 · · · dxN Ψ(x′1 , . . . , xN )∗ Ψ(x1 , . . . , xN )
2 x′1 =x1
Z
1
= dx1 − ∇21 γ(x′1 ; x1 ) .
2 x′1 =x1
(7.5)
One may be tempted to try to minimize this functional over ρ1 and ρ2 . However, this solution
is not as simple as it looks. Without any constraint, the minimization of this functional gives
energies that significantly below those of FCI. The issue is the fact that by integrating out
the majority of the electrons we loose information regarding the sign of the wave function. In
other words, not all ρ1 and ρ2 that we can conceive correspond to a N -electron wave function.
One needs to augment the energy minimization with N -representability constraints, that is,
constraints that make sure that ρ1 and ρ2 come from a N -electron wave function.
7.3. THEORETICAL FOUNDATIONS OF DFT: THE HOHENBERG–KOHN THEOREMS 77
Thomas and Fermi were the first to propose a way to compute the energy of a system
of electrons via a functional of the electron density ρ(x1 ) alone. By assuming the uniform
electron gas model, they derived the following kinetic energy functional:
Z
T ≈ TTF [ρ(x1 )] = CF dr ρ5/3 (r), (7.10)
Note that T is difficult to approximate, and TTF is generally not very accurate. The Thomas–
Fermi model has problems describing molecular systems. It cannot describe bonding.
This functional was later improved by Dirac by adding electron exchange effects:
Z
Ex [ρ] = −Cx dr ρ4/3 (r). (7.12)
The combination of the Thomas–Fermi functional with Dirac exchange is usually called the
local density approximation (LDA) because it assumes a uniform electron density.
The HK theorems are both trivial and nontrivial, and they do not provide a recipe to find F [ρ].
Currently, the exact F [ρ] is unknown and it is likely that we will not be able to write a closed
form expression for it. Contrast this with the functionals based on the one- and two-electron
density matrices where the functionals are exactly known. Here we will not give a proof of the
HK theorems. In the next section we will take a look at a different proof proposed by Levy.
Let us introduce a functional that searches all possible wave functions Ψ̃ of N electrons and
minimizes the energy, subject to the constraint that the density of Ψ̃ is equal to a given density
ρ:
E[ρ] = min ⟨Ψ̃| Ĥ |Ψ̃⟩ = min ⟨Ψ̃| T̂ + V̂ee |Ψ̃⟩ + Vex [ρ]. (7.17)
Ψ̃→ρ Ψ̃→ρ
78 CHAPTER 7. DENSITY FUNCTIONAL THEORY
Note that the last term, Vex [ρ], is the only component that includes information about the
arrangement of the atoms. If we identify the first term with the universal functional:
30
Here we use the infimum, which we may write the ground state energy as:30
is similar to the concept of mini-
mum. The infimum is the great- Z
est lower bound of a set. This con- E = inf F [ρ] + drvex (r)ρ(r) . (7.19)
cept is useful, for example, when ρ
considering the positive real num-
bers R>0 = {x ∈ R : x >
0}. This set has no minimum but
Note that E[ρ] is only defined for ground-state densities obtained by minimizing the expec-
it does have a infimum, the num- tation value of an antisymmetric wave function. Thus, this approach does not suffer from the
ber 0. problem of N -representability. The Levy approach also solves the v-representability prob-
lem and guarantees that there are not other densities that are not v-representable and yield
the same ground state energy.
1X 2 X X
N N N
Ĥs = − ∇i + vs (xi ) = ĥs (xi ). (7.21)
2 i i i
Ĥs Ψs = Es Ψs , (7.22)
where Ψs is a single Slater determinant, |Ψs ⟩ = |ψ1 · · · ψN ⟩, and the orbitals ψi satisfy one-
electron Schrödinger equations:
The energy of the auxiliary system is given by the sum of the one-electron energies:
X
N
Es = ⟨Ψs | Ĥs |Ψs ⟩ = ϵi . (7.24)
i
7.5. THE KOHN–SHAM METHOD 79
X
N
ρs (x) = |ψi (x)|2 , (7.25)
i
Kohn and Sham now make the following assumption. They impose that the exact ground state
density and the density of the auxiliary system are equal
ρ = ρs . (7.27)
The above condition is satisfied by finding an appropriate external potential vs . Note, that
this does not imply that the density matrix of the real and auxiliary systems coincide, that is,
in general we will have that γ ̸= γs .
For the auxiliary we can write the energy as a functional of the Kohn–Sham orbitals:
Z
E[{ψi }] = Ts [{ψi }] + dx vs (x)ρ(x), (7.28)
X
N
1
Ts [{ψi }] = ⟨Ψs | T̂ |Ψs ⟩ = ⟨ψi | − ∇2 |ψi ⟩ . (7.29)
i
2
Note, that the exact kinetic energy T [ρ] = ⟨Ψ| T̂ |Ψ⟩ in general differs from Ts [{ψi }]. So
why do we introduce the auxiliary system? So that we can approximate T [ρ] using an ex-
pression that depends on orbitals. Functionals of the kinetic energy that contain orbitals are
significantly more accurate than approximations bases only on the electron density.
Now we can manipulate the energy expression to explicitly contain the KS kinetic energy.
Starting from the exact density functional we rewrite it in terms of quantities that are known
plus an unknown exchange-correlation functional (Exc [{ψi }]):
X
occ
LKS [{ψi }] = EKS [{ψi }] − λji (⟨ψi |ψj ⟩ − δij ). (7.31)
ij
At this point, we can derive Hartree–Fock-like equations by impose that this Lagrangian
is stationary with respect to changes in the orbitals. One new aspect in this derivation is that
80 CHAPTER 7. DENSITY FUNCTIONAL THEORY
we have to compute the variation in the exchange-correlation functional when the orbitals
are changed from {ψi } to {ψi + δψi }
XZ δExc XZ δExc δρ
Exc [{ψi + δψi }] − Exc [{ψi }] = dx ∗ δψi = dx δψi
i
δψ i i
δρ δψi∗
XZ (7.32)
= dx vxc (x)ψi (x)δψi ,
i
where we have introduced the exchange-correlation potential (vxc ), which is the functional
derivative of Exc with respect to the density.
If we impose the stationarity condition on LKS [{ψi }] we obtain the following eigenvalue
equation
Xocc
1 2
− ∇ + vex (r) + J(r) + vxc (r) ψi (x) =
ˆ λji ψj (x), (7.33)
2 j
which, after performing a rotation to the basis that diagonalizes the matrix λij yields the
canonical Kohn–Sham equations
1 2
− ∇ + vex (r) + J(r) + vxc (r) ψi (x) = ϵi ψi (x).
ˆ (7.34)
2
We can collect all the potential energy terms together and identify them with the potential of
the auxiliary system (vs )
ˆ + vxc (r).
vs (r) = vex (r) + J(r) (7.35)
where ϵx (ρ) and ϵc (ρ) are the exchange and correlation energy density. The exchange contri-
bution derived by Dirac is given by
ϵDirac
x (ρ) = −Cx ρ1/3 , (7.37)
while for the correlation part a common choice is the fuctional by Vosko–Wilk–Nusair (VWN),
ϵVWN
c , which was obtained by fitting the energy of the homogeneous electron gas computed
with highly-accurate quantum Monte Carlo methods.
Further improvements can be introduced if one considers the derivatives of the density.
A common way to introduce the first derivative is via the dimensionless quantity
|∇ρ|
x= . (7.38)
ρ
Wave function methods are a class of quantum chemistry approaches that try to build the
exact solution to the electronic Schrödinger equation in a systematic way. Most methods start
with the assumption that the Hartree–Fock solution (Φ0 ) is accurate and that the exact wave
function (Ψ) can be obtained from it by adding a small correction.
To be more rigorous we have to introduce the notion of correlation energy. One defini-
tion of electron correlation is the difference between the FCI (EFCI ) and Hartree–Fock (EHF )
energy
Ecorr = EFCI − EHF ≤ 0. (8.1)
This difference is negative because the Hartree–Fock energy is variational, that is, EHF ≥
EFCI . Note that the HF energy is exact (EHF = EFCI ) only for one electron systems, or certain
other special cases. The correlation energy is further divided into two components
• Dynamical or weak correlation. These are correlation effects that result from the
short-range Coulomb repulsion of electrons and long-range London dispersion inter-
actions.
• Static, strong, or nondynamical correlation. These are correlation effects that arise
from the interaction of near-degenerate electronic states, for example, when breaking a
bond, in excited states, in systems with coupled localized spins. In this case the HF single
determinant approximation is qualitatively incorrect and one needs to use a multiref-
erence approach.
We have already seen an example where static correlation plays an important role: the
dissociation of the H2 molecule. In this case, we saw that the ground state requires two de-
terminants to correctly describe the covalent configuration in the dissociation limit (H• +H• ).
Another case where the HF approximation fails is symmetry forbidden processes. For example
consider a reaction in which the reactant is well described by the Slater determinant
where ϕS is a symmetric orbital, while the product is described by the Slater determinant
81
82 CHAPTER 8. SINGLE-REFERENCE WAVE FUNCTION METHODS
the we write a determinant where the spin orbital ψi has been replaced with ψa as
In this notation, the indices at the bottom show the spin orbitals from which electrons have
been removed, while the indices at the top show the empty orbitals that have been occupied by
electrons. At times we may use an alternative notation based on second quantization. In this
equivalent approach, we introduce two new operators, â†p and âp , that create and annihilate an
electron in a generic spin orbital ψp . In second quantization we can write a singly-substituted
determinant from the HF reference by first removing and electron from ψi via the operator
âi and then creating an electron in ψa via â†a
where we use the convention that upper indices are read from left to right, while lower indices
are read from right to left.
Using this notation the FCI wave function may be written as
X
occ X
vir
1 X X ab ab
occ vir
1 X X abc abc
occ vir
|ΨFCI ⟩ = |Φ0 ⟩ + cai |Φai ⟩ + cij |Φij ⟩ + cijk |Φijk ⟩ + . . . .
i a
4 ij 36
ab ijk abc
(8.9)
Note that we assigned a coefficient equal to one to the HF determinant Φ0 . This way of writing
ΨFCI is called intermediate normalization. Instead of being normalized, ΨFCI is such that
the projection of the HF determinant onto ΨFCI is one, that is,
To derive equations for the energy we can simply project the Schrödinger equation on the left
with Φ0
⟨Φ0 | Ĥ |ΨFCI ⟩ = ⟨Φ0 |ΨFCI ⟩ E = E. (8.11)
Therefore, E = ⟨Φ0 | Ĥ |ΨFCI ⟩. We can rewrite this condition in a more detailed way by
plugging in the expansion of ΨFCI in terms of determinants
Xocc X
vir
1 X
occ X
vir
E = ⟨Φ0 | Ĥ |Φ0 ⟩ + cai |Φai ⟩ + ij |Φij ⟩ + . . .
cab ab
i a
4 ij ab
X
occ X
vir
1 X X ab
occ vir (8.12)
= ⟨Φ0 | Ĥ |Φ0 ⟩ + cai ⟨Φ0 | Ĥ |Φai ⟩ + cij ⟨Φ0 | Ĥ |Φab
ij ⟩ .
| {z } a
4 ij
i ab
EHF | {z }
Ecorr
Here we can directly identify the Hartree–Fock energy and the contribution due to electron
correlation. Although the FCI expansion includes triply and higher-substituted determinants,
8.1. TRUNCATED CONFIGURATION INTERACTION 83
Slater’s rules imply that the matrix elements between the reference and a generic excited deter-
ij··· ⟩), ⟨Φ0 | Ĥ |Φij··· ⟩, are nonzero only when a determinant is at most a double
minant (|Φab··· ab···
substitution with respect to Φ0 . This is what determines this expansion to terminate. This re-
sult is interesting because depending on what we want to compute, we may not need to know
the entire wave function. In the case of the energy, knowledge of only the coefficients for
the singly- and doubly-substituted determinants is sufficient to recover the exact correlation
energy. This result, however, does not mean that the FCI wave function can be approximates
only with singles and doubles, but it does suggest that such an approximation may be accurate.
In truncated CI methods, the FCI expansion is truncated at a certain excitation level n. If
we introduce the excitation operator Ĉ
1 X X ab··· † †
occ vir
Ĉk = cij··· âa âb · · · âj âi , (8.14)
(k!)2 ij···
ab···
| {z }
k-fold excitation
The simplest truncated CI approach is CI singles (CIS), which consists in the approxima-
tion Ĉ ≈ Ĉ1
occ X
X vir
|ΨCIS ⟩ = |Φ0 ⟩ + cai |Φai ⟩ . (8.16)
i a
To solve for the CI coefficients cai one has to diagonalize the Hamiltonian in the basis of the ref-
erence and all of its singly substituted determinant. Interestingly, if the reference is a Hartree–
Fock determinant, the CIS wave function one finds that the matrix elements between the
reference and any single is a matrix element of the Fock operator
X
occ
⟨Φai | Ĥ |Φ0 ⟩ = hai + ⟨ak||ik⟩ ≡ fai = ⟨ψa | fˆ |ψi ⟩ . (8.17)
k
Now recall that the Hartree–Fock equation could be written in the most general form as
X
occ
fˆ |ψi ⟩ = |ψi ⟩ λji . (8.18)
j
This equation shows that when we apply fˆ onto an occupied orbital we get back a linear
combination of orbitals that are occupied. Therefore, when we left-project a virtual orbital
ψa onto fˆ |ψi ⟩ , the matrix element is null
X
occ
⟨ψa | fˆ |ψi ⟩ = ⟨ψa |ψi ⟩ λji = 0. (8.19)
j
This result is known as Brillouin’s theorem, and it implies that for HF orbitals ⟨Φai | Ĥ |Φ0 ⟩ =
0. Consequently, the Hamiltonian matrix in CIS does not have off-diagonal blocks that cou-
ple the reference to the singles and the CIS ground state energy is simply the Hartree–Fock
energy. Although CIS does not improve the description of the Hartree–Fock solution for the
ground state, it can be used to compute excited states.
84 CHAPTER 8. SINGLE-REFERENCE WAVE FUNCTION METHODS
1 X X ab ab
occ vir
|ΨCID ⟩ = |Φ0 ⟩ + cij |Φij ⟩ . (8.20)
4 ij
ab
When both singles and doubles are included in CI, we have CI with singles and doubles (CISD).
Once can also define higher order approximations that additionally include triples (CISDT)
and quadruples (CISDTQ).
Despite its simplicity, truncated CI suffers from a major deficiency: the energy and other
properties lack size consistency, that is, they are not additive for separate fragments. Consider
two fragments, A and B. We say that a computational method is size consistent if a computa-
tion of the energy of A and B held at infinite distance [E(A · · · B)] is equal to the sum of the
energy of A [E(A)] plus the energy of B [E(B)]
For truncated CIs, in general one has that E(A · · · B) ̸= E(A) + E(B), and we therefore say
that they lack size consistency. The reason why truncated CI is not size consistent is that the
wave function is not multiplicatively separable. For FCI, if we know the wave function of
molecules A (ΨA ) and B (ΨB ), the total wave function for A · · · B should be the product
To see why this is not the case consider a CID wave function for both systems, for example,
|ΨA
CID ⟩ = (1 + Ĉ2 ) |Φ0 ⟩, and similarly for B. A CID computation on A · · · B still includes
A A
|ΨAB
CID ⟩ = (1 + Ĉ2 + Ĉ2 ) |Φ0 Φ0 ⟩ .
A B A B
(8.23)
|ΨA
CID ΨCID ⟩ = (1 + Ĉ2 )(1 + Ĉ2 ) |Φ0 Φ0 ⟩ = (1 + Ĉ2 + Ĉ2 + Ĉ2 Ĉ2 ) |Φ0 Φ0 ⟩ . (8.24)
B A B A B A B A B A B
ply substituted determinants. These quadruples are missing from the CID wave function of
A · · · B and that is why this method does not yield the correct energy. Size consistency is
a necessary requirement of any electronic structure theory, because otherwise we could not
compute consistent reaction energies. Note that corrections for the lack of size consistency in
truncated CI do exists, for example, Davidson’s size-consistency correction.
where the quantities hpq = ⟨ψp | ĥ |ψq ⟩ are matrix elements of the one-body operator while
⟨pq||rs⟩ are two-electron integrals.
In the Møller–Plesset formalism, the Hamiltonian is first partitioned into a zeroth-order
part (Ĥ0 ) plus a perturbation (V̂ ), and Ĥ0 is chosen to be the Fock operator
X
Ĥ0 = fˆ = ϵp â†p âp . (8.26)
p
8.2. MØLLER–PLESSET PERTURBATION THEORY 85
Here we assume that the orbitals have been variationally optimized via the Hartree–Fock
method and rotated to the canonical basis (so that fpq = ϵp δpq ). It follows that the per-
turbation contains one- and two-body terms
X 1X
V̂ = Ĥ − fˆ = (hpq − ϵp δpq )â†p âq + ⟨pq||rs⟩ â†p â†q âs âr . (8.27)
pq
4 pqrs
(1)
X
occ
1X
occ
E0 = ⟨Φ0 | V̂ |Φ0 ⟩ = EHF − ϵi = − ⟨ij||ij⟩ , (8.29)
i
2 ij
so that the sum of zeroth- and first-order energies is the Hartree–Fock energy
(0) (1)
E0 + E0 = ⟨Φ0 | Ĥ0 + V̂ |Φ0 ⟩ = ⟨Φ0 | Ĥ |Φ0 ⟩ = EHF . (8.30)
Recall that in standard perturbation theory, the first-order correction to the wave function
(1)
[Ψ0 ] is given by the sum
(1)
X (0)
(0)
⟨Ψk |V̂ |Ψ0 ⟩
(0)
|Ψ0 ⟩ = |Ψk ⟩ (0) (0)
(8.31)
k̸=0 E 0 − Ek
(0)
where the set of states {Ψk } are eigenstates of the zeroth-order Hamiltonian. In our case
(0)
the state that we want to improve is the Hartree–Fock determinant, that is, Ψ0 = Φ0 ,
(0)
and the other states in the set {Ψk } are all singly, doubly, etc., substituted determinants,
(0)
{Φai , Φab a
ij , . . .}. The matrix element between singly excited determinants [Ψk = Φi ] and
the reference are zero
⟨Φai |V̂ |Φ0 ⟩ = 0. (8.32)
(0) (0)
The only matrix elements ⟨Ψk |V̂ |Ψ0 ⟩ that are nonzero are those between the doubles
(0) (0)
[Ψk = Φabij ] and the reference [Ψ0 = Φ0 ]
⟨Φab
ij |V̂ |Φ0 ⟩ = ⟨ab||ij⟩ . (8.33)
(0) (0)
The denominator E0 − Ek is the difference of the zeroth-order energy of these states and
may be written as
(0) (0)
E0 − Ek = ⟨Φ0 | Ĥ0 |Φ0 ⟩ − ⟨Φab
ij | Ĥ0 |Φij ⟩ = ϵi + ϵj − ϵa − ϵb .
ab
(8.34)
This gives us the following expression for the first-order wave function correction
1 X X ab
occ vir
(1) ⟨ab||ij⟩
|Ψ0 ⟩ = |Φij ⟩ , (8.35)
4 ij ϵi + ϵj − ϵa − ϵb
ab
where in the sum we include a factor 14 to avoid double counting determinants. note that the
first-order wave function coefficients
ab,(1) ⟨ab||ij⟩
cij = , (8.36)
ϵi + ϵj − ϵa − ϵb
86 CHAPTER 8. SINGLE-REFERENCE WAVE FUNCTION METHODS
(2)
X | ⟨Ψ(0) |V̂ |Ψ(0) ⟩ |2
0 k
E0 = (0) (0)
, (8.37)
k̸=0 E0 − Ek
• The MP2 energy can be generalized to higher orders (MP3, MP4, etc.). However, this
series is not guaranteed to converge. Cases are known where the correlation energy
diverges after a certain point.
• MP perturbation theory is well suited for the weak correlation regime and cannot be
applied to bond-breaking processes. In this case it is necessary to use a multireference
perturbation theory (MRPT).
where the operator T̂ is defined as the Ĉ operator encountered in CI as the sum of one-, two-,
and higher-body operators up to order n
1 X X ab··· † †
occ vir
T̂k = tij··· âa âb · · · âj âi . (8.43)
(k!)2 ij···
ab···
| {z }
k-fold excitation
The product wave function can then be shown to be equivalent to the CC wave function for
the system A · · · B
A B A
+T̂ B
|ΨA
CC ΨCC ⟩ = e
B T̂
eT̂ |ΦA
0 Φ0 ⟩ = e
B T̂
|ΦA
0 Φ0 ⟩ = |ΨCC ⟩ ,
B AB
(8.46)
due to the fact that exp(T̂ A ) exp(T̂ B ) = exp(T̂ A + T̂ B ). Note that in general, given two
operators X̂ and Ŷ , we have that
only if X̂ and Ŷ commute, that is, [X̂, Ŷ ] = 0. Since cluster operators involve only substitu-
tions from occupied to virtual orbitals, they always commute, and in particular [T̂ A , T̂ B ] = 0.
Problem 8.3.1 Show that the single (T̂1 ) and double (T̂2 ) excitation operators defined as
follow commute, that is, [T̂1 , T̂2 ] = 0.
X
occ X
vir
T̂1 = tai â†a âi . (8.48)
i a
X
occ X
vir
† †
T̂2 = tab
ij âa âb âj âi . (8.49)
ij ab
Remember to avoid using the same dummy index in two independent quantities.
Equations for the CC energy and amplitudes are obtained by plugging in the CC wave
function into the Schrödinger equation
while the CC amplitudes are obtained from projections onto excited determinants ⟨Φab···
ij··· |
−T̂
0 = ⟨Φab···
ij··· | e ĤeT̂ |Φ0 ⟩ . (8.53)
88 CHAPTER 8. SINGLE-REFERENCE WAVE FUNCTION METHODS
How can we compute the term e−T̂ ĤeT̂ ? Well, there is a very useful formula named after
Baker–Campbell–Hausdorff (BCH) that says that this expression can be rewritten as a series
of commutators
1 1 1
e−T̂ ĤeT̂ =Ĥ + [Ĥ, T̂ ] + [[Ĥ, T̂ ], T̂ ] + [[[Ĥ, T̂ ], T̂ ], T̂ ] + [[[[Ĥ, T̂ ], T̂ ], T̂ ], T̂ ] + . . .
2! 3! 4!
∞
X
=Ĥ + [. . . [[Ĥ, T̂ ], T̂ ] . . .] .
| {z }
k=1
k nested commutators
(8.54)
This series is infinite, but because of the structure of the Hamiltonian and the fact that all
the components in the T̂ operators commute, we have that the BCH series truncates after
the four-fold commutator term, and this result is independent on the number of electrons.
This is important because it means that we can find all the terms in the CC equations and so
we do not need to approximate the expression for e−T̂ ĤeT̂ .
To develop a hierarchy of coupled cluster methods that approximate the FCI energy better
and better we can truncate the T̂ operator to some low particle number (2-4). Because we do
not use any argument from perturbation theory to come up with these schemes, we usually
think of CC theory as a nonperturbative method. The simplest CC approximation is CC
with doubles (CCD), which uses the approximation T̂ ≈ T̂2 . In this case the Schrödinger
equation reads
ĤeT̂2 |Φ0 ⟩ = ECCD eT̂2 |Φ0 ⟩ , (8.55)
which upon projection onto Φ0 gives
1 XX
occ vir
= ⟨Φ0 |Φ0 ⟩ + ⟨Φ0 |Φab (8.57)
ij ⟩ + . . . = 1,
4 ij
ab
| {z }
=0
where the orthogonality of determinants makes all contributions but the first null. Similarly,
the right hand side of Eq. (8.56) gives
⟨Φ0 | ĤeT̂2 |Φ0 ⟩ = ⟨Φ0 | Ĥ 1 + T̂2 + . . . |Φ0 ⟩
1 X X ab
occ vir
= ⟨Φ0 |Ĥ|Φ0 ⟩ + ⟨Φij |Ĥ|Φ0 ⟩
4 ij (8.58)
ab
1 XX
occ vir
= EHF + ⟨ij||ab⟩ tab
ij .
4 ij
ab
It is interesting to analyze the CCD wave function to understand its structure. Expanding the
exponential we get
1 2 1
|ΨCCD ⟩ = eT̂2 |Φ0 ⟩ = |Φ0 ⟩ + T̂2 |Φ0 ⟩ + T̂2 |Φ0 ⟩ + T̂23 |Φ0 ⟩ + . . . . (8.59)
2!
| 3!
{z }
unlinked quadruples, hextuples, …
The leading term in ΨCCD is the Hartree–Fock determinant, followed by double substitutions
generated by the T̂2 operator. However, we also see contributions involving quadruple and
8.3. COUPLED CLUSTER THEORY 89
higher substitutions that arise from the higher powers of the cluster operator (T̂22 , etc.). These
terms generate the unlinked excitations that were missing in the CID method. The cost of
CCD is proportional to O2 V 4 , where O and V are the number of occupied and virtual orbitals,
respectively.
Because add singles does not change the computational scaling, the CC with singles and
doubles (CCSD) approximation is generally preferred to CCD. Note that singles play a special
role in CC theory. Consider, for example, the CCSD wave function
where we have used the fact that T̂1 and T̂2 commute to rearrange this expression. Thouless’
theorem says that the effect of exp(T̂1 ) applied onto a determinant Φ0 is to generate another
(nonnormalized) determinant
eT̂1 |Φ0 ⟩ = C |Φ̃0 ⟩ , (8.61)
where C is a normalization constant. Therefore, the operator T̂1 in CC theory may be viewed
as creating a different Slater determinant than the reference. This observation explains the
reason why CCSD is also very insensitive to the choice of orbitals used to form the reference
determinant.
Note that the CC energy can be written in general as
1
⟨Φ0 | ĤeT̂ |Φ0 ⟩ = ⟨Φ0 | Ĥ 1 + T̂1 + T̂2 + T̂12 + . . . |Φ0 ⟩
2
occ X
X vir
1 XX
occ vir
1 XX
occ vir
= EHF + fia tai + ⟨ij||ab⟩ tab
ij + ⟨ij||ab⟩ tai tbj .
i a
4 ij 2 ij
ab ab
(8.62)
The CCSD method is generally insufficient to achieve chemical high accuracy, defined as
an error of 1 kcal mol−1 or less. To improve the accuracy of CCSD one can add triples and
arrive at the CCSDT method defined by the wave function
The cost to evaluate ECCSD(T) scales as O3 V 4 and this method yields results that are generally
of equal or better quality than CCSDT.
In perturbation theory we expand a given exact state Ψ0 and its corresponding eigenvalue
E0 as a power series in λ
∞
X
(0) (1) (2) (n)
|Ψ0 (λ)⟩ = |Ψ0 ⟩ + λ |Ψ0 ⟩ + λ |Ψ0 ⟩ + ... = λn |Ψ0 ⟩ , (8.66)
n=0
∞
X
(0) (1) (2) (n)
E0 (λ) = E0 + λE0 + λE0 + . . . = λn E0 , (8.67)
n=0
(n) (n)
where |Ψ0 ⟩ and E0 indicate a generic n-th order correction to the wave function and
energy, respectively. For convenience, we assume that the zeroth-order wave function is nor-
malized,
(0) (0)
⟨Ψ0 |Ψ0 ⟩ = 1, (8.68)
(0)
and that all corrections are orthogonal to |Ψ0 ⟩, that is
(0) (n)
⟨Ψ0 |Ψ0 ⟩ = 0, n = 1, 2, . . . . (8.69)
This implies that the wave function satisfies the intermediate normalization convention,
that is
(0)
⟨Ψ0 |Ψ0 ⟩ = 1. (8.70)
To obtain order-by-order expressions for perturbation theory we plug in the definition of
|Ψ0 (λ)⟩ and E0 (λ) into the Schrödinger equation
[Ĥ (0) + λĤ (1) ] |Ψ0 (λ)⟩ = E0 (λ) |Ψ0 (λ)⟩ , (8.71)
more explicitly
(0) (1) (0) (1) (0) (1)
(Ĥ (0) +λĤ (1) ) |Ψ0 ⟩ + λ |Ψ0 ⟩ + . . . = E0 + λE0 + . . . |Ψ0 ⟩ + λ |Ψ0 ⟩ + . . . .
(8.72)
Next, we proceed to collect terms with the same value of λ. For example, we can explicitly
identify the terms of order zero and one
(0) (1) (0) (0) (0) (1) (0) (0) (1)
Ĥ (0) |Ψ0 ⟩+λ(Ĥ (0) |Ψ0 ⟩+Ĥ (1) |Ψ0 ⟩)+. . . = E0 |Ψ0 ⟩+λ(E0 |Ψ0 ⟩+E0 |Ψ0 ⟩)+. . . .
(8.73)
This equation must be valid for every value of λ, so terms with equal power of λ must be equal
for each order of perturbation. The zeroth-order terms on both sides of this equation are
(0) (0) (0)
Ĥ (0) |Ψ0 ⟩ = E0 |Ψ0 ⟩ . (8.74)
(0)
This equation tells us two things. First, it tells us that the zeroth-order wave function |Ψ0 ⟩
(0)
must be an eigenfunction of Ĥ (0) . Second, if we apply ⟨Ψ0 | from the left we get an expression
(0)
for E0
(0) (0) (0) (0) (0) (0)
⟨Ψ0 | Ĥ (0) |Ψ0 ⟩ = ⟨Ψ0 | E0 |Ψ0 ⟩ = E0 . (8.75)
The first-order terms give the equation
(1) (0) (1) (0) (0) (1)
Ĥ (0) |Ψ0 ⟩ + Ĥ (1) |Ψ0 ⟩ = E0 |Ψ0 ⟩ + E0 |Ψ0 ⟩ . (8.76)
(0)
Applying ⟨Ψ0 | from the left to both sides we get
(0) (1) (0) (0) (1) (0) (0) (0) (0) (1)
⟨Ψ0 | Ĥ (0) |Ψ0 ⟩ + ⟨Ψ0 | Ĥ (1) |Ψ0 ⟩ = E0 ⟨Ψ0 |Ψ0 ⟩ +E0 ⟨Ψ0 |Ψ0 ⟩, (8.77)
| {z } | {z }
=1 =0
8.3. COUPLED CLUSTER THEORY 91
(0) (1)
and the last term drops because we assumed that ⟨Ψ0 |Ψ0 ⟩ = 0. The first term of this
(0) (1) (0) (0) (1)
expression is zero as well because ⟨Ψ0 | Ĥ (0) |Ψ0 ⟩ = E0 ⟨Ψ0 |Ψ0 ⟩ = 0. Therefore,
we get (after rearranging terms)
(1) (0) (0)
E0 = ⟨Ψ0 | Ĥ (1) |Ψ0 ⟩ . (8.78)
So we can compute the first-order energy correction from the zeroth-order wave function!
That is like getting something for nothing.
To derive the first-order wave function we rearrange the terms in the Schrödinger equa-
tion
(0) (1) (1) (0)
(Ĥ (0) − E0 ) |Ψ0 ⟩ = (E0 − Ĥ (1) ) |Ψ0 ⟩ . (8.79)
Next we will project this expression onto the basis of eigenfunctions of Ĥ (0) . These are the
(0)
states |Ψk ⟩ with k = 0, 1, . . . that satisfy
(0) (0) (0)
Ĥ (0) |Ψk ⟩ = Ek |Ψk ⟩ . (8.80)
After projection we need to simplify the expression a bit using Eq. (8.80)
(0) (0) (1) (0) (1) (0)
⟨Ψk | (Ĥ (0) − E0 ) |Ψ0 ⟩ = ⟨Ψk | (E0 − Ĥ (1) ) |Ψ0 ⟩
(0) (0) (0) (1) (0) (0)
⟨Ψk | (Ek − E0 ) |Ψ0 ⟩ = − ⟨Ψk | Ĥ (1) |Ψ0 ⟩ (8.81)
(0) (0) (0) (1) (0) (0)
(Ek − E0 ) ⟨Ψk |Ψ0 ⟩ = − ⟨Ψk | Ĥ (1) |Ψ0 ⟩ .
(0) (0)
Assuming that Ψ0 is not degenerate with any of the excited states Ψk , all of the terms with
(0) (0)
k ̸= 0 of the last expression can be divided by (Ek − E0 ) to obtain
(0) (0)
(0) (1) ⟨Ψk | Ĥ (1) |Ψ0 ⟩
⟨Ψk |Ψ0 ⟩ = (0) (0)
. (8.82)
E0 − Ek
(0) (1)
Note, however, that when k = 0, the term ⟨Ψ0 |Ψ0 ⟩ = 0 because we assumed it so when
we choose to use intermediate normalization.
With this information we can now construct the first-order
P∞ wave function correction. If
(1) (0) (0)
we multiply |Ψ0 ⟩ by the resolution of the identity 1̂ = k=0 |Ψk ⟩ ⟨Ψk | we can write
(0) (1)
[using Eq. (8.82) to express ⟨Ψk |Ψ0 ⟩]
∞
X
(1) (0) (0) (1)
|Ψ0 ⟩ = |Ψk ⟩ ⟨Ψk |Ψ0 ⟩
k=0
X∞ (0) (0)
⟨Ψk | Ĥ (1) |Ψ0 ⟩ (0)
= (0) (0)
|Ψk ⟩ (8.83)
k>0 E0 − Ek
X∞
(1) (0)
= Ck |Ψk ⟩ .
k>0
(0) (1)
Note that the sum excludes the term with k = 0 because ⟨Ψ0 |Ψ0 ⟩ = 0. The quantity
(0) (1)
⟨Ψk |Ψ0 ⟩ that appears in this expression is nothing else that the first-order wave function
(1) (0)
coefficient Ck for determinant |Ψk ⟩
(0) (0)
(1) ⟨Ψk | Ĥ (1) |Ψ0 ⟩
Ck = (0) (0)
. (8.84)
E0 − Ek
This quantity is proportional to the off-diagonal matrix element of the first-order Hamiltonian
(0) (0) (0) (0)
⟨Ψk | Ĥ (1) |Ψ0 ⟩ (also called a coupling term), and inversely proportional to E0 − Ek ,
which is an energy difference between two zeroth-order states.
92 CHAPTER 8. SINGLE-REFERENCE WAVE FUNCTION METHODS
(1)
The last term was obtained using the expression for |Ψ0 ⟩ we derived before.
CHAPTER 9
9.1 Choosing the right units, geometry input, and selecting the
electronic state
As you setup a computation you will have several options for providing information about
your molecule to a quantum chemistry program. The molecular geometry is usually specified
by either in xyz or zmat format. In the xyz format you provide a list of the element symbol
(AtomI) and the (x, y, z) coordinates (XI YI ZI):
Atom1 X1 Y1 Z1
Atom2 X2 Y2 Z2
...
AtomM XM YM ZM
In the zmat format, the geometry is specified via internal coordinates, that is, via bond lengths
(RIJ), bond angles (ThetaIJK), and , bond dihedral angles (PhiIJKL):
Atom1
Atom2 I R2I
Atom3 I R3I J Theta3IJ
Atom4 I R4I J Theta4IJ K Phi4IJK
Atom5 I R5I J Theta5IJ K Phi5IJK
...
Most programs assume that the (x, y, z) coordinates are given in units of Ångstrom (Å), but
keywords may be provided to switch to atomic units (bohr).
Other important inputs to quantum chemistry computations are the charge and spin mul-
tiplicity of the electronic state we are interested in studying. The charge determines the total
number of electrons, since
X
M
charge = Zα − Nα − Nβ , (9.1)
α=1
93
94CHAPTER 9. GENERAL ADVICE FOR RUNNING QUANTUM CHEMISTRY COMPUTATIONS
Next, you will have to select a basis set to perform your computations. In this case it is impor-
tant to distinguish between DFT and wave function computations.
• For DFT and Hartree–Fock computations common choices include Pople’s 6-31G* and
6-31G** basis sets but I would discourage you from using these because they are now
obsolete. Better, use the Karlsruhe basis sets (def2-SVP, def2-TZVP, def2-QZVP, def2-
TZVPD, …). For routine computations, a def2-TZVP basis should provide nearly-
converged results.
• For wave function methods like MP2, CCSD, CCSD(T), etc., Dunning’s correlation-
consistent basis sets (cc-pVXZ, with X = D, T, Q, 5) are recommended. These are
families of basis sets that systematically converge to the full basis set limit. The cc-
pVDZ basis is likely to provide reasonable, but potentially inaccurate results. A cc-
pVTZ basis is generally recommended for routine applications. For highly-accurate
benchmark results it is necessary to use the cc-pVQZ or cc-pV5Z basis sets.
• Note that to describe Rydberg states, long range interactions, and anions it is generally
recommended to add diffuse functions to you basis. For the Pople/Karlsruhe basis sets
you will see the symbols “+” and “D”, while the correlation-consistent basis sets add the
“aug-” prefix to the basis set name.
• Dunning’s basis sets (cc-pVXZ) are meant to correlate only valence electrons, so core
electrons should be frozen in computations that use them. To account for core electron
correlation it is necessary to use a core-valence set, usually indicated with the symbol
cc-pCVXZ, or the newer weighted core-valence sets (cc-pwCVXZ).
• Other variants of Dunning’s basis sets include basis extended for period 3 atoms [cc-
pV(X+d)Z], pseudopotential basis sets for heavy elements [cc-pVXZ-PP], and basis
sets for relativistic computations [cc-pVXZ-DK].
Note, that is common practice in Quantum Chemistry to cite the original refer-
ence to the basis sets employed in a computation. Crafting a basis set for a large num-
ber of elements in the periodic table takes a lot of work and the people that invested
the time doing so deserve credit.
A great resource for finding the definition of basis sets and the papers in which they were
introduced is the Basis Set Exchange website:
https://www.basissetexchange.org.
9.3. METHODS FOR GROUND-STATE COMPUTATIONS 95
• Increase the number of SCF iterations that the code is allowed to perform.
• Change the initial orbital guess. A program may have different ways to guess the initial
orbitals used in a SCF procedure.
• Change the basis set. In general large basis sets or basis sets with diffuse functions will
be more problematic to converge. Another solution is to converge the SCF procedure
with a small basis set and then to use these orbitals as guesses for a computation with a
larger basis.
96CHAPTER 9. GENERAL ADVICE FOR RUNNING QUANTUM CHEMISTRY COMPUTATIONS
• In certain cases, it is possible to use other convergence tricks. One is level shift, which
consists in shifting the energy of unoccupied orbitals. This works for certain cases (di-
radicals and triplet states). Some code may implement some more refined methods to
improve convergence, check the manual.
9.5.2 Finding the right SCF solution, symmetry, and orbital occupation
Even if the SCF procedure converges, sometimes the solution found is wrong. This can hap-
pen if you are stretching bonds, or have many diffuse functions, or if the molecule has symme-
try. When you find a wrong SCF solution you may find some strange outcomes, for example,
the energy of the LUMO may be lower than that of the HOMO, like in the calculation below
on N2 :
==> Iterations <==
Total Energy Delta E RMS |[F,P]|
@DF-RHF iter 1: -104.05276519746047 -1.04053e+02 1.84072e-01 DIIS
...
@DF-RHF iter 8: -107.84466525047402 -3.15438e-10 2.08693e-07 DIIS
Energy and wave function converged.
Doubly Occupied:
Virtual:
Doubly Occupied:
Virtual:
You can see that the final energy in this computation is significantly smaller than the previous
one (−108.953488 vs. −107.844665250 Eh ).
In other cases, changing occupation numbers is not enough and you might have to run a
stability analysis. A stability analysis is similar to a harmonic frequency analysis, but instead
of moving the nuclei around we test if the SCF solution is stable with respect to changes in the
orbitals. In a stability analysis one computes the Hessian with respect to orbital rotations. The
presence of one or more negative eigenvalues of the Hessian indicate that the SCF solution is
unstable. Some programs allow you to use the information from a stability analysis to find a
better (more stable) solution.
As a first example consider the following computation on O2
molecule h2 {
0 1
O
O 1 1.22220
symmetry c1
}
We can tell this is a singlet solution because the expectation value of Ŝ 2 is practically zero.
However, this solution is not stable as the following analysis reveals
Wavefunction unstable!
Lowest UHF->UHF stability eigenvalues:
A -0.111172
insufficient we need a method like CASSCF that uses a linear combination of determinants to
describe an electronic state.
As an example, consider H2 with a bond distance (1.5 Å) that is twice the equilibrium one.
If we run a restricted Hartree–Fock (RHF) computation we get
==> Iterations <==
* 1 -0.951333 |20>
* 2 0.308164 |02>
In the last two lines of this output we can read the coefficients associated with the determinants
included in the CASSCF procedure, where the determinants are indicated as a list of number
with the occupation of each orbital (0, +, −, or 2 for empty, single alpha electron, single beta
electron, and doubly occupied, respectively).
• Do not use an optimized geometry unless you have verified that the geometry optimiza-
tion has actually converged.
• Geometry optimization can be slow to converge, especially if the atoms have to “go
around a corner”. Since geometry optimizers usually estimate the Hessian instead of
computing the exact one, starting the optimization with the full Hessian may improve
convergence. This is particularly true of transition state searches. Most of the time, you
can significantly reduce the cost to compute the Hessian by using smaller basis or less
expensive method.
• Some molecules are hard to optimize due to the choice of optimization coordinates.
Sometimes you can improve convergence by using more complex coordinate systems.
• Once you have found a stationary point, run a frequency analysis to verify that the ge-
ometry is a local minimum or a transition state.
• Another aspect to keep in mind is that in certain codes, geometry optimization cannot
proceed from a geometry with high symmetry (e.g., D2h ) to low symmetry (e.g., C1 ). If
you want to find a minimum, it is generally a good idea to start with a guess geometry
that has no symmetry and let the optimizer determine if the minimum is symmetric.
Again, this is why it is important to run a frequency analysis after a geometry optimiza-
tion to verify the nature of the stationary point.
CHAPTER 10
In this chapter we will discuss methods can be used to treat bond break process and electron-
ically excited states.
in which both the orbitals (ϕi ) from which the determinants (ΦI ) and the determinant coeffi-
cients (CI ) are variationally optimized. Note, that in general, that they determinants employed
to expand the MCSCF wave function (understood as occupation patters of electrons) are not
changed in the optimization process.
The MCSCF energy is obtained by performing a simultaneous optimization of the orbitals
and the determinant coefficients
101
102 CHAPTER 10. MULTIREFERENCE WAVE FUNCTION METHODS
• Active orbitals (A). These orbitals have variable occupation in the CASSCF wave
function. In most cases these are chosen to be valence orbital associated with the par-
ticular aspect we want to model. For example, for to describe the wave function for the
single bond-breaking reaction A–B −−→ A+B it is sufficient to include a single pair of
A–B bonding/antibonding orbitals.
• Virtual orbitals (V). These orbitals are not occupied in the CASSCF wave function.
These are usually high-energy orbitals that are typically irrelevant to description of a
chemical reaction or an excited state.
Virtual (V)
Particle (P = A ∪ V)
e, f, . . .
a, b, . . .
Active (A)
u, v, . . .
Hole (H = C ∪ A)
i, j, . . .
Core (C)
m, n, . . .
In CASSCF one generates determinants with the core orbitals occupied and nA active elec-
trons distributed in the active orbitals. Therefore, a generic CASSCF determinant has the form
Y
C
|ΦI ⟩ = â†u â†v · · · â†z â†m |⟩ . (10.3)
| {z } m
active
In the literature, and active space in which a given number of electrons (nel ) is distributed in
a given number of active orbitals (nA ) is abbreviated with CAS(nel e,nA o). For example, an
active space with two electrons in two orbitals is denoted CAS(2e,2o).
In CASSCF, the space of determinants is complete, in the sense that all possible determi-
nants are included. One consequence of this choice is that the wave function is invariant with
respect to unitary transformations of the active orbitals. This means that if we transform the
active orbitals via a unitary transformation U
X
ψu → ψu′ = ψv Uvu , (10.4)
v
there is a corresponding set of CI coefficients CI′ such that the transformed CASSCF wave
function is identical to the original one
X X
CI |ΦI ⟩ = CI′ |Φ′I ⟩ . (10.6)
I I
This property implies that we are free to work with any unitarily transformed version of the
CASSCF active orbitals (e.g., delocalized, localized).
Note that since the CAS space includes all possible determinants, its size grows very quickly
with the number of active orbitals and electrons. If we ignore point group symmetry, we can
compute the number of determinants with a given number of alpha (nα ) and beta (nβ ) elec-
trons in the active orbitals just by counting how many arrangements of electrons are possible
for the two spin cases. For each spin case, the number of ways we can distribute n electrons is
10.2. MULTIREFERENCE CONFIGURATION INTERACTION 103
given by the number of ways we can choose n spin orbitals to occupy and taking into account
that permutations are equivalent. This quantity is given by the binomial coefficient
n n!
= , (10.7)
k (n − k)!k!
As shown in Tab. 10.1, this quantity grows very quickly, and therefore, practical computations
are typically limited to active spaces of size CAS(16e,16o).
Table 10.1: Number of determinants in a CAS space at half filling. These numbers ignore the
reduction in size due to point group symmetry.
Note that one can also define a corresponding CASCI approach in which one generated a
CAS space and performs the variational optimization of the CI coefficients without optimiz-
ing the orbitals.
where the space of external determinants is generated by applying the operators â†a âi and
â†a â†b âj âi to all the CAS determinants
EXT = {ΦA |ΦA = â†a âi |ΦI ⟩ , â†a â†b âj âi |ΦI ⟩ ∀ΦI ∈ CAS}. (10.10)
104 CHAPTER 10. MULTIREFERENCE WAVE FUNCTION METHODS
In defining the EXT space, the orbital indices i, j run over the core and active orbitals, while
a, b run over the active and virtuals. The u-MRCISD method provides very accurate results,
but it has a computational scaling that becomes quickly unmanageable when the number of
CAS determinants is large. The approximate number of u-MRCISD determinants depends on
the number of hole (NH ) and particle (NP ) orbitals and is proportional to
u-MRCISD
Ndet = dNH2 NP2 , (10.11)
which is limited both by the size of the active space (d) and the number of electrons and basis
functions used.
A more economical alternative to the u-MRCISD method is the internally-contracted
MRCISD (ic-MRCISD) method. In ic-MRCISD, one still generates all possible singles and
doubles out of a CAS; however, an excitation operator is applied to the entire CAS reference
all at once. For example, internally-contracted singly excited configurations are defined as
X
|Ψai ⟩ = â†a âi |ΨCAS ⟩ = CI â†a âi |ΦI ⟩ . (10.12)
I
The right hand side of this equation shows that |Ψai ⟩ is a linear combination of excited deter-
minants (â†a âi |ΦI ⟩) with coefficients fixed to the value of the CAS wave function (CI ).
The ic-MRCISD wave function is defined by the following wave function guess
where fpq are matrix elements of the average Fock matrix and are computed as
where P̂CAS and P̂SD are projection operators defined in terms of the CAS determinants and
all singly and doubly excited determinants generated from the CAS (SD). For example, P̂CAS
is defined as
X
CAS
P̂CAS = |ΦI ⟩ ⟨ΦI | . (10.18)
I
Another way to define F̂ is using the Dyall partitioning, which includes two-body terms in
the active space. Note that it was also found beneficial to apply a shift to the active orbital
energies (IPEA shift) to
Like for MRCI, one can develop various contracted variants of MRPT: uncontracted,
partially-contracted, internally-contracted, and strongly contracted. All of these differ in the
number of variational degrees of freedom. The most common formulation of MRPT are the
second-order CAS perturbation theory (CASPT2) and second-order n-electron valence per-
turbation theory (NEVPT2). Describing these approaches in detail is beyond the scope of these
notes and the interested reader is urged to consult the primary literature.
Many variants of multireference coupled cluster (MRCC) methods have also been pro-
posed. These can be divided in four broad categories. The first is Fock-space multireference
coupled cluster methods, which aim to describe states with different number of electrons.
The second one is Jeziorksi–Monkhorst methods, which employ an uncontracted ansatz of
the form
X
CAS
I
|ΨJM-MRCC ⟩ = CI eT̂ |ΦI ⟩ . (10.19)
I
The third type is internally-contracted MRCC theories, which generalize the ic-MRCI method
and are based on the following ansatz
X
CAS
|Ψic-MRCC ⟩ = eT̂ |ΨCAS ⟩ = CI eT̂ |ΦI ⟩ . (10.20)
I
There are also a large number of methods that use the single-reference CC ansatz but modify
it so that it can describe some type of wave functions with mild multideterminantal character.
Some example include methods-of-moment CC, equation-of-motion spin-flip CC, and the
CCSDt/CCSDtq methods.