Explicitly correlated Gaussian functions have been extensively used in quantum-mechanical variational calculations in atomic, molecular, and nuclear physics. This book is an attempt to collect the relevant information about this established tool in computational quantum mechanics.
Often we describe bound-state and scattering problems in nuclear and atomic physics through the Shrödinger equation.
Unfortunately modern quantum physics offers problems that we can not solve analytically. Luckily the availability of powerful computers is shifting the emphasis from the analytical computation of the solution toward numerical analysis. During the last century numerous methods were developed in order to approximate solutions numerically, e.g. Monte-Carlo simulation, Hypershperical expansion, variational methods with different trial wave functions etc.
In this section we discuss variational method with trial function in the form of correlated Gaussians which is widely used in the modern physics.
Mathematically it is based on the Ritz theorem, that states that for an arbitrary function Ψ from the state space
the expectation value of the Hamiltonian (<Ψ|H|Ψ>/<Ψ|Ψ>) is larger then the ground state energy. So choosing different trial wave functions and calculating mean values of the Hamiltonian for this functions allows us to get an upper bound for the ground state energy.
To show the idea of the method we consider two particles in 1 dimension interacting through the oscillator potential
It is really simple textbook problem with the ground state solution
where we assumed for simplicity, that the total momentum is equal to zero.
To show how the method works we choose the trial wave function in the form of just one gaussian
where we have only one real positive parameter which has to minimize the energy. The idea of the method is to pick this parameter stochastically using just generator of the real numbers. We find out that independently of the seed after 50 attempts we found value of α that gives the ground state energy with 5 significant digits.
Convergence of the expectation value of the Hamiltonian (different curves corresponds to different starting seeds for the generator of the random numbers).
Of course it's really simple example and we can establish energy with high precision just because we are working in the space that contains ground state wave function of the Hamiltonian.
In order to establish excited states it's not enough to use just one Gaussian so we pick trial wave function in more general form
As above we pick parameters stochastically and then determine linear parameters ci demanding minimal expectation value of the Hamiltonian. Using N=25 and one random set of the parameters (we assume that ) we get first 6 eigenstates with 5 significant digits.
From this simple example we learn that we are able to approximate solution of the Schrödinger equation without any preliminary knowledge about the system using only random search. The main problem is to estimate how good this approximation is.
Hamiltonian under consideration[edit | edit source]
We shall consider a non-relativistic quantum mechanical -particle (N-body) system described by a Hamiltonian
where and are the mass and the coordinate of particle number ; the first term is the kinetic energy operator; the second term is the one-body force, like the external field (often an oscillator, ), for example Magneto-optical trap; and the third term is the two-body forces, the inter-particle interactions.
This Hamiltonian can describe a system of atoms in a trap or a nucleus.
In general one can also consider spin-dependent, three-body, non-local and other types of interactions which we shall introduce when needed.
Basis expansion of the Schrödinger equation[edit | edit source]
We are going to solve the Schrödinger equation
where is the Hamiltonian of a quantum few-body system, and are the eigenfunction and the eigenvalue to be found.
We shall expand the wave-function in terms of a set of basis functions ,
Inserting the expansion into the Schrödinger equation and multiplying from the left with gives
Or, in the matrix notation
where and are correspondingly the Hamiltonian and the overlap matrices with the matrix elements
The matrix equation is called the generalized eigenvalue problem. There exist established routines to solve this problem, for example the function "eig" in Octave, or the function "gsl_eigen_gensymm" in the GNU Scientific Library.
If the basis function are orthogonal and normalized the overlap matrix equals the unity matrix, , and the generalized eigenvalue problem reduces to the ordinary eigenvalue problem, .
We shall use Gaussians --- which are not orthogonal --- as basis functions, therefore we shall deal with the generalized eigenvalue problem.
Gaussians as basis functions[edit | edit source]
We shall use the so-called Correlated Gaussians (or Explicitly Correlated Gaussians) as the basis functions. For a system of particles with coordinates , the Correlated Gaussian is defined as
where denotes the dot-product of the two vectors; and where , a symmetric positive-defined matrix, and , the shift-vectors, are (cleverly chosen) parameters of the Gaussian.
In matrix notation,
where is the column of the coordinates and is the column of the shift-vectors ,
Matrix elements with Gaussians[edit | edit source]
We represent the few-body wave-function as a linear combination of correlated Gaussians,
We shall discuss the choice of matrices in these Gaussians later and here only calculate the matrix elements.
(where , and ) can be calculated making an orthogonal transformation, , where , to the basis where the matrix is diagonal,
Let us consider a more general form of the kinetic energy operator,
where is a constant matrix, for example, for the Hamiltonian above
The matrix element is given as
Integrating by parts with respect to the left derivative gives
Calculating term after term,
The matrix element of the interaction potential between, say, particles and ,
can be written in a more general form,
where is a size-N vector with all components equal zero except for and .
A one-body interaction, , has the same form, , where and all other components equal zero.
For a central potential in a Gaussian form,
the matrix element between shifted Gaussians can be calculated straightforwardly,
The rank-1 updates of the determinant, , and the matrix inverse, , can be efficiently calculated using the following formulae,
For a general form-factor central potential one way to calculate the matrix element is through the Fourier transform of the potential, ,
where is the Fourier transform of the potential , and .
The last integral can be also written via the potential itself,
where and .
Here are integrals for some popular potentials,
- Coulomb, ;
- Oscillator, .
In nuclear physics the tensor potential between two nucleons has the form
where is the form-factor; ; and are the coordinates of the nucleons; , are the Pauli matrices related to the spins, and , of the two nucleons: , .
One often adds the term to make sure that the potential has no central component (that is, the average of the potential over all directions is zero). Without this extra term the above tensor potential has a central spin-spin component ).
Again introducing the column with , , , and the vector-columns , , the potential can be written in a convenient general form,
The matrix element of this operator between shifted Gaussians is given as
This can be calculated analytically for a Gaussian form-factor, ,
Finally, for a Gaussian tensor potential,
The spin-orbit potential between two nucleons --- with coordinates and and spins and --- is usually written in the form
where ; is the total spin of the two nucleons,
and is the relative orbital momentum between the two nucleons
where denotes vector-product of two vectors.
The orbital momentum can be rewritten --- using the column with , , --- in a convenient general form,
For a Gaussian form-factor, , the corresponding matrix element between shifted Gaussians can be calculated analytically,
where , , .
Mathematical formulation for the ground state problem[edit | edit source]
Let us consider a time-independent physical system whose Hamiltonian H is Hermitian and bounded from below. We want to approximate the discrete eigenvalues of H and its wave functions
where we ordered eigenvalues s.t.
It means that we would like to find such square integrable functions , that , with some . Unfortunately in practice we don't know exact eigenvalues of the Hamiltonian, so first we have to find approximation to the energy . The following theorem gives us the receipt. Here we would like to restrict ourselves to the ground state, but using the Min-max theorem one can extend to the whole discrete spectrum of the Hamiltonian
The expectation value of the Hamiltonian for any from the state space is equal or larger then the ground state energy .
Apparently the function can be decomposed in the orthogonal basis : . With this decomposition we write the mean value of the Hamiltonian: , from which follows that .
This statement is often called Ritz theorem and might be seen as a corollary of the Min-max theorem.
This result allows us compute an upper bound for the ground state energy.
The following theorem according to Weinstein allows us to rewrite our initial demand that in terms of the variance
There exist at least one eigenvalue in the interval .
We write in the basis, and get . There exist integer , s.t. . With this we rewrite variance
This result might be useful if and only if the lower bound can be calculated as close as possible to the ground state energy.
With these theorems we see the way to proceed:
1. Take convenient basis in the state space of the Hamiltonian.
2. Cut the basis size to some finite number.
3. Minimise expectation value of the Hamiltonian in this basis.
4. Enlarge basis and do step 3.
5. Do steps 3,4 as long as needed to insure convergence of the ground state energy.
6. Calculate variance.
7. If variance is larger than some precision value than enlarge basis size and do 3,4,5,6 again, otherwise we are done.
In practice steps 3,4,5 alone can give accurate value of energy. Steps 6,7 are needed for approximation of the wave function.
This is due to the following theorem
The expectation value of the Hamiltonian is stationary in the neighbourhood of the discrete eigenvalues.
So in general it is easier to get accurate approximation to the energy than to other observables.
We want to start with the first step: take some convenient basis. We would like to define convenient for our problem
1. Simple transformation from one system of coordinates to another.
2. Possibility to eliminate the centre of mass.
3. Easy computations for the overlap and kinetic energy.
It is of advantage to introduce rescaled coordinates,
where is a conveniently chosen mass scale. Indeed the kinetic energy and the harmonic trap have a more symmetric form in the rescaled coordinates,
The Jacobian of the transformation from to is equal
A further suitable linear transformation to a new set of coordinates is possible,
or, in matrix notation,
where is the transformation matrix.
If the transformation matrix is unitary, , the diagonal form of the kinetic energy and the harmonic trap is preserved in the new coordinates,
Last transformation is of particular use if new system has the coordinate
which can be seen as a center of mass coordinate. It allows us to work with a wave function in the form
where is the ground state wave function for the oscillator potential.
First we consider trial wave function in the basis of completely general shifted Gaussian, which can be used to describe a system in the external field with anisotropic inter-particle interaction
, a symmetric positive-defined matrix, and , a shift vector, are the non-linear parameters of the Gaussian and n=N-1.
With this definition we have non-linear variational parameters. To find those one can use deterministic methods (e.g. Powell's method) or methods based on a stochastic search. We use latter approach though we find
linear variational parameters through a full minimization with respect to a given set of non-linear parameters.
The matix elements can be determined analytically either by diagonalizing the matrix or Cholesky decompose
(since it is positive definite) and change basis to coordinates where the matrix is either diagonal or unit. This way many integrals can be determined
by iterative integration.
Correlated Gaussians are generally non-orthogonal and the overlap is therefore non-diagonal,
where we defined
Here we calculate kinetic energy
where we defined
for with to be the identity matrix, one can get simpler expression after noticing that
To proceed one has to derive the following identities:
which we need to calculate variance we have to calculate the following matrix element
Here we calculate matrix element
In general we can not write analytical expression for this integral, but we can reduce it to D dimensional integrals. For example, consider
just one term from the sum
to simplify this integral we have to make transformation from Jacobi set , where matrices A,A',s,s' are defined to the Jacobi set , where
. transformation between those sets are provided through the orthogonal matrix U: x=Uy.
With this we write
can be found analytically.
If we can write potential as a sum of Gaussians
then the integral can be found analytically in the same way as we found overlap.
To consider particles with spin we add spin part to the trial wave function
where for particles with spin = 1/2 function is just an array of N elements. Each element is an eigenfunction of the spin's projection on the predefined axis. For example defines the system with all particles have spin in the same direction. Next we define the spin operator that acts on the particle number in the following way
and zero otherwise.
Here we discuss spin-orbit potential of the form , where
and - relative angular momentum, where - Levi-Chevita symbol, and . We have to calculate following matrix element
again we are making transformation from the Jacobi set to the Jacobi set using a transformation matrix U: x=Uy.
In the previous section we considered completely general setup, which is suitable for any inter-particle potentials and external fields.
This approach is far from optimal if for example we are interested in ground state of N bosons with isotropic pairwise interaction, because
in this case we know that our ground state must have zero orbital momentum, with this in mind we write the trial wave function in the smaller variational basis:
If we put shift vectors to be zeros , then the trial wave function treats Cartesian components of vectors equivalently, which leads to zero angular momentum, otherwise the wave function will contain all possible angular momentum and we need an effective procedure to build an eigenstate for a given angular momentum. Matrix elements for this trial wave functions can be obtained from the general case, but we write it explicitly.
where we defined
where we defined
We consider the matrix element of the operator , choice of can give the total angular momentum or an angular momentum for appropriate relative coordinate.
First we calculate matrix elements of the form
and now we can calculate matrix element for operator
We define total angular momentum to be . If we make transformation to the Jacobi set, than we obtain , where is the linear momentum corresponding to the coordinate. So if we assume that the system as a whole is at rest s.t. than the following matrix element defines total angular momentum
After simplification (first we rotate to the set of coordinate, where matrix takes a diagonal form, than we rotate back and rotate to the set where matrix takes a diagonal form, and again rotate back) we obtain
We take the following integral
where - is a Kronecker delta.
With this we write total angular momentum
Appendix A: some integrals with 3D Gaussians[edit | edit source]
Here we need to indroduce a notation to simplify the expressions:
Appendix B: some integrals with 1D Gaussians[edit | edit source]
First the overlap (denoted N and used in the following expressions)
for any matrix F with the right dimensions including .
Exponential and sinusoidal terms[edit | edit source]
Due to the Euler formulae for (Co)Sine and the linearity we can express these matrix elements
as the (Real) Imaginary part of the above matrix element with complex vector
This can be calculated as the limiting case of the previous matrix element