The free field theories that we’ve discussed so far are very special: we can determine their spectrum, but nothing interesting then happens. They have particle excitations, but these particles don’t interact with each other.
Here we’ll start to examine more complicated theories that include interaction terms. These will take the form of higher order terms in the Lagrangian. We’ll start by asking what kind of small perturbations we can add to the theory. For example, consider the Lagrangian for a real scalar field,
(3.202) |
The coefficients are called coupling constants. What restrictions do we have on to ensure that the additional terms are small perturbations? You might think that we need simply make “”. But this isn’t quite right. To see why this is the case, let’s do some dimensional analysis. Firstly, note that the action has dimensions of angular momentum or, equivalently, the same dimensions as . Since we’ve set , using the convention described in the introduction, we have . With , and , the Lagrangian density must therefore have
(3.203) |
What does this mean for the Lagrangian (3.202)? Since , we can read off the mass dimensions of all the factors to find,
(3.204) |
So now we see why we can’t simply say we need , because this statement only makes sense for dimensionless quantities. The various terms, parameterized by , fall into three different categories
: For this term, the dimensionless parameter is , where has dimensions of mass. Typically in quantum field theory, is the energy scale of the process of interest. This means that is a small perturbation at high energies , but a large perturbation at low energies . Terms that we add to the Lagrangian with this behavior are called relevant because they’re most relevant at low energies (which, after all, is where most of the physics we see lies). In a relativistic theory, , so we can always make this perturbation small by taking .
: this term is small if . Such perturbations are called marginal.
for : The dimensionless parameter is , which is small at low-energies and large at high energies. Such perturbations are called irrelevant.
As you’ll see later, it is typically impossible to avoid high energy processes in quantum field theory. (We’ve already seen a glimpse of this in computing the vacuum energy). This means that we might expect problems with irrelevant operators. Indeed, these lead to “non-renormalizable” field theories in which one cannot make sense of the infinities at arbitrarily high energies. This doesn’t necessarily mean that the theory is useless; just that it is incomplete at some energy scale.
Let me note however that the naive assignment of relevant, marginal and irrelevant is not always fixed in stone: quantum corrections can sometimes change the character of an operator.
Typically in a quantum field theory, only the relevant and marginal couplings are important. This is basically because, as we’ve seen above, the irrelevant couplings become small at low-energies. This is a huge help: of the infinite number of interaction terms that we could write down, only a handful are actually needed (just two in the case of the real scalar field described above).
Let’s look at this a little more. Suppose that we some day discover the true superduper “theory of everything unimportant” that describes the world at very high energy scales, say the GUT scale, or the Planck scale. Whatever this scale is, let’s call it . It is an energy scale, so . Now we want to understand the laws of physics down at our puny energy scale . Let’s further suppose that down at the energy scale , the laws of physics are described by a real scalar field. (They’re not of course: they’re described by non-Abelian gauge fields and fermions, but the same argument applies in that case so bear with me). This scalar field will have some complicated interaction terms (3.202), where the precise form is dictated by all the stuff that’s going on in the high energy superduper theory. What are these interactions? Well, we could write our dimensionful coupling constants in terms of dimensionless couplings , multiplied by a suitable power of the relevant scale ,
(3.205) |
The exact values of dimensionless couplings depend on the details of the high-energy superduper theory, but typically one expects them to be of order 1: . This means that for experiments at small energies , the interaction terms of the form with will be suppressed by powers of . This is usually a suppression by many orders of magnitude. (e.g for the energies explored at the LHC, ). It is this simple argument, based on dimensional analysis, that ensures that we need only focus on the first few terms in the interaction: those which are relevant and marginal. It also means that if we only have access to low-energy experiments (which we do!), it’s going to be very difficult to figure out the high energy theory (which it is!), because its effects are highly diluted except for the relevant and marginal interactions. The discussion given above is a poor man’s version of the ideas of effective field theory and Wilson’s renormalization group, about which you can learn more in the “Statistical Field Theory” course.
In this course we’ll study only weakly coupled field theories i.e.
ones that can truly be considered as small perturbations of the
free field theory at all energies. In this section, we’ll look at
two types of interactions
1) theory:
(3.206) |
with . We can get a hint for what the effects of this extra term will be. Expanding out in terms of and , we see a sum of interactions that look like
(3.207) |
These will create and destroy particles. This suggests that the
Lagrangian describes a theory in which particle number is
not conserved. Indeed, we could check that the number operator
now satisfies .
2) Scalar Yukawa Theory
(3.208) |
with . This theory couples a complex scalar to a real scalar . While the individual particle numbers of and are no longer conserved, we do still have a symmetry rotating the phase of , ensuring the existence of the charge defined in (2.150) such that . This means that the number of particles minus the number of anti-particles is conserved. It is common practice to denote the anti-particle as .
The scalar Yukawa theory has a slightly worrying aspect: the potential has a stable local minimum at , but is unbounded below for large enough . This means we shouldn’t try to push this theory too far.
In this course we restrict attention to weakly coupled field theories where we can use perturbative techniques. The study of strongly coupled field theories is much more difficult, and one of the major research areas in theoretical physics. For example, some of the amazing things that can happen include
Charge Fractionalization: Although electrons have electric charge 1, under the right conditions the elementary excitations in a solid have fractional charge (where ). For example, this occurs in the fractional quantum Hall effect.
Confinement: The elementary excitations of quantum chromodynamics (QCD) are quarks. But they never appear on their own, only in groups of three (in a baryon) or with an anti-quark (in a meson). They are confined.
Emergent Space: There are field theories in four dimensions which at strong coupling become quantum gravity theories in ten dimensions! The strong coupling effects cause the excitations to act as if they’re gravitons moving in higher dimensions. This is quite extraordinary and still poorly understood. It’s called the AdS/CFT correspondence.
There’s a useful viewpoint in quantum mechanics to describe situations where we have small perturbations to a well-understood Hamiltonian. Let’s return to the familiar ground of quantum mechanics with a finite number of degrees of freedom for a moment. In the Schrödinger picture, the states evolve as
(3.209) |
while the operators are independent of time.
In contrast, in the Heisenberg picture the states are fixed and the operators change in time
(3.210) |
The interaction picture is a hybrid of the two. We split the Hamiltonian up as
(3.211) |
The time dependence of operators is governed by , while the time dependence of states is governed by . Although the split into and is arbitrary, it’s useful when is soluble (for example, when is the Hamiltonian for a free field theory). The states and operators in the interaction picture will be denoted by a subscript and are given by,
(3.212) |
This last equation also applies to , which is time dependent. The interaction Hamiltonian in the interaction picture is,
(3.213) |
The Schrödinger equation for states in the interaction picture can be derived starting from the Schrödinger picture
(3.214) | |||||
So we learn that
(3.215) |
“Well, Birmingham has much the best theoretical physicist to work with, Peierls; Bristol has much the best experimental physicist, Powell; Cambridge has some excellent architecture. You can make your choice.”
Oppenheimer’s advice to Dyson on which university position to accept.
We want to solve (3.215). Let’s write the solution as
(3.216) |
where is a unitary time evolution operator such that and . Then the interaction picture Schrödinger equation (3.215) requires that
(3.217) |
If were a function, then we could simply solve this by
(3.218) |
But there’s a problem. Our Hamiltonian is an operator, and we have ordering issues. Let’s see why this causes trouble. The exponential of an operator is defined in terms of the expansion,
(3.219) |
But when we try to differentiate this with respect to , we find that the quadratic term gives us
(3.220) |
Now the second term here looks good, since it will give part of
the that we need on the right-hand side of (3.217).
But the first term is no good since the sits the
wrong side of the integral term, and we can’t commute it through
because when .
So what’s the way around this?
Claim: The solution to (3.217) is given by Dyson’s
Formula. (Essentially first figured out by Dirac, although the
compact notation is due to Dyson).
(3.221) |
where stands for time ordering where operators evaluated at later times are placed to the left
(3.222) |
Expanding out the expression (3.221), we now have
Actually these last two terms double up since
(3.223) | |||||
where the range of integration in the first expression is over , while in the second expression it is which is, of course, the same thing. The final expression is the same as the second expression by a simple relabelling. This means that we can write
(3.224) |
Proof: The proof of Dyson’s formula is simpler than explaining what all the notation means! Firstly observe that under the sign, all operators commute (since their order is already fixed by the sign). Thus
(3.225) | |||||
since , being the upper limit of the integral, is the latest time so can be pulled out to the left.
Before moving on, I should confess that Dyson’s formula is rather formal. It is typically very hard to compute time ordered exponentials in practice. The power of the formula comes from the expansion which is valid when is small and is very easily computed.
Let us now apply the interaction picture to field theory, starting with the interaction Hamiltonian for our scalar Yukawa theory,
(3.226) |
Unlike the free theories discussed in Section 2, this interaction doesn’t conserve particle number, allowing particles of one type to morph into others. To see why this is, we use the interaction picture and follow the evolution of the state: , where is given by Dyson’s formula (3.221) which is an expansion in powers of . But contains creation and annihilation operators for each type of particle. In particular,
: This operator can create or destroy particles. Let’s call them mesons.
: This operator can destroy particles through , and create anti-particles through . Let’s call these particles nucleons. Of course, in reality nucleons are spin particles, and don’t arise from the quantization of a scalar field. But we’ll treat our scalar Yukawa theory as a toy model for nucleons interacting with mesons.
: This operator can create nucleons through , and destroy anti-nucleons through .
Importantly, remains conserved in the presence of . At first order in perturbation theory, we find terms in like . This kills a meson, producing a nucleon-anti-nucleon pair. It will contribute to meson decay .
At second order in perturbation theory, we’ll have more complicated terms in , for example . This term will give contributions to scattering processes . The rest of this section is devoted to computing the quantum amplitudes for these processes to occur.
To calculate amplitudes we make an important, and slightly dodgy, assumption:
Initial and final states are eigenstates of the free theory |
This means that we take the initial state at , and the final state at , to be eigenstates of the free Hamiltonian . At some level, this sounds plausible: at , the particles in a scattering process are far separated and don’t feel the effects of each other. Furthermore, we intuitively expect these states to be eigenstates of the individual number operators , which commute with , but not . As the particles approach each other, they interact briefly, before departing again, each going on its own merry way. The amplitude to go from to is
(3.227) |
where the unitary operator is known as the S-matrix. (S is for scattering). There are a number of reasons why the assumption of non-interacting initial and final states is shaky:
Obviously we can’t cope with bound states. For example, this formalism can’t describe the scattering of an electron and proton which collide, bind, and leave as a Hydrogen atom. It’s possible to circumvent this objection since it turns out that bound states show up as poles in the S-matrix.
More importantly, a single particle, a long way from its neighbors, is never alone in field theory. This is true even in classical electrodynamics, where the electron sources the electromagnetic field from which it can never escape. In quantum electrodynamics (QED), a related fact is that there is a cloud of virtual photons surrounding the electron. This line of thought gets us into the issues of renormalization — more on this next term in the “AQFT” course. Nevertheless, motivated by this problem, after developing scattering theory using the assumption of non-interacting asymptotic states, we’ll mention a better way.
Consider the relativistically normalized initial and final states,
(3.228) |
The initial state contains a single meson of momentum ; the final state contains a nucleon-anti-nucleon pair of momentum and . We may compute the amplitude for the decay of a meson to a nucleon-anti-nucleon pair. To leading order in , it is
(3.229) |
Let’s go slowly. We first expand out using (2.159). (Remember that the in this formula is in the interaction picture, which is the same as the Heisenberg picture of the free theory). The piece will turn into something proportional to , while the piece will turn into a two meson state. But the two meson state will have zero overlap with , and there’s nothing in the and operators that lie between them to change this fact. So we have
(3.230) | |||||
where, in the second line, we’ve commuted past , picking up a delta-function which kills the integral. We now similarly expand out and . To get non-zero overlap with , only the and contribute, for they create the nucleon and anti-nucleon from . We then have
(3.231) | |||||
and so we get our first quantum field theory amplitude.
Notice that the -function puts constraints on the possible decays. In particular, the decay only happens at all if . To see this, we may always boost ourselves to a reference frame where the meson is stationary, so . Then the delta function imposes momentum conservation, telling us that and .
Later you will learn how to turn this quantum amplitude into something more physical, namely the lifetime of the meson. The reason this is a little tricky is that we must square the amplitude to get the probability for decay, which means we get the square of a -function. We’ll explain how to deal with this in Section 3.6 below, and again in next term’s “Standard Model” course.
From Dyson’s formula, we want to compute quantities like , where and are eigenstates of the free theory. The ordering of the operators is fixed by , time ordering. However, since the ’s contain certain creation and annihilation operators, our life will be much simpler if we can start to move all annihilation operators to the right where they can start killing things in . Recall that this is the definition of normal ordering. Wick’s theorem tells us how to go from time ordered products to normal ordered products.
Let’s start simple. Consider a real scalar field which we decompose in the Heisenberg picture as
(3.232) |
where
(3.233) |
where the signs on make little sense, but apparently you have Pauli and Heisenberg to blame. (They come about because , which is sometimes called the positive frequency piece, while is the negative frequency piece). Then choosing , we have
where the last line is normal ordered, and for our troubles we have picked up the extra term which is the propagator we met in (2.165). So for we have
(3.235) |
Meanwhile, for , we may repeat the calculation to find
(3.236) |
So putting this together, we have the final expression
(3.237) |
where is the Feynman propagator defined in (2.168), for which we have the integral representation
(3.238) |
Let me reiterate a comment from Section 2: although
and are both operators, the
difference between them is a c-number function, .
Definition: We define the contraction of a pair of
fields in a string of operators to mean replacing those operators with the
Feynman propagator, leaving all other operators untouched. We use the
notation,
(3.239) |
to denote contraction. So, for example,
(3.240) |
A similar discussion holds for complex scalar fields. We have
(3.241) |
prompting us to define the contraction
(3.242) |
For any collection of fields , , etc, we have
(3.243) |
To see what the last part of this equation means, let’s look at an example. For , the equation reads
(3.244) | |||||
Proof: The proof of Wick’s theorem proceeds by induction and a little thought. It’s true for . Suppose it’s true for and now add . We’ll take for all . Then we can pull out to the left of the time ordered product, writing
(3.245) |
The term stays where it is since it is already normal ordered. But in order to write the right-hand side as a normal ordered product, the term has to make its way past the crowd of operators. Each time it moves past , we pick up a factor of from the commutator. (Try it!)
Let’s look at scattering. We have the initial and final states
(3.246) |
We can then look at the expansion of . In fact, we really want to calculate since we’re not interested in situations where no scattering occurs. At order we have the term
(3.247) |
Now, using Wick’s theorem we see there is a piece in the string of operators which looks like
(3.248) |
which will contribute to the scattering because the two fields annihilate the particles, while the two fields create particles. Any other way of ordering the and fields will give zero contribution. This means that we have
(3.249) | |||||
where, in going to the third line, we’ve used the fact that for relativistically normalized states,
(3.250) |
Now let’s insert this into (3.247), to get the expression for at order ,
(3.251) |
where the expression in square brackets is (3.249), while the final integral is the propagator which comes from the contraction in (3.248). Now the terms double up with the others to cancel the factor of out front. Meanwhile, the and integrals give delta-functions. We’re left with the expression
(3.252) | |||||
Finally, we can trivially do the integral using the delta-functions to get
In fact, for this process we may drop the terms since the denominator is never zero. To see this, we can go to the center of mass frame, where and, by momentum conservation, . This ensures that the 4-momentum of the meson is , so . We therefore have the end result,
(3.253) |
We will see another, much simpler way to reproduce this result shortly using Feynman diagrams. This will also shed light on the physical interpretation.
This calculation is also relevant for other scattering processes, such as , . Each of these comes from the term (3.249) in Wick’s theorem. However, we will never find a term that contributes to scattering , for this would violate the conservation of charge.
If we want to compute scattering at order , we would need to pick out the term
(3.254) |
and a similar term with and exchanged. Once more, this term also contributes to similar scattering processes, including and .
“Like the silicon chips of more recent years, the Feynman diagram was bringing computation to the masses.”
Julian Schwinger
As the above example demonstrates, to actually compute scattering amplitudes using Wick’s theorem is rather tedious. There’s a much better way. It requires drawing pretty pictures. These pictures represent the expansion of and we will learn how to associate numbers (or at least integrals) to them. These pictures are called Feynman diagrams.
The object that we really want to compute is , since we’re not interested in processes where no scattering occurs. The various terms in the perturbative expansion can be represented pictorially as follows
Draw an external line for each particle in the initial state and each particle in the final state . We’ll choose dotted lines for mesons, and solid lines for nucleons. Assign a directed momentum to each line. Further, add an arrow to solid lines to denote its charge; we’ll choose an incoming (outgoing) arrow in the initial state for . We choose the reverse convention for the final state, where an outgoing arrow denotes .
Join the external lines together with trivalent vertices
Each such diagram you can draw is in 1-1 correspondence with the terms in the expansion of .
To each diagram we associate a number, using the Feynman rules
Add a momentum to each internal line
To each vertex, write down a factor of
(3.255) |
where is the sum of all momenta flowing into the vertex.
For each internal dotted line, corresponding to a particle with momentum , we write down a factor of
(3.256) |
We include the same factor for solid internal lines, with replaced by the nucleon mass .
Let’s apply the Feynman rules to compute the amplitudes for various processes. We start with something familiar:
Let’s look at how this works for the scattering at order . We can write down the two simplest diagrams contributing to this process. They are shown in Figure 9.
Applying the Feynman rules to these diagrams, we get
(3.257) |
which agrees with the calculation (3.252) that we performed earlier. There is a nice physical interpretation of these diagrams. We talk, rather loosely, of the nucleons exchanging a meson which, in the first diagram, has momentum . This meson doesn’t satisfy the usual energy dispersion relation, because : the meson is called a virtual particle and is said to be off-shell (or, sometimes, off mass-shell). Heuristically, it can’t live long enough for its energy to be measured to great accuracy. In contrast, the momentum on the external, nucleon legs all satisfy , the mass of the nucleon. They are on-shell. One final note: the addition of the two diagrams above ensures that the particles satisfy Bose statistics.
There are also more complicated diagrams which will contribute to the scattering process at higher orders. For example, we have the two diagrams shown in Figures 10 and 11,
![]() |
![]() |
and similar diagrams with and exchanged. Using the Feynman rules, each of these diagrams translates into an integral that we will not attempt to calculate here. And so we go on, with increasingly complicated diagrams, all appearing at higher order in the coupling constant .
Our final result for the nucleon scattering amplitude at order was
The -function follows from the conservation of 4-momentum which, in turn, follows from spacetime translational invariance. It is common to all S-matrix elements. We will define the amplitude by stripping off this momentum-conserving delta-function,
(3.258) |
where () is the sum of the initial (final) 4-momenta, and the factor of out front is a convention which is there to match non-relativistic quantum mechanics. We can now refine our Feynman rules to compute the amplitude itself:
Draw all possible diagrams with appropriate external legs and impose 4-momentum conservation at each vertex.
Write down a factor of at each vertex.
For each internal line, write down the propagator
Integrate over momentum flowing through each loop .
This last step deserves a short explanation. The diagrams we’ve computed so far have no loops. They are tree level diagrams. It’s not hard to convince yourself that in tree diagrams, momentum conservation at each vertex is sufficient to determine the momentum flowing through each internal line. For diagrams with loops, such as those shown in Figures 10 and 11, this is no longer the case.
Let’s now look at the amplitude for a nucleon-anti-nucleon pair to annihilate into a pair of mesons: . The simplest Feynman diagrams for this process are shown in Figure 12
where the virtual particle in these diagrams is now the nucleon rather than the meson . This fact is reflected in the denominator of the amplitudes which are given by
(3.259) |
As in (3.253), we’ve dropped the from the propagators as the denominator never vanishes.
For the scattering of a nucleon and an anti-nucleon, , the Feynman diagrams are a little different. At lowest order, they are given by the diagrams of Figure 13. It is a simple matter to write down the amplitude using the Feynman rules,
(3.260) |
Notice that the momentum dependence in the second term is different from that of nucleon-nucleon scattering (3.257), reflecting the different Feynman diagram that contributes to the process. In the center of mass frame, , the denominator of the second term is . If , then this term never vanishes and we may drop the . In contrast, if , then the amplitude corresponding to the second diagram diverges at some value of . In this case it turns out that we may also neglect the term, although for a different reason: the meson is unstable when , a result we derived in (3.231). When correctly treated, this instability adds a finite imaginary piece to the denominator which overwhelms the . Nonetheless, the increase in the scattering amplitude which we see in the second diagram when is what allows us to discover new particles: they appear as a resonance in the cross section. For example, the Figure 14 shows the cross-section (roughly the amplitude squared) plotted vertically for scattering from the ALEPH experiment in CERN. The horizontal axis shows the center of mass energy. The curve rises sharply around 91 GeV, the mass of the Z-boson.
For , the simplest diagram we can write down has a single loop, and momentum conservation at each vertex is no longer sufficient to determine every momentum passing through the diagram. We choose to assign the single undetermined momentum to the right-hand propagator. All other momenta are then determined. The amplitude corresponding to the diagram shown in the figure is
These integrals can be tricky. For large , this integral goes as , which is at least convergent as . But this won’t always be the case!
We see that in many of the amplitudes above — in particular those that include the exchange of just a single particle — the same combinations of momenta are appearing frequently in the denominators. There are standard names for various sums and differences of momenta: they are known as Mandelstam variables. They are
(3.261) | |||||
where, as in the examples above, and are the momenta of the two initial particles, and and are the momenta of the final two particles. We can define these variables whether the particles involved in the scattering are the same or different. To get a feel for what these variables mean, let’s assume all four particles are the same. We sit in the center of mass frame, so that the initial two particles have four-momenta
(3.262) |
The particles then scatter at some angle and leave with momenta
(3.263) |
Then from the above definitions, we have that
(3.264) |
The variable measures the total center of mass energy of the collision, while the variables and are measures of the momentum exchanged between particles. (They are basically equivalent, just with the outgoing particles swapped around). Now the amplitudes that involve exchange of a single particle can be written simply in terms of the Mandelstam variables. For example, for nucleon-nucleon scattering, the amplitude (3.257) is schematically . For the nucleon-anti-nucleon scattering, the amplitude (3.260) is . We say that the first case involves “t-channel” and “u-channel” diagrams. Meanwhile the nucleon-anti-nucleon scattering is said to involve “t-channel” and “s-channel” diagrams. (The first diagram indeed includes a vertex that looks like the letter “T”).
Note that there is a relationship between the Mandelstam variables. When all the masses are the same we have . When the masses of all 4 particles differ, this becomes .
So far we’ve computed the quantum amplitudes for various scattering processes. But these quantities are a little abstract. In Section 3.6 below (and again in next term’s “Standard Model” course) we’ll see how to turn amplitudes into measurable quantities such as cross-sections, or the lifetimes of unstable particles. Here we’ll instead show how to translate the amplitude (3.253) for nucleon scattering into something familiar from Newtonian mechanics: a potential, or force, between the particles.
Let’s start by asking a simple question in classical field theory that will turn out to be relevant. Suppose that we have a fixed -function source for a real scalar field , that persists for all time. What is the profile of ? To answer this, we must solve the static Klein-Gordon equation,
(3.265) |
We can solve this using the Fourier transform,
(3.266) |
Plugging this into (3.265) tells us that , giving us the solution
(3.267) |
Let’s now do this integral. Changing to polar coordinates, and writing , we have
(3.268) | |||||
We compute this last integral by closing the contour in the upper half plane , picking up the pole at . This gives
(3.269) |
The field dies off exponentially quickly at distances , the Compton wavelength of the meson.
Now we understand the profile of the field, what does this have to do with the force between particles? We do very similar calculations to that above in electrostatics where a charged particle acts as a -function source for the gauge potential: , which is solved by . The profile for then acts as the potential energy for another charged (test) particle moving in this background. Can we give the same interpretation to our scalar field? In other words, is there a classical limit of the scalar Yukawa theory where the particles act as -function sources for , creating the profile (3.269)? And, if so, is this profile then felt as a static potential? The answer is essentially yes, at least in the limit . But the correct way to describe the potential felt by the particles is not to talk about classical fields at all, but instead work directly with the quantum amplitudes.
Our strategy is to compare the nucleon scattering amplitude (3.253) to the corresponding amplitude in non-relativistic quantum mechanics for two particles interacting through a potential. To make this comparison, we should first take the non-relativistic limit of (3.253). Let’s work in the center of mass frame, with and . The non-relativistic limit means which, by momentum conservation, ensures that . In fact one can check that, for this particular example, this limit doesn’t change the scattering amplitude (3.253): it’s given by
(3.270) |
How do we compare this to scattering in quantum mechanics? Consider two particles, separated by a distance , interacting through a potential . In non-relativistic quantum mechanics, the amplitude for the particles to scatter from momentum states into momentum states can be computed in perturbation theory, using the techniques described in Section 3.1. To leading order, known in this context as the Born approximation, the amplitude is given by
(3.271) |
There’s a relative factor of that arises in comparing the quantum field theory amplitude to , that can be traced to the relativistic normalization of the states . (It is also necessary to get the dimensions of the potential to work out correctly). Including this factor, and equating the expressions for the two amplitudes, we get
(3.272) |
where we’ve introduced the dimensionless parameter . We can trivially invert this to find,
(3.273) |
But this is exactly the integral (3.267) we just did in the classical theory. We have
(3.274) |
This is the Yukawa potential. The force has a range , the Compton wavelength of the exchanged particle. The minus sign tells us that the potential is attractive.
Notice that quantum field theory has given us an entirely new perspective on the nature of forces between particles. Rather than being a fundamental concept, the force arises from the virtual exchange of other particles, in this case the meson. In Section 6 of these lectures, we will see how the Coulomb force arises from quantum field theory due to the exchange of virtual photons.
We could repeat the calculation for nucleon-anti-nucleon scattering. The amplitude from field theory is given in (3.260). The first term in this expression gives the same result as for nucleon-nucleon scattering with the same sign. The second term vanishes in the non-relativisitic limit (it is an example of an interaction that doesn’t have a simple Newtonian interpretation). There is no longer a factor of in (3.271), because the incoming/outgoing particles are not identical, so we learn that the potential between a nucleon and anti-nucleon is again given by (3.274). This reveals a key feature of forces arising due to the exchange of scalars: they are universally attractive. Notice that this is different from forces due to the exchange of a spin 1 particle — such as electromagnetism — where the sign flips when we change the charge. However, for forces due to the exchange of a spin 2 particle — i.e. gravity — the force is again universally attractive.
Let’s briefly look at the Feynman rules and scattering amplitudes for the interaction Hamiltonian
(3.275) |
The theory now has a single interaction vertex, which comes with a factor of , while the other Feynman rules remain the same. Note that we assign to the vertex rather than . To see why this is, we can look at scattering, which has its lowest contribution at order , with the term
(3.276) |
Any one of the fields can do the job of annihilation or creation. This gives different contractions, which cancels the sitting out front.
Feynman diagrams in the theory sometimes come with extra combinatoric factors (typically 2 or 4) which are known as symmetry factors that one must take into account. For more details, see the book by Peskin and Schroeder.
Using the Feynman rules, the scattering amplitude for is simply . Note that it doesn’t depend on the angle at which the outgoing particles emerge: in theory the leading order two-particle scattering occurs with equal probability in all directions. Translating this into a potential between two mesons, we have
(3.277) |
So scattering in theory is due to a -function potential. The particles don’t know what hit them until it’s over.
We’ve seen how one can compute scattering amplitudes by writing down all Feynman diagrams and assigning integrals to them using the Feynman rules. In fact, there are a couple of caveats about what Feynman diagrams you should write down. Both of these caveats are related to the assumption we made earlier that “initial and final states are eigenstates of the free theory” which, as we mentioned at the time, is not strictly accurate. The two caveats which go some way towards ameliorating the problem are the following
We consider only connected Feynman diagrams, where every part of the diagram is connected to at least one external line. As we shall see shortly, this will be related to the fact that the vacuum of the free theory is not the true vacuum of the interacting theory. An example of a diagram that is not connected is shown in Figure 17.
We do not consider diagrams with loops on external lines, for example the diagram shown in the Figure 18. We will not explain how to take these into account in this course, but you will discuss them next term. They are related to the fact that the one-particle states of the free theory are not the same as the one-particle states of the interacting theory. In particular, correctly dealing with these diagrams will account for the fact that particles in interacting quantum field theories are never alone, but surrounded by a cloud of virtual particles. We will refer to diagrams in which all loops on external legs have been cut-off as “amputated”.
![]() |
![]() |
So far we’ve learnt to compute the quantum amplitudes for particles decaying or scattering. As usual in quantum theory, the probabilities for things to happen are the (modulus) square of the quantum amplitudes. In this section we will compute these probabilities, known as decay rates and cross sections. One small subtlety here is that the S-matrix elements all come with a factor of , so we end up with the square of a delta-function. As we will now see, this comes from the fact that we’re working in an infinite space.
Let’s start with something familiar and recall how to derive Fermi’s golden rule from Dyson’s formula. For two energy eigenstates and , with , we have to leading order in the interaction,
(3.278) | |||||
where . This gives us the probability for the transition from to in time , as
(3.279) |
The function in brackets is plotted in Figure 19 for fixed . We see that in time , most transitions happen in a region between energy eigenstates separated by . As , the function in the figure starts to approach a delta-function. To find the normalization, we can calculate
Consider now a transition to a cluster of states with density . In the limit , we get the transition probability
(3.280) | |||||
which gives a constant probability for the transition per unit time for states around the same energy .
(3.281) |
This is Fermi’s Golden Rule.
In the above derivation, we were fairly careful with taking the limit as . Suppose we were a little sloppier, and first chose to compute the amplitude for the state at to transition to the state at . Then we get
(3.282) |
Now when squaring the amplitude to get the probability, we run into the problem of the square of the delta-function: . Tracking through the previous computations, we realize that the extra infinity is coming because is the probability for the transition to happen in infinite time . We can write the delta-functions as
(3.283) |
where is shorthand for (we used a very similar trick when looking at the vacuum energy in (2.100)). We now divide out by this power of to get the transition probability per unit time,
(3.284) |
which, after integrating over the density of final states, gives us back Fermi’s Golden rule. The reason that we’ve stressed this point is because, in our field theory calculations, we’ve computed the amplitudes in the same way as (3.282), and the square of the -functions will just be re-interpreted as spacetime volume factors.
Let’s now look at the probability for a single particle of momentum (I=initial) to decay into some number of particles with momentum and total momentum . This is given by
(3.285) |
Our states obey the relativistic normalization formula (2.140),
(3.286) |
where we have replaced by the volume of 3-space. Similarly,
(3.287) |
If we place our initial particle at rest, so and , we get the probability for decay
(3.288) |
where, as in the second derivation of Fermi’s Golden Rule, we’ve exchanged one of the delta-functions for the volume of spacetime: . The amplitudes are, of course, exactly what we’ve been computing. (For example, in (3.231), we saw that for a single meson decaying into two nucleons). We can now divide out by to get the transition function per unit time. But we still have to worry about summing over all final states. There are two steps: the first is to integrate over all possible momenta of the final particles: . The factors of spatial volume in this measure cancel those in (3.288), while the factors of in (3.288) conspire to produce the Lorentz invariant measure for 3-momentum integrals. The result is an expression for the density of final states given by the Lorentz invariant measure
(3.289) |
The second step is to sum over all final states with different numbers (and possibly types) of particles. This gives us our final expression for the decay probability per unit time, .
(3.290) |
is called the width of the particle. It is equal to the reciprocal of the half-life .
Collide two beams of particles. Sometimes the particles will hit and bounce off each other; sometimes they will pass right through. The fraction of the time that they collide is called the cross section and is denoted by . If the incoming flux is defined to be the number of incoming particles per area per unit time, then the total number of scattering events per unit time is given by,
(3.291) |
We would like to calculate from quantum field theory. In fact, we can calculate a more sensitive quantity known as the differential cross section which is the probability for a given scattering process to occur in the solid angle . More precisely
(3.292) |
where we’ve used the expression for probability per unit time that we computed in the previous subsection. and are the energies of the incoming particles. We now need an expression for the unit flux. For simplicity, let’s sit in the center of mass frame of the collision. We’ve been considering just a single particle per spatial volume , meaning that the flux is given in terms of the 3-velocities as . This then gives,
(3.293) |
If you want to write this in terms of momentum, then recall from your course on special relativity that the 3-velocities are related to the momenta by .
Equation (3.293) is our final expression relating the S-matrix to the differential cross section. You may now take your favorite scattering amplitude, and compute the probability for particles to fly out at your favorite angles. This will involve doing the integral over the phase space of final states, with measure . Notice that different scattering amplitudes have different momentum dependence and will result in different angular dependence in scattering amplitudes. For example, in theory the amplitude for tree level scattering was simply . This results in isotropic scattering. In contrast, for nucleon-nucleon scattering we have schematically . This gives rise to angular dependence in the differential cross-section, which follows from the fact that, for example, , where is the angle between the incoming and outgoing particles.
So far we’ve learnt to compute scattering amplitudes. These are nice and physical (well – they’re directly related to cross-sections and decay rates which are physical) but there are many questions we want to ask in quantum field theory that aren’t directly related to scattering experiments. For example, we might want to compute the viscosity of the quark gluon plasma, or the optical conductivity in a tentative model of strange metals, or figure out the non-Gaussianity of density perturbations arising in the CMB from novel models of inflation. All of these questions are answered in the framework of quantum field theory by computing elementary objects known as correlation functions. In this section we will briefly define correlation functions, explain how to compute them using Feynman diagrams, and then relate them back to scattering amplitudes. We’ll leave the relationship to other physical phenomena to other courses.
We’ll denote the true vacuum of the interacting theory as . We’ll normalize such that
(3.294) |
and . Note that this is different from the state we’ve called which is the vacuum of the free theory and satisfies . Define
(3.295) |
where is in the Heisenberg picture of the full theory, rather than the interaction
picture that we’ve been dealing with so far. The are called correlation functions, or
Green’s functions. There are a number of different ways of looking at these objects which tie
together nicely. Let’s start by asking how to compute using Feynman diagrams. We prove
the following result
Claim: We use the notation , and write
to denote the field in the Heisenberg picture, and
to denote the field in the interaction picture. Then
(3.296) |
where the operators on the right-hand side are evaluated on
, the vacuum of the free theory.
Proof: Take . Then we can drop the
and write the numerator of the right-hand side as
We’ll use the factors of to convert each of the into and we choose operators in the two pictures to be equal at some arbitrary time . Then we can write
Now let’s deal with the two remaining at either end of the string of operators. Consider an arbitrary state and look at
(3.297) |
where is the Schrödinger evolution operator, and the equality above follows because . Now insert a complete set of states, which we take to be energy eigenstates of ,
(3.298) | |||||
But the last term vanishes. This follows from the Riemann-Lebesgue lemma which says that for any well-behaved function
(3.299) |
Why is this relevant? The point is that the in (3.298) is really an integral , because all states are part of a continuum due to the momentum. (There is a caveat here: we want the vacuum to be special, so that it sits on its own, away from the continuum of the integral. This means that we must be working in a theory with a mass gap – i.e. with no massless particles). So the Riemann-Lebesgue lemma gives us
(3.300) |
(Notice that to derive this result, Peskin and Schroeder instead send in a slightly imaginary direction, which also does the job). We now apply the formula (3.300), to the top and bottom of the right-hand side of (3.296) to find
(3.301) |
which, using the normalization , gives us the left-hand side, completing the proof. .
We’re getting closer to our goal of computing the Green’s functions since we can compute both and using the same methods we developed for S-matrix elements; namely Dyson’s formula and Wick’s theorem or, alternatively, Feynman diagrams. But what about dividing one by the other? What’s that all about? In fact, it has a simple interpretation. For the following discussion, we will work in theory. Since there is no ambiguity in the different types of line in Feynman diagrams, we will represent the particles as solid lines, rather than the dashed lines that we used previously. Then we have the diagramatic expansion for .
(3.302) |
These diagrams are called vacuum bubbles. The combinatoric factors (as well as the symmetry factors) associated with each diagram are such that the whole series sums to an exponential,
(3.303) |
So the amplitude for the vacuum of the free theory to evolve into itself is . A similar combinatoric simplification occurs for generic correlation functions. Remarkably, the vacuum diagrams all add up to give the same exponential. With a little thought one can show that
(3.304) |
where “connected” means that every part of the diagram is connected to at least one of the external legs. The upshot of all this is that dividing by has a very nice interpretation in terms of Feynman diagrams: we need only consider the connected Feynman diagrams, and don’t have to worry about the vacuum bubbles. Combining this with (3.296), we learn that the Green’s functions can be calculated by summing over all connected Feynman diagrams,
(3.305) |
As a simple example, let’s look at the four-point correlation function in theory. The sum of connected Feynman diagrams is given by,
All of these are connected diagrams, even though they don’t look that connected! The point is that a connected diagram is defined by the requirement that every line is joined to an external leg. An example of a diagram that is not connected is shown in the figure. As we have seen, such diagrams are taken care of in shifting the vacuum from to .
The Feynman diagrams that we need to calculate for the Green’s functions depend on . This is rather different than the Feynman diagrams that we calculated for the S-matrix elements, where we were working primarily with momentum eigenstates, and ended up integrating over all of space. However, it’s rather simple to adapt the Feynman rules that we had earlier in momentum space to compute . For theory, we have
Draw external points , connected by the usual propagators and vertices. Assign a spacetime position to the end of each line.
For each line from to write down a factor of the Feynman propagator .
For each vertex at position , write down a factor of .
Having described how to compute correlation functions using Feynman diagrams, let’s now relate them back to the S-matrix elements that we already calculated. The first step is to perform the Fourier transform,
(3.306) |
These are very closely related to the S-matrix elements that we’ve computed above. The difference is that the Feynman rules for , effectively include propagators for the external legs, as well as the internal legs. A related fact is that the 4-momenta assigned to the external legs is arbitrary: they are not on-shell. Both of these problems are easily remedied to allow us to return to the S-matrix elements: we need to simply cancel off the propagators on the external legs, and place their momentum back on shell. We have
Each of the factors vanishes once the momenta are placed on-shell. This means that we only get a non-zero answer for diagrams contributing to which have propagators for each external leg.
So what’s the point of all of this? We’ve understood that ignoring the unconnected diagrams is related to shifting to the true vacuum . But other than that, introducing the Green’s functions seems like a lot of bother for little reward. The important point is that this provides a framework in which to deal with the true particle states in the interacting theory through renormalization. Indeed, the formula (3.7.2), suitably interpreted, remains true even in the interacting theory, taking into account the swarm of virtual particles surrounding asymptotic states. This is the correct way to consider scattering. In this context, (3.7.2) is known as the LSZ reduction formula. You will derive it properly next term.