Quantum field theory is the best method we have for describing particles and forces in a way that takes both quantum mechanics and special relativity into account. It makes many wonderfully accurate predictions. And yet, it has embroiled physics in some remarkable problems: struggles with infinities!

I want to sketch some of the key issues in the case of quantum electrodynamics, or ‘QED’. The history of QED has been nicely told here:

• Silvian Schweber, *QED and the Men who Made it: Dyson, Feynman, Schwinger, and Tomonaga*, Princeton U. Press, Princeton, 1994.

Instead of explaining the history, I will give a very simplified account of the current state of the subject. I hope that experts forgive me for cutting corners and trying to get across the basic ideas at the expense of many technical details. The nonexpert is encouraged to fill in the gaps with the help of some textbooks.

QED involves just one dimensionless parameter, the fine structure constant:

Here is the electron charge, is the permittivity of the vacuum, is Planck’s constant and is the speed of light. We can think of as a dimensionless version of the electron charge. It says how strongly electrons and photons interact.

Nobody knows why the fine structure constant has the value it does! In computations, we are free to treat it as an adjustable parameter. If we set it to zero, quantum electrodynamics reduces to a free theory, where photons and electrons do not interact with each other. A standard strategy in QED is to take advantage of the fact that the fine structure constant is small and expand answers to physical questions as power series in This is called ‘perturbation theory’, and it allows us to exploit our knowledge of free theories.

One of the main questions we try to answer in QED is this: if we start with some particles with specified energy-momenta in the distant past, what is the probability that they will turn into certain other particles with certain other energy-momenta in the distant future? As usual, we compute this probability by first computing a complex amplitude and then taking the square of its absolute value. The amplitude, in turn, is computed as a power series in

The term of order in this power series is a sum over Feynman diagrams with vertices. For example, suppose we are computing the amplitude for two electrons wth some specified energy-momenta to interact and become two electrons with some other energy-momenta. One Feynman diagram appearing in the answer is this:

Here the electrons exhange a single photon. Since this diagram has two vertices, it contributes a term of order The electrons could also exchange two photons:

giving a term of A more interesting term of order is this:

Here the electrons exchange a photon that splits into an electron-positron pair and then recombines. There are infinitely many diagrams with two electrons coming in and two going out. However, there are only finitely many with vertices. Each of these contributes a term proportional to to the amplitude.

In general, the external edges of these diagrams correspond to the experimentally observed particles coming in and going out. The internal edges correspond to ‘virtual particles’: that is, particles that are not directly seen, but appear in intermediate steps of a process.

Each of these diagrams is actually a notation for an integral! There are systematic rules for writing down the integral starting from the Feynman diagram. To do this, we first label each edge of the Feynman diagram with an energy-momentum, a variable The integrand, which we shall not describe here, is a function of all these energy-momenta. In carrying out the integral, the energy-momenta of the external edges are held fixed, since these correspond to the experimentally observed particles coming in and going out. We integrate over the energy-momenta of the internal edges, which correspond to virtual particles, while requiring that energy-momentum is conserved at each vertex.

However, there is a problem: the integral typically diverges! Whenever a Feynman diagram contains a loop, the energy-momenta of the virtual particles in this loop can be arbitrarily large. Thus, we are integrating over an infinite region. In principle the integral could still converge if the integrand goes to zero fast enough. However, we rarely have such luck.

What does this mean, physically? It means that if we allow virtual particles with arbitrarily large energy-momenta in intermediate steps of a process, there are ‘too many ways for this process to occur’, so the amplitude for this process diverges.

Ultimately, the continuum nature of spacetime is to blame. In quantum mechanics, particles with large momenta are the same as waves with short wavelengths. Allowing light with arbitrarily short wavelengths created the ultraviolet catastrophe in classical electromagnetism. Quantum electromagnetism averted that catastrophe—but the problem returns in a different form as soon as we study the interaction of photons and charged particles.

Luckily, there is a strategy for tackling this problem. The integrals for Feynman diagrams become well-defined if we impose a ‘cutoff’, integrating only over energy-momenta in some bounded region, say a ball of some large radius In quantum theory, a particle with momentum of magnitude greater than is the same as a wave with wavelength less than Thus, imposing the cutoff amounts to ignoring waves of short wavelength—and for the same reason, ignoring waves of high frequency. We obtain well-defined answers to physical questions when we do this. Unfortunately the answers depend on and if we let they diverge.

However, this is not the correct limiting procedure. Indeed, among the quantities that we can compute using Feynman diagrams are the charge and mass of the electron! Its charge can be computed using diagrams in which an electron emits or absorbs a photon:

Similarly, its mass can be computed using a sum over Feynman diagrams where one electron comes in and one goes out.

The interesting thing is this: to do these calculations, we must start by assuming some charge and mass for the electron—but the charge and mass we *get out* of these calculations do not equal the masses and charges we *put in!*

The reason is that virtual particles affect the observed charge and mass of a particle. Heuristically, at least, we should think of an electron as surrounded by a cloud of virtual particles. These contribute to its mass and ‘shield’ its electric field, reducing its observed charge. It takes some work to translate between this heuristic story and actual Feynman diagram calculations, but it can be done.

Thus, there are two different concepts of mass and charge for the electron. The numbers we put into the QED calculations are called the ‘bare’ charge and mass, and Poetically speaking, these are the charge and mass we would see if we could strip the electron of its virtual particle cloud and see it in its naked splendor. The numbers we get out of the QED calculations are called the ‘renormalized’ charge and mass, and These are computed by doing a sum over Feynman diagrams. So, they take virtual particles into account. These are the charge and mass of the electron clothed in its cloud of virtual particles. It is these quantities, not the bare quantities, that should agree with experiment.

Thus, the correct limiting procedure in QED calculations is a bit subtle. For any value of and any choice of and we compute and The necessary integrals all converge, thanks to the cutoff. We choose and so that and agree with the experimentally observed charge and mass of the electron. The bare charge and mass chosen this way depend on so call them and

Next, suppose we want to compute the answer to some other physics problem using QED. We do the calculation with a cutoff using and as the bare charge and mass in our calculation. Then we take the limit

In short, rather than simply fixing the bare charge and mass and letting we cleverly adjust the bare charge and mass as we take this limit. This procedure is called ‘renormalization’, and it has a complex and fascinating history:

• Laurie M. Brown, ed., *Renormalization: From Lorentz to Landau (and Beyond)*, Springer, Berlin, 2012.

There are many technically different ways to carry out renormalization, and our account so far neglects many important issues. Let us mention three of the simplest.

First, besides the classes of Feynman diagrams already mentioned, we must also consider those where one photon goes in and one photon goes out, such as this:

These affect properties of the photon, such as its mass. Since we want the photon to be massless in QED, we have to adjust parameters as we take to make sure we obtain this result. We must also consider Feynman diagrams where nothing comes in and nothing comes out—so-called ‘vacuum bubbles’—and make these behave correctly as well.

Second, the procedure just described, where we impose a ‘cutoff’ and integrate over energy-momenta lying in a ball of radius is not invariant under Lorentz transformations. Indeed, any theory featuring a smallest time or smallest distance violates the principles of special relativity: thanks to time dilation and Lorentz contractions, different observers will disagree about times and distances. We could accept that Lorentz invariance is broken by the cutoff and hope that it is restored in the limit, but physicists prefer to maintain symmetry at every step of the calculation. This requires some new ideas: for example, replacing Minkowski spacetime with 4-dimensional Euclidean space. In 4-dimensional Euclidean space, Lorentz transformations are replaced by rotations, and a ball of radius is a rotation-invariant concept. To do their Feynman integrals in Euclidean space, physicists often let time take imaginary values. They do their calculations in this context and then transfer the results back to Minkowski spacetime at the end. Luckily, there are theorems justifying this procedure.

Third, besides infinities that arise from waves with arbitrarily short wavelengths, there are infinities that arise from waves with arbitrarily *long* wavelengths. The former are called ‘ultraviolet divergences’. The latter are called ‘infrared divergences’, and they afflict theories with massless particles, like the photon. For example, in QED the collision of two electrons will emit an infinite number of photons with very long wavelengths and low energies, called ‘soft photons’. In practice this is not so bad, since any experiment can only detect photons with energies above some nonzero value. However, infrared divergences are conceptually important. It seems that in QED any electron is inextricably accompanied by a cloud of soft photons. These are real, not virtual particles. This may have remarkable consequences.

Battling these and many other subtleties, many brilliant physicists and mathematicians have worked on QED. The good news is that this theory has been proved to be ‘perturbatively renormalizable’:

• J. S. Feldman, T. R. Hurd, L. Rosen and J. D. Wright, *QED: A Proof of Renormalizability*, Lecture Notes in Physics **312**, Springer, Berlin, 1988.

• Günter Scharf, *Finite Quantum Electrodynamics: The Causal Approach*, Springer, Berlin, 1995

This means that we can indeed carry out the procedure roughly sketched above, obtaining answers to physical questions as power series in

The bad news is we do not know if these power series converge. In fact, it is widely believed that they diverge! This puts us in a curious situation.

For example, consider the magnetic dipole moment of the electron. An electron, being a charged particle with spin, has a magnetic field. A classical computation says that its magnetic dipole moment is

where is its spin angular momentum. Quantum effects correct this computation, giving

for some constant called the gyromagnetic ratio, which can be computed using QED as a sum over Feynman diagrams with an electron exchanging a single photon with a massive charged particle:

The answer is a power series in but since all these diagrams have an even number of vertices, it only contains integral powers of The lowest-order term gives simply In 1948, Julian Schwinger computed the next term and found a small correction to this simple result:

By now a team led by Toichiro Kinoshita has computed up to order This requires computing over 13,000 integrals, one for each Feynman diagram of the above form with up to 10 vertices! The answer agrees very well with experiment: in fact, if we also take other Standard Model effects into account we get agreement to roughly one part in

This is the most accurate prediction in all of science.

However, as mentioned, it is widely believed that this power series *diverges!* Next time I’ll explain why physicists think this, and what it means for a divergent series to give such a good answer when you add up the first few terms.

Thanks for this interesting article! It clarified several things for me. However, there is a basic aspect I don’t understand. I feel embarrassed to ask, but: what is the role of space in this? Suppose I want to understand what happens, at particle level, in one cubic meter of space. Would I then dice that space in many tiny cubes and apply the methods from your post to each tiny cube? Or have I got this totally wrong?

I thought about this more and realized: my uncertainty comes from not knowing the mathematical structure of the phase space. In particular since the number of particles can apparently vary over time. It this the Fock space thing? My “dicing” remark was inspired by your phrase “external edges are held fixed”…

I should emphasize that there are lots of different viewpoints on quantum field theory, so when you start learning it you feel like you’re talking to blind men who are experts on elephants: everyone will give you a different story, and it’s all very confusing. After a decade of study things get clearer.

In this post I’m telling just one version of the story: how to use path integrals to perturbatively compute the scattering matrix, also known as the ‘S-matrix’. The S-matrix says: if you shoot in a collection of particles with specified energies and momenta at past infinity, what is the amplitude for a collection of particles with specified energies and momenta to come out at future infinity?

The S-matrix is what people doing particle accelerator experiments care about, so it’s what you typically learn how to compute in a first course in quantum field theory. It is

quite hardto take the S-matrix and read off the answers to the questions you’re talking about.Your questions are about the

localbehavior of quantum field theory: e.g, what’s going on in one hypercubic meter of spacetime. (You said ‘cubic meter of space’, but since everything here is relativistic, a spacetime viewpoint would be more appropriate.) The S-matrix is a purelyglobalthing: computing it involves integrals over all of Minkowski spacetime. To make matters worse (from your viewpoint), these integrals are easiest to do with the help of a Fourier transform. So, the calculations actually involve integrals over particle energy-momentanot locations of events

Luckily, there is more to quantum field theory than the S-matrix! There

isa local story. It can again be studied using path integrals (among several other methods). Now, however, instead of integrating over all of spacetime, we should integrate over a finite 4-volume: for example, a hypercube.This makes Fourier transforms somewhat less attractive. So, instead of integrating over energy-momenta, one for each edge of a Feynman diagram, we integrate over spacetime points, one for each vertex of our Feynman diagram! These vertices are ‘events’: points where an interaction occurs.

If you do this, you can fix the spacetime locations of the incoming and outgoing particles rather than their momenta. That is, you fix where the particles enter and leave your hypercube. These correspond to the ‘loose ends’ of your Feynman diagram.

Technically, this sort of calcuation is a lot harder than computing an S-matrix, so you’ll never see it in an introductory textbook. Conceptually it is easier.

Thanks, your explanation about scattering was crucial for me! I’ve read up a little on scattering in general, and in particle physics in particular, including the quite interesting hardware: bubble and cloud and spark and drift chambers. So, mathematically, it’s all about before-after asymptotics.

I had secretly hoped I could come up with a computer visualization of some low-dimensional toy QFT, where particles flit around in front of me as probability clouds. But apparently, achieving that would start from a different story than the one in your blog post.

For what you want, it would be better to take the Hamiltonian for the theory a 2-dimensional grid—one space dimension and one time dimension. If you do a nonrelativistic version (that is, take the limit) you’d get the ordinary Schrödinger equation for massive particles on a line, discretized, but with a certain amplitude for particles to interact. This interaction can turn 2 particles into 2, 3 into 1, 1 into 3, 4 into 0 or 0 into 4.

It would take a little while to explain it well enough to code it up. But I don’t think it’s very easy to “see” the wavefunction of a multiparticle state! It’ll be a function of many variables: the positions of all the particles in your collection.

(Yes, this is the Fock space business.)

Thanks again, this helps!

Thank you very much for this. The whole series on continuum is wonderful.

Last time I sketched how physicists use quantum electrodynamics, or ‘QED’, to compute answers to physics problems as power series in the fine structure constant, which is

I concluded with a famous example: the magnetic moment of the electron. With a truly heroic computation, physicists have used QED to compute this quantity up to order If we also take other Standard Model effects into account we get agreement to roughly one part in

However, if we continue adding up terms in this power series, there is no guarantee that the answer converges!

(1) Nice post. Thank-you! You might perhaps want to indicate the direction in which the diagrams are to be read. As you describe it, the first one should be read from bottom to top. However, the left to right reading also makes sense I hope (e- e+ to photon and then back to e- e+).

(2) Is it necessary that the divergences arise from the hypothesis of continuous space-time? Don’t string theorists claim to be able tame the divergences by replacing point particles with more extended objects?

(1) Feynman diagrams are always read bottom to top.

(2) String theory seems to eliminate the ultraviolet divergences in any single Feynman diagram, but the sum over Feynman diagrams seems to still diverge. I say “seems to” because it seems hard to find precise theorems about this lore.