*Waiting for the other shoe to drop.*

This is a figure of speech that means ‘waiting for the inevitable consequence of what’s come so far’. Do you know where it comes from? You have to imagine yourself in an apartment on the floor below someone who is taking off their shoes. When you hear one, you know the next is coming.

There’s even an old comedy routine about this:

A guest who checked into an inn one night was warned to be quiet because the guest in the room next to his was a light sleeper. As he undressed for bed, he dropped one shoe, which, sure enough, awakened the other guest. He managed to get the other shoe off in silence, and got into bed. An hour later, he heard a pounding on the wall and a shout: “When are you going to drop the other shoe?”

When we were working on math together, James Dolan liked to say “the other shoe has dropped” whenever an inevitable consequence of some previous realization became clear. There’s also the mostly British phrase *the penny has dropped*. You say this when someone finally realizes the situation they’re in.

But sometimes one realization comes after another, in a long sequence. Then it feels like it’s raining shoes!

I guess that’s a rather strained metaphor. Perhaps *falling like dominoes* is better for these long chains of realizations.

This is how I’ve felt in my recent research on the interplay between quantum mechanics, stochastic mechanics, statistical mechanics and extremal principles like the principle of least action. The basics of these subjects should be completely figured out by now, but they aren’t—and a lot of what’s known, nobody bothered to tell most of us.

So, I was surprised to rediscover that the Maxwell relations in thermodynamics are formally identical to Hamilton’s equations in classical mechanics… though in retrospect it’s obvious. Thermodynamics obeys the principle of maximum entropy, while classical mechanics obeys the principle of least action. Wherever there’s an extremal principle, symplectic geometry, and equations like Hamilton’s equations, are sure to follow.

I was surprised to discover (or maybe rediscover, I’m not sure yet) that just as statistical mechanics is governed by the principle of maximum entropy, quantum mechanics is governed by a principle of maximum ‘quantropy’. The analogy between statistical mechanics and quantum mechanics has been known at least since Feynman and Schwinger. But this basic aspect was never explained to me!

I was also surprised to rediscover that simply by replacing amplitudes by probabilities in the formalism of quantum field theory, we get a nice formalism for studying stochastic many-body systems. This formalism happens to perfectly match the ‘stochastic Petri nets’ and ‘reaction networks’ already used in subjects from population biology to epidemiology to chemistry. But now we can systematically borrow tools from quantum field theory! All the tricks that particle physicists like—annihilation and creation operators, coherent states and so on—can be applied to problems like the battle between the AIDS virus and human white blood cells.

And, perhaps because I’m a bit slow on the uptake, I was surprised when yet another shoe came crashing to the floor the other day.

Because quantum field theory has, at least formally, a nice limit where Planck’s constant goes to zero, the same is true for for stochastic Petri nets and reaction networks!

In quantum field theory, we call this the ‘classical limit’. For example, if you have a really huge number of photons all in the same state, quantum effects sometimes become negligible, and we can describe them using the *classical* equations describing electromagnetism: the classical Maxwell equations. In stochastic situations, it makes more sense to call this limit the ‘large-number limit’: the main point is that there are lots of particles in each state.

In quantum mechanics, different observables don’t commute, so the so-called commutator matters a lot:

These commutators tend to be proportional to Planck’s constant. So in the limit where Planck’s constant goes to zero, observables commute… but commutators continue to have a ghostly existence, in the form of Poisson bracket:

Poisson brackets are a key part of symplectic geometry—the geometry of classical mechanics. So, this sort of geometry naturally shows up in the study of stochastic Petri nets!

Let me sketch how it works. I’ll start with a section reviewing stuff you should already know if you’ve been following the network theory series.

### The stochastic Fock space

Suppose we have some finite set . We call its elements **species**, since we think of them as different kinds of things—e.g., kinds of chemicals, or kinds of organisms.

To describe the probability of having any number of things of each kind, we need the **stochastic Fock space**. This is the space of real formal power series in a bunch of variables, one for each element of It won’t hurt to simply say

Then the stochastic Fock space is

this being math jargon for the space of formal power series with real coefficients in some variables one for each element of

We write

and use this abbreviation:

We use to describe a state where we have things of the first species, of the second species, and so on.

More generally, a **stochastic state** is an element of the stochastic Fock space with

where

and

We use to describe a state where is the probability of having things of the first species, of the second species, and so on.

The stochastic Fock space has some important operators on it: the **annihilation operators** given by

and the **creation operators** given by

From these we can define the **number operators**:

Part of the point is that

This says the stochastic state is an eigenstate of all the number operators, with eigenvalues saying how many things there are of each species.

The annihilation, creation, and number operators obey some famous commutation relations, which are easy to check for yourself:

The last two have easy interpretations. The first of these two implies

This says that if we start in some state create a thing of type and then count the things of that type, we get *one more* than if we counted the number of things before creating one. Similarly,

says that if we annihilate a thing of type and then count the things of that type, we get *one less* than if we counted the number of things before annihilating one.

### Introducing Planck’s constant

Now let’s introduce an extra parameter into this setup. To indicate the connection to quantum physics, I’ll call it which is the usual symbol for Planck’s constant. However, I want to emphasize that we’re not doing quantum physics here! We’ll see that the limit where is very interesting, but it will correspond to a limit where there are many things of each kind.

We’ll start by defining

and

Here stands for ‘annihilate’ and stands for ‘create’. Think of as a *rescaled* annihilation operator. Using this we can define a rescaled number operator:

So, we have

and this explains the meaning of the parameter The idea is that instead of counting things one at time, we count them in bunches of size

For example, suppose Then we’re counting things in dozens! If we have a state with

then there are 36 things of the *i*th kind. But this implies

so there are 3 dozen things of the *i*th kind.

Chemists don’t count in dozens; they count things in big bunches called moles. A mole is approximately the number of carbon atoms in 12 grams: Avogadro’s number, 6.02 × 10^{23}. When you count things by moles, you’re taking to be 1.66 × 10^{-24}, the reciprocal of Avogadro’s number.

So, while in quantum mechanics Planck’s constant is ‘the quantum of action’, a unit of action, here it’s ‘the quantum of quantity’: the amount that corresponds to *one thing*.

We can easily work out the commutation relations of our new rescaled operators:

These are just what you see in quantum mechanics! The commutators are all proportional to

Again, we can understand what these relations mean if we think a bit. For example, the commutation relation for and says

This says that if we start in some state create a thing of type and then count the things of that type, we get * more* than if we counted the number of things before creating one. This is because we are counting things not one at a time, but in bunches of size

You may be wondering why I defined the rescaled annihilation operator to be times the original annihilation operator:

but left the creation operator unchanged:

I’m wondering that too! I’m not sure I’m doing things the best way yet. I’ve also tried another more symmetrical scheme, taking and This gives the same commutation relations, but certain other formulas become more unpleasant. I’ll explain that some other day.

Next, we can take the limit as and define **Poisson brackets** of operators by

To make this rigorous it’s best to proceed algebraically. For this we treat as a formal variable rather than a specific number. So, our number system becomes the algebra of polynomials in . We define the **Weyl algebra** to be the algebra over generated by elements and obeying

We can set in this formalism; then the Weyl algebra reduces to the algebra of polynomials in the variables and This algebra is commutative! But we can define a **Poisson bracket** on this algebra by

It takes a bit of work to explain to algebraists exactly what’s going on in this formula, because it involves an interplay between the algebra of polynomials in and which is commutative, and the Weyl algebra, which is not. I’ll be glad to explain the details if you want. But if you’re a physicist, you can just follow your nose and figure out what the formula gives. For example:

Similarly, we have:

I should probably use different symbols for and after we’ve set since they’re really different now, but I don’t have the patience to make up more names for things!

Now, we can think of and as coordinate functions on a 2*k*-dimensional vector space, and all the polynomials in and as functions on this space. This space is what physicists would call a ‘phase space’: they use this kind of space to describe the position and momentum of a particle, though here we are using it in a different way. Mathematicians would call it a ‘symplectic vector space’, because it’s equipped with a special structure, called a symplectic structure, that lets us define Poisson brackets of smooth functions on this space. We won’t need to get into that now, but it’s important—and it makes me happy to see it here.

### More

There’s a lot more to do, but not today. My main goal is to understand, in a really elegant way, how the master equation for a stochastic Petri net reduces to the rate equation in the large-number limit. What we’ve done so far is start thinking of this as a limit. This should let us borrow ideas about classical limits in quantum mechanics, and apply them to stochastic mechanics.

Stay tuned!