*Waiting for the other shoe to drop.*

This is a figure of speech that means ‘waiting for the inevitable consequence of what’s come so far’. Do you know where it comes from? You have to imagine yourself in an apartment on the floor below someone who is taking off their shoes. When you hear one, you know the next is coming.

There’s even an old comedy routine about this:

A guest who checked into an inn one night was warned to be quiet because the guest in the room next to his was a light sleeper. As he undressed for bed, he dropped one shoe, which, sure enough, awakened the other guest. He managed to get the other shoe off in silence, and got into bed. An hour later, he heard a pounding on the wall and a shout: “When are you going to drop the other shoe?”

When we were working on math together, James Dolan liked to say “the other shoe has dropped” whenever an inevitable consequence of some previous realization became clear. There’s also the mostly British phrase *the penny has dropped*. You say this when someone finally realizes the situation they’re in.

But sometimes one realization comes after another, in a long sequence. Then it feels like it’s raining shoes!

I guess that’s a rather strained metaphor. Perhaps *falling like dominoes* is better for these long chains of realizations.

This is how I’ve felt in my recent research on the interplay between quantum mechanics, stochastic mechanics, statistical mechanics and extremal principles like the principle of least action. The basics of these subjects should be completely figured out by now, but they aren’t—and a lot of what’s known, nobody bothered to tell most of us.

So, I was surprised to rediscover that the Maxwell relations in thermodynamics are formally identical to Hamilton’s equations in classical mechanics… though in retrospect it’s obvious. Thermodynamics obeys the principle of maximum entropy, while classical mechanics obeys the principle of least action. Wherever there’s an extremal principle, symplectic geometry, and equations like Hamilton’s equations, are sure to follow.

I was surprised to discover (or maybe rediscover, I’m not sure yet) that just as statistical mechanics is governed by the principle of maximum entropy, quantum mechanics is governed by a principle of maximum ‘quantropy’. The analogy between statistical mechanics and quantum mechanics has been known at least since Feynman and Schwinger. But this basic aspect was never explained to me!

I was also surprised to rediscover that simply by replacing amplitudes by probabilities in the formalism of quantum field theory, we get a nice formalism for studying stochastic many-body systems. This formalism happens to perfectly match the ‘stochastic Petri nets’ and ‘reaction networks’ already used in subjects from population biology to epidemiology to chemistry. But now we can systematically borrow tools from quantum field theory! All the tricks that particle physicists like—annihilation and creation operators, coherent states and so on—can be applied to problems like the battle between the AIDS virus and human white blood cells.

And, perhaps because I’m a bit slow on the uptake, I was surprised when yet another shoe came crashing to the floor the other day.

Because quantum field theory has, at least formally, a nice limit where Planck’s constant goes to zero, the same is true for for stochastic Petri nets and reaction networks!

In quantum field theory, we call this the ‘classical limit’. For example, if you have a really huge number of photons all in the same state, quantum effects sometimes become negligible, and we can describe them using the *classical* equations describing electromagnetism: the classical Maxwell equations. In stochastic situations, it makes more sense to call this limit the ‘large-number limit': the main point is that there are lots of particles in each state.

In quantum mechanics, different observables don’t commute, so the so-called commutator matters a lot:

These commutators tend to be proportional to Planck’s constant. So in the limit where Planck’s constant goes to zero, observables commute… but commutators continue to have a ghostly existence, in the form of Poisson bracket:

Poisson brackets are a key part of symplectic geometry—the geometry of classical mechanics. So, this sort of geometry naturally shows up in the study of stochastic Petri nets!

Let me sketch how it works. I’ll start with a section reviewing stuff you should already know if you’ve been following the network theory series.

### The stochastic Fock space

Suppose we have some finite set . We call its elements **species**, since we think of them as different kinds of things—e.g., kinds of chemicals, or kinds of organisms.

To describe the probability of having any number of things of each kind, we need the **stochastic Fock space**. This is the space of real formal power series in a bunch of variables, one for each element of It won’t hurt to simply say

Then the stochastic Fock space is

this being math jargon for the space of formal power series with real coefficients in some variables one for each element of

We write

and use this abbreviation:

We use to describe a state where we have things of the first species, of the second species, and so on.

More generally, a **stochastic state** is an element of the stochastic Fock space with

where

and

We use to describe a state where is the probability of having things of the first species, of the second species, and so on.

The stochastic Fock space has some important operators on it: the **annihilation operators** given by

and the **creation operators** given by

From these we can define the **number operators**:

Part of the point is that

This says the stochastic state is an eigenstate of all the number operators, with eigenvalues saying how many things there are of each species.

The annihilation, creation, and number operators obey some famous commutation relations, which are easy to check for yourself:

The last two have easy interpretations. The first of these two implies

This says that if we start in some state create a thing of type and then count the things of that type, we get *one more* than if we counted the number of things before creating one. Similarly,

says that if we annihilate a thing of type and then count the things of that type, we get *one less* than if we counted the number of things before annihilating one.

### Introducing Planck’s constant

Now let’s introduce an extra parameter into this setup. To indicate the connection to quantum physics, I’ll call it which is the usual symbol for Planck’s constant. However, I want to emphasize that we’re not doing quantum physics here! We’ll see that the limit where is very interesting, but it will correspond to a limit where there are many things of each kind.

We’ll start by defining

and

Here stands for ‘annihilate’ and stands for ‘create’. Think of as a *rescaled* annihilation operator. Using this we can define a rescaled number operator:

So, we have

and this explains the meaning of the parameter The idea is that instead of counting things one at time, we count them in bunches of size

For example, suppose Then we’re counting things in dozens! If we have a state with

then there are 36 things of the *i*th kind. But this implies

so there are 3 dozen things of the *i*th kind.

Chemists don’t count in dozens; they count things in big bunches called moles. A mole is approximately the number of carbon atoms in 12 grams: Avogadro’s number, 6.02 × 10^{23}. When you count things by moles, you’re taking to be 1.66 × 10^{-24}, the reciprocal of Avogadro’s number.

So, while in quantum mechanics Planck’s constant is ‘the quantum of action’, a unit of action, here it’s ‘the quantum of quantity': the amount that corresponds to *one thing*.

We can easily work out the commutation relations of our new rescaled operators:

These are just what you see in quantum mechanics! The commutators are all proportional to

Again, we can understand what these relations mean if we think a bit. For example, the commutation relation for and says

This says that if we start in some state create a thing of type and then count the things of that type, we get * more* than if we counted the number of things before creating one. This is because we are counting things not one at a time, but in bunches of size

You may be wondering why I defined the rescaled annihilation operator to be times the original annihilation operator:

but left the creation operator unchanged:

I’m wondering that too! I’m not sure I’m doing things the best way yet. I’ve also tried another more symmetrical scheme, taking and This gives the same commutation relations, but certain other formulas become more unpleasant. I’ll explain that some other day.

Next, we can take the limit as and define **Poisson brackets** of operators by

To make this rigorous it’s best to proceed algebraically. For this we treat as a formal variable rather than a specific number. So, our number system becomes the algebra of polynomials in . We define the **Weyl algebra** to be the algebra over generated by elements and obeying

We can set in this formalism; then the Weyl algebra reduces to the algebra of polynomials in the variables and This algebra is commutative! But we can define a **Poisson bracket** on this algebra by

It takes a bit of work to explain to algebraists exactly what’s going on in this formula, because it involves an interplay between the algebra of polynomials in and which is commutative, and the Weyl algebra, which is not. I’ll be glad to explain the details if you want. But if you’re a physicist, you can just follow your nose and figure out what the formula gives. For example:

Similarly, we have:

I should probably use different symbols for and after we’ve set since they’re really different now, but I don’t have the patience to make up more names for things!

Now, we can think of and as coordinate functions on a 2*k*-dimensional vector space, and all the polynomials in and as functions on this space. This space is what physicists would call a ‘phase space’: they use this kind of space to describe the position and momentum of a particle, though here we are using it in a different way. Mathematicians would call it a ‘symplectic vector space’, because it’s equipped with a special structure, called a symplectic structure, that lets us define Poisson brackets of smooth functions on this space. We won’t need to get into that now, but it’s important—and it makes me happy to see it here.

### More

There’s a lot more to do, but not today. My main goal is to understand, in a really elegant way, how the master equation for a stochastic Petri net reduces to the rate equation in the large-number limit. What we’ve done so far is start thinking of this as a limit. This should let us borrow ideas about classical limits in quantum mechanics, and apply them to stochastic mechanics.

Stay tuned!

If I can ask, why are you posing ? As far as I remember, even in QM, it’s usually 1. is the commutator of x and p. Of course it’s perhaps just a convention.

Also, your definition of A and C on reminds me more of the use of x and p in QM. In that case, it would be natural to multiply p times , since it’s how it’s defined in usual QM. However, you lose , which I suppose it was the reason you defined them like that.

Martino wrote:

I so often set in my thinking on quantum mechanics that I don’t know whether the ‘usual’ commutator of annihilation and creation operators is 1 or

But I know that to get Poisson brackets in the classical limit I need a bunch of quantities whose commutators are proportional to (perhaps plus higher-order corrections).

We could define quantities and as linear combinations of and , and work with those. However, in the stochastic context I don’t know what those quantities mean! Since the Hamiltonian for the master equation is built using creation and annihilation operators, and my goal is to study the large-number limit of the master equation, I’d rather work with those.

It’s a bit bizarre that we don’t know what and mean in the stochastic context! This is something I should correct sometime… I think I can sense yet another shoe dropping!

However, at present I have no understanding of these quantities in stochastic mechanics, and I have another goal in mind, so I’ll postpone thinking about them for a while. Thanks for reminding me to do it someday.

Interesting. Following is some off-the-cuff intuition on why one might want to scale the annihilation operator but not the creation operator. It may be way out there, but I just thought I’d run it by you. Anyway, here goes:

For pure states, we have that so that the creation operator maps pure states to pure states (i.e., in this case from a state having probability one of having things to one having probability one of having things). However, the annihilation operator satisfies so it typically maps a pure state to a non-normalized state (i.e., the coefficient is typically greater than one). So, it makes sense to rescale the annihilation operator by the inverse of the “expected” maximum “system size.” Then, both operators will typically map pure states to pure states (and hence mixed states to mixed states), as long as the number of things being annihilated at any given time is less than . If we want the possibility of an infinite “system size,” then we should take the limit .

At any rate, whether that makes any sense or not, I am looking forward to the next post on this to see where you’re going with it.

I’m glad you’re enjoying my attempts…

I have a few reasons for rescaling the annihilation and creation operators the way I did. First, as you mention, it’s nice to retain the property that hitting a normalized state with a creation operator gives a normalized state. Second, the master equation seems to look simpler in terms of rescaled annihilation and creation operators if we do what I did, instead of the more symmetrical-looking

Third, some things about coherent states seem to work better.

But I’ll have to keep doing calculations to know for sure what’s the best setup.

It is all elegant, beautiful and thorough.

I don’t like only the loss of the symmetry in the and operator definition.

I am thinking that the classical limit can be obtained using a product of creation, or annihilation:

and

in the limit of great n, could be possible that is obtained the classical limit? Or the Poisson bracket?

The operator number is:

but the commutation rules are very complex.

If we want to keep the symmetry we can use

If I were wanting to build self-adjoint Hamiltonians I would probably want to do this, since then and it’s easy to tell which combinations of and are self-adjoint.

But in stochastic mechanics we want infinitesimal stochastic Hamiltonians, with

for all using my usual definition of the angle bracket:

for

In this situation it’s very nice that the creation operators obey a simple rule:

which implies

I have chosen my rescaled annihilation, creation and number operators

so they still obey these rules:

which implies

along with simple commutation relations.

Using powers of annihilation and creation operators makes things very complicated, and I don’t think it fits the idea of ‘rescaling’ that’s built into the classical limit.

It is a little more complex, for two particles:

where is the number operator for a single particle, and

so on.

A dead end.

There is a simple result:

so that (quantum mechanic analogy with [x,p]):

and

to obtain higher degree derivative

and

is it possible with these commutators to obtain annihilation of N particles?

I’ve been telling you a lot about stochastic mechanics, which is like quantum mechanics but with probabilities replacing amplitudes. In Part 1 of this mini-series I started telling you about the ‘large-number limit’ in stochastic mechanics. Now let’s see how coherent states get into the act.