If a butterfly flaps its wings in China today, it may cause a tornado in America next week.
Most of you will be familiar with this “Butterfly Effect” that is frequently used to illustrate
a typical behavior of chaotic systems: Even smallest disturbances can grow and have big
consequences. The name “Butterfly Effect” was popularized
by James Gleick in his 1987 book “Chaos” and is usually attributed to the meteorologist
Edward Lorenz. But I recently learned that this is *not what Lorenz actually meant by
I learned this from a paper by Tim Palmer, Andreas Döring, and Gregory Seregin called
“The Real Butterfly Effect” and that led me to dig up Lorenz’ original paper from
1969. Lorenz, in this paper, does not write about butterfly wings. He instead refers to
a sea gull’s wings, but then attributes that to a meteorologist whose name he can’t
recall. The reference to a butterfly seems to have come from a talk that Lorenz gave
in 1972, which was titled “Does the Flap of a Butterfly’s Wings in Brazil set off
a Tornado in Texas?” The title of this talk was actually suggested
by the session chair, a meteorologist by name Phil Merilees. In any case, it was the butterfly
that stuck instead of the sea gull. And what was the butterfly talk about? It was a summary
of Lorentz 1969 paper. So what’s in that paper?
In that paper, Lorenz made a much stronger claim than that a chaotic system is sensitive
to the initial conditions. The usual butterfly effect says that any small inaccuracy in the
knowledge that you have about the initial state of the system will eventually blow up
and make a large difference. But if you did precisely know the initial state, then you
could precisely predict the outcome, and if only you had good enough data you could make
predictions as far ahead as you like. It’s chaos, alright, but it’s still deterministic.
Now, in the 1969 paper, Lorenz looks at a system that has an even worse behavior. He
talks about weather, so the system he considers is the Earth, but that doesn’t really matter,
it could be anything. He says, let us divide up the system into pieces of equal size. In
each piece we put a detector that makes a measurement of some quantity. That quantity
is what you need as input to make a prediction. Say, air pressure and temperature. He further
assumes that these measurements are arbitrarily accurate. Clearly unrealistic, but that’s
just to make a point.
How well can you make predictions using the data from your measurements? You have data
on that finite grid. But that does not mean you can generally make a good prediction on
the scale of that grid, because errors will creep into your prediction from scales *smaller
than the grid. You expect that to happen of course because that’s chaos; the non-linearity
couples all the different scales together and the error on the small scales doesn’t
stay on the small scales.
But you can try to combat this error by making the grid smaller and putting in more measurement
devices. For example, Lorenz says, if you have a typical grid of some thousand kilometers,
you can make a prediction that’s good for, say, 5 days. After these 5 days, the errors
from smaller distances screw you up. So then you go and decrease your grid length by a
factor of two. Now you have many more measurements and much more data. But, and here comes the
important point: Lorenz says this may only increase the time for which you can make a
good prediction by half of the original time. So now you have 5 days plus 2 and a half days.
Then you can go and make your grid finer again. And again you will gain half of the time.
So now you have 5 days plus 2 and half plus 1 and a quarter. And so on.
Most of you will know that if you sum up this series all the way to infinity it will converge
to a finite value, in this case that’s 10 days. This means that even if you have an
arbitrarily fine grid and you know the initial condition precisely, you will only be able
to make predictions for a finite amount of time.
And this is the real butterfly effect. That a chaotic system may be deterministic and
yet still be non-predictable beyond a finite amount of time .
This of course raises the question whether there actually is any system that has such
properties. There are differential equations which have such a behavior. But whether the
real butterfly effect occurs for any equation that describes nature is unclear. The Navier-Stokes
equation, which Lorenz was talking about may or may not suffer from the “real” butterfly
effect. No one knows. This is presently one of the big unsolved problems in mathematics.
However, the Navier-Stokes equation, and really any other equation for macroscopic systems,
is strictly speaking only an approximation. On the most fundamental level it’s all particle
physics and, ultimately, quantum mechanics. And the equations of quantum mechanics do
not have butterfly effects because they are linear. Then again, no one would use quantum
mechanics to predict the weather, so that’s a rather theoretical answer.
The brief summary is that even in a deterministic system predictions may only be possible for
a finite amount of time and *that is what Lorenz really meant by “Butterfly Effect.”
You find the references to the papers in the information below the video. Thanks for watching.