Markov
Chains
I thought I’d
try to explain what a Markov Chain is here.
Hopefully without too much math.
Anyone reading this, please correct me on anything I say here that isn’t
right. It’s been a while since I did
this in school.
Basically,
a Markov chain is a model of a process that has a certain number of “states” it
can be in, and changes from state to state at discrete intervals (every dice
roll, or every day, or every at-bat) with certain fixed probabilities.
For
instance, consider moving around a Monopoly board (ignoring jail and doubles
and Chance cards and such). There are
36 squares on the board. These can be
the “states” – call them state 0 to state 35.
To get from
“Go” to “Baltic Avenue”, you have to roll a 3.
The chance of that happening is 1 in 18, and so the transition
probability from Go (state 0) to Baltic Avenue (state 3) is .0555. With 36 states, there are 36 times 36 = 1296
different transition probabilities.
Most of these are zero – it’s impossible to move from “Go” to
“Boardwalk” in one roll.
That’s not
too hard, to write down all 1,296 probabilities. But it gets complicated when there’s more than one dice
roll. What is the chance of
transitioning from Go to Marvin Gardens (29 squares) in three rolls? Well, you could roll 10-10-9, or you could
roll 9-10-10, or you could roll 12-6-11 … and once you’ve figured out all
the combinations, you could add up all the probabilities to get your
answer. But there are a lot of
combinations, and this is only three rolls – what happens when you want to
figure it out for 10 rolls, or 20, or 1,000?
As it turns
out, if you have a computer and the first set of 1,296 probabilities, it’s easy
to figure out the probabilities for 10 rolls, just by writing the
1,296 probabilities in a matrix, and multiplying that matrix by itself 10
times (a linear algebra thing that computers can do quickly). So if you want to know what the probability
is of winding up on Park Place after exactly 87 rolls, that can actually
be calculated pretty easily.
The concept
extends pretty easily to baseball. Suppose
you’re wondering how many runs the Seattle Mariners would score in a game if
you put together a batting order alphabetically (so Beltre bats first
and [Ichiro] Suzuki bats last).
To do that, you start by defining the states. There are going to be a lot more than 36 of them, though. You’ve got combinations of inning (9 possibilities), runners on base (8), how many outs (3), and who’s batting (9 batting slots). That’s 9 times 8 times 3 times 9, or 1944 states. Multiply that by itself, and you get almost 4 million probabilities. (Or more -- you might need to consider the number of runs scored, too.)
Most of
those probabilities are zero – you can’t go from two outs in the fifth to one
out in the seventh in one plate appearance.
The others, you can calculate using reasonable methods. If the state is “third inning, one out,
runner on first and second, Ibanez at bat,” what’s the transition probability
to “third inning, one out, nobody on, Lopez at bat”? It’s pretty much just the probability of Ibanez hitting a
home run, which is maybe 20/685, going by his 2005 stats. Basically, from any state, you can only get
to a couple of handfuls of other states, and the probabilities aren’t that hard
to figure out.
You do have
to make a few assumptions. For
instance, suppose the chance that Ibanez hits a double is 5%. But does the runner on first score? If you figure the runner on first will score
half the time, you can give a 2.5% chance of transitioning to “third inning,
one out, runners on second and third, Lopez at bat” and the other 2.5% to
“third inning, one out, runner on second, Lopez at bat.” But the point is that you only have to worry
about the single plate appearance, not the whole game. The matrix multiplication takes care
of figuring the game.
That is,
the Markov Chain technique says, that (a) if you can come up with a bunch of
states; and (b) if you can figure the probabilties of any possible single
change of state; then (c) a good computer can tell you all kinds of stuff about
what happens over a whole bunch of state changes instead of just one.
And for the
lineup question, (a) we know all the states, which consist of
inning/outs/bases/batter; (b) we can make reasonable assumptions, like an APBA
game, of what all the single probabilities are for an at-bat; which means that
(c) the computer can tell us the probabilities of what happens over a whole
game.
The master
of applying the Markov technique is Mark
Pankin. Mark has published numerous studies on the topic, and the
Philadelphia Daily News recently ran an article on his recommendations
for the Phillies batting order (including rebuttal comments from Manager
Charlie Manuel).
By the way,
I don’t think Mark actually uses the “4 million transition probabilities”
method – I think he found a way to simplify it into something more
manageable.