By drawing these circles around each player’s preferred policy mix we learn something important. We see that these circles overlap. The areas of overlap show us policy combinations that improve on the status quo for a coalition of two of the three players. For instance, the lined oblong area tilting toward the upper left of the figure depicts policies that improve the well-being of Democrats and Republicans (ah, a bipartisan foreign policy opposed by independent blue-collar workers). The gray petal-shaped area improves the interests of Democrats and blue-collar independents (at the expense of Republicans), and the bricked-over area provides a mix of trade and defense spending that benefit the Republicans and blue-collar independents (to the chagrin of Democrats).
Because we assumed that each of the three voting blocs is equal in size, each overlapping area identifies defense and trade policies that command the support of two-thirds of the electorate. Here’s the rub, then, when it comes to talking about the national interest. One coalition wants more free trade and less defense spending. Another wants less free trade and less defense spending. The third wants less free trade and more defense spending. So, we can assemble a two-thirds majority for more defense spending and also for less. We can find a two-thirds coalition for more free trade or for higher tariffs or (in the politically charged rhetoric of trade debate) for more fair trade. In fact, there are loads of ways to allocate effort between defense spending and trade policy to make better off whichever coalition forms. 7
What, then, is the national interest? We might have to conclude that except under the direst circumstances there is no such thing as “the national interest,” even if the term refers to what a large majority favors. That is surprising, perhaps, but it follows logically from the idea that people will align themselves behind policies that are closer to what they want against policies that are farther from what they advocate. It just happens that any time there are trade-offs between alternative ways to spend money or to exert influence, there are likely to be many different spending or influence combinations that beat the prevailing view. None can be said to be a truer reflection of the national interest than another; that reflection is in the eyes of the beholder, not in some objective assessment of national well-being. So much for the venerable notion that our leaders pursue the national interest, or, for that matter, that business executives single-mindedly foster shareholder value. I suppose, freed as they are to build a coalition that wants whatever it is they also want, that our leaders really are free to pursue their own interests and to call that the national interest or the corporate interest.
WHAT IS THE OTHER GUY’S BEHAVIOR?
(DOES HE HAVE GOOD CARDS OR NOT?)
To understand how interests frame so many of the questions we have at stake, game theory still requires that people behave in a logically consistent way within those interests. That does not mean that people cannot behave in surprising ways, for surely they can. If you’ve ever played the game Mastermind, you’ve confronted the difficulties of logic directly. In Mastermind—a game I’ve used with students to teach them about really probing their beliefs—one player sets up four (or, in harder versions, more) colored pegs selected from among six colors in whatever order he or she chooses. The rest of the players cannot see the pegs. They propose color sequences of pegs and are told that yes, they got three colors right, or no, they didn’t get any right, or yes, they got one color in the right position but none of the others. In this way, information accumulates from round to round. By keeping careful track of the information about what is true and what is false, you gradually eliminate hypotheses and converge on a correct view of what order the colored pegs are in. This is the point behind a game like Mastermind, Battleship, or charades. It is also one point behind the forecasting games I designed and use to predict and engineer events.
The key to any of these games is sorting out the difference between knowledge and beliefs. Different players in any game are likely to start out with different beliefs because they don’t have enough information to know the true lay of the land. It is fine to sustain beliefs that could be consistent with what’s observed, but it’s not sensible to hold on to beliefs after they have been refuted by what is happening around us. Of course, sorting out when beliefs and actions are inconsistent requires working out the incentives people have to lie, mislead, bluff, and cheat.
In Mastermind this is easy to do because the game has rules that stipulate the order of guessing and that require the person who placed the pegs to respond honestly to proposed color sequences suggested by other players. There is no point to the game if the person placing the pegs lies to everyone else. But even when everyone tells the truth, it is easy to slip into serious lapses in logic that can lead to entirely wrong beliefs. That is something to be careful about.
Slipping into wrong beliefs is a problem for many of us. It is easy to look at facts selectively and to reach wrong conclusions. That is a major problem, for instance, with the alleged police practice of profiling, or some people’s judgment about the guilt or innocence of others based on thin evidence that is wrongly assessed. There are very good reasons why the police and we ordinary folk ought not to be too hasty in jumping to conclusions.
Let me give an example to help flesh out how easily we can slip into poor logical thinking. Baseball is beset by a scandal over performance-enhancing drugs. Suppose you know that the odds someone will test positive for steroids are 90 percent if they actually used steroids. Does that mean when someone tests positive we can be very confident that they used steroids? Journalists seem to think so. Congress seems to think so. But it just isn’t so. To formulate a policy we need an answer to the question, How likely is it that someone used steroids if they test positive? It is not enough to know how likely they are to test positive if they use steroids. Unfortunately, we cannot easily know the answer to the question we really care about. We can know whether someone tested positive, but that could be a terrible basis for deciding whether the person cheated. A logically consistent use of probabilities—working out the real risks—can help make that clear.
Imagine that out of every 100 baseball players (only) 10 cheat by taking steroids (game theory notwithstanding, I am an optimist) and that the tests are accurate enough that 9 out of every 10 cheaters test positive. To evaluate the likelihood of guilt or innocence we still need to know how many honest players test positive—that is, how often the tests give us a false positive answer. Tests are, after all, far from perfect. Just imagine that while 90 out of every 100 players do not cheat, 10 percent of the honest players nevertheless test (falsely) positive. Looking at these numbers it’s easy to think, well, hardly anyone gets a false positive (only 10 percent of the innocent) and almost every guilty party gets a true positive (90 percent of the guilty), so knowing whether a person tested positive must make us very confident of their guilt. Wrong! 8
With the numbers just laid out, 9 out of 10 cheaters test positive and 9 out of 90 innocent ball players also test positive. So, 9 out of 18 of the positive test results include cheaters and 9 out of 18 include absolutely innocent baseball players. In this example, the odds that a player testing positive actually uses steroids are fifty-fifty, just the flip of a coin. That is hardly enough to ruin a person’s career and reputation. Who would want to convict so many innocents just to get the guilty few? It is best to take seriously the dictum “innocent until proven guilty.”
Читать дальше