Pages

Thursday, October 02, 2008

Economics is Not a Science, Nor is Basketball Statistical Analysis

"Economists cannot predict tomorrow's economy; they cannot agree on the state of the economy today; they cannot even arrive at a consensus on why the economy behaved as it did in the past...We may not ever be able to build a positive science of economics based on empirical knowledge but that is no reason to wrap the little we know in a pseudoscientific fog of superstition."--Walter Russell Mead, 1993.

Considering how the U.S. financial markets are crumbling right before our eyes, don't those words from 15 years ago send a chill up your spine? Economists want the general public--and especially their employers in the academic world--to think that they are practicing science but sadly that is not the case. What does this have to do with basketball? In recent years, many people who trained academically as economists and/or statisticians have tried to use "econometric" style models to make evaluations about basketball players and teams. Frankly, I'm not sure if it would be more frightening if these guys devoted all of their energies to confusing the general public about basketball or if they abandoned this effort and resumed their failed attempts to understand our past, present and future economy. In either case, they will no doubt continue to make bold pseudoscientific declarations with complete confidence that they understand the issues better than anyone else. 

Roger Lowenstein's review of Nassim Nicholas Taleb's book The Black Swan does not mention basketball at all but the way that he debunks the idea of economics as science provides a good template for explaining much of what is conceptually wrong with "econometric" based basketball statistical analysis. Lowenstein summarizes Taleb's thesis simply and bluntly: "His heavy artillery is aimed at financial academics and economists, the latter because they try to forecast such stuff as market moves and interest rates. He calls them frauds. It’s as good a word as any." Taleb used to be a derivatives trader on Wall Street, so he has firsthand experience with just how poorly economists understand what they are talking about. Lowenstein says, "The academics who drive him to tears are the ones who have explained—or misexplained—his old profession. They think that markets are from Mediocristan when in fact they inhabit Extremistan." Lowenstein explains, "Mediocristan is the terrain of the ordinary, the part of the world that conforms to the bell curve. It answers to statistics and knowable probabilities. Height resides in Mediocristan. You may find one 7-footer on your block, almost certainly not two...Personal wealth, however, is from Extremistan. For instance, the average wealth of 1,000 people will be very different if one of those people is Bill Gates. This distinction is potent. In Extremistan, past events are a faulty guide to projecting the future. Gates may be the world’s richest person, but it isn’t unthinkable that someday, someone (at Google, perhaps?) will be twice as rich. Wars also reside in Extremistan. Prior to World War II, the planet had never experienced a conflict as terrible. Then we did. Suppose you frequent a pond. Day after day you see swans—always white. Naturally (but incorrectly) you presume that all swans are white. World War II was a black swan—horrific and unpredictable. Market crashes are black swans. Winning at blackjack is not one. The odds in casino games are known. The finance profession has badly mischaracterized markets in such a way as to overlook the possibility of black swans. Business schools teach that risk is quantifiable—that markets resemble a casino. You will draw the bad queen once in a deck but never twice. That is why securities analysts presume to define the 'riskiness' of stocks in precise, arithmetic terms. They model the future on the past. But stocks, alas, are from Extremistan. Taleb makes much of the example of Long-Term Capital Management, a subject about which I wrote a book. The spectacular meltdown of the hedge fund, run by Nobel Prize-certified economists and intellectual heavyweights, was a primo example of faulty precision, of modeling markets according to past events. The fund’s genius managers couldn’t predict the black swan of the Russian debt default; they drew a run of bad queens, and down they went. And L.T.C.M. is merely emblematic; it is the entire profession of finance, its edifice of modern portfolio theory, and virtually every tool that financial consultants regularly rely on, that Taleb identifies as wrongheaded." 

It is precisely that kind of wrongheaded thinking--and arrogance--that leads to much of the nonsense that is spewed by people who think (or at least claim) that they are scientifically analyzing basketball (I'm not sure if these guys know better but enjoy selling books and getting a lot of publicity for themselves or if they really believe what they are saying). Unfortunately, the general public suffers not only from illiteracy (or at least poor reading comprehension) but also from "Innumeracy." If I make a skill set-based comparison of two players based on my informed opinion--i.e., an opinion based not only on watching a lot of basketball but also on interacting with professionals who make their living evaluating basketball players--every Tom, Dick and Harry thinks that his opinion is just as valid and informed but if some guy invents a formula, gives it a catchy name and says that player x is worth 30.2 but player y is worth 28.7 then Tom, Dick and Harry are ready to bow down to those numbers as if they are the Golden Calf. Guess what--all those numbers reflect are the knowledge (or lack thereof) and bias of the person who created that formula; the numbers may be 90% correct or 90% incorrect but most people don't understand math or statistics so they don't feel comfortable challenging the numbers, or else they only lash out at the numbers that speak poorly of "their guy" but they love the numbers that elevate "their guy" and/or downgrade "their guy's" rival. One tell-tale sign that these numbers are not the products of science is that you do not hear their creators speak of margin of error. If a scientific formula spits out the number 30.4 as a player value the reality is that there is a certain probability that the actual value is somewhat higher or lower than that--but the "stat gurus" rarely if ever mention this and they certainly don't emphasize this point as much as it should be emphasized; they like to promote the idea that their numbers are "exact" while observations by seasoned professionals (scouts and other talent evaluators) are subjective. It is true that observations are subjective but so are the stat formulas; that is why intelligent people understand that you have to combine observation with statistics and that you have to watch games in order to figure out what the numbers really mean. For instance, who is charged with a turnover is not nearly as significant as what really caused the turnover. If Pete Maravich throws a great pass to a stiff who fumbles the ball out of bounds, Maravich may get a turnover in the boxscore but anyone who understands the game realizes that the problem is that his teammate can't catch the ball; on the other hand, if Maravich carelessly throws the ball away or makes a pass that no one could reasonably be expected to catch, then that reflects badly on him regardless of how the scorekeeper officially documented that play. 

"Stat gurus" plainly do not want to discuss or consider the fact that some of their most precious numbers--the raw data that they plug into their formulas, stats like assists, steals, blocked shots and turnovers--are subjectively recorded. During last season's playoffs, I did a detailed post demonstrating that Chris Paul's supposedly record setting playoff assist totals were in fact inflated by generous scorekeeping. Shouldn't that be of interest to the "stat gurus"? Isn't that claim something that they seriously need to investigate on their own to either confirm or reject? I provided very specific information so that anyone could watch a tape of the game and find the exact plays that I described and thus judge for themselves whether or not each of those assists should have been awarded. Yet I see no indication that the "stat gurus" are the slightest bit concerned about the fact that a lot of their basic data is seriously flawed. A lot of these guys spent a good portion of the season pumping up Chris Paul as the MVP and it is highly likely that they did so on the basis of bogus assist numbers. Based on a skill-set evaluation of Paul's game, I consider him to be the best point guard in the NBA and a top five MVP candidate but that is not the point; the point is that if you are basing your whole analysis of the NBA purely on numbers and some of the basic numbers you are using are not right then your whole analysis is bogus. If a real scientist finds out that the raw data he has gathered is flawed then he understands that he has to gather new, accurate data. Unfortunately, many of the basketball "stat gurus" are not scientists; they are "mad scientists" at best. 

No one should misinterpret what I am saying to mean that I am some kind of Luddite who is against using basketball statistics; what I am against is the misuse of basketball statistics, just like I am against the misuse of media platforms by people who spout hype and biased commentary as opposed to communicating information in a fact based, objective manner. Dan Rosenbaum and Dean Oliver are two welcome exceptions to the above critique of basketball statistical analysis; everything that I have seen of their work indicates that they understand the limitations of what their statistical analysis can show and that they are working hard to improve what their models can do as opposed to acting like they have everything figured out already. In an insightful post titled Using statistics in basketball: the bar is higher, Rosenbaum writes, "Statistical analysis can play a critical role in basketball decision-making, but it can also be misleading if the complexities of the game of basketball (and the statistical issues generated by those complexities) are not well understood. In other words, the bar is higher for statistical analysis in basketball than it is in baseball. Ultimately this will greatly benefit the teams that incorporate skilled statistical analysts in the right way, because the greater complexities in basketball will mean that it will be harder for other teams to ever catch up with the first teams that get this right. It will be fascinating seeing how this all plays out over the next few years."

5 comments:

  1. David,

    You pretty much summed up my misgivings about heavy reliance on statistical analysis here. I find it amusing when people say (for example) player X is better than player Y, and it as a fact because the "stats" say so. No, the fact is that player X has a higher (insert player rating tool here) than player Y. Those tools are not objective because they are subject to the subjective opinion of the person who created them on how the statistical data should be analysed.

    On top of that, basketball is such a complex game that it is impossible to capture a lot of what happens with any statistical tool. (For example, I have yet to find a satisfactory way to statistically evaluate "clutch" play or individual defense.) So if the opinions of scouts and talent evaluators are considered "subjective", it's because what they are doing is not as simple as plugging numbers into a formula. Unless they make subjective assessments they will miss a lot of what is happening on the court in their analysis.

    Far be it from me to completely dismiss the usefulness of some of these tools and statistics in general, but the way some people throw the numbers around you would think they are scientific fact and have all the answers. I guess the people who get paid to scout players and teams in person are just wasting their time since all the information they need is already contained in the stats.

    ReplyDelete
  2. First of all, while I agree that there are good arguments about the limitations of basketball statistical analysis, you’ve chosen exactly the wrong critique. ‘Black Swans’ are not an issue in basketball statistics (not at the NBA, at least), because there are no catastrophic, paradigm-changing shifts in the game. There are incremental changes, of course, as rules change, training techniques and tactics evolve and so forth, but a ‘Black Swan’ event would be suddenly using a tennis ball instead of a basketball, moving the hoop up to 25 feet, and allowing blocking and tackling. In that case, a player’s statistics before the change wouldn’t be much use in predicting their statistics afterwards. But, see, that doesn’t happen (unless you think MJ’s career pro stats shouldn’t include just his NBA ones, I guess ).

    Your example about who’s responsible for a dropped pass isn’t very strong either: sure, some good passes will be dropped, but the whole point of keeping statistics is that some bad passes will be saved; and over enough passes this should even out.

    There IS a good argument to be made about the difficulty of isolating any particular variable, particularly when looking at individuals – are a particular player’s scoring totals good instead of average because a teammate is creating double- and triple-teams for the defense, or are they only good instead of great because the teammate is the first option? And in basketball this means statistics need to be carefully designed and carefully used in evaluating players, and put in the context of the team. Thoughtful analysts are looking for ways to do that in a somewhat objective manner, which is the whole point.

    After all, the point of basketball is winning, and that's decided by a statistic, not a basketball insider's personal opinion about who won, so it seems plausible that other statistics might have something interesting to say about the game.

    ReplyDelete
  3. Anonymous:

    Barring a major rules change, you are right that there will not be a "Black Swan" in the NBA equivalent to World War II or some other catastrophic real world crisis but the point of Taleb's book and Lowenstein's review is that the methodology employed by economists is not scientific and not highly effective at understanding the past or present nor at predicting the future.

    Rosenbaum's quote--and his post that I linked to--explains why basketball stat analysis is harder to do in a meaningful way than baseball stat analysis. Basically, in basketball there are more moving parts--certainly there are more parts moving simultaneously--so it is difficult to come up with metrics that accurately explain what is happening/has happened. So even if there are not "Black Swans" in the NBA, basketball is still difficult to measure using "econometric" models. I see a value in attempting to do this anyway but I reject assertions that this goal has already been accomplished, at least by any of the methods that have been so well publicized in various books and websites.

    ReplyDelete
  4. David - Thoughts....

    First, basketball is played in Mediocristan. Quintessentially so. The Black Swan argument doesn't apply. It's not a game where you have odds that are 99/1 or 999/1 against. It's a game much more akin to coinflipping, albeit with a moderately weighted coin.

    Taleb wouldn't have an issue with basketball statistics. At the end of the day, basketball is about wins, and wins are about who scores more points. It's statistical by nature, unlike say, gymnastics. So it's a perfectly reasonable, even ideal subject for an academic to create a statistical model for.

    Some more thoughts.

    First, it's important to note that Berri's purpose is not to try to concoct a high falutin way of winning a locker room argument. His main purpose is to create a model in order to explore the same quirks in human behavior that Taleb writes chapter and verse about. Second, you ought to read the WOW at this point, given how many posts you have written about it directly or indirectly. I will buy you a copy if you send me your address. Third, if you did read the book, you would find a wealth of stuff in there not related to the individual performance metric you find so infuriating. For instance, there is an major analysis of the idea that highly paid superstars drive gate attendance. He does a lot of work showing that it's team wins that is most responsible for filling seats and that highly paid superstars only fill seats when they actually win games, which not all of them do. That is a very interesting conclusion that no one I know of disputes, and it might have been an interesting insight to incorporate into your post about Arenas and his new contract. There is also a lot of work on salary in the NBA that doesn't involve Wins Produced and which you might find interesting.

    Finally, Berri, for his part, has written with high praise on several occasions about Taleb's books. So, if you think Taleb threatens Berri, well, it just isn't so.

    What Taleb would have a huge problem with is claims about the future. For instance, if someone bilked pensioners out of their money by convincing them he had found in the Wages of Wins a foolproof way to make money betting on basketball. And I can assure Berri would have a big problem with that too for the same reasons.

    I love both Taleb and Berri and can assure you, there is no significant point of disagreement between the two. A major foundation of the WOW is the idea that people in general (and nba executives and commentators in particular) are fooled by randomness and misled by heuristics and cognitive biases . This is totally in line with Taleb's thinking.

    Really, the people Taleb is aiming at in the basketball context are not the economists but people like sports journalists, the "historians," who try , as Lowenstein says, to "make a sensible yarn of what might, in reality, be unpredictable or random."

    It's ironic really, this whole post, since half the posts I have written here are directly inspired by Taleb and the other half by Berri. But anyway....

    Owen

    ReplyDelete
  5. Owen:

    It is an oversimplification to say that basketball is played in "Mediocristan" merely because there are only two possible outcomes for a team in a given game. In order to evaluate/compare players there are many factors that have to be considered and the current statistical tools/methods are not adequate to do this without also applying contextual information based on the input of knowledgeable basketball observers.

    I don't object to academics or anyone else trying to create statistical models for basketball. I actually have always been quite enthusiastic about the idea and I have bought many books on the subject over the past two decades, from Martin Manley's to Dave Heeren's to Berri's to many others. What I object to is when someone who has created a stat model acts as though he has devised the definitive evaluation method and that his model supersedes the need for actually watching the games.

    Taleb certainly fired shots in many directions, including at historians and journalists but, as the review noted, he expressed particular disdain for economists who claim that they can accurately model the future. The Mead and Rosenbaum quotes are significant parts of the post and their insights should be given heavy weight.

    ReplyDelete