57 pages • 1-hour read
Charles WheelanA modern alternative to SparkNotes and CliffsNotes, SuperSummary offers high-quality Study Guides with detailed chapter summaries and analysis of major themes, characters, and more.
The author reflects on his uneasy relationship with mathematics, particularly when he couldn’t see the real-world applications of concepts he was learning. During his senior year of high school, he took an AP Calculus final exam unprepared. Finding the questions completely unfamiliar, he approached his teacher, Carol Smith, with whom he had a contentious relationship. She suggested he should have studied. After another student, Brian Arbetter, quietly spoke with her, Mrs. Smith discovered she had mistakenly distributed the second-semester exam and canceled the test. The author ranks this vindication as a top-five life moment.
Despite hating calculus, Wheelan enjoyed physics and probability because they had practical applications. He introduces the Monty Hall problem from Let’s Make a Deal, where contestants face three doors and must decide whether to switch their choice after one door is revealed to contain a goat.
During graduate school math camp, classmate Will Warshauer struggled with understanding how an infinite series could converge to a finite number. The author explained it intuitively: Repeatedly moving half the remaining distance to a wall means getting infinitely close but never reaching it, with the total distance converging to a finite limit.
The author argues statistics can be misused despite their power. He compares them to a weapon—helpful when used correctly but dangerous otherwise. He imagines a hypothetical study linking taking work breaks to developing cancer, which misleadingly ignores that the breaks are often for smoking. The book aims to make statistical concepts intuitive and accessible while teaching readers to recognize misuse. He cites Swedish mathematician Andrejs Dunkels’s observation that, while statistics can be used to deceive, telling the truth without them is even harder.
The author observes that students who find statistics confusing often discuss statistical concepts comfortably in everyday contexts like batting averages. The NFL passer rating exemplifies a useful but imperfect descriptive statistic: Comparing Jay Cutler’s playoff rating of 31.8 to Aaron Rodgers’s 55.4 efficiently explains why the Chicago Bears lost to the Green Bay Packers in 2011. Similarly, the Gini index measures income inequality on a scale from zero to one, enabling comparisons across countries. Sweden’s index of .23 indicates greater equality than that of the United States at .45 or South Africa at .65. Statistics process data to transform raw information into meaningful insights.
Statistics serve four main functions. First, description and comparison summarize complex information using tools like averages. Second, inference uses samples to make informed conjectures about larger populations, as with political polling or studies of sexual behavior. Third, assessing risk and probability allows casinos and insurance companies to manage uncertainty. For instance, the firm Caveon Test Security uses probability to detect cheating by identifying statistically unlikely patterns of identical wrong answers on exams. Finally, identifying important relationships through statistical detective work allows researchers to isolate associations while controlling for confounding factors. For example, Princeton economist Alan Krueger’s terrorism research found that terrorists tend to be well educated and from middle- or high-income families, contradicting common assumptions. His data also showed more terrorism in countries with greater political repression. Krueger hypothesizes that because terrorists are motivated by political goals, the educated and affluent have the strongest incentives to force societal change.
The author acknowledges that statistical analysis rarely reveals absolute truth and that honest people can disagree about interpretations. Former Secretary of Defense Donald Rumsfeld’s aphorism about working with available resources applies equally to research with imperfect data. Statistics can be deliberately misused or lead to inadvertent errors, making it essential to recognize both legitimate findings and manipulation.
Wheeler opens by asking two questions related by the answers’ use of statistics. First, what is the financial health of the US middle class? And second, who is the greatest baseball player ever? Descriptive statistics summarize large data sets by finding different kinds of middle points. Derek Jeter’s career batting average of .313 encapsulates 17 seasons of performance. In contrast, using per capita income (another kind of average) to measure middle-class well-being misleads because this information is not adjusted for inflation and can be skewed by extreme wealth.
Two fundamental measures of central tendency, or the middle, are the mean and median. The mean is the average of a set of numbers, calculated by adding up the values and dividing by the number of separate values in the set. The mean is sensitive to outliers. For example, if Microsoft founder Bill Gates enters a bar where 10 patrons each earn $35,000 dollars annually, his billion-dollar income raises the patrons’ mean income to $91 million. In contrast, the median is derived by arranging a set of values in ascending order and finding the middle one. In the same bar example, even with Bill Gates in the room, the patrons’ median income remains $35,000 (the middle of the set of 11 numbers). Even if Warren Buffett joins them, the median still remains $35,000 (the middle of the set of 12 numbers).
In a printer quality comparison, one company’s products are seemingly vastly inferior: They have a mean of 9.1 defects versus a competitor’s 2.8. But the median reveals a different story. A frequency distribution shows that the first company actually has a lemon problem—most printers work well, but a few have numerous defects that inflate the mean.
Standard deviation measures dispersion around the mean, or how far most pieces of data will be from the average. For normally distributed data, 68.2 percent of observations fall within one standard deviation of the mean, and 95.4 percent within two. This property makes the normal distribution—which readers may know as the bell curve—foundational to statistics.
Percentages require careful interpretation. Wheeler poses a quick math problem to readers: If a dress is originally priced at $100, then marked down 25%, and then marked up 25%, how much does it cost? The answer is counter-intuitive: The dress now costs $93.75, not $100 (because the markup is calculating 25% of $75). In a real-world example that illustrates a similar slipperiness with percentages, Illinois’s income tax increase from 3% to 5% can be correctly described by proponents as a two-percentage-point increase and by opponents a 67% increase.
The chapter ends by returning to the questions posed at its beginning. To evaluate the greatest baseball player, Steve Moyer, president of Baseball Info Solutions, suggests using mean calculations: on-base percentage, slugging percentage, and at bats. These statistics point to Babe Ruth as the greatest player due to his unique ability to both hit and pitch. In contrast, economists Jeff Grogger and Alan Krueger recommend examining median wages, adjusted for inflation, to assess middle-class economic health. A graph of 1979 to 2011 shows that the median wage stagnated, while the 90th percentile grew significantly.
The chapter explores how technically accurate statistics can mislead. Precision differs from accuracy. Precision reflects mathematical exactitude, while accuracy measures consistency with truth and fact. For instance, Senator Joseph McCarthy’s 1950 claim about possessing a list of 205 communists in the State Department was precise, in that he named a specific number, but false, in that he had no such names. The author’s golf range finder provided precise measurements but because it was set to meters instead of yards and sometimes aimed at the wrong targets, this precise information was useless. Finally, Wall Street risk models before the 2008 crisis were precise, but built on flawed assumptions, making their conclusions harmful.
Ambiguity about what is being measured can create contradictory but simultaneously true statements. For example, data from The Economist show US manufacturing output rising while employment falls. This is because the units of analysis differ: With greater automation, productivity can increase while fewer workers are being hired. Likewise, examining the quality of schools as a whole versus the preparedness of students individually can yield opposite conclusions. Similarly, global inequality appears to have gotten worse over time when countries are the unit of analysis. However, when people are the unit, we can see that global inequality has actually been falling, because a high proportion of the world’s poor live in fast-growing nations like China and India.
The mean and median can be selectively cited to support opposing narratives. The George W. Bush administration promoted tax cuts by citing a mean reduction exceeding $1,000 per household. But this average hid the fact that the median for most households was actually under $100—a huge tax break for a small number of very rich people skewed the mean. Conversely, the median can obscure important information. Evolutionary biologist Stephen Jay Gould survived for 20 years after being diagnosed with a cancer with a median survival time of eight months. The median does not account for the extent and frequency of occurrence of outliers.
Inflation adjustment is crucial for comparisons of money across time. Hollywood reports box office using nominal dollars, without inflation adjustment, making recent films like Avatar (2009) appear more successful than older ones. When adjusted for inflation, Gone with the Wind (1939) tops the list, while Avatar falls to 14th. The US minimum wage’s nominal value has risen seemingly dramatically, but in reality, its real purchasing power has declined.
Manipulating timeframes also alters narratives. The author’s professor used identical defense spending data to praise either President Ronald Reagan or President Jimmy Carter depending on the political leanings of his audience; his graph of defense spending over the years could be zoomed in to show Reagan as increasing spending, or zoomed out to show Carter as the originator of a defense boom. Percentages can exaggerate when calculated from small bases. A seemingly enormous annual 527% tax increase for a Tuberculosis Sanitarium District actually cost homeowners less than a sandwich because it was increasing an initial value of $1.15 to $6.70.
Measuring the wrong thing creates perverse incentives. For example, using test scores alone to evaluate schools created not a better educational system, but a strong push for administrators to cheat the tests. In Houston, Texas, schools under Superintendent Rod Paige manipulated dropout statistics by reclassifying departing students and kept weak students from taking benchmark exams to inflate averages. Other testing scandals involved teachers supplying correct answers to students or correcting tests themselves before submitting them. Similarly, New York’s cardiologist scorecards about surgery outcomes led doctors to simply avoid operating on the sickest patients to protect their mortality statistics.
Measuring the wrong thing can also mislead about cause and effect. Chicago newspapers rank selective enrollment schools as best based on test scores, but admission to these schools requires high scores from applicants. This means that the kids who get in are likely to continue testing well. Praising schools for this is the logical equivalent of praising a basketball team for producing tall students.
Wheelan ends the chapter by critiquing the U.S. News & World Report college rankings, which use arbitrary weighting and focus on inputs rather than outcomes. Critics like Michael McPherson note the rankings lack genuine precision. Despite these criticisms, Bard College president Leon Botstein acknowledges people crave simple answers—this is why rankings will never go away. The chapter concludes that statistical malfeasance stems from poor judgment (accuracy) rather than mathematical errors (precision).
Charles Wheelan establishes his authorial persona as a relatable pragmatist skeptical of abstraction. The Introduction begins with a confession about hating math in high school and a funny personal anecdote about a high school calculus exam. Wheelan’s adolescent rebellion and vindication against a seemingly pointless academic exercise put him on the side of readers who may have been similarly discomfited by complex math courses like calculus. He contrasts this with his enjoyment of physics and probability, which have clear, practical purposes. The juxtaposition foregrounds the book’s argument: Statistics are valuable because of their application, not their theoretical elegance. This personal, narrative-driven approach is intended to disarm the reader’s potential math anxiety. By positioning himself as a fellow victim of abstract mathematics, Wheelan builds rapport and credibility. This narrative strategy aligns the book with popular science writing that prioritizes accessibility and intuition over technical rigor, presenting statistics as an applied tool for interpreting the world.
Wheelan’s informal, first-person narrative voice is a deliberate choice that reinforces his intent to make statistics accessible. The text is filled with casual asides, humorous scenarios like Bill Gates walking into a bar to illustrate statistical outliers, and self-deprecating admissions of his own mathematical shortcomings. This conversational and non-threatening tone contrasts sharply with that of a traditional textbook, subverting reader expectations that a book about statistics must be difficult and boring, and demonstrating that rigorous intellectual concepts can be conveyed effectively through a popular, accessible medium.
To underscore the dual nature of data analysis, Wheelan consistently employs the metaphor of statistics as a powerful but potentially dangerous tool. In the introduction, he compares statistics to “a high-caliber weapon: helpful when used correctly and potentially disastrous in the wrong hands” (xvi). This metaphor is developed throughout the opening chapters, culminating in Chapter 3’s focus on statements that are true but grossly misleading. The chapter is structured around examples of how Statistics Can Mislead or Be Manipulated, whether through cherry-picking the mean over the median in the debate over tax cuts, ignoring inflation in Hollywood box office reports, or creating perverse incentives as seen in the Houston schools dropout scandal. The weapon/tool metaphor transforms the abstract concept of statistical misuse into a tangible and memorable warning. It shifts the reader’s focus from the mechanics of calculation to the ethics and judgment involved in application, framing statistical literacy as a form of self-defense against manipulation.
The initial chapters illustrate statistics for a general audience by juxtaposing complex, academic concepts with familiar, everyday examples. This technique of analogy serves as the primary pedagogical tool for demystifying the discipline, highlighting the idea that Statistical Literacy is Empowering and not a rarefied skill. Chapter 1 explicitly connects the intimidating Gini index, a measure of income inequality, to the commonly understood NFL passer rating. Both are presented as functionally identical: descriptive statistics that condense “complex information into a single number” (2). Batting averages, grade point averages, and the memorable story of Bill Gates walking into a bar are used to ground and explain core concepts like mean, median, and the distorting effect of outliers. This method breaks down the barrier between “statistics” as a formal subject and statistical reasoning as a daily cognitive process. By demonstrating that readers already engage with statistical ideas in contexts like sports, Wheelan reframes the learning process as one of translating existing intuition into a more formal vocabulary.
The conceptual distinction between precision and accuracy is introduced as a critical theme in Chapter 3, functioning as a recurring motif that organizes the chapter’s various examples of statistical deception. Wheelan defines precision as exactitude and accuracy as truthfulness, illustrating the difference with anecdotes like Senator Joseph McCarthy’s specific but false claim about having a list of 205 communists and Wheelan’s own golf range finder set to the wrong units. This distinction then becomes the analytical lens for the rest of the chapter. Wall Street’s complex risk models were precise but fatally inaccurate, failing to stem the 2008 worldwide financial collapse. Hollywood’s nominal box office records are precise but inaccurate measures of commercial success. The U.S. News & World Report college rankings are criticized for creating a “level of precision that those data just don’t support” (56). By structuring the chapter around this motif, Wheelan provides a unifying framework for understanding different forms of statistical misuse. The argument is that the problem is often not bad math but a misplaced faith in numbers that seem exact, teaching the reader to question the underlying assumptions behind a statistic.



Unlock all 57 pages of this Study Guide
Get in-depth, chapter-by-chapter summaries and analysis from our literary experts.