Part VII: Probability & Confirmation
Chapters 19–21: Bayesian Confirmation Theory, The Problem of Induction, and Statistical Inference in Science
Science aspires to be more than mere speculation — it seeks confirmed theories, supportedhypotheses, and warranted beliefs about the natural world. But what does it mean for evidence to confirm a hypothesis? How should we update our beliefs in light of new data? And can we ever justify the inductive leap from observed regularities to universal laws?
Part VII of this course examines the philosophical foundations of probability and confirmation theory — the formal frameworks that attempt to make precise the informal notion of “evidential support.” At the centre of these debates stands Bayesianism, which has become the dominant account of confirmation in contemporary philosophy of science. The Bayesian framework provides an elegant, mathematically rigorous apparatus for understanding how evidence bears on hypotheses — yet it faces deep philosophical objections concerning the nature and origin of prior probabilities.
Underlying all of confirmation theory is the ancient problem of induction, first articulated with devastating clarity by David Hume in the 18th century and given a new twist by Nelson Goodman in the 20th. If we cannot justify induction, then the entire edifice of empirical science rests on foundations that philosophy cannot secure. Various responses — from Strawson’s dissolution to Reichenbach’s pragmatic vindication to Bayesian approaches — attempt to address this challenge.
Finally, we examine the statistical methods that scientists actually use in practice — frequentist hypothesis testing, significance testing, Bayesian statistics, and model selection — and the philosophical assumptions that underpin them. The ongoing “replication crisis” in the sciences has brought these methodological questions from the seminar room to the front pages, making this chapter among the most practically urgent in the entire course.
The Central Question
The core problem can be stated simply: What is the relationship between evidence and hypothesis? When a scientist performs an experiment and obtains a result, in what sense does that result “support” or “confirm” the theory being tested? The answer turns out to be far more subtle than common sense would suggest.
Consider a simple example. A theory predicts that a certain chemical reaction will produce a blue precipitate. You perform the experiment and observe a blue precipitate. Does this confirm the theory? Intuitively, yes — but notice that many different theories might predict a blue precipitate. The observation is logically consistent with infinitely many hypotheses. The question is how to quantify the degree to which the evidence favours one hypothesis over another.
“Probability is the very guide of life.”— Joseph Butler, The Analogy of Religion (1736)
The Bayesian answer, which has come to dominate philosophy of science, is that confirmation should be understood in terms of conditional probability. Evidence $E$ confirms hypothesis $H$ just in case the probability of $H$ given $E$ is greater than the prior probability of $H$:
$$P(H|E) > P(H)$$
This deceptively simple formula, derived from Bayes’ theorem, gives rise to a rich and powerful theory of confirmation — one that can handle the ravens paradox, the grue problem, and many other traditional puzzles. But it also raises new problems: Where do the prior probabilities come from? How do we handle old evidence? Can Bayesian agents with different priors ever converge?
These three chapters trace the arc from the abstract mathematics of probability to the concrete practice of statistical inference in the sciences, showing how philosophical assumptions permeate every stage of the scientific process.
Historical Context
The mathematical theory of probability emerged in the 17th century from correspondence between Pascal and Fermat about games of chance. But its application to scientific reasoning has a longer pedigree. Thomas Bayes, an 18th-century Presbyterian minister, posthumously published the theorem that bears his name in 1763. Pierre-Simon Laplace independently developed and greatly extended Bayesian methods, applying them to astronomical problems and articulating the “classical” interpretation of probability as a ratio of favourable to possible outcomes.
The 20th century saw an explosion of work on probability and confirmation. Rudolf Carnap attempted to develop a purely logical theory of confirmation, while Karl Popper rejected confirmation entirely in favour of falsification. The Bayesian revival, led by figures like Frank Ramsey, Bruno de Finetti, L.J. Savage, and later philosophers like Richard Jeffrey, Jon Dorling, and Colin Howson, restored probabilistic confirmation to the centre of the field.
Meanwhile, the practice of statistics in science developed largely independently of philosophical concerns. R.A. Fisher, Jerzy Neyman, and Egon Pearson developed the frequentist methods that dominated 20th-century science. The tension between frequentist and Bayesian approaches — and the philosophical assumptions underlying each — remains one of the most important and practically consequential debates in the philosophy of science.
Chapters in Part VII
Bayesian Confirmation Theory
How Bayes’ theorem provides a general framework for understanding how evidence confirms hypotheses. We examine the distinction between subjective and objective Bayesianism, the problems of old evidence and the priors, and how Bayesian confirmation handles classic paradoxes like the ravens and grue.
The Problem of Induction
Hume’s devastating argument that induction cannot be rationally justified, and the centuries of philosophical responses it has provoked. From Strawson’s dissolution to Goodman’s “new riddle” to Popper’s deductivism, we examine the deepest challenge to empirical knowledge.
Statistical Inference in Science
The philosophical foundations of the statistical methods scientists actually use: frequentist hypothesis testing, p-values, Bayesian statistics, likelihoodism, and model selection. We examine the replication crisis and what it reveals about the logic of statistical reasoning.
Key Themes Across Part VII
The Nature of Probability
What are probabilities? Are they objective frequencies in the world, subjective degrees of belief, logical relations between propositions, or propensities of physical systems? The answer profoundly shapes how we understand scientific confirmation.
Subjectivity in Science
Bayesianism appears to introduce subjective elements — prior probabilities — into the heart of scientific reasoning. Is this a fatal flaw, or does the machinery of conditionalization ensure that subjective starting points wash out in the long run?
Theory and Practice
The gap between philosophical theories of confirmation and the statistical methods scientists actually use is both wide and consequential. The replication crisis reveals what happens when practice drifts from sound epistemological foundations.
The Limits of Formal Methods
Can the relationship between evidence and hypothesis be fully captured by a mathematical formalism? Or does scientific reasoning involve irreducibly informal elements of judgment — what Kuhn called the “essential tension” between tradition and innovation?
Key Philosophers in Part VII
| Philosopher | Key Contribution | Chapter |
|---|---|---|
| Thomas Bayes | Bayes’ theorem for inverse probability | 19 |
| David Hume | The original problem of induction | 20 |
| Nelson Goodman | The new riddle of induction (grue) | 19, 20 |
| Carl Hempel | The ravens paradox of confirmation | 19 |
| Rudolf Carnap | Logical probability and degree of confirmation | 19, 20 |
| R.A. Fisher | Significance testing and p-values | 21 |
| Deborah Mayo | Error statistics and severe testing | 21 |
| Hans Reichenbach | Pragmatic vindication of induction | 20 |
Formal Foundations
Part VII makes extensive use of the probability calculus. The three axioms of probability (Kolmogorov axioms) are:
$$1.\; P(A) \geq 0 \text{ for all } A$$
$$2.\; P(\Omega) = 1$$
$$3.\; \text{If } A \cap B = \emptyset, \text{ then } P(A \cup B) = P(A) + P(B)$$
From these axioms we derive conditional probability and Bayes’ theorem:
$$P(H|E) = \frac{P(E|H) \cdot P(H)}{P(E)}$$
This single equation is the engine of Bayesian confirmation theory. Understanding its components — the prior $P(H)$, the likelihood $P(E|H)$, the marginal likelihood $P(E)$, and the posterior $P(H|E)$ — is essential for the chapters that follow.