Sample Spaces, Events, and Conditioning
sample space, event, conditional probability, multiplication rule, independence
1 Role
This page is the real starting point of probability.
It explains how probability models are built, what an event actually is, and why conditioning is the basic way we update probabilities when some information becomes known.
2 First-Pass Promise
Read this page first.
If you stop here, you should still understand:
- what a sample space is
- what an event is
- what a probability assignment is trying to describe
- why conditioning means
shrink the world and renormalize
3 Why It Matters
Probability does not begin with formulas for means or Gaussians. It begins with model design:
- what outcomes are possible?
- what subsets of outcomes matter?
- what information is known?
- how should the probabilities change after that information arrives?
Those questions keep returning in statistics, ML, control, and reliability:
- probability of failure given a sensor state
- probability of a label given observed features
- probability of an event given partial evidence
- probability of a future outcome given current state
If the sample space or conditioning event is unclear, later formulas become meaningless.
4 Prerequisite Recall
- a set is a collection of objects
- a subset chooses some elements from a larger set
- intersections mean
both happen, unions meanat least one happens, and complements meandoes not happen
5 Intuition
A probability model starts by deciding what the world of possible outcomes is.
That world is the sample space.
An event is then a question you ask about that world. Since an event may be true for many outcomes, it is naturally represented as a subset of the sample space.
Conditioning is the first major change of viewpoint in probability. Once you learn that some event has happened, you no longer care about the full original world. You restrict attention to the smaller world where that event is true, and then recompute probabilities inside it.
So the right mental model is:
sample space -> events -> conditioning as restricted sample space
6 Formal Core
Definition 1 (Definition) A sample space, usually denoted \(\Omega\), is the set of all possible outcomes of an experiment or random study.
An event is a subset \(A \subseteq \Omega\).
Definition 2 (Probability Rule) A probability assignment gives each event \(A\) a number \(P(A)\) between \(0\) and \(1\) such that:
- \(P(\Omega)=1\)
- \(P(A) \ge 0\) for every event \(A\)
- disjoint events add:
\[ P(A \cup B) = P(A) + P(B) \qquad \text{when } A \cap B = \varnothing \]
In finite equally likely settings, probability is often computed as
\[ P(A) = \frac{|A|}{|\Omega|}. \]
Definition 3 (Conditional Probability) If \(P(B) > 0\), the conditional probability of \(A\) given \(B\) is
\[ P(A \mid B) = \frac{P(A \cap B)}{P(B)}. \]
This means: restrict attention to the event \(B\), then ask how much of that restricted world also lies in \(A\).
Proposition 1 (Key Statement) If \(P(B)>0\), then
\[ P(A \cap B) = P(B)\,P(A \mid B). \]
This is the multiplication rule. It is one of the main bridges from event language to later probabilistic calculations.
7 Worked Example
Roll two fair dice and record the ordered outcome \((d_1,d_2)\).
The sample space is
\[ \Omega = \{(d_1,d_2): d_1,d_2 \in \{1,2,3,4,5,6\}\}, \]
so \(|\Omega|=36\).
Define the events
\[ A = \{(d_1,d_2): d_1 + d_2 \ge 10\} \]
and
\[ B = \{(d_1,d_2): d_1 = 6\}. \]
First compute \(P(A)\).
The outcomes with sum at least \(10\) are
\[ (4,6), (5,5), (6,4), (5,6), (6,5), (6,6), \]
so
\[ P(A)=\frac{6}{36}=\frac{1}{6}. \]
Now condition on \(B\).
If the first die is known to be \(6\), the relevant world shrinks to
\[ B = \{(6,1),(6,2),(6,3),(6,4),(6,5),(6,6)\}, \]
which has \(6\) equally likely outcomes.
Inside that restricted world, the event \(A\) becomes
\[ A \cap B = \{(6,4),(6,5),(6,6)\}, \]
so
\[ P(A \mid B)=\frac{3}{6}=\frac{1}{2}. \]
This example shows the point clearly:
- the original probability was \(1/6\)
- the conditional probability became \(1/2\)
- new information changed the relevant sample space
Conditioning does not just “plug into a formula.” It changes which world you are measuring inside.
8 Computation Lens
Many probability calculations become much easier once you translate words into sets.
Typical moves are:
- define the sample space explicitly
- write the event as a subset of that space
- use intersections for
both, unions foreither, complements fornot - if information is revealed, condition on that event and work inside the restricted space
In finite settings, this often reduces to careful counting.
In more advanced settings, the same structure remains, but counting gets replaced by densities, integrals, or conditional distributions.
9 Application Lens
Conditioning is the language of information.
In statistics and ML, we almost never ask about an outcome in total isolation. We ask about it given data, features, labels, state, or evidence:
- failure probability given an alarm
- class probability given an input
- state estimate given observations
- risk given a subgroup or condition
So even this beginner topic is already the first bridge to inference and prediction. Once you are comfortable with conditioning as “restrict the world, then renormalize,” later topics like Bayes, conditional expectation, and graphical models become much easier.
10 Stop Here For First Pass
If you can now explain:
- what the sample space is supposed to capture
- why events are subsets
- why conditional probability restricts attention to a smaller world
- why new information can change a probability dramatically
then this page has done its main job.
11 Go Deeper
The best next move is to continue the probability spine with Random Variables and Distributions, where outcomes stop being tracked directly and are mapped into numerical quantities.
12 Optional Paper Bridge
- MIT RES.6-012 conditional probability lecture materials -
Second pass- an official MIT bridge into conditioning as probability revision after evidence. Checked2026-04-24. - Translating Between Probability and Sets -
Paper bridge- a short Harvard note that helps decode event notation quickly when reading theory-heavy material. Checked2026-04-24.
13 Optional After First Pass
If you want more practice before moving on:
- redraw a sample space after conditioning
- compute probabilities using complements instead of direct counting
- compare conditional probability with unconditional probability on the same event
14 Common Mistakes
- choosing a sample space that is too vague or inconsistent
- forgetting that an event is a subset, not a single outcome
- conditioning on an event with probability zero
- treating \(P(A \mid B)\) as the same thing as \(P(B \mid A)\)
- using a formula for conditioning without updating the underlying world
15 Exercises
- Toss three fair coins. Define the sample space and the event
exactly two heads. - Roll a fair die once. Let \(A\) be
evenand \(B\) begreater than 3. Compute \(P(A)\), \(P(B)\), \(P(A \cap B)\), and \(P(A \mid B)\). - Explain in words why conditioning on an event can change a probability even when the original experiment has not changed.
16 Sources and Further Reading
- Statistics 110: Probability -
First pass- a strong official course hub for introductory probability. Checked2026-04-24. - Penn State STAT 414 -
First pass- current official notes with clear pages on sample spaces, events, and conditional probability. Checked2026-04-24. - CMU OLI Probability and Statistics -
Second pass- useful for structured practice and alternative explanations of events and probability rules. Checked2026-04-24. - MIT RES.6-012 conditional probability lecture materials -
Second pass- official MIT reinforcement for conditioning and evidence. Checked2026-04-24. - Translating Between Probability and Sets -
Paper bridge- a compact note that helps with notation fluency for later theorem reading. Checked2026-04-24.
Sources checked online on 2026-04-24:
- Harvard Stat 110
- Penn State STAT 414 overview
- Penn State sample spaces and conditional probability notes
- CMU OLI Probability and Statistics
- MIT RES.6-012 conditional probability lecture materials