Discrete Probability Bridge
discrete probability, sample space, indicator variable, expectation, linearity of expectation
1 Role
This page is the fourth step of the discrete-math module.
It is the bridge from finite structure into probability:
- outcomes become a sample space
- counting becomes probability on uniform spaces
- indicator variables turn events into algebra
- expectation becomes a counting tool instead of only a distribution summary
2 First-Pass Promise
Read this page after Graphs and Trees.
If you stop here, you should still understand:
- how counting and finite probability fit together
- what an indicator variable is
- why linearity of expectation is so useful
- why probability often starts from modeling, not from formulas
- how this page hands off naturally to the main Probability module
3 Why It Matters
Many first probability problems are really counting problems with uncertainty layered on top.
That is especially true when:
- the sample space is finite
- all outcomes are equally likely
- we care about collisions, occupancy, or random choices
- we want an expected count rather than a full distribution
This page matters because it turns ideas from discrete math into probability habits without asking for the whole continuous probability toolkit yet.
It is where learners first see that:
counting and expectation are often more powerful than trying to enumerate every case directly
4 Prerequisite Recall
- Counting and Combinatorics already built the habit of describing finite objects precisely
- Graphs and Trees already treated finite structure as a set of objects plus relationships
- probability on a finite uniform sample space often reduces to
favorable outcomes / total outcomes
5 Intuition
Suppose a process chooses one outcome uniformly from a finite set \(\Omega\).
Then probability is just normalized counting:
\[ \mathbb{P}(A) = \frac{|A|}{|\Omega|}. \]
That already explains why discrete probability belongs near combinatorics.
But there is a second major idea:
Instead of trying to compute the whole distribution of a quantity, we often define indicator variables for simple events and add them.
That turns “how many collisions happen?” or “how many vertices have a property?” into expectation calculations.
This is the first place where probability starts to feel like algebra on random variables rather than only fraction-of-outcomes bookkeeping.
6 Formal Core
Definition 1 (Definition: Finite Probability Space) If \(\Omega\) is a finite sample space and all outcomes are equally likely, then for any event \(A \subseteq \Omega\),
\[ \mathbb{P}(A) = \frac{|A|}{|\Omega|}. \]
This is the cleanest bridge from combinatorics into probability.
Definition 2 (Definition: Indicator Variable) For an event \(A\), the indicator variable of \(A\) is
\[ \mathbf{1}_A = \begin{cases} 1, & \text{if } A \text{ occurs},\\ 0, & \text{if } A \text{ does not occur}. \end{cases} \]
Its expectation is
\[ \mathbb{E}[\mathbf{1}_A] = \mathbb{P}(A). \]
Theorem 1 (Theorem Idea: Linearity of Expectation) For random variables \(X_1,\dots,X_n\),
\[ \mathbb{E}[X_1 + \cdots + X_n] = \mathbb{E}[X_1] + \cdots + \mathbb{E}[X_n]. \]
This holds whether or not the variables are independent.
That is why indicator decompositions are so useful: they let us count expected totals one small event at a time.
Theorem 2 (Theorem Idea: Counting Through Indicators) If a random quantity \(X\) can be written as
\[ X = \mathbf{1}_{A_1} + \cdots + \mathbf{1}_{A_m}, \]
then
\[ \mathbb{E}[X] = \mathbb{P}(A_1) + \cdots + \mathbb{P}(A_m). \]
So expected counts often become much easier than exact distributions.
Theorem 3 (Theorem Idea: Uniform Finite Modeling) In a finite uniform setting, many probabilities are easier to compute by:
- defining the right outcome space
- counting all outcomes
- counting favorable outcomes
- simplifying the ratio
This is the main modeling workflow of discrete probability.
7 Worked Example
Suppose we choose a random binary string of length \(5\), with each of the \(2^5 = 32\) strings equally likely.
What is the expected number of 1s in the string?
7.1 Step 1: Define indicator variables
For each position \(i \in \{1,2,3,4,5\}\), let
\[ X_i = \begin{cases} 1, & \text{if the } i\text{th bit is }1,\\ 0, & \text{if the } i\text{th bit is }0. \end{cases} \]
Then the total number of 1s is
\[ X = X_1 + X_2 + X_3 + X_4 + X_5. \]
7.2 Step 2: Compute each expectation
Each bit is equally likely to be 0 or 1, so
\[ \mathbb{E}[X_i] = \mathbb{P}(X_i=1) = \frac12. \]
7.3 Step 3: Use linearity of expectation
Therefore
\[ \mathbb{E}[X] = \mathbb{E}[X_1] + \cdots + \mathbb{E}[X_5] = 5 \cdot \frac12 = \frac52. \]
This example matters because it shows the bridge clearly:
- the sample space is finite
- the model is uniform
- the quantity is counted through indicators
- expectation is easier than the full distribution
8 Computation Lens
In first-pass discrete probability, the main tactics are:
- define the sample space carefully
- check whether the space is uniform
- count total outcomes and favorable outcomes when possible
- use indicator variables for expected counts
- prefer linearity of expectation over case explosion
This is one of the first places where “don’t brute-force the whole distribution” becomes a mathematical strategy.
9 Application Lens
This page is the clean handoff from discrete math into the main probability module.
It already supports:
- occupancy and collision intuition
- random graph or random string counts
- expected number of successes or matches
- randomized-algorithm sanity checks
The same ideas later scale into:
- binomial and hypergeometric models
- expectation and variance calculations
- concentration arguments
- randomized graph and network reasoning
10 Stop Here For First Pass
If you can now explain:
- why finite probability often starts with counting
- what an indicator variable does
- why \(\mathbb{E}[\mathbf{1}_A] = \mathbb{P}(A)\)
- why linearity of expectation is useful even without independence
then this page has done its job.
11 Go Deeper
The next page in this module is:
The strongest adjacent bridges are:
12 Optional Deeper Reading After First Pass
- MIT 6.1200J introduction to probability lecture - official MIT lecture PDF that introduces probability through rigorous discrete modeling. Checked
2026-04-25. - Penn State STAT 414 counting techniques - official open course notes reinforcing the combinatorial side of finite probability. Checked
2026-04-25. - Penn State STAT 414 mathematical expectation - official open course notes reinforcing expectation in discrete settings. Checked
2026-04-25.
13 Sources and Further Reading
- MIT 6.1200J introduction to probability lecture -
First pass- official lecture PDF introducing discrete probability through precise modeling. Checked2026-04-25. - MIT 6.1200J Mathematics for Computer Science -
Second pass- official course hub connecting counting, recurrences, graph theory, and discrete probability. Checked2026-04-25. - MIT 6.042J Mathematics for Computer Science -
Second pass- classic official MIT course page for the broader discrete-math sequence. Checked2026-04-25. - Penn State STAT 414 counting techniques -
First pass- official open notes on counting methods that naturally feed finite probability. Checked2026-04-25. - Penn State STAT 414 mathematical expectation -
First pass- official open notes on expectation in discrete random-variable settings. Checked2026-04-25.
Sources checked online on 2026-04-25:
- MIT 6.1200J lecture 18
- MIT 6.1200J course page
- MIT 6.042J course page
- Penn State STAT 414 Lesson 03
- Penn State STAT 414 Lesson 08