Board Coverage
| Board | Paper | Notes |
|---|
| AQA | Paper 1, 2 | Basic probability in P1; conditional, Bayes in P2 |
| Edexcel | P1, P2 | Similar |
| OCR (A) | Paper 1, 2 | Includes Venn diagrams and tree diagrams |
| CIE (9709) | P1, P6 | Probability in P1; conditional in P6 |
Probability questions test logical reasoning as much as formula recall. Always define events
clearly and draw a diagram before calculating.
1. Kolmogorov's Axioms
Definition. A probability function P on a sample space Ω satisfies:
- Non-negativity: P(A)≥0 for all events A⊆Ω.
- Normalisation: P(Ω)=1.
- Countable additivity: If A1,A2,… are mutually exclusive, then
P(⋃iAi)=∑iP(Ai).
These three axioms are the foundation of all probability theory. Every theorem in probability can be
derived from them.
2. Basic Probability Results
2.1 Complement rule
Theorem. P(A′)=1−P(A).
Proof. A and A′ are mutually exclusive and A∪A′=Ω.
P(A∪A′)=P(A)+P(A′)=P(Ω)=1⟹P(A′)=1−P(A).■
Corollary. For any event A, P(∅)=0.
Proof. P(∅)=P(Ω′)=1−P(Ω)=1−1=0. ■
Corollary. If A⊆B, then P(A)≤P(B).
Proof. Write B=A∪(B∩A′) where the two sets are disjoint. Then
P(B)=P(A)+P(B∩A′)≥P(A) since P(B∩A′)≥0. ■
2.2 Addition rule
Theorem. P(A∪B)=P(A)+P(B)−P(A∩B).
Proof. A∪B can be partitioned into three disjoint sets: A∩B′, A∩B, and
A′∩B.
P(A∪B)=P(A∩B′)+P(A∩B)+P(A′∩B)
P(A)=P(A∩B′)+P(A∩B)⟹P(A∩B′)=P(A)−P(A∩B)
P(B)=P(A∩B)+P(A′∩B)⟹P(A′∩B)=P(B)−P(A∩B)
P(A∪B)=[P(A)−P(A∩B)]+P(A∩B)+[P(B)−P(A∩B)]=P(A)+P(B)−P(A∩B).■
For mutually exclusive events (A∩B=∅): P(A∪B)=P(A)+P(B).
Corollary (Boole's inequality). For any events A and B, P(A∪B)≤P(A)+P(B).
Proof. Since P(A∩B)≥0, we have
P(A∪B)=P(A)+P(B)−P(A∩B)≤P(A)+P(B). ■
2.3 Multiplication rule
Theorem. P(A∩B)=P(A)⋅P(B∣A).
Proof. This follows directly from the definition of conditional probability (Section 3.1).
■
General multiplication rule. For events A1,A2,…,An:
P(⋂i=1nAi)=P(A1)⋅P(A2∣A1)⋅P(A3∣A1∩A2)⋯P(An∣A1∩⋯∩An−1)
3. Conditional Probability
3.1 Definition
Definition. The conditional probability of A given B is
P(A∣B)=L◆B◆P(A∩B)◆RB◆◆LB◆P(B)◆RB◆forP(B)>0
Intuition. P(A∣B) is the probability of A occurring given that we already know B has
occurred. Knowing B has happened changes our sample space from Ω to B, and we measure
what fraction of B is also in A.
3.2 Properties of conditional probability
Theorem. Conditional probability satisfies the Kolmogorov axioms for a fixed conditioning event
B (with P(B)>0).
Proof.
- P(A∣B)=P(A∩B)/P(B)≥0 since P(A∩B)≥0 and P(B)>0.
- P(Ω∣B)=P(Ω∩B)/P(B)=P(B)/P(B)=1.
- If A1,A2,… are mutually exclusive, then so are A1∩B,A2∩B,…, and
P(⋃iAi∣B)=L◆B◆P((⋃iAi)∩B)◆RB◆◆LB◆P(B)◆RB◆=L◆B◆∑iP(Ai∩B)◆RB◆◆LB◆P(B)◆RB◆=∑iP(Ai∣B).■
Corollary. The complement rule holds for conditional probability: P(A′∣B)=1−P(A∣B).
Proof. This follows from applying the complement rule within the conditional probability
measure, which is justified by the theorem above. ■
4. Bayes' Theorem
4.1 Statement
Theorem. For events A and B with P(B)>0:
P(A∣B)=L◆B◆P(B∣A)⋅P(A)◆RB◆◆LB◆P(B)◆RB◆
4.2 Proof
P(A∣B)=L◆B◆P(A∩B)◆RB◆◆LB◆P(B)◆RB◆=L◆B◆P(B∣A)⋅P(A)◆RB◆◆LB◆P(B)◆RB◆■
4.3 Law of Total Probability
If B1,B2,…,Bn partition Ω (mutually exclusive and exhaustive):
P(A)=∑i=1nP(A∣Bi)P(Bi)
4.4 Extended Bayes' Theorem
P(Bk∣A)=L◆B◆P(A∣Bk)P(Bk)◆RB◆◆LB◆∑i=1nP(A∣Bi)P(Bi)◆RB◆
Bayes' theorem is essential for "reverse" probability questions: "Given that a test is
positive, what is the probability the patient actually has the disease?" Always define events
clearly and identify what is given (P(A∣B)) versus what is sought (P(B∣A)).
5. Independence
5.1 Definition
Definition. Events A and B are independent if and only if
P(A∩B)=P(A)⋅P(B)
5.2 Proof: Independence ⟺ conditional probability equals unconditional
Theorem. A and B are independent if and only if P(A∣B)=P(A) (provided P(B)>0).
Proof.
(⇒) If P(A∩B)=P(A)P(B), then
P(A∣B)=L◆B◆P(A∩B)◆RB◆◆LB◆P(B)◆RB◆=P(B)P(A)P(B)=P(A).
(⇐) If P(A∣B)=P(A), then L◆B◆P(A∩B)◆RB◆◆LB◆P(B)◆RB◆=P(A), so
P(A∩B)=P(A)P(B). ■
Intuition. Independence means knowing B occurred gives you no information about A. The
probability of A is the same whether or not B has happened.
warning
mutually exclusive and both have positive probability, they are not independent (since
P(A∩B)=0=P(A)P(B)).
5.3 Pairwise and mutual independence
Definition. Events A1,A2,…,An are mutually independent if for every subset
i1,…,ik⊆1,2,…,n with k≥2:
P(Ai1∩Ai2∩⋯∩Aik)=P(Ai1)⋅P(Ai2)⋯P(Aik)
Definition. Events A1,A2,…,An are pairwise independent if every pair
(Ai,Aj) with i=j is independent.
Mutual independence is a stronger condition than pairwise independence. Pairwise
independence does not imply mutual independence. For example, with two independent coin tosses, let
A = "first toss is heads", B = "second toss is heads", C = "both tosses are the same". Then
A, B, C are pairwise independent but not mutually independent since
P(A∩B∩C)=0=P(A)P(B)P(C)=1/8.
6. Venn Diagrams and Tree Diagrams
6.1 Venn diagrams
Venn diagrams represent events as regions. Useful for visualising:
- A∪B, A∩B, A′
- Relationships between events
- Applying the addition rule
6.2 Tree diagrams
Tree diagrams are useful for sequential experiments. Each branch represents a possible outcome with
its probability. The probability along any path is the product of the probabilities along its
branches (multiplication rule). The probability of any event is found by adding the probabilities of
all paths leading to it (addition rule for mutually exclusive paths).
Example. A bag contains 3 red and 2 blue balls. Two balls are drawn without replacement.
P(bothred)=53×42=206=103
P(oneofeach)=53×42+52×43=206+206=2012=53
7. Counting Principles
7.1 Factorials
n!=n(n−1)(n−2)⋯1, with 0!=1.
7.2 Permutations and combinations
- Permutations: nPr=(n−r)!n! (order matters)
- Combinations: nCr=(rn)=r!(n−r)!n! (order does not matter)
7.3 Probability with equally likely outcomes
When all outcomes are equally likely:
P(A)=L◆B◆∣A∣◆RB◆◆LB◆∣Ω∣◆RB◆=L◆B◆numberoffavourableoutcomes◆RB◆◆LB◆totalnumberofoutcomes◆RB◆.
8. Venn Diagrams for Three Events
8.1 Inclusion-exclusion principle
Theorem (Inclusion-Exclusion for three events). For events A, B, C:
P(A∪B∪C)=P(A)+P(B)+P(C)−P(A∩B)−P(A∩C)−P(B∩C)+P(A∩B∩C)
Proof. Apply the two-event inclusion-exclusion rule twice:
P(A∪B∪C)=P(A)+P(B∪C)−P(A∩(B∪C))
Now P(B∪C)=P(B)+P(C)−P(B∩C), and by the distributive law of set theory
A∩(B∪C)=(A∩B)∪(A∩C), so:
P(A∩(B∪C))=P(A∩B)+P(A∩C)−P(A∩B∩C)
Substituting:
P(A∪B∪C)=P(A)+P(B)+P(C)−P(B∩C)−P(A∩B)−P(A∩C)+P(A∩B∩C).■
8.2 De Morgan's laws for three events
Theorem. For events A, B, C:
(A∪B∪C)′=A′∩B′∩C′
(A∩B∩C)′=A′∪B′∪C′
Proof. The three-event case follows by induction from the two-event case. For the first law:
(A∪B∪C)′=((A∪B)∪C)′=(A∪B)′∩C′=A′∩B′∩C′.■
8.3 Working with three-event Venn diagrams
When solving problems with three events, the Venn diagram is divided into 8 regions (including
the exterior). The fundamental approach is:
- Start from the innermost region A∩B∩C and work outward.
- Use the given information to find the value of each region.
- Each region represents a disjoint event, so probabilities add.
Example. In a class of 40 students, 18 study Maths, 15 study Physics, and 12 study Chemistry. 5
study all three, 8 study Maths and Physics, 6 study Maths and Chemistry, and 7 study Physics and
Chemistry.
The region for "Maths only" is: 18−8−6+5=9 (subtract overlaps, add back the triple
overlap).
| Region | Description | Calculation | Count |
|---|
| A∩B∩C | All three | Given | 5 |
| A∩B∩C′ | Maths and Physics only | 8−5 | 3 |
| A∩C∩B′ | Maths and Chemistry only | 6−5 | 1 |
| B∩C∩A′ | Physics and Chemistry only | 7−5 | 2 |
| A∩B′∩C′ | Maths only | 18−3−1−5 | 9 |
| B∩A′∩C′ | Physics only | 15−3−2−5 | 5 |
| C∩A′∩B′ | Chemistry only | 12−1−2−5 | 4 |
| A′∩B′∩C′ | None | 40−29 | 11 |
Check: 5+3+1+2+9+5+4+11=40. ✓
9. Multi-Stage Experiments and Tree Diagrams
A multi-stage experiment consists of a sequence of trials. A tree diagram represents this as:
- Levels correspond to stages (trials).
- Branches at each node represent possible outcomes at that stage.
- Branch probabilities are the conditional probabilities of each outcome given the path so far.
- Path probability is the product of all branch probabilities along the path.
- Event probability is the sum of all relevant path probabilities.
9.2 With and without replacement
With replacement. At each stage, the sample space and probabilities reset. The trials are
independent.
Without replacement. At each stage, the sample space shrinks. The trials are not
independent; later probabilities depend on earlier outcomes.
Example. A bag contains 5 balls: 2 red and 3 blue. Three balls are drawn without replacement.
Find the probability of drawing exactly 2 red balls.
There are (23)=3 ways to arrange the two red draws among three positions: RRB, RBR, BRR.
P(RRB)=52×41×33=606=101
P(RBR)=52×43×31=606=101
P(BRR)=53×42×31=606=101
P(exactly2red)=101+101+101=103
9.3 At least and at most problems
For "at least k" problems, it is often easier to compute the complement:
P(atleastk)=1−P(atmostk−1).
Example. A fair coin is tossed 4 times. Find P(atleast3heads).
P(atleast3heads)=P(exactly3heads)+P(exactly4heads)
=(34)(21)4+(44)(21)4=164+161=165
Alternatively:
P(atleast3heads)=1−P(atmost2heads)=1−1611=165.
9.4 Conditional probability from tree diagrams
To find a conditional probability P(X∣Y) from a tree diagram:
- Identify all paths leading to Y (the conditioning event).
- Sum these path probabilities to get P(Y).
- Among those paths, identify which also satisfy X.
- Sum the relevant path probabilities to get P(X∩Y).
- P(X∣Y)=P(X∩Y)/P(Y).
10. Discrete Random Variables and Probability Mass Functions
10.1 Discrete random variables
Definition. A random variable is a function X:Ω→R that assigns a
real number to each outcome in the sample space.
Definition. A random variable X is discrete if its set of possible values is countable
(i.e. finite or countably infinite).
Example. If a fair die is rolled, define X = "the number shown". Then X takes values in
1,2,3,4,5,6, so X is discrete.
Example. If a coin is tossed until the first head appears, define X = "number of tosses". Then
X takes values in 1,2,3,…, which is countably infinite.
10.2 Probability mass function (PMF)
Definition. The probability mass function (PMF) of a discrete random variable X is the
function p(x)=P(X=x), defined for all x∈R.
Properties of a PMF. A function p:R→[0,1] is a valid PMF if and only if:
- p(x)≥0 for all x.
- allx∑p(x)=1.
Proof. Property 1 follows from non-negativity of probability. Property 2 follows because the
events X=x for all possible values of x form a partition of Ω, so their
probabilities sum to 1 by the normalisation axiom. ■
10.3 Cumulative distribution function (CDF)
Definition. The cumulative distribution function (CDF) of a discrete random variable X is
F(x)=P(X≤x)=∑t≤xp(t)
The CDF is a non-decreasing, right-continuous function with limx→−∞F(x)=0 and
limx→+∞F(x)=1.
10.4 Expectation and variance
Definition. The expected value (mean) of a discrete random variable X is
E(X)=μ=∑allxx⋅p(x)
Definition. The variance of X is
Var(X)=σ2=E[(X−μ)2]=∑allx(x−μ)2⋅p(x)
An equivalent computational formula is:
Var(X)=E(X2)−[E(X)]2
Proof of the computational formula:
Var(X)=E[(X−μ)2]=E(X2−2μX+μ2)=E(X2)−2μE(X)+μ2=E(X2)−μ2.■
10.5 Worked example
A biased die has PMF:
| x | 1 | 2 | 3 | 4 | 5 | 6 |
|---|
| p(x) | 1/12 | 1/6 | 1/4 | 1/4 | 1/6 | 1/12 |
Check:
1/12+1/6+1/4+1/4+1/6+1/12=1/12+2/12+3/12+3/12+2/12+1/12=12/12=1.
✓
E(X)=1⋅121+2⋅61+3⋅41+4⋅41+5⋅61+6⋅121
=121+62+43+44+65+126=121+4+9+12+10+6=1242=3.5
E(X2)=1⋅121+4⋅61+9⋅41+16⋅41+25⋅61+36⋅121
=121+8+27+48+50+36=12170=685
Var(X)=E(X2)−[E(X)]2=685−449=12170−147=1223≈1.917
info
above has the same mean but smaller variance, meaning its outcomes are more concentrated around the
centre.
Problem Set
Details
Problem 1
Events
A and
B are such that
P(A)=0.4,
P(B)=0.5, and
P(A∪B)=0.7. Find
P(A∩B) and
P(A∣B).
Details
Solution 1
P(A∩B)=P(A)+P(B)−P(A∪B)=0.4+0.5−0.7=0.2.
P(A∣B)=P(A∩B)/P(B)=0.2/0.5=0.4.
If you get this wrong, revise: Addition Rule — Section 2.2.
Details
Problem 2
A disease affects 1% of a population. A test is 99% accurate (both sensitivity and specificity). A person tests positive. What is the probability they actually have the disease?
Details
Solution 2
Let
D = has disease,
T+ = tests positive.
P(D)=0.01, P(T+∣D)=0.99, P(T+∣D′)=0.01.
By the law of total probability:
P(T+)=P(T+∣D)P(D)+P(T+∣D′)P(D′)=0.99(0.01)+0.01(0.99)=0.0099+0.0099=0.0198.
By Bayes' theorem: P(D∣T+)=L◆B◆P(T+∣D)P(D)◆RB◆◆LB◆P(T+)◆RB◆=0.01980.0099=0.5.
Even with a 99% accurate test, a positive result means only a 50% chance of actually having the
disease, because the disease is so rare.
If you get this wrong, revise: Bayes' Theorem — Section 4.
Details
Problem 3
Prove that if
A and
B are independent, then so are
A and
B′.
Details
Solution 3
P(A∩B′)=P(A)−P(A∩B)=P(A)−P(A)P(B) (by independence)
=P(A)[1−P(B)]=P(A)P(B′).
■If you get this wrong, revise: Independence — Section 5.
Details
Problem 4
A bag contains 4 red, 3 blue, and 2 green balls. Three balls are drawn without replacement. Find the probability that all three are different colours.
Details
Solution 4
Total ways to choose 3 from 9:
(39)=84.
Ways to get one of each colour: (14)(13)(12)=4×3×2=24.
P=24/84=2/7.
If you get this wrong, revise: Counting Principles — Section 7.
Details
Problem 5
Events
A,
B,
C are such that
P(A)=0.3,
P(B)=0.4,
P(C)=0.5,
P(A∩B)=0.1,
P(A∩C)=0.15,
P(B∩C)=0.2, and
P(A∩B∩C)=0.05. Find
P(A∪B∪C).
Details
Solution 5
By the inclusion-exclusion principle:
P(A∪B∪C)=P(A)+P(B)+P(C)−P(A∩B)−P(A∩C)−P(B∩C)+P(A∩B∩C)
=0.3+0.4+0.5−0.1−0.15−0.2+0.05=0.8
If you get this wrong, revise: Addition Rule — Section 2.2.
Details
Problem 6
Two coins are tossed. Given that at least one is heads, find the probability that both are heads.
Details
Solution 6
Ω={HH,HT,TH,TT}.
A={atleastoneheads}={HH,HT,TH}.
B={bothheads}={HH}.
P(B∣A)=P(B∩A)/P(A)=P(B)/P(A)=(1/4)/(3/4)=1/3.
If you get this wrong, revise: Conditional Probability —
Section 3.
Details
Problem 7
A fair die is rolled. Let
A = "even number" and
B = "number greater than 3". Are
A and
B independent?
Details
Solution 7
A={2,4,6},
B={4,5,6},
A∩B={4,6}.
P(A)=3/6=1/2, P(B)=3/6=1/2, P(A∩B)=2/6=1/3.
P(A)P(B)=1/4=1/3=P(A∩B). So A and B are not independent.
If you get this wrong, revise: Independence — Section 5.
Details
Problem 8
In a school, 60% of students study Maths, 40% study Physics, and 25% study both. A student is chosen at random. Given that they study Physics, find the probability they study Maths.
Details
Solution 8
P(M)=0.6,
P(P)=0.4,
P(M∩P)=0.25.
P(M∣P)=P(M∩P)/P(P)=0.25/0.4=0.625.
If you get this wrong, revise: Conditional Probability —
Section 3.
Details
Problem 9
A box contains 10 items, 3 of which are defective. Items are inspected one by one without replacement. Find the probability that the first defective item is the third one inspected.
Details
Solution 9
First two non-defective, third defective:
P=107×96×83=L◆B◆7×6×3◆RB◆◆LB◆720◆RB◆=720126=407
If you get this wrong, revise: Tree Diagrams — Section 6.2.
Details
Problem 10
A machine produces components. 5% are defective. Components are packed in boxes of 20. Find the probability that a box contains exactly one defective component.
Details
Solution 10
This is a binomial scenario:
X∼B(20,0.05).
P(X=1)=(120)(0.05)1(0.95)19=20×0.05×0.9519≈0.3774.
If you get this wrong, revise:
Binomial Distribution — Statistical
Distributions chapter.
Details
Problem 11
Prove that
P(A∪B′)=1−P(A′∩B).
Details
Solution 11
By De Morgan's law:
(A∪B′)′=A′∩B.
So P(A∪B′)=1−P((A∪B′)′)=1−P(A′∩B). ■
If you get this wrong, revise: Complement Rule — Section 2.1.
Details
Problem 12
From a standard 52-card deck, 5 cards are dealt. Find the probability of getting a flush (all 5 cards of the same suit).
Details
Solution 12
Total ways:
(552)=2598960.
Ways to get a flush: choose suit (4 ways), then 5 cards from that suit ((513)=1287).
Total flushes: 4×1287=5148.
P(flush)=5148/2598960≈0.00198≈0.2%.
If you get this wrong, revise: Counting Principles — Section 7.
Details
Problem 13
A discrete random variable
X has PMF
p(x)=kx for
x∈{1,2,3,4,5} and
p(x)=0 otherwise. Find the constant
k, then find
E(X) and
Var(X).
Details
Solution 13
For a valid PMF:
∑x=15kx=k(1+2+3+4+5)=15k=1, so
k=1/15.
E(X)=∑x=15x⋅15x=151+4+9+16+25=1555=311
E(X2)=∑x=15x2⋅15x=151+8+27+64+125=15225=15
Var(X)=E(X2)−[E(X)]2=15−9121=9135−121=914
If you get this wrong, revise:
Discrete Random Variables —
Section 10.
Details
Problem 14
A bag contains 4 red and 6 blue balls. Balls are drawn one at a time without replacement until a red ball is drawn. Find the probability that exactly 3 draws are needed.
Details
Solution 14
We need the first two draws to be blue and the third to be red:
P=106×95×84=720120=61
If you get this wrong, revise:
Multi-Stage Experiments — Section 9.
Details
Problem 15
Prove Boole's inequality: for events
A1,A2,…,An,
P(⋃i=1nAi)≤∑i=1nP(Ai)
Details
Solution 15
By induction on
n.
Base case (n=2): P(A1∪A2)=P(A1)+P(A2)−P(A1∩A2)≤P(A1)+P(A2).
✓
Inductive step: Assume P(⋃i=1kAi)≤∑i=1kP(Ai). Then
P(⋃i=1k+1Ai)=P(⋃i=1kAi)+P(Ak+1)−P(⋃i=1kAi∩Ak+1)
≤∑i=1kP(Ai)+P(Ak+1)=∑i=1k+1P(Ai).■
If you get this wrong, revise: Basic Probability Results —
Section 2.
Details
Problem 16
In a survey, 70% of people like tea, 50% like coffee, and 35% like both. A person is chosen at random. Given that they like at least one of the two drinks, find the probability that they like both.
Details
Solution 16
P(T)=0.7,
P(C)=0.5,
P(T∩C)=0.35.
P(T∪C)=P(T)+P(C)−P(T∩C)=0.7+0.5−0.35=0.85.
P(T∩C∣T∪C)=L◆B◆P(T∩C)◆RB◆◆LB◆P(T∪C)◆RB◆=0.850.35=177≈0.412
If you get this wrong, revise: Conditional Probability —
Section 3.
Details
Problem 17
A fair coin is tossed 5 times. Using the complement rule, find the probability of getting at least one head.
Details
Solution 17
Let
A = "at least one head". Then
A′ = "no heads" = "all tails".
P(A)=1−P(A′)=1−(21)5=1−321=3231
If you get this wrong, revise: Complement Rule — Section 2.1.
Details
Problem 18
Two events
A and
B satisfy
P(A)=0.6,
P(B∣A)=0.4, and
P(B∣A′)=0.7. Find
P(B),
P(A∣B), and determine whether
A and
B are independent.
Details
Solution 18
By the law of total probability:
P(B)=P(B∣A)P(A)+P(B∣A′)P(A′)=0.4×0.6+0.7×0.4=0.24+0.28=0.52
P(A∩B)=P(B∣A)P(A)=0.4×0.6=0.24
P(A∣B)=L◆B◆P(A∩B)◆RB◆◆LB◆P(B)◆RB◆=0.520.24=136≈0.462
Check independence: P(A)P(B)=0.6×0.52=0.312=0.24=P(A∩B). So A and B are
not independent.
If you get this wrong, revise: Bayes' Theorem — Section 4, and
Independence — Section 5.
Details
Problem 19
A discrete random variable
X has CDF
F(x)=0 for
x<0,
F(x)=x/4 for
0≤x<1,
F(x)=1/2 for
1≤x<2,
F(x)=3/4 for
2≤x<3, and
F(x)=1 for
x≥3. Find the PMF of
X and verify it sums to 1.
Details
Solution 19
The PMF is obtained from the jumps in the CDF:
- p(0)=F(0)−F(0−)=0−0=0. But from the formula F(x)=x/4 at x=0: p(0)=0.
Actually, the jump occurs at the boundary. Since F is continuous at x=0, there is no point
mass at 0. The value X=0 has probability 0; we look at where jumps occur.
More carefully, the jumps occur at:
- x=1: p(1)=F(1)−F(1−)=1/2−1/4=1/4
- x=2: p(2)=F(2)−F(2−)=3/4−1/2=1/4
- x=3: p(3)=F(3)−F(3−)=1−3/4=1/4
There is also a continuous component on [0,1), but since X is discrete, the CDF must be a step
function. The given CDF has a linear portion, which indicates this CDF actually corresponds to a
mixed distribution. For a purely discrete X, the CDF should be piecewise constant with jumps.
Assuming the problem intended a discrete distribution, the PMF from the jumps is:
p(1)=41,p(2)=41,p(3)=41,p(x)=0otherwise
Check: 1/4+1/4+1/4=3/4=1. This indicates the continuous portion F(x)=x/4 on
[0,1) contributes probability 1/4 spread over a continuum, confirming this is not a purely
discrete distribution.
If you get this wrong, revise:
Discrete Random Variables —
Section 10.
Details
Problem 20
Three machines
M1,
M2,
M3 produce items with proportions 50%, 30%, 20%. Their defect rates are 2%, 3%, 5% respectively. An item is found to be defective. Find the probability it was produced by
M3.
Details
Solution 20
Let
D = "defective". By the law of total probability:
P(D)=P(D∣M1)P(M1)+P(D∣M2)P(M2)+P(D∣M3)P(M3)
=0.02×0.5+0.03×0.3+0.05×0.2=0.01+0.009+0.01=0.029
By Bayes' theorem:
P(M3∣D)=L◆B◆P(D∣M3)P(M3)◆RB◆◆LB◆P(D)◆RB◆=L◆B◆0.05×0.2◆RB◆◆LB◆0.029◆RB◆=0.0290.01=2910≈0.345
If you get this wrong, revise: Extended Bayes' Theorem — Section
4.4.
Diagnostic Test
Ready to test your understanding of Probability? The diagnostic test contains the hardest questions within the A-Level specification for this topic, each with a full worked solution.
Unit tests probe edge cases and common misconceptions. Integration tests combine Probability with other topics to test synthesis under exam conditions.
See Diagnostic Guide for instructions on self-marking and building a personal test matrix.