1. Sympathetic Magic
- mistake of confusing r.v. with its distribution.
- adding r.v.s are not the same as adding pmfs.
- P(X=x)+P(Y=y)
- can be bigger than 1.
- you can’t do things like this.
- “Word is not the thing, the map is not the territory.”
- r.v. = random house, dist. = blueprint
- you can have many r.v.s with the same dist.
- blueprint : specifying all the probabilities for randomly choosing the color of doors of the house.
2. Poisson Distribution
- Most important discrete distribution.
- PMF
- P(X=k)=k!e−λλk,k∈{0,1,2,…}
- λ is the rate parameter. (λ>0) → param of poisson dist.
- X∼Pois(λ)
- Valid?
- ∑k=0∞k!e−λλk=e−λeλ=1 ( adding PMF = 1)
- ∑k=0∞k!λk=eλ by taylor series. check!!
- expected value
- E(X)=e−λ∑k=1∞k k!λk (when k=0, it’s 0 anyway)
- =e−λ∑k=1∞ (k−1)!λk=λe−λ∑k=1∞ (k−1)!λk−1=λ e−λeλ=λ
Why do we care about Poisson?
- It’s the most used dist. as a model for discrete data in the real world.
1) General Applications
- Counting number of successes where there are a large number of trials each with small prob. of success.
- Examples ( Not exactly, but approximately Poisson)
- number of emails in an hour
- Number of chocolate chips in a chocolate chip cookie
- Number of earthquakes in a year in some region
- There is no upper bound in k∈{0,1,2,…} as it goes towards infinity, and the prob. being small is not that clear (how small?)
- so it is useful to approximate a real world dist. with a poisson dist.
2) Poisson Paradigm (Poisson Approximation)
-
Event A1,A2,…,An, P(Aj)=pj, n large, pj’s small.
- large number of trials, each unlikely
-
Events are independent or weakly dependent, then number of Aj’s that occur is approximately Poisson. → Pois(λ)
- weakly dependent is hard to define.
- we could get a little bit of information.
- if we know A1,A2 happened, maybe it helps a little bit to know if A3 happened.
-
Expected Number of events occuring = λ=∑j=1npj
- If all events are independent & pj’s are all the same, we have Bernoulli trials
3) Connection between X∼Bin(n,p) & X∼Pois(λ)
- Let n→∞ (n is large number), λ=np is held constant.
- Each are expected values for Binomial & Poisson
- Find what happens to the PMF P(X=k)=(kn)pk(1−p)n−k, k fixed.
- p=λ/n, so plug it in.
- (kn)=k!n(n−1)...(n−k+1) ( 조합의 정의)
- P(X=k)=k! nkn(n−1)...(n−k+1)λk(1−nλ)n(1−nλ)−k
→ 이제 limit을 취하자. (n→∞,p→0)

- (1−nλ)n은 e−λ로 감.
- =k!λk e−λ (limit 취하고 남은 값.)
→ Poisson PMF at k.
- So, Binomial converges to Poisson when n→∞,p→0.
Raindrop Example
- We have a piece of paper, and we want to know how many raindrops hit this paper in 1 minute.
- let’s break the paper into millions of tiny squares. ( n→∞,p→0)
- each square is unlikely to be hit by a raindrop.
- If we assume all squares are independent, probability is all the same, binomial is what we should use.
- Probably not exactly independent.
- Also, binomial only takes Y/N.
- what if a square gets hit 2 times? → no way to express this by binomial
- Poisson Dist. seems better.
Triple birthday problem
- Have n people, find the approx. prob. that there are 3 people that have the same birthday.
- Assume n is reasonably large number.
- (3n) triplets of people, indicator r.v. for each, Iijk,i<j<k
- Iijk=1 if they all have the same birthday.
- E(number of triple matches)=(3n)36521
- 1st person can have whatever birthday, and the 2nd person should match the 1st person (1/365), and the 3rd person should match the 2nd person.(1/365)
- Let X=number of triple matches → this is approximately Pois(λ),λ=(3n)36521,
- each triplets matching birthdays is unlikely, n is fairly large.
- We don’t exactly have independence.
- I123,I124 are not completely independent. (weak dependence)
- if I123=1, we already know that 1,2 have the same birthday when we compute I124
- P(X≥1)=1−P(X=0)≈1−e−λ λ0/0!=1−e−λ (prob. of at least one triple match)
→ poisson is very useful for getting rough approximations.