In this post we will explore the duality between uniform and exponential distributions, and try to see how it can be used as a clever trick to solve probability problems.
Suppose you have a problem that mentions “i.i.d. Uniform” and asks about order statistics, typically minimum and/or maximum. Instead of doing tedious arithmetic/calculus, what if there was another way?
Recall the Poisson Process. This counting process has independent arrivals and it is a well-known fact that inter-arrival times are exponentially distributed. In addition, the less-known feature is that arrivals are also uniform for a finite interval. This is where the intuition comes from.
We skip the rigorous proof here, but an alternative way to sample uniformly distributed random variables and then sort them is to generate exponential random variables, representing intervals in-between, draw another one representing the last interval and eventually normalize the entire interval.
Let be independent Exponential(1) random variables. Denote
.
Then are jointly distributed exactly the same as Uniform[0;1] order statistics.
Couple of useful facts (and links) about random variables:
Expression | Distribution | Mean | Variance |
---|---|---|---|
wiki | wiki | ||
Sum of i.i.d. wiki | |||
Min of i.i.d. wiki | |||
Mix of indep.wiki: | wiki | ||
Joint | Dirichletwiki with |
Now let’s demonstrate the application of the uniform-exponential for getting succinct proofs.
Problem 1 (Easy). Given i.i.d. Uniform [0;1] random variables, find the probability that the sum of their maximum and minimum is greater than 1.
Let be independent Exponential(1) random variables. Denote minimum and maximum as . Then, by the machinery above, it holds that
and .
We want to evaluate . After substituting , this becomes simply the probability of and equals to 1/2 by symmetry.
The alternative approach is to notice that maximum is one minus minimum and play with it. This evolves bits of creativity and the standard machinery helps when “noticing” was not successful.
Problem 2 (Hard). Given i.i.d. Uniform [0;1] random variables, find the correlation between their minimum and maximum.
Let be independent Exponential(1) random variables. Denote minimum and maximum as . Then, by the machinery above, it holds that
and .
After this the magic happens when one recognizes that for every two variables , by linearity, it holds that
Denote for . Note that we can drop the index , since by symmetry all these are the same.
Let’s substitute minimum and maximum into the “magic” observation. Then simply , where indices are and . The original correlation after opening the sum can be expressed as
,
where variance comes from . Hence, we have that , which is also a covariance! Finally, for correlation, noticing that variance for minimum and maximum are equal (same machinery),
.
Problem 3 (Hard). Given i.i.d. Uniform [0;1] random variables, find the expected value of the smallest distance between them.
Use the machinery above. Denote the denominator of exponentials as . Then the distances are conveniently and we are looking to find the expectation of their minimum. They share the common denominator that can be moved out, such as
NB: random variables and are not inside the minimum function (see image in the beginning). From the table above, we know that . Now the idea is to find the Beta distribution in the form Gamma’s fraction.
Since and are not independent, we can’t use Beta distribution without justification. Recall that the exponential distribution is memoryless and remain the same independently distributed exponential after the minimum time passed.
Let’s condition the expectation on the minimum being achieved for some index , and we want to use the law of total expectation . The denominator consists of:
- and , independent from the minimum ;
- the minimum itself;
- exponential variables greater or equal , indices from to skipping index .
By memoryless property hence it can be written as , where has exactly the same exponential distribution and independent of . Now that are obviously measurable and other variables are independent, we can remove the condition and make enumerator Gamma by multiplying and dividing by , to receive the general expectation
The denominator has the minimum times, and reminder is exponentials independent from the minimum, with their sum being . Now we recall that has Gamma distribution and we can use the Beta ratio from the table:
0 Comments