Strategic practice and homework 7 Prof. Joe Blitzstein Department of Statistics, Harvard University PDF

Title Strategic practice and homework 7 Prof. Joe Blitzstein Department of Statistics, Harvard University
Author Manya Narwal
Course Propability
Institution Harvard University
Pages 19
File Size 407.7 KB
File Type PDF
Total Downloads 109
Total Views 124

Summary

Stat 110 Strategic Practice 7, Fall 2011
Prof. Joe Blitzstein (Department of Statistics, Harvard University)
Joint, Conditional, and Marginal Distributions
Stat 110 Homework 7 Solutions...


Description

Stat 110 Strategic Practice 7, Fall 2011 Prof. Joe Blitzstein (Department of Statistics, Harvard University)

1

Joint, Conditional, and Marginal Distributions 1. A random point (X, Y, Z ) is chosen uniformly in the ball B = {(x, y, z) : x2 + y 2 + z 2 ≤ 1}. (a) Find the joint PDF of X, Y, Z . (b) Find the joint PDF of X, Y . (c) Find an expression for the marginal PDF of X, as an integral. 2. Let X and Y be i.i.d. Unif(0, 1). Find the expected value and the standard deviation of the distance between X and Y . 3. Let U1 , U 2 , U3 be i.i.d. Unif(0, 1), and let L = min(U1 , U2 , U3 ), M = max(U1 , U2 , U3 ). (a) Find the marginal CDF and marginal PDF of M, and the joint CDF and joint PDF of L, M. Hint: for the latter, start by considering P (L ≥ l, M ≤ m). (b) Find the conditional PDF of M given L. 4. A group of n ≥ 2 people decide to play an exciting game of Rock-PaperScissors. As you may recall, Rock smashes Scissors, Scissors cuts Paper, and Paper covers Rock (despite Bart Simpson saying “Good old rock, nothing beats that!”). Usually this game is played with 2 players, but it can be extended to more players as follows. If exactly 2 of the 3 choices appear when everyone reveals their choice, say a, b ∈ {Rock, P aper, Scissors} where a beats b, the game is decisive: the players who chose a win, and the players who chose b lose. Otherwise, the game is indecisive and the players play again. For example, with 5 players, if one player picks Rock, two pick Scissors, and two pick Paper, the round is indecisive and they play again. But if 3 pick Rock and 2 pick Scissors, then the Rock players win and the Scissors players lose the game.

1

Assume that the n players independently and randomly choose between Rock, Scissors, and Paper, with equal probabilities. Let X, Y, Z be the number of players who pick Rock, Scissors, Paper, respectively in one game. (a) Find the joint PMF of X, Y, Z. (b) Find the probability that the game is decisive. Simplify your answer (it should not involve a sum of many terms). (c) What is the probability that the game is decisive for n = 5? What is the limiting probability that a game is decisive as n → ∞? Explain brief ly why your answer makes sense. 5. A chicken lays n eggs. Each egg independently does or doesn’t hatch, with probability p of hatching. For each egg that hatches, the chick does or doesn’t survive (independently of the other eggs), with probability s of survival. Let N ∼ Bin(n, p) be the number of eggs which hatch, X be the number of chicks which survive, and Y be the number of chicks which hatch but don’t survive (so X + Y = N ). Find the marginal PMF of X, and the joint PMF of X and Y . Are they independent?

2

Stat 110 Strategic Practice 7 Solutions, Fall 2011 Prof. Joe Blitzstein (Department of Statistics, Harvard University)

1

Joint, Conditional, and Marginal Distributions 1. A random point (X, Y, Z ) is chosen uniformly in the ball B = {(x, y, z) : x2 + y 2 + z 2 ≤ 1}. (a) Find the joint PDF of X, Y, Z .

Just as in 2 dimensions uniform in a region means that probability is proportional to length, here probability is proportional to volume. That is, P ((X, Y, Z) ∈ A) = c · Volume(A) if A is contained in B, where c is a constant. Letting A = B, we have that c1 is 34 π, the volume of the ball. So the joint PDF of (X, Y, Z) is f (x, y, z) =

(

3 , 4π

if x2 + y 2 + z 2 ≤ 1; 0, otherwise.

(b) Find the joint PDF of X, Y . We just need to “integrate out” the z from the joint PDF of X, Y, Z. The limits of integration are found by noting that for any (x, y), we need to have z satisfy x2 + y 2 + z 2 ≤ 1. Z ∞ f (x, y, z)dz fX,Y (x, y) = −∞ Z √1−x2 −y2 3 = dz 4π −√1−x2 −y2 3 p = 1 − x2 − y 2 , 2π for x2 + y 2 ≤ 1 (and the PDF is 0 otherwise).

(c) Find an expression for the marginal PDF of X, as an integral.

1

We can integrate out y, z from the joint PDF of X, Y, Z , or integrate out y from the joint PDF of X, Y . Using the result of (b), we have for −1 ≤ x ≤ 1 that the marginal PDF of X is 3 fX (x) = 2π

Z

√ 1−x2 √ − 1−x2

p

1 − x2 − y 2 dy.

2. Let X and Y be i.i.d. Unif(0, 1). Find the expected value and the standard deviation of the distance between X and Y . Let W = |X − Y |. By 2-D LOTUS, Z 1Z E(W ) = 0

1 0

|x − y |dxdy.

Split this into two parts (to get rid of the absolute values): x < y and x ≥ y (i.e., break the square into two triangles). By symmetry the integral over x < y equals the integral over x > y, so Z 1Z y Z 1 2 1 y E(W ) = 2 (y − x)dxdy = 2 dy = . 3 0 0 2 0 Next, we find E(W 2 ). This can either be done by computing the double integral Z 1Z 1 2 E(W ) = (x − y)2 dxdy, 0

0

or by writing E (W 2 ) = E (X − Y )2 = EX 2 + EY 2 − 2E (XY ), which is

1 2E(X 2 ) − 2(EX)2 = 2Var(X) = , 6 since E(XY ) = E(X )E (Y ) for X, Y independent, and E(X) = E(Y ) and E(X 2 ) = E(Y 2 ) (as X and Y have the same distribution). Thus, E(W ) = 1/3, Var(W ) = E(W 2 ) − (E(W ))2 =

1 , 18

1 and the standard deviation of the distance between X and Y is √18 =

2

1 . √ 3 2

3. Let U1 , U 2 , U3 be i.i.d. Unif(0, 1), and let L = min(U1 , U2 , U3 ), M = max(U1 , U2 , U3 ). (a) Find the marginal CDF and marginal PDF of M, and the joint CDF and joint PDF of L, M. Hint: for the latter, start by considering P (L ≥ l, M ≤ m). The event M ≤ m is the same as the event that all 3 of the Uj are at most m, so the CDF of M is FM (m) = m3 and the PDF is fM (m) = 3m2 , for 0 ≤ m ≤ 1. The event L ≥ l, M ≤ m is the same as the event that all 3 of the Uj are between l and m (inclusive), so P (L ≥ l, M ≤ m) = (m − l)3 for m ≥ l with m, l ∈ [0, 1]. By the axioms of probability, we have P (M ≤ m) = P (L ≤ l, M ≤ m) + P (L > l, M ≤ m). So the joint CDF is P (L ≤ l, M ≤ m) = m3 − (m − l)3 , for m ≥ l with m, l ∈ [0, 1]. The joint PDF is obtained by differentiating this with respect to l and then with respect to m (or vice versa): f (l, m) = 6(m − l), for m ≥ l withR m, l ∈ [0, 1]. As a check, note that getting the marginal PDF of m M by finding 0 f (l, m)dl does recover the PDF of M (the limits of integration are from 0 to m since the min can’t be more than the max). (b) Find the conditional PDF of M given L. The marginal PDF of L is fL (l) = 3(1 − l)2 for 0 ≤ l ≤ 1 since P (L > l) = P (U1 > l, U2 > l, U 3 > l) = (1 − l)3 (alternatively, use the PDF of M together with the symmetry that 1 − Uj has the same distribution as Uj , or integrate out m in the joint PDF of L, M). So the conditional PDF of M given L is fM |L (m|l) =

f (l, m) 2(m − l) = , fL (l) (1 − l)2

for all m, l ∈ [0, 1] with m ≥ l. 3

4. A group of n ≥ 2 people decide to play an exciting game of Rock-PaperScissors. As you may recall, Rock smashes Scissors, Scissors cuts Paper, and Paper covers Rock (despite Bart Simpson saying “Good old rock, nothing beats that!”). Usually this game is played with 2 players, but it can be extended to more players as follows. If exactly 2 of the 3 choices appear when everyone reveals their choice, say a, b ∈ {Rock, P aper, Scissors} where a beats b, the game is decisive: the players who chose a win, and the players who chose b lose. Otherwise, the game is indecisive and the players play again. For example, with 5 players, if one player picks Rock, two pick Scissors, and two pick Paper, the round is indecisive and they play again. But if 3 pick Rock and 2 pick Scissors, then the Rock players win and the Scissors players lose the game. Assume that the n players independently and randomly choose between Rock, Scissors, and Paper, with equal probabilities. Let X, Y, Z be the number of players who pick Rock, Scissors, Paper, respectively in one game. (a) Find the joint PMF of X, Y, Z. The joint distribution of X, Y, Z is n! P (X = a, Y = b, Z = c) = a!b!c!

✓ ◆a+b+c 1 3

where a, b, c are any nonnegative integers with a + b + c = n, since (1/3)a+b+c is the probability of any specific configuration of choices for each player with the right numbers in each category, and the coefficient in front counts the number of distinct ways to permute such a configuration. Alternatively, we can write the joint PMF as P (X = a, Y = b, Z = c) = P (X = a)P (Y = b|X = a)P (Z = c|X = a, Y = b), where for a + b + c = n, P (X = a) can be found from the Bin(n, 1/3) PMF, P (Y = b|X = a) can be found from the Bin(n − a, 1/2) PMF, and P (Z = c|X = a, Y = b) = 1. This is a Multinomial(n, (31, 31, 13 )). distribution.

4

(b) Find the probability that the game is decisive. Simplify your answer (it should not involve a sum of many terms). Hint: using symmetry, the probability can be written as 3 timesP a certain n sum. n To do the summation, use the binomial theorem or the fact that k=0 k = 2n .

The game is decisive if and only if exactly one of X, Y, Z is 0. These cases are disjoint so by symmetry, the probability is 3 times the probability that X is zero and Y and Z are nonzero. Note that if X = 0 and Y = k, then Z = n − k. This gives n−1 X

n! P (decisive) = 3 0!k!(n − k)! k=1 ✓ ◆n X n−1 ✓ ◆ n 1 =3 3 k k=1

✓ ◆n 1 3

2n − 2 3n−1 n P Pn n  n since n−1 k=0 k = 2 − 2 (by the binomial theorem or the k=1 k = −1 − 1 + n fact that a set with n elements has 2 subsets). As a check, when n = 2 this reduces to 2/3, which makes sense since for 2 players, the game is decisive if and only if the two players do not pick the same choice. =

(c) What is the probability that the game is decisive for n = 5? What is the limiting probability that a game is decisive as n → ∞? Explain brief ly why your answer makes sense. For n = 5, the probability is (25 − 2)/34 = 30/81 ≈ 0.37. As n → ∞, (2 n − 2)/3n−1 → 0, which make sense since if the number of players is very large, it is very likely that there will be at least one of each of Rock, Paper, and Scissors. 5. A chicken lays n eggs. Each egg independently does or doesn’t hatch, with probability p of hatching. For each egg that hatches, the chick does or doesn’t survive (independently of the other eggs), with probability s of survival. Let N ∼ Bin(n, p) be the number of eggs which hatch, X be the number of chicks which survive, and Y be the number of chicks which hatch but don’t survive (so X + Y = N ). Find the marginal PMF of X, and the joint PMF of X and Y . Are they independent? 5

Marginally we have X ∼ Bin(n, ps), as shown on a previous homework problem using a story proof (the eggs can be thought of as independent Bernoulli trials with probability ps of success for each). Here X and Y are not independent, unlike in the chicken-egg problem from class (where N was Poisson). This follows immediately from thinking about an extreme case: if X = n, then clearly Y = 0. So they are not independent: P (Y = 0) < 1, while P (Y = 0|X = n) = 1. To find the joint distribution, condition on N and note that only the N = i + j term is nonzero: for any nonnegative integers i, j with i + j ≤ n, P (X = i, Y = j) = P (X = i, Y = j |N = i + j )P (N = i + j) = P (X = i|N = i + j)P (N = i + j) ◆ ✓ ◆ ✓ n i+j i j pi+j (1 − p)n−i−j s (1 − s) = i+j i n! (ps)i (p(1 − s))j (1 − p)n−i−j . = i!j!(n − i − j)! If we let Z be the number of eggs which don’t hatch, then from the above we have that (X, Y, Z ) has a Multinomial(n, (ps, p(1 − s), (1 − p)) distribution, which makes sense intuitively since each egg independently falls into 1 of 3 categories: hatch-and-survive, hatch-and-don’t-survive, and don’t-hatch, with probabilities ps, p(1 − s), 1 − p respectively.

6

Stat 110 Homework 7, Fall 2011 Prof. Joe Blitzstein (Department of Statistics, Harvard University) 1. (a) A stick is broken into three pieces by picking two points independently and uniformly along the stick, and breaking the stick at those two points. What is the probability that the three pieces can be assembled into a triangle? Hint: a triangle can be formed from 3 line segments of lengths a, b, c if and only if a, b, c ∈ (0, 1/2). The probability can be interpreted geometrically as proportional to an area in the plane, avoiding all calculus, but make sure for that approach that the distribution of the random point in the plane is Uniform over some region. (b) Three legs are positioned uniformly and independently on the perimeter of a round table. What is the probability that the table will stand? 2. Let (X, Y ) be a uniformly random point in the triangle in the plane with vertices (0, 0), (0, 1), (1, 0). Find the joint PDF of X and Y , the marginal PDF of X , and the conditional PDF of X given Y . 3. Let X, Y be i.i.d. Expo(λ). Find E|X − Y | in two ways: (a) using a 2-D LOTUS and (b) using the memoryless property without any calculus. 4. Two students, A and B, are working independently on homework (not necessarily for the same class). Student A takes Y1 ∼ Expo(λ1 ) hours to f inish his or her homework, while B takes Y2 ∼ Expo(λ2 ) hours. (a) Find the CDF and PDF of YY21, the ratio of their problem-solving times. (b) Find the probability that A finishes his or her homework before B does. 5. The bus company from Blissville decides to start service in Blotchville, sensing a promising business opportunity. Meanwhile, Fred has moved back to Blotchville, inspired by a close reading of I Had Trouble in Getting to Solla Sollew. Now when Fred arrives at the bus stop, either of two independent bus lines may come by (both of which take him home). The Blissville company’s bus arrival times are exactly 10 minutes apart, whereas the time from one Blotchville company bus to the next is Expo(101 ). Fred arrives at a uniformly random time on a certain day. (a) What is the probability that the Blotchville company bus arrives first? Hint: one good way is to use the continuous Law of Total Probability. (b) What is the CDF of Fred’s waiting time for a bus? 1

6. Emails arrive in an inbox according to a Poisson process with rate λ (so the number of emails in a time interval of length t is distributed as Pois(λt), and the numbers of emails arriving in disjoint time intervals are independent). Let X, Y, Z be the numbers of emails that arrive from 9 am to noon, noon to 6 pm, and 6 pm to midnight (respectively) on a certain day. (a) Find the joint PMF of X, Y, Z . (b) Find the conditional joint PMF of X, Y, Z given that X + Y + Z = 36. (c) Find the conditional PMF of X + Y given that X + Y + Z = 36, and f ind E(X + Y |X + Y + Z = 36) and Var(X + Y |X + Y + Z = 36) (conditional expectation and conditional variance given an event are defined in the same way as expectation and variance, using the conditional distribution given the event in place of the unconditional distribution). 7. Shakespeare wrote a total of 884647 words in his known works. Of course, many words are used more than once, and the number of distinct words in Shakespeare’s known writings is 31534 (according to one computation). This puts a lower bound on the size of Shakespeare’s vocabulary, but it is likely that Shakespeare knew words which he did not use in these known writings. More specifically, suppose that a new poem of Shakespeare were uncovered, and consider the following (seemingly impossible) problem: give a good prediction of the number of words in the new poem that do not appear anywhere in Shakespeare’s previously known works. The statisticians Ronald Thisted and Bradley Efron studied this problem in a paper called “Did Shakespeare write a newly-discovered poem?”, which performed statistical tests to try to determine whether Shakespeare was the author of a poem discovered by a Shakespearean scholar in 1985. A simplified version of their method is developed in the problem below. The method was originally invented by Alan Turing (the founder of computer science) and I.J. Good as part of the effort to break the German Enigma code during World War II. Let N be the number of distinct words that Shakespeare knew, and assume these words are numbered from 1 to N . Suppose for simplicity that Shakespeare wrote only two plays, A and B. The plays are reasonably long and they are of the same length. Let Xj be the number of times that word j appears in play A, and Yj be the number of times it appears in play B, for 1 ≤ j ≤ N . (a) Explain why it is reasonable to model Xj as being Poisson, and Yj as being Poisson with the same parameter as Xj . 2

(b) Let the numbers of occurrences of the word “eyeball” (which was coined by Shakespare) in the two plays be independent Pois(λ) r.v.s. Show that the probability that “eyeball” is used in play B but not in play A is e−λ (λ − λ2 /2! + λ3 /3! − λ4 /4! + . . . ). (c) Now assume that λ from (b) is unknown and is itself taken to be a random variable to reflect this uncertainty. So let λ have a PDF f0 . Let X be the number of times the word “eyeball” appears in play A and Y be the corresponding value for play B. Assume that the conditional distribution of X, Y given λ is that they are independent Pois(λ) r.v.s. Show that the probability that “eyeball” is used in play B but not in play A is the alternating series P (X = 1) − P (X = 2) + P (X = 3) − P (X = 4) + . . . . Hint: condition on λ and use (b). (d) Assume that every word’s numbers of occurrences in A and B are distributed as in (c), where λ may be different for different words but f0 is fixed. Let Wj be the number of words that appear exactly j times in play A. Show that the expected number of distinct words appearing in play B but not in play A is E (W1 ) − E(W2 ) + E (W3 ) − E(W4 ) + . . . . (This shows that W1 − W2 + W3 − W4 + . . . is an unbiased predictor of the number of distinct words appearing in play B but not in play A: on average it is correct. Moreover, it can be computed just from having seen play A, without needing to know f0 or any of the λj . This method can be extended in various ways to give predictions for unobserved plays based on observed plays.)

3

Stat 110 Homework 7 Solutions, Fall 2011 Prof. Joe Blitzstein (Department of Statistics, Harvard University) 1. (a) A stick is broken into three pieces by picking two points independently and uniformly along the stick, and breaking the stick at those two points. What is the probability that the three pieces can be assembled into a triangle? Hint: a triangle can be formed from 3 line segments of lengths a, b, c if and only if a, b, c 2 (0, 1/2). The probability can be interpreted geometrically as proportional to an area in the plane, avoiding all calculus, but make sure for that approach that the distribution of the random point in the plane is Uniform over some region. We can assume the length is 1 (in some choice of units, the length will be 1, and the choice of units for length does not affect whether a triangle can be formed). So let X, Y be i.i.d. Unif(0,1) random variables. Let x and y be the observed values of X and Y respectively. If x < y, then the side lengths are x, y  x, and 1  y , and a triangle can be formed if and only if y > 12 , y < x + 21, x < 12 . Similarly, if x > y, then a triangle can be formed if and only if x > 21, x < y + 12 , y < 21 . Since (X, Y ) is Uniform over the square 0  x  1, 0  y  1, the probability of a subregion is proportional to its area. The region given by y > 1/2, y < x+ 1/2, x < 1/2 is a triangle with area 1/8, as is the region given by x > 1/2, x < y+1/2, y < 1/2, as illustrated in the picture below. Thus, the probability that a triangle can be formed is 1/8 + 1/8 = 1/4. 1

0.9

0.8

0.7

y

0.6

0.5

0.4

0.3

0.2

0.1

0 0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

x

Note that the idea of interpreting probabilities as areas works here because (X, Y ) 1

is Uniform on the square. For other distributions, in general we would need to find the joint PDF of X, Y and integrate over the appr...


Similar Free PDFs