Moments and Deviations

Prev: Game-Theoretic Techniques Next: Tail Inequalities

Problems

3.1

Consider an occupancy problem in which balls are independently and uniformly distributed in bins. Show that, for large , the expected number of empty bins approaches , where is the base of the natural logarithm. What is the expected number of empty bins when balls are thrown into bins? See Theorem 4.18.


3.2

Suppose balls are thrown into bins. Give the best bound you can on to ensure that the probability of there being a bin containing at least two balls is at least .


3.3

A parallel computer consists of processors and memory modules. During a step, each processor sends a memory request to one of the memory modules. A memory module that receives either one or two requests can satisfy its request(s); modules that receive more than two requests will satisfy two requests and discard the rest.

  1. Assuming that each processor chooses a memory module independently and uniformly at random, what is the expected number of processors whose requests are satisfied? Use the approximation if necessary.
  2. Repeat the computation for the case where each memory module can satisfy only one request during a step.

3.4

Consider the following experiment, which proceeds in a sequence of rounds. For the first round, we have balls, which are thrown independently and uniformly at random into bins. After round , for , we discard every ball that fell into a bin by itself in round . The remaining balls are retained for round , in which they are thrown independently and uniformly at random into the bins. Show that there is a constant such that with probability , the number of rounds is at most .


3.5

Let be a random variable with expectation and standard deviation .

  1. Show that for any ,

This version of the Chebyshev inequality is sometimes referred to as the Chebyshev-Cantelli bound. 2. Prove that

Under what circumstances does this give a better bound than the Chebyshev inequality?


3.6

Let be a non-negative integer-valued random variable with positive expectation. Prove the following inequalities.

  1. Explain why the second inequality always gives a stronger bound than the first inequality.

3.7

Let and be chosen independently and uniformly at random from , where is a prime. Suppose we generate pseudo-random numbers from by choosing , for . For any , show that there is a choice of the witness set such that and the probability that none of the ‘s lie in the set is at least .


3.8

Suggest a scheme for four-point sampling from the range where is a prime. For samples using this scheme, give an upper bound on the probability that all attempts fail to discover a witness given and compare this with the bound of that the naive use of four samples would yield. En route, derive an upper bound on the fourth central moment of the sum of four-way independent random variables.


3.9

Due to D. R. Karger and R. Motwani [233].

  1. Let be two disjoint subsets of a universe such that . Suppose we select a random set by independently sampling each element of with probability . We say that the random sample is good if the following two conditions hold: and . Show that for , the probability that is good is larger than some positive constant.
  2. Suppose now that the random set is chosen by sampling the elements of with only pairwise independence. Show that for a suitable choice of the value of , the probability that is good is larger than some positive constant.

3.10

The sharp threshold result in the coupon collector’s problem does not imply that the probability of needing more than trials goes to zero at a doubly exponential rate if were not a constant, but were allowed to grow with . Let the probability of requiring more than trials be . For constant , show that can be bounded from above and below by polynomials in .


3.11

Consider the extension of the coupon collector’s problem to that of collecting at least copies of each coupon type. Show that the sharp threshold for the number of selections required, denoted , is centered at . In other words, for any positive integer and constant , prove that


3.12

Consider the following process related to the coupon collector problem. There are bins and players, and each player has an infinite supply of balls. The bins are all initially empty. We have a sequence of rounds: in each round, each player throws a ball into an empty bin chosen independently at random from all currently empty bins. Let the random variable be the number of rounds before every bin is non-empty. Determine the expected value of . What can you say about the tail of ‘s distribution?


3.13

Let be a random bipartite graph on two independent sets of vertices and , each with vertices. For each pair of vertices and , the probability that the edge between them is present is , and the presence of any edge is independent of all other edges. Let for some .

  1. Show that the probability that contains an isolated vertex is asymptotically equal to .
  2. Suggest and prove a generalization of this to random non-bipartite graphs.

3.14

Due to R. M. Karp.

Consider a bin containing balls chosen at random, without replacement, from a collection of distinct balls. Without being able to see or count the balls in the bin, we would like to simulate random sampling with replacement from the original set of balls. Our only access to the balls is that we can sample without replacement from the bin.

Consider the following strategy. Suppose that balls have been drawn from the bin so far. Flip a coin with the probability of HEADS being . If HEADS appears, then pick one of the previously drawn balls uniformly at random; otherwise, draw a random ball from the bin. Show that each choice is independently and uniformly distributed over the space of the original balls. How many times can we repeat the sampling?


3.15

Due to D. Angluin and L. G. Valiant [28].

Let denote a random bipartite graph with vertices in each of the vertex sets and . Each possible edge, independently, is present with probability . Consider the following algorithm for constructing a perfect matching, see Section 7.3, in such a random graph. Modify the Proposal Algorithm of Section 3.5 as follows. Each can propose only to adjacent . A vertex always accepts a proposal, and if a proposal causes a divorce, then the newly divorced is the next to propose. The sampling procedure outlined in Problem 3.14 helps implement the Principle of Deferred Decisions. How small can you make the value of and still have the algorithm succeed with high probability? The following fact concerning the degree of a vertex in proves useful: