Tail Inequalities

Prev: Moments and Deviations Next: The Probabilistic Method

Problems

4.1

Suppose you are given a biased coin that has , for some fixed , without being given any other information about . Devise a procedure for estimating by a value such that you can guarantee that

for any choice of the constants . Let be the number of times you need to flip the biased coin to obtain the estimate. What is the smallest value of for which you can still give this guarantee?


4.2

Let be a random variable. Define the th factorial moment of , , as the expected value of

Let denote a random graph on vertices where each edge independently is present with probability , and denote a graph on vertices that has edges chosen uniformly at random. Let denote the number of isolated vertices in , and let be the number of isolated vertices in . Consider the case

and

for a real value . Prove that and are asymptotically equal to , where .


4.3

For in the range , use (4.1) to obtain a closed-form upper bound for , as a function of and , that is within a constant factor of the best possible.


4.4

Let be independent geometrically distributed random variables each having expectation , each of the is an independent experiment counting the number of tosses of an unbiased coin up to and including the first HEADS. Let

and be a positive real constant. Use moment generating functions and the Chernoff technique to derive the best upper bound you can on


4.5

The result of Theorem 4.2 bounds the probability of the sum of Poisson trials deviating far below its expectation. Use this to give a bound on the probability of the sum of independent geometric random variables deviating above its expectation, thus providing an alternative approach to that in Problem 4.4.


4.6

Hoeffding’s Bound [202].

Suppose are independent Poisson trials such that . Let

Our goal is to show that from the standpoint of deviations from the mean, the worst case is when the ‘s are all equal. Let be the sum of independent Bernoulli trials each having probability of assuming the value . Then, for any and any , show that

and


4.7

Due to W. Hoeffding [202].

This problem deals with a useful generalization of the Hoeffding bound in Problem 4.6.

  1. A function is said to be convex if for any , and , the following inequality is satisfied:

Show that the function is convex for any . What can you say when ? 2. Let be a random variable that assumes values in the interval , and let . Define the Bernoulli random variable such that and . Show that for any convex function , . 3. Let be independent and identically distributed random variables over , and define

Using parts (1) and (2), derive upper and lower tail bounds for the random variable using the Chernoff bound technique. In particular, show that

Remark: While the results in this problem hold for continuous random variables, they may be a bit easier to prove in the case where take on a discrete set of values in the interval . Also, it should be easy to generalize this to distributions defined over arbitrary intervals . See also Problem 4.21.


4.8

Consider a BPP algorithm that has an error probability of , for some polynomially bounded function of the input size . Using the Chernoff bound on the tail of the binomial distribution, show that a polynomial number of independent repetitions of this algorithm suffice to reduce the error probability to .


4.9

Consider now the following variant of the bit-fixing algorithm. Each packet randomly orders the bit positions in the label of its source, and then corrects the mismatched bits in that order. Show that there is a permutation for which with high probability this algorithm requires steps to complete the routing.


4.10

Suppose we run Valiant’s scheme on an -node network in which every node is of degree ; each packet first goes to a random destination chosen uniformly from all the nodes and then on to its final destination. Show that the expected number of steps for the completion of the first phase is


4.11

The lattice approximation problem is an extension of the set-balancing problem, Example 4.5. As before, we are given an matrix all of whose entries are or . In addition, we are given a column vector with entries, all of which are in the interval . We wish to find a column vector with entries, all of which are from the set , so as to minimize

We think of the vector as an integer approximation to the given real vector , in the sense that is close to in every component. This has applications to approximating certain integer programs given solutions to their linear programming relaxations, along the lines of Section 4.3. Derive a bound on assuming that were derived from using randomized rounding.


4.12

Consider the global wiring problem of Section 4.3. We wish to approximate the best possible solution without the restriction that only one-bend routes are used. Adapt the approach in Section 4.3 to devise an algorithm running in time polynomial in the number of gates and nets, achieving an approximation similar to that in Theorem 4.8.


4.13

The set-cover problem is the following: given sets over a universe , find the smallest set such that for , . An alternative formulation of this problem is the following: given a - matrix , find a - column vector such that the dot product of each row of with is positive while minimizing . The matrix has rows, and the th row is the incidence vector of the set .

Given a matrix , let denote the size of the smallest set-cover for . Let be the number of rows in . Show that we can adapt the technique of linear programming followed by randomized rounding to find a set-cover of size times .


4.14

Show that the RandQS algorithm of Chapter 1 runs in time with high probability.


4.15

Redesign the parameters of the LazySelect algorithm of Chapter 3 and invoke the Chernoff bound to show that with high probability it finds the th smallest of elements in

steps, with probability .


4.16

Prove Lemmas 4.9 and 4.10. Also, formulate and prove their generalizations to the case where the conditioning is done on more than one random variable. Finally, using these, prove Lemma 4.11.


4.17

Prove Theorem 4.12.


4.18

Prove Lemma 4.14.


4.19

Using Lemma 4.14, prove Theorem 4.13.


4.20

Derive the tail bounds described in Problem 4.7(3) by applying Azuma’s inequality, Corollary 4.17, to the Doob martingale sequence obtained from by setting

and, for ,

How does this bound compare with the one obtained in Problem 4.7?


4.21

Prove Azuma’s inequality, Theorem 4.16, for the case where for all . Note that this is the same as Corollary 4.17 with . Do you see how to generalize this to the case of arbitrary ‘s? Hint: Concentrate on the upper tail bound, since the lower tail bound can be obtained by negating the random variables. Consider the martingale difference sequence obtained by setting , and note that

You can essentially mimic the proof of Theorem 4.1, but be careful to use conditional expectations and the martingale property in going from the analog of equation (4.2) to that of equation (4.3). Since the random variables could have arbitrary distributions over the interval , you will also have to make use of an argument similar to that in Problem 4.7.


4.22

Due to A. Kamath, R. Motwani, K. Palem, and P. Spirakis [228].

Consider again the issue of tail bounds on the number of empty bins studied in Theorem 4.18. In this setting, let be the indicator variable whose value is if and only if bin is empty, and define

as the number of empty bins.

Define

and let be mutually independent Bernoulli random variables that take value with probability and value with probability ; note that the sum

has the binomial distribution with parameters and .

  1. Show that for all ,

Conclude that any Chernoff bound on the upper tail of ‘s distribution also applies to the upper tail of ‘s distribution, even though the Bernoulli variables are not mutually independent. The point is that their correlation is negative and only helps to reduce the tail probability. How does the resulting bound on the upper tail of ‘s distribution compare with the bound given in Theorem 4.18? 2. Can you show that for all ,

Repeat the exercise in part (1) for the lower tail.