In 1952, Chernoff in [3] introduced a technique that gave sharp upper bounds on the tail of the distribution of binary independent random variables. Since then, Chernoff bounds are coming to be the fundamental tools used in bounding the tail probabilities of the sums of bounded and the distribution of binary independent random variables. And the bounds have many different expressions under distinct settings. As the bound is so important, we list it as a theorem.
Theorem A Let $X=\{X_1, X_2, \cdots, X_n\}$ be a set of $n$ mutually independent binary random variables. Let $S=\sum\limits_{i=1}^nX_i$ and $\mu=E[S]$. For $0<\varepsilon$,
and
By the Chernoff's technique, one can obtain a better and more efficient evaluation of the bounds on the tail probabilities. The following theorem can be found in Alon and Spencer's book [1].
Theorem B Let $S$ be a random variable with Poisson distribution and $E(S)=\mu$. For any $\varepsilon>0$, we have the following inequalities:
By Theorem B, one can find good applications if a good bound on the tail probability of $S$ is obtained. For instance, Schmidt, Siegel and Srinivasan [9] obtained a bound on the tail probability of $S$, when the $X_i$'s are $k$-wise independent for $k$ smaller than a certain function of $n$, $\varepsilon$ and $\mu$. $k$-wise independence means that any $k$-subset of $X$ contains mutually independent random variables. An important idea for deriving the bounds on the tail probability of $S$ when the $X_i$'s are not mutually independent involved the theory of martingales by the so-called Azuma-Hoeffding inequality.
A dependence graph $G(V, E)$ for a set of random variables $X=\{X_1, X_2, \cdots, X_n\}$ is a graph with vertex set $V=\{X_1, X_2, \cdots, X_n\}$ and the edge set $E$ such that for each edge $X_iX_j$, $X_i$ and $X_j$ are not mutually independent. For any non-negative integer $d$, we say that the exhibit $d$-bounded dependence, if $\{X_1, X_2, \cdots, X_n\}$ have a dependence graph with maximum vertex degree $d$. The notion of dependence graph is used widely in random graphs and random algorithms, for more detailed information, refer to [1, 5, 6, 10]. Recall that the notion of a dependence graph of random variables is also used from the hypothesis of the Lovász's local lemma.
Let $g^{+}(\mu, \varepsilon)=e^{\mu \varepsilon}\cdot (1+\varepsilon)^{-\mu(1+\varepsilon)}$ and $g^{-}(\mu, \varepsilon)=e^{-\mu\varepsilon^2/2}$.
Pemmaraju showed the following results in [7, 8].
Theorem C Let $X=\{X_1, X_2, \cdots, X_n\}$ be a set of $n$ identically distributed binary random variables with a $d$-bounded dependence graph. $S=\sum\limits_{i=1}^{n}X_i$ and $\mu=E(S)$. For any $\varepsilon$, $0<\varepsilon\le 1$, the following holds:
An equitable $t$-coloring of a graph $G$ is a proper $t$-coloring for which any two color class differ in size by at most one. The notion of equitable colorings of graphs is used in the proof of the above result. Lemma 1 is a conjecture of P. Erdös and showed by Hajnal and Szemerédi in [4]. Lemma 2 is showed by B. Bollabós and Guy in [2].
Lemma 1 A graph $G$ with maximum degree $\Delta$ has a $\Delta+1$-equitable coloring.
Lemma 2 A tree $T$ with $n$ vertices is equitably 3-colorable if $n\ge 3\Delta(T)- 8$ or $n = 3\Delta(T) - 10$.
Here we also use Lemma 1 as in [7] to derive bounds on the tail probabilities of the sums of binary independent random variables. The structure of the paper is as follows. In Section 2, we give two results on the bounds on the tail probabilities of the sums of binary independent random variables. Section 3 contains the main result and its proof. With an application of Lemma 2, a good bound will be followed.
Theorems 1 and 2 are two results on the bounds on the tail probabilities of the sums of binary independent random variables. Theorem 1 is stronger than the Chernoff bounds in some sense and independent of $n$. Theorem 2 is stronger than the Chernoff bounds for some $k$ and $\lambda$ and easy for computation. Both will be used in next section.
Lemma 3 Let $\lambda$ be a non-negative real number and $k$ be a non-negative integer such that $(k-\lambda)^2> k$. If $x\ge k$ and $x>\lambda> 0$, then we have
Proof Let $A(x)=(x+1-\lambda)/(x+1-k)$, $B(x)=(\frac{x}{x+1})^x$ and $C(x)=(\frac{x+1-\lambda}{x-\lambda})^{(x-k)}$. And then
So, we have the following:
by $\ln\frac{x}{1+x}+\ln(1+\frac{1}{x-\lambda})=\ln(1+\frac{\lambda}{(x+1)(x-\lambda)})$.
Let
By $x\ge 0$ and $\ln(1+x)\le x$, we have
By $k< (\lambda-k)^2$, we have the following:
For $0<\lambda< x$, we have
So $F'(x)< 0$.
Note that $\lim\limits_{x\rightarrow +\infty}A(x)=1$, $\lim\limits_{x\rightarrow +\infty}B(x)=e^{-1}$ and $\lim\limits_{x\rightarrow +\infty}C(x)=e^{1}$, and then $\lim\limits_{x\rightarrow +\infty}F(x)=1$, so we have $F(x)> 1$.
Lemma 4 Let ${\rm BIN}(n, \lambda/n)$ be the sum of $n$ independent Bernoulli variables, each of which is equal to 1 with probability $\lambda/n$ and 0 otherwise. Let $k$ be a non-negative integer such that $(k-\lambda)^2> k$. Then we have
where ${\rm BIN}(k; n, \lambda/n)=\Pr({\rm BIN}(n, \lambda/n)=k)$.
Proof By Lemma 3, we have
Note that $\lim\limits_{n\rightarrow+\infty}{\rm BIN}(k; n, \lambda/n)=\frac{\lambda^ke^{-\lambda}}{k!}$. And then we have the conclusion.
Lemma 5 Let ${\rm BIN}(n, \lambda/n)$ be the sum of $n$ independent Bernoulli variables, each of which is equal to 1 with probability $\lambda/n$ and 0 otherwise. Let $(k-\lambda)^2>k$, where $k$ is a non-negative integer and $0<\lambda<n$. Then we have the following:
1. If $k<\lambda$, then
2. If $k>\lambda$, then
Proof It follows immediately from Lemma 4.
Lemma 6 Let ${\rm BIN}(n, \lambda/n)$ be the sum of $n$ independent Bernoulli variables, each of which is equal to 1 with probability $\lambda/n$ and 0 otherwise. Let $(k-\lambda)^2>k$, where $k$ is a non-negative integer and $0<\lambda<n$. Then we have the following:
Proof 1. If $n>\lambda>k\ge m$, then
By Lemma 4, for $0\le i\le k$,
And so
2. If $0<\lambda<k\le m$, then
By Lemma 4, for $k\le i\le n$,
And then,
Clearly, the results of Theorems 1 and 2 are stronger than those of Theorems A and B and easier to handle.
Let $X=\{X_1, X_2, \cdots, X_n\}$ be a set of $n$ binary random variables and $G(V, E)$ be the dependence graph associated with $X$. Assume that $G$ admits a $t$-equitable coloring and $C_1, C_2, \cdots, C_t$ be the color classes of the coloring. Then we have two different ways to add restrictions with the elements of $X$, which are Models 1 and 2.
Model 1 The sum of the expected value of $X_i$'s is a constant. For any color class $C_i$, $1\le i\le n$, elements of $C_i$ are identical distribution.
Model 2 The sum of the expected value of $X_i$'s is a constant and $E(X_i)=E(X_j)$ for $1\le i<j\le n$.
Lemma 7 Let $X=\{X_1, X_2, \cdots, X_n\}$ be a set of $n$ binary random variables with dependence graph $G(V, E)$. Suppose that $G$ can be colored equitably with at most $t$ colors. For $(k-t-\mu)^2>kt-t^2$, $k-t>\mu>0$, we have upper tail probability
For $(k+t-\mu)^2>kt+t^2$, $k+t<\mu<\lceil n/t\rceil$, we have the lower tail probability
Proof We distinguish two cases.
Model 1 Let $C_1, C_2, \cdots, C_t$ be the $t$ colors classes in a $t$-equitable-coloring of $G$. For each $j\in\{1, 2, \cdots, t\}$ and $i\in C_j$, let $E(\sum_{i\in C_j}X_i)=\mu/t$, and $E(X_i)=\mu/(t{|C_j|})$. We now rewrite the event $S\ge k$ as follows.
$S\ge k$ is equivalent to $S\ge \frac{k}{n}\cdot n$, which is equivalent to
For
Hence
for
So
for $(k-t-\mu)^2>kt-t^2$, $k-t>\mu>0$, by Theorem 1, we have
So $\Pr(S\ge k)<\sum_{i\ge \frac{k}{t}-1}\frac{\mu^ie^{-\mu/t}}{t^{i-1}i!}$.
Model 2 Let $C_1, C_2, \cdots, C_t$ be the $t$ colors classes in a $t$-equitable-coloring of $G$. Let
for $1\le i\le n$.
We now rewrite the event $S\ge k$ as follows:
Let $m=\lfloor n/t\rfloor$, we have
Since $\sum\limits_{j\in C_i, |C_i|=m}X_j\sim b(m, \mu/n)$, for $k>\mu/t$, we have
So $\Pr(S\ge k)<\sum\limits_{i\ge \frac{k}{t}-1}\frac{\mu^ie^{-\mu/t}}{t^{i-1}i!}$.
This completes the proof of upper tail probability bound. The proof of the lower tail probability bound is similar and omitted.
Lemma 8 Let $X=\{X_1, X_2, \cdots, X_n\}$ be a set of $n$ binary random variables with dependence graph $G(V, E)$. Suppose that $G$ can be colored equitably with at most $t$ colors. If $n/t$ is an integer, Then for $(k-\mu)^2>kt$, $k>\mu>0$, we have upper tail probability
For $(k-\mu)^2>kt$, $k<\mu<n/t$, we have lower tail probability
Theorem 1 Let $X=\{X_1, X_2, \cdots, X_n\}$ be a set of $n$ binary random variables with dependence graph $G(V, E)$. Let $d$ be the maximum degree of $G$. If
and $k-d-1>\mu>0$, then the upper tail probability is as follows:
If $(k+d+1-\mu)^2>(d+1)(k+d+1)$ and $k+d+1<\mu<\lceil n/(d+1)\rceil$, then the lower tail probability is as follows:
Actually, if $n/(d+1)$ is an integer, for $(k-\mu)^2>k(d+1)$ and $k>\mu>0$, then
For $(k-\mu)^2>k(d+1)$ and $k<\mu<n/(d+1)$, then
Proof Let $G$ be the dependence graph of with maximum vertex degree $d$ such that a dependence graph exists by definition of d-bounded dependence. By Lemma 1, $G$ has a $d+1$-equitable coloring. Replacing $t$ by $(d+1)$ in Lemmas 7 and 8 yields the desired bounds.
Note that Lemmas 5 and 6 are special cases of Theorem 1.
The strength of the equitable coloring technique in deriving tail probability bounds lies in the fact that it allows us to focus closely on the structure of the dependence graph. In particular, a small equitable chromatic number for a dependence graph leads to sharp tail probability bounds. Not much seems to be known about the equitable chromatic number of different graph classes. The connection between equitable colorings and tail probability bounds presented.
The following result comes ready-made from Bollobás and Guy [2]. Lemma 2 essentially implies that if $\Delta(T) \le n/3$, then $T$ can be equitably 3-colored. This immediately translates to the following tail probability bounds on tree structured dependence graphs.
Theorem 2 Let $X=\{X_1, X_2, \cdots, X_n\}$ be a set of $n$ identical random variables that have a tree structured dependence graph with maximum degree no greater than $n/3$. Then we have the following bounds of the tail probabilities of the sum $S=\sum\limits_{i=1}^{n}X_i$ of the $X_i$'s as follows:
If $(k-3-\mu)^2>3(k-3)$, $k-3>\mu>0$, then
If $(k+3-\mu)^2>3(k+3)$, $k+3<\mu<\lceil n/3\rceil$, then