Electronic Colloquium on Computational Complexity, Revision 2 of Report No. 132 (2015)
On Percolation and NP-Hardness
Huck Bennett ∗ Daniel Reichman † Igor Shinkar ∗
July 13, 2016
Abstract The edge-percolation and vertex-percolation random graph models start with an arbitrary graph G, and randomly delete edges or vertices of G with some fixed probability. We study the computational hardness of problems whose inputs are obtained by applying percolation to worst-case instances. Specifically, we show that a number of classical NP-hard graph problems remain essentially as hard on percolated instances as they are in the worst-case (assuming NP * BPP). We also prove hardness results for other NP-hard problems such as Constraint Satisfaction Problems and Subset-Sum, with suitable definitions of random deletions. We focus on proving the hardness of the Maximum Independent Set problem and the Graph Coloring problem on percolated instances. To show this we establish the robustness of the corresponding parameters α(·) and χ(·) to percolation, which may be of independent interest. Given a graph G, let G0 be the graph obtained by randomly deleting edges of G. We show that if α(G) is small, then α(G0) remains small with probability at least 0.99. Similarly, we show that if χ(G) is large, then χ(G0) remains large with probability at least 0.99.
1 Introduction
The theory of NP-hardness suggests that we are unlikely to find optimal solutions to NP-hard problems in polynomial time. This theory applies to the worst-case setting where one considers the worst running-time over all inputs of a given size. It is less clear whether these hardness results apply to “real-life” instances. One way to address this question is to examine to what extent known NP-hardness results are stable under random perturbations, as it seems reasonable to assume that a given instance of a problem may be subjected to noise. Recent work has studied the effect of random perturbations of the input on the runtime of algorithms. In their seminal paper Spielman and Teng [33] introduced the idea of smoothed analy- sis to explain the superior performance of algorithms in practice compared with formal worst-case bounds. Roughly speaking, smoothed analysis studies the running time of an algorithm on a per- turbed worst-case instance. In particular, they showed that subjecting the weights of an arbitrary linear program to Gaussian noise yields instances on which the simplex algorithm runs in expected polynomial time, despite the fact that there are pathological linear programs for which the simplex algorithm requires exponential time. Since then smoothed analysis has been applied to a number of other problems [13, 34].
∗Department of Computer Science, Courant Institute of Mathematical Sciences, New York University. Email: [email protected], [email protected] †Electrical Engineering and Computer Science, University of California Berkeley, Berkeley, CA, USA. Email: [email protected].
1
ISSN 1433-8092 In contrast to smoothed analysis, we study when worst-case instances of problems remain hard under random perturbations. Specifically, we study to what extent NP-hardness results are ro- bust when instances are subjected to random deletions. Previous work is mainly concerned with Gaussian perturbations of weighted instances. Less work has examined the robustness of hardness results of unweighted instances with respect to discrete noise. We focus on two forms of percolation on graphs. Given a graph G = (V,E) and a parameter 0 p ∈ (0, 1), we define Gp,e = (V,E ) as the probability space of graphs on the same set of vertices, 0 where each edge e ∈ E is contained in E independently with probability p. We say that Gp,e 0 0 is obtained from G by edge percolation. We define Gp,v = (V ,E ) as the probability space of 0 graphs, in which every vertex v ∈ V is contained in V independently with probability p, and Gp,v 0 is the subgraph of G induced by the vertices V . We say that Gp,v is obtained from G by vertex percolation. We also study appropriately defined random deletions applied to instances of other NP-hard problems, such as 3-SAT and Subset-Sum. Throughout we refer to instances that are subjected to random deletions as percolated instances. Our main question is whether such percolated instances remain hard to solve by polynomial-time algorithms assuming NP * BPP . The study of random discrete structures has resulted with a wide range of mathematical tools which have proven instrumental in proving rigorous results regarding such structures [5, 17, 19, 26]. One reason for studying percolated instances is that it may offer the opportunity to apply these methods to a broader range of distributions of instances of NP-hard problems beyond random graphs and random formulas. Furthermore, percolated instances are studied in host of different disciplines and are frequently used to explain properties of the the WWW [1,8], hence it is of interest to understand the computational complexity of instances arising from percolated networks.
1.1 A first example – 3-Coloring Consider the 3-Coloring problem, where given a graph G = (V,E) we need to decide whether G is 3-colorable. Suppose that given a graph G we sample a random subgraph G0 of G, by deleting 1 each edge of G independently with probability p = 2 , and ask whether the resulting graph is 3- colorable. Is there a polynomial time algorithm that can decide with high probability whether G0 is 3-colorable? We demonstrate that a polynomial-time algorithm for deciding whether G0 is 3-colorable is impossible assuming NP * BPP . We show this by considering the following polynomial time reduction from the 3-Coloring problem to itself. We will need the following definition. Definition 1.1. Given a graph G, the R-blowup of G is a graph G0 = (V 0,E0), where every vertex v is replaced by an independent set ve of size R, which we call the cloud corresponding to v. We then connect the clouds ue and ve by a complete R × R bipartite graph (u, v) ∈ E. Given an n-vertex graph H the reduction outputs a graph G that is an R-blow-up of H for R = Cplog(n), where C > 0 is large enough. It is easy to see that H is 3-colorable if and only if G is 3-colorable. In fact, the foregoing reduction satisfies a stronger robustness property for random subgraphs 0 1 G of G obtained by deleting edges of G with probability 2 . Namely, if H is 3-colorable, then G is 3-colorable, and hence G0 is also 3-colorable with probability 1. On the other hand, if H is not 3-colorable, then G is not 3-colorable, and with high probability G0 is not 3-colorable either.
2 Indeed, for any edge (v1, v2) in H let U1,U2 be two clouds in G corresponding to v1 and v2. 0 0 Fixing two arbitrary sets U1 ⊆ U1 and U2 ⊆ U2 each of size R/3, the probability that there is 0 −R2/9 no edge in G connecting a vertex from U1 to a vertex in U2 is 2 . By union bounding over R 2 R2/9 0 0 0 the |E| · R/3 2 choices of U1,U2 we get that there is at least one edge between U1 and 0 0 U2 with high probability. When this holds we can decode any 3-coloring of G to a 3-coloring of H by coloring each vertex v of H with the color that appears the largest number of times in the coloring of the corresponding cloud in G0, breaking ties arbitrarily. Therefore, a polynomial time algorithm for deciding the 3-colorability of G0 implies a polynomial time algorithm for determining the 3-colorability of H with high probability, and hence unless N P ⊆ coRP there is no polynomial time algorithm that given a 3-colorable graph G finds a 3-coloring of a random subgraph of G.1
Toward a stronger notion of robustness The example above raises the question of whether the blow-up described above is really necessary. Na¨ıvely, one could hope for stronger hardness of the 3-Coloring problem, namely, that for any graph H if H is not 3-colorable, then with high probability a random subgraph H0 of H is not 3-colorable either. However, this is not true in general, as H can be a 3-critical graph, i.e., a 3-colorable graph such that deletion of any edge of H decreases its chromatic number (consider for example the case of an odd cycle). Nonetheless, if random deletions do not decrease the chromatic number of a graph by much, then one could use hardness of approximation results on the chromatic number to deduce hardness results for coloring percolated graphs. This naturally leads to the following question. Question. Let G be an arbitrary graph, and let G0 be a random subgraph of G obtained from G by deleting each edge of G with probability 1/2. Is it true that if χ(G) is large, then χ(G0) is also large with probability at least 0.99?2 In this paper we give a positive answer to this question and show that in some sense the chro- matic number of any graph is robust against random deletions. We also consider the question of robustness for other graph parameters. For independent sets we demonstrate that if the inde- pendence number of G is small, then with high probability the independence number of a random subgraph of G is small as well. Hardness results are derived for other graph-theoretic problems such as Minimum Vertex Cover and Hamiltonian Cycle. Similarly, we show that for CSP formulas that are sufficiently dense randomly deleting its clauses does not change the maximum possible frac- tion of clauses that can be satisfied simultaneously. In particular, this implies that these problems remain essentially as hard on percolated instances as they are on worst-case instances.
Remark. It is worth noting that there are graph parameters for which hardness on percolated instances differs significantly from hardness on the original instance. For example, standard results in random graph theory imply that for every n-vertex graph G, with high probability the size of the 0 1 largest clique in the graph G obtained by edge percolation with p = 2 is O(log n). In particular, a 1Note that in the foregoing example, if we start with a bounded degree graph H, we can reduce it to a bounded degree graph G by using an R × R bipartite expander instead of the complete bipartite graph. 2We note that if instead of choosing the subgraph G0 at random, we choose an arbitrary subgraph of G with |E|/2 edges, then it is possible that χ(G0) is much smaller than χ(G). For example, consider the n-vertex graph G = (V,E) that consists of a clique of size n/3 and a complete bipartite graph with n/3 edges on each side. Then χ(G) = n/3, whereas if we remove all the edges of the n/3-clique, the graph becomes 2-colorable, while the number of removed `n/3´ 2 edges is 2 < n /18 < |E|/2.
3 maximum clique in G0 can be found in time nO(log n), which is significantly faster than the fastest known algorithm for finding a maximum clique in the worst-case.
1.2 Robustness of NP-hard problems under percolation In proving hardness results for percolated instances we use the concept of robust reductions which we explain below. It will be convenient to consider promise problems3. We start by introducing the following definition.
Definition 1.2. Let A = (AYES,ANO) and B = (BYES,BNO) be two promise problems. For each y ∈ {0, 1}∗ (an instance of the problem B) let noise(y) be a distribution on {0, 1}∗, that is samplable in time that is polynomial in |y|.
• A polynomial time reduction R from A to B is said to be noise-robust if
1. For all x ∈ AYES it holds that R(x) ∈ BYES, and Pr[noise(R(x)) ∈ BYES] > 0.99.
2. For all x ∈ ANO it holds that R(x) ∈ BNO, and Pr[noise(R(x)) ∈ BNO] > 0.99.
• If in the first item we have Pr[noise(R(x)) ∈ BYES] = 1, then we say that R is a noise-robust coRP-reduction. Similarly, if in the second item we have Pr[noise(R(x)) ∈ BNO] = 1, then we say that R is a noise-robust RP-reduction.
• The problem B = (BYES,BNO) is said to be NP-hard under a noise-robust reduction if there exists a noise-robust reduction from an NP-hard problem to B.
• We say that the problem A is strongly-noise-robust to B if
1. For all x ∈ AYES it holds that x ∈ BYES, and Pr[noise(x) ∈ BYES] > 0.99.
2. For all x ∈ ANO it holds that x ∈ BNO, and Pr[noise(x) ∈ BNO] > 0.99. We use the term noise-robust to avoid confusion with other notions of robust reductions that have appeared in the literature. In order to ease readability, we will often write robust reductions instead, always referring to noise-robust reductions as defined above. Note that in the last item of Definition 1.2 there is no reduction involved. Instead, we think of the problem A as a relaxation of B with AYES ⊆ BYES and ANO ⊆ BNO, and hence any algorithm that solves B in particular solves A. However, the relaxation is also robust: after applying noise to a YES-instance (resp. NO-instance) of A, it stays a YES-instance (resp. NO-instance) of B with high probability.
Proposition 1.3. Let L = (LYES,LNO) be a promise problem, and for each y instance of L, let noise(y) be a distribution on instances of L that is samplable in time that is polynomial in |y|. If L is NP-hard under a noise-robust reduction, then there is no polynomial time algorithm that when given an input y decides with high probability whether noise(y) ∈ LYES or noise(y) ∈ LNO, unless N P ⊆ BPP.
3Recall, that a promise problem is a generalization of a decision problem, where for the problem L there are two disjoint subsets LYES and LNO, such that an algorithm that solves L must accept all the inputs in LYES and reject all inputs in LNO. If the input does not belong to LYES ∪ LNO, there is no requirement on the output of the algorithm.
4 Indeed, the example given in Section 1.1 gives a noise-robust reduction from the 3-Coloring problem to itself, where noise refers to random deletions of the edges in a given graph (edge percolation). Therefore, the 3-Coloring problem is NP-hard under a noise-robust reduction.
1.3 Our results In this paper we show that a number of NP-hard problems remain hard to solve even after random deletions, i.e., they are NP-hard under noise-robust reductions. Furthermore, we show that some gap NP-hard problems are, in fact, strongly-noise-robust to the same problems with a smaller gap. Specifically, we focus on showing these results for the gap versions of the maximum independent set and chromatic number problems. As technical tools, we prove a number of combinatorial results about the independence number and the chromatic number of percolated graphs that might be of independent interest.
Maximum Independent Set and Percolation
Theorem 1.4. Let G = (V,E) be an n-vertex graph. Then, with high probability α(Gp,e) ≤ α(G) O p log(np) . We observe that in general, the upper bound above cannot be improved, as it is well known log(np) that the independence number of G(n, p) is Ω p with high probability (see, e.g., [5]). In the Coloring-vs-MIS(q, a) problem, given a graph G the goal is to distinguish between the YES-case where χ(G) ≤ q and the NO-case where α(G) ≤ a. By using Theorem 1.4 together with inapproximability results of Feige and Kilian [14] we obtain the following hardness result.
Theorem 1.5. For any q, a the Coloring-vs-MIS(q, a) problem is strongly-noise-robust to a Coloring-vs-MIS(q, O p log(np) ), where n denotes the number of vertices in the given graph, and noise is edge-percolation with probability p. In particular, for any constant ε > 0, there is no polynomial time algorithm that given an n- 1 1−2ε vertex graph G approximates either α(Gp,e) or χ(Gp,e) within a pn1−2ε (resp. pn ) factor for 1 any p > n1−2ε unless N P ⊆ BPP. We also prove analogous theorems for vertex percolation.
Graph Coloring and Percolation Theorem 1.4 says that it is hard to approximate the chro- matic number of a percolated graph within a n1−ε factor, but says nothing about hardness of coloring percolated graphs with small (constant) chromatic number. We address this question be- low by proving lower bounds4 on the chromatic number of percolated graphs. To do this we use results from additive combinatorics and discrete Fourier analysis.
Theorem 1.6. Let G = (V,E) √be an n-vertex graph, and let k = χ(G). Then, for every λ > 0 it −λ2/2 holds that Pr[χ(G 1 ) ≥ k/2 − λ k] > 1 − e . 2 ,v Theorem 1.7. Let G = (V,E) be a graph with m edges, and let k = χ(G). Then, for every 1/3 1/4 α ∈ (0, 1) it holds that Pr[χ(G 1 ) ≥ max{Ωα(k ), Ωα(k/m )}] > 1 − α. 2 ,e
4 The notation Oα(f(n)) (resp. Ωα(f(n))) means that O(f(n)) (resp. Ω(f(n))) holds for fixed α.
5 1/4 3/8 In Theorem 1.7, the Ωα(χ(G)/m ) lower bound is better when χ(G) = ω(m ), and the 1/3 3/8 Ωα(χ(G) ) lower bound is better when χ(G) = o(m ). 1 We also obtain lower bounds on the chromatic numbers of Gp,v and Gp,e for p < 2 by composing the bounds in Theorems 1.6 and 1.7 dlog2(1/p)e times. In the Gap-Coloring(q, Q) problem we are given an n-vertex graph G and the goal is to distin- guish between the YES-case where G is q-colorable, and the NO-case where the chromatic number of G is at least Q. There is a large body of work proving hardness results for this problem [18, 23, 21] including stronger results assuming variants of the Unique Games Conjecture [10, 12]. The strongest NP-hardness result of this form, due to Huang [21], shows that Gap-Coloring(q, exp(Ω(q1/3))) is NP-hard. Combining this with Theorem 1.7 we obtain an analogous hardness result under noise- robust reductions for this problem.
Theorem 1.8. For all q < Q the Gap-Coloring(q, Q) problem is strongly-noise-robust to the 1/3 1 Gap-Coloring(q, Ω(Q )) problem, where noise is 2 -edge-percolation applied to the graph. In particular, for any sufficiently large constant q given a q-colorable graph G it is NP-hard to 1/3 find a exp(Ω(q ))-coloring of G 1 with high probability. 2 ,e
Satisfiability and other CSP’s We prove hardness of approximating the value of percolated k-CSP instances. An instance Φ of k-CSP over some alphabet Σ (e.g. Σ = {0, 1}) is a formula consisting of a collection of clauses C1, ..., Cm over n variables x1, ..., xn taking values in Σ, where k each clause is associated with some k-ary predicate f :Σ → {0, 1} over variables xi1 , . . . , xik . An instance Φ is said to be simple if all clauses in Φ are distinct. Given an assignment σ :
{x1, ..., xn} → Σ we say that the constraint C on the variables xi1 , . . . , xik is a satisfied by σ if fC (σ(xi1 ), ..., σ(xik )) = 1, where fC is the predicate corresponding to C. Given a formula Φ, and an assignment σ to its variables the value of Φ with respect to the assignment σ, denoted by valσ(Φ), is the fraction of constraints of Φ satisfied by σ. The value of Φ is defined as val(Φ) = maxσ valσ(Φ). If val(Φ) = 1 we say that Φ is satisfiable. We are typically interested in CSP where constraints belong to some fixed family of predicates F. Wk For example, in the k-SAT problem, the constraints are all of the form f(z1, . . . , zk) = i=1(zi = bi), for b1, . . . , bk ∈ {0, 1}. We assume that k, the arity of the constraints, is some fixed constant that does not depend on the number of variables n. These definitions give rise to the following optimization problem. Given a CSP instance Φ find an assignment that maximizes the value of Φ. We refer to this maximization problem as Max-CSP-F, where F denotes the family of predicates constraints are taken from. For 0 < s < c ≤ 1, Gap-CSP-F(c, s) is the promise problem where YES-instances are formulas Φ such that val(Φ) ≥ c, and NO-instances are formulas Φ such that val(Φ) ≤ s. Here we assume the constraints of CSP instances are restricted to be in the family F. We study two models of percolation on instances of CSP, clause percolation and variable perco- c lation. Given an instance Φ of CSP its clause percolation is a random formula Φp over the same set of variables, that is obtained from Φ by keeping each clause of Φ independently with probability p.
Theorem 1.9. Let ε, δ ∈ (0, 1) be fixed constants. There is a polynomial time reduction such that given a simple unweighted instance Φ of Max-CSP-F outputs a simple unweighted instance Ψ of 1 Max-CSP-F on n variables, such that val(Ψ) = val(Φ), and for any p > nk−1−δ the following holds. c 1. If val(Φ) = 1, then val(Ψp) = 1 with probability 1.
6 c 2. If val(Φ) < 1, then with high probability |val(Ψp) − val(Φ)| < ε. This immediately implies the following corollary.
Corollary 1.10. Let F be a collection of Boolean constraints of arity k, and suppose that for some 0 < s < c ≤ 1 the problem Gap-CSP-F(c, s) is NP-hard. Then Gap-CSP-F(c−ε, s+ε) is NP-hard 1 under a noise-robust reduction, where noise is p-clause percolation with p > nk−1−δ , with n denoting the number of variables in a given formula, and ε, δ > 0 arbitrary constants.
One ingredient of the proof of Theorem 1.9 that may be of independent interest is establishing that k-CSP does not admit a non-trivial approximation on formulas with nk−η clauses, where η > 0 is an arbitrary small positive constant. We also consider variable percolation. Given an instance Φ of CSP we consider a random v formula Φp whose set of variables is a subset S of the variables of Φ, where each variable of Φ is v in S independently with probability p ∈ (0, 1) and the clauses of Φp are all clauses of Φ induced by S. In other words, a clause C of Φ survives if and only if all variables of C are in S. Using ideas similar to those used for clause percolation we show that p-vertex percolated instance are 1 essentially as hard as in the worst case for p > n1−δ for any δ ∈ (0, 1).
Other Problems For the Minimum Vertex Cover problem, we prove that for any α and δ > 0 an algorithm that gives α approximation for percolated instances implies also a α − δ approximation algorithm for worst-case instances. Our results hold for both edge and vertex percolation, where the edges or 1 the vertices of a given graph remain with probability p > n1−ε for some ε ∈ (0, 1). In particular, assuming the Unique Games Conjecture, the result of [24] implies that 2 − δ approximation for the Minimum Vertex Cover problem is NP-hard under a noise-robust reduction for any constant δ > 0. For the Hamiltonicity problem, we prove hardness results for percolated instances of both directed and undirected graphs with respect to edge percolation. We show that the problem where one needs to determine whether a graph contains a Hamiltonian cycle is also hard for percolated 1 graphs, where each edge of a given graph is kept in the graph with probability p > n1−ε for any ε ∈ (0, 1). We additionally study percolation on instances of the Subset-Sum problem, where each item of the set is deleted with probability 1 − p. We show that the problem remains hard as long as p = Ω( 1 ), where ε ∈ (0, 1/2) is an arbitrary constant, and n is the number of items in the n1/2−ε given instance.
1.4 Related Work There is a wide body of work on random discrete structures that has produced a wide range of mathematical tools [5, 17, 19, 26]. Randomly subsampling subgraphs by including each edge independently in the sample with probability p has been studied extensively in works concerned with cuts and flows (e.g., [22]). More recently, sampling subgraphs has been used to find independent sets [15]. The effect of subsampling variables in mathematical relaxations of constraint satisfaction problems on the value of these relaxations was studied in [4]. Edge-percolated graphs have been also used to construct hard instances for specific algorithms. For example, Kuˇcera[25] proved that the well-known greedy coloring algorithm performs poorly on the complete r-partite graph in which every edge is removed independently with probability 1/2
7 and r = nε for ε > 0. Namely, for this graph G, even if vertices are considered in a random order by n the greedy algorithm, with high probability Ω( log n ) colors are used to color the percolated graph whereas χ(G) ≤ nε. Misra [27] studied edge percolated instances of the Max-Cut problem. He proves that assuming N P= 6 BPP there is no polynomial time algorithm that computes the size of the maximum cut 1+ε in Gp,e for any p > d−1 in graphs of maximal degree d. This result is tight in the sense that it 1−ε is easy to show that once p < d−1 , with high probability Gp,e breaks into connected components of logarithmic size, and hence can be solved optimally in polynomial time. The techniques used in [27] differ from ours and rely on a recent hardness result for counting independent sets in sparse graphs [32]. Bukh [7] has considered coloring edge-percolated graphs, and states the question of whether E[χ(G 1 )] = Ω(χ(G)/ log(χ(G))) as an “interesting problem.” Bukh observed that the chromatic 2 ,e number of G 1 ,e has the same distribution as the chromatic number of the complement of G 1 ,e, 2 p 2 and therefore E[χ(G 1 )] ≥ χ(G). However, it is not clear how to leverage the lower bound on 2 ,e the expectation to obtain a lower bound on χ(G 1 ) with high probability, which is required for 2 ,e our noise-robust reductions. For instance, standard martingale methods do not seem to imply that p χ(G 1 ) ≥ Ω( χ(G)) holds with high probability. 2 ,e
1.5 Preliminaries An independent set in a graph G = (V,E) is a set of vertices that spans no edges. The independence number α(G) denotes the maximum size of an independent set in G.A legal coloring of a graph G is an assignment of colors to vertices such that no two adjacent vertices share the same color. The chromatic number χ(G) denotes the minimum number of colors needed for a legal coloring of G. Note that in a legal coloring of G each color class forms an independent set, and hence χ(G) · α(G) ≥ n. A vertex cover in a graph G = (V,E) is a set of vertices S ⊆ V such that every edge e ∈ E is incident to at least one vertex in S. Note that a subset of the vertices S ⊆ V is an independent set in G if and only if V \ S is a vertex cover. In particular, G contains a vertex cover of size k if and only if it contains an independent set of size n − k. We will always measure the running time of algorithms in terms of the size of the percolated instance. Since G and Gp,e have the same number of vertices, this generally does not affect the size of the instance by more than a polynomial factor. On the other hand, Gp,v may be much smaller than G for very small values of p. However, in this work we will be only dealing with the case where p = 1 , hence with high probability the size of the vertex percolated and original graphs n1−Ω(1) are polynomially related as well. We will use the following version of the Chernoff bound.
Lemma 1.11 (Chernoff bound, Theorem 7.3.2 in [20]). Let x1, . . . , xn be independent Bernoulli Pn 2 trials with Pr[xi = 1] = p, and let µ = E[ i=1 xi] = pn. Let r ≥ e . Then Pn Pr[ i=1 xi > (1 + r)µ] < exp(−(µr/2) ln r).
8 2 Maximum Independent Set and Percolation
In this section we demonstrate the hardness of approximating α(G) and χ(G) in both edge perco- lated and vertex percolated graphs. We base our results on a theorem of Feige and Kilian, saying that for every fixed ε > 0 the problem Coloring-vs-MIS(nε, nε) is NP-hard. Theorem 2.1 ([14]). For every ε > 0 it is NP-hard to decide whether a given n-vertex graph G satisfies χ(G) ≤ nε or α(G) ≤ nε.
Edge percolation Below we prove Theorem 1.4. We will use the following lemma, due to Turan (see, e.g. [2]). Lemma 2.2. Every graph H with l vertices and e edges contains an independent set of size at least l2 2e+l . As a corollary we observe that if a graph contains no large independent sets, then it can also cannot contain large subsets of the vertices that span a small number of edges. Corollary 2.3. Let G = (V,E) be an n-vertex graph satisfying α(G) ≤ k. Then every set of vertices of size l ≥ k spans at least l(l − k)/2k edges. Proof. Let H be a subgraph of G induced by l vertices, and suppose that H spans e edges. Then, l2 l2 by Lemma 2.2 we have α(H) ≥ 2e+l . On the other hand, α(H) ≤ α(G) ≤ k, and hence 2e+l ≤ k, as required.
We are now ready to prove Theorem 1.4 saying that for any n-vertex graph G = (V,E) it holds α(G) that with high probability α(Gp,e) ≤ O p log(np) .
Proof of Theorem 1.4. For a given graph G, let k = α(G) + 1 and let C > 0 be a large enough α(G) l(l−k) constant. By Corollary 2.3, every subset of size l = C p log(np) spans at least 2k edges in G. Hence, by taking union bound over all subsets of size l, the probability there exists a set of size l in Gp,e that spans no edge is at most
l n l(l−k) en l(l − k) −Ω(l) · (1 − p) 2k < · exp −p · < (np) , l l 2k