Quick viewing(Text Mode)

Anders Skjäl: Complexity Theory for the Global Optimizer I a Method Classification by Neumaier (2004)

OSE SEMINAR 2013 Complexity theory for the global optimizer

Anders Skjäl CENTER OF EXCELLENCE IN OPTIMIZATION AND SYSTEMS ENGINEERING ÅBO AKADEMI UNIVERSITY

ÅBO, NOVEMBER 15 2013 I Many of the converge asymptotically to a global optimum with probability 1 . No way to know when it happens I A method classification by Neumaier (2004) . Incomplete: clever heuristics . Asymptotically complete: converges eventually . Complete: converges and knows when prescribed tolerance is reached . Rigorous: converges despite rounding errors (floating point arithmetic) I “Deterministic global optimization”

Introduction 2 26 | What is Global Optimization? Several interpretations: I Many trials increase the chance of catching the global optimum . Multistart, scatter search . : genetic algorithms, simulated annealing, particle swarm, ant colony...

Anders Skjäl: Complexity theory for the global optimizer I A method classification by Neumaier (2004) . Incomplete: clever heuristics . Asymptotically complete: converges eventually . Complete: converges and knows when prescribed tolerance is reached . Rigorous: converges despite rounding errors (floating point arithmetic) I “Deterministic global optimization”

Introduction 2 26 | What is Global Optimization? Several interpretations: I Many trials increase the chance of catching the global optimum . Multistart, scatter search . Metaheuristics: genetic algorithms, simulated annealing, particle swarm, ant colony... I Many of the algorithms converge asymptotically to a global optimum with probability 1 . No way to know when it happens

Anders Skjäl: Complexity theory for the global optimizer Introduction 2 26 | What is Global Optimization? Several interpretations: I Many trials increase the chance of catching the global optimum . Multistart, scatter search . Metaheuristics: genetic algorithms, simulated annealing, particle swarm, ant colony... I Many of the algorithms converge asymptotically to a global optimum with probability 1 . No way to know when it happens I A method classification by Neumaier (2004) . Incomplete: clever heuristics . Asymptotically complete: converges eventually . Complete: converges and knows when prescribed tolerance is reached . Rigorous: converges despite rounding errors (floating point arithmetic) I “Deterministic global optimization”

Anders Skjäl: Complexity theory for the global optimizer I Results are sometimes negative . n n chess/checkers/go cannot be solved in polynomial time × . Conjecture A implies that problem B is C-hard I The terminology is well-suited for studying global optimization (in the strict sense)

I will discuss some concepts in complexity theory and their implications for our field I The NP class, approximation complexity, randomized algorithms

Introduction 3 26 | What is Complexity Theory?

I Aims to establish absolute limits for how fast a problem can be solved I -independent

Anders Skjäl: Complexity theory for the global optimizer I The terminology is well-suited for studying global optimization (in the strict sense)

I will discuss some concepts in complexity theory and their implications for our field I The NP class, approximation complexity, randomized algorithms

Introduction 3 26 | What is Complexity Theory?

I Aims to establish absolute limits for how fast a problem can be solved I Algorithm-independent I Results are sometimes negative . n n chess/checkers/go cannot be solved in polynomial time × . Conjecture A implies that problem B is C-hard

Anders Skjäl: Complexity theory for the global optimizer Introduction 3 26 | What is Complexity Theory?

I Aims to establish absolute limits for how fast a problem can be solved I Algorithm-independent I Results are sometimes negative . n n chess/checkers/go cannot be solved in polynomial time × . Conjecture A implies that problem B is C-hard I The terminology is well-suited for studying global optimization (in the strict sense)

I will discuss some concepts in complexity theory and their implications for our field I The NP class, approximation complexity, randomized algorithms

Anders Skjäl: Complexity theory for the global optimizer I Also Ò (asymptotically from below) and Ê (asymptotically from above and below) I Sorting algorithms . Naïve sort: O(n2) operations . Quicksort (1960): O(n logn) operations on average . Merge sort (1945): O(n logn) operations in worst case

Algorithm Runtimes 4 26 | Algorithm Analysis

I First contact with runtimes I Big O notation . f(n) = O(g(n)) if f is asymptotically bounded from above by a constant times g

Anders Skjäl: Complexity theory for the global optimizer I Sorting algorithms . Naïve sort: O(n2) operations . Quicksort (1960): O(n logn) operations on average . Merge sort (1945): O(n logn) operations in worst case

Algorithm Runtimes 4 26 | Algorithm Analysis

I First contact with runtimes I Big O notation . f(n) = O(g(n)) if f is asymptotically bounded from above by a constant times g I Also Ò (asymptotically from below) and Ê (asymptotically from above and below)

Anders Skjäl: Complexity theory for the global optimizer Algorithm Runtimes 4 26 | Algorithm Analysis

I First contact with runtimes I Big O notation . f(n) = O(g(n)) if f is asymptotically bounded from above by a constant times g I Also Ò (asymptotically from below) and Ê (asymptotically from above and below) I Sorting algorithms . Naïve sort: O(n2) operations . Quicksort (1960): O(n logn) operations on average . Merge sort (1945): O(n logn) operations in worst case

Anders Skjäl: Complexity theory for the global optimizer I ...or not . Coppersmith-Winograd algorithm (1987), O(n2.38) more efficient only for enormous matrices

Algorithm Runtimes 5 26 | Improved Runtimes

I Asymptotic improvements may have practical significance . Fast Fourier transform, O(n logn) . Matrix multiplication – Trivial bounds: O(n3), Ò(n2) – Strassen algorithm (1969): O(n2.81)

Anders Skjäl: Complexity theory for the global optimizer Algorithm Runtimes 5 26 | Improved Runtimes

I Asymptotic improvements may have practical significance . Fast Fourier transform, O(n logn) . Matrix multiplication – Trivial bounds: O(n3), Ò(n2) – Strassen algorithm (1969): O(n2.81) I ...or not . Coppersmith-Winograd algorithm (1987), O(n2.38) more efficient only for enormous matrices

Anders Skjäl: Complexity theory for the global optimizer I Objections: . O(2n ) is ok for small n . O(20.000001n ) is better than O(n100) for moderate n (artificial) I Polynomial on what machine? . Any classical computer . Turing machines, random-access machines and other theoretical machines can simulate each other with only polynomial slow-down . Not true for quantum computers as far as we know

Algorithm Runtimes 6 26 | Tractability

Polynomial runtime is sometimes considered synonymous with “reasonable algorithm” I Define P . the class of decision problems with a O(nk ) algorithm

Anders Skjäl: Complexity theory for the global optimizer . O(20.000001n ) is better than O(n100) for moderate n (artificial) I Polynomial on what machine? . Any classical computer . Turing machines, random-access machines and other theoretical machines can simulate each other with only polynomial slow-down . Not true for quantum computers as far as we know

Algorithm Runtimes 6 26 | Tractability

Polynomial runtime is sometimes considered synonymous with “reasonable algorithm” I Define P . the class of decision problems with a O(nk ) algorithm I Objections: . O(2n ) is ok for small n

Anders Skjäl: Complexity theory for the global optimizer I Polynomial on what machine? . Any classical computer . Turing machines, random-access machines and other theoretical machines can simulate each other with only polynomial slow-down . Not true for quantum computers as far as we know

Algorithm Runtimes 6 26 | Tractability

Polynomial runtime is sometimes considered synonymous with “reasonable algorithm” I Define P . the class of decision problems with a O(nk ) algorithm I Objections: . O(2n ) is ok for small n . O(20.000001n ) is better than O(n100) for moderate n (artificial)

Anders Skjäl: Complexity theory for the global optimizer . Turing machines, random-access machines and other theoretical machines can simulate each other with only polynomial slow-down . Not true for quantum computers as far as we know

Algorithm Runtimes 6 26 | Tractability

Polynomial runtime is sometimes considered synonymous with “reasonable algorithm” I Define P . the class of decision problems with a O(nk ) algorithm I Objections: . O(2n ) is ok for small n . O(20.000001n ) is better than O(n100) for moderate n (artificial) I Polynomial on what machine? . Any classical computer

Anders Skjäl: Complexity theory for the global optimizer Algorithm Runtimes 6 26 | Tractability

Polynomial runtime is sometimes considered synonymous with “reasonable algorithm” I Define P . the class of decision problems with a O(nk ) algorithm I Objections: . O(2n ) is ok for small n . O(20.000001n ) is better than O(n100) for moderate n (artificial) I Polynomial on what machine? . Any classical computer . Turing machines, random-access machines and other theoretical machines can simulate each other with only polynomial slow-down . Not true for quantum computers as far as we know

Anders Skjäl: Complexity theory for the global optimizer I A satisfiability problem, 3-SAT I Naïve algorithm: O(2n ) I Best known: O(1.439n )

Hard Problems 7 26 | Example: Satisfiability

Does any truth assignment of the Boolean variables x,y,z satisfy the expression:

(x y z) (¯x y¯ z) (x y¯ z¯) (¯x y z¯) ∨ ∨ ∧ ∨ ∨ ∧ ∨ ∨ ∧ ∨ ∨

Anders Skjäl: Complexity theory for the global optimizer Hard Problems 7 26 | Example: Satisfiability

Does any truth assignment of the Boolean variables x,y,z satisfy the expression:

(x y z) (¯x y¯ z) (x y¯ z¯) (¯x y z¯) ∨ ∨ ∧ ∨ ∨ ∧ ∨ ∨ ∧ ∨ ∨ I A satisfiability problem, 3-SAT I Naïve algorithm: O(2n ) I Best known: O(1.439n )

Anders Skjäl: Complexity theory for the global optimizer . Formal definitions with a Turing machine (checking the proof) or a non-deterministic Turing machine (guessing the proof) . P NP ⊂ I Examples . Decision versions of many optimization problems, discrete and continuous . Is minx X f(x) < M? If it is, then a point x0 X,f(x0) < M is a proof ∈ . Graph isomorphism,∈ traveling salesman, quadratic assignment, longest path, bin packing, knapsack, ... . Games like Battleships, Mastermind, Minesweeper, ...

Hard Problems 8 26 | “Easy to verify”

I Many difficult problems have this in common: . A ’yes’ answer can be verified quickly by checking a candidate (proof/certificate/witness) I Define the class NP of problems for which any ’yes’ instance has a proof which can be checked in polynomial time

Anders Skjäl: Complexity theory for the global optimizer I Examples . Decision versions of many optimization problems, discrete and continuous . Is minx X f(x) < M? If it is, then a point x0 X,f(x0) < M is a proof ∈ . Graph isomorphism,∈ traveling salesman, quadratic assignment, longest path, bin packing, knapsack, ... . Games like Battleships, Mastermind, Minesweeper, ...

Hard Problems 8 26 | “Easy to verify”

I Many difficult problems have this in common: . A ’yes’ answer can be verified quickly by checking a candidate (proof/certificate/witness) I Define the class NP of problems for which any ’yes’ instance has a proof which can be checked in polynomial time . Formal definitions with a Turing machine (checking the proof) or a non-deterministic Turing machine (guessing the proof) . P NP ⊂

Anders Skjäl: Complexity theory for the global optimizer . Games like Battleships, Mastermind, Minesweeper, ...

Hard Problems 8 26 | “Easy to verify”

I Many difficult problems have this in common: . A ’yes’ answer can be verified quickly by checking a candidate (proof/certificate/witness) I Define the class NP of problems for which any ’yes’ instance has a proof which can be checked in polynomial time . Formal definitions with a Turing machine (checking the proof) or a non-deterministic Turing machine (guessing the proof) . P NP ⊂ I Examples . Decision versions of many optimization problems, discrete and continuous . Is minx X f(x) < M? If it is, then a point x0 X,f(x0) < M is a proof ∈ . Graph isomorphism,∈ traveling salesman, quadratic assignment, longest path, bin packing, knapsack, ...

Anders Skjäl: Complexity theory for the global optimizer Hard Problems 8 26 | “Easy to verify”

I Many difficult problems have this in common: . A ’yes’ answer can be verified quickly by checking a candidate (proof/certificate/witness) I Define the class NP of problems for which any ’yes’ instance has a proof which can be checked in polynomial time . Formal definitions with a Turing machine (checking the proof) or a non-deterministic Turing machine (guessing the proof) . P NP ⊂ I Examples . Decision versions of many optimization problems, discrete and continuous . Is minx X f(x) < M? If it is, then a point x0 X,f(x0) < M is a proof ∈ . Graph isomorphism,∈ traveling salesman, quadratic assignment, longest path, bin packing, knapsack, ... . Games like Battleships, Mastermind, Minesweeper, ...

Anders Skjäl: Complexity theory for the global optimizer I Definition: C belongs to the class of NP-hard problems if . every problem in NP can be reduced to C in polynomial time I Definition: C belongs to the class of NP-complete problems if . C is NP-hard, and . C NP ∈

Hard Problems 9 26 | Complete Problems

Some problems in NP are as hard as any other NP problem I If one of these problems can be solved in polynomial time, then so can any NP problem I Cook and Levin proved that any NP problem can be reduced to (reformulated as) a satisfiability problem I Karp (1972) listed 21 problems with the property

Anders Skjäl: Complexity theory for the global optimizer Hard Problems 9 26 | Complete Problems

Some problems in NP are as hard as any other NP problem I If one of these problems can be solved in polynomial time, then so can any NP problem I Cook and Levin proved that any NP problem can be reduced to (reformulated as) a satisfiability problem I Karp (1972) listed 21 problems with the property I Definition: C belongs to the class of NP-hard problems if . every problem in NP can be reduced to C in polynomial time I Definition: C belongs to the class of NP-complete problems if . C is NP-hard, and . C NP ∈

Anders Skjäl: Complexity theory for the global optimizer I Restriction: 0-1 Linear Programming is NP-complete MINLP is NP-hard ⇒ I Local replacement: Satisfiability to 3-SAT . A local replacement of long clauses by clauses with three literals

x1 x2 x3 x4 x5 ∨ ∨ ∨ ∨ ↓ (x1 x2 y1) (y ¯1 x3 y2) (y ¯2 x4 x5) ∨ ∨ ∧ ∨ ∨ ∧ ∨ ∨

Hard Problems 10 26 | Proving NP-hardness

If B is NP-complete and B reduces to C (in polynomial time), then C is NP-hard I Three common techniques: restriction, local replacement, component design

Anders Skjäl: Complexity theory for the global optimizer I Local replacement: Satisfiability to 3-SAT . A local replacement of long clauses by clauses with three literals

x1 x2 x3 x4 x5 ∨ ∨ ∨ ∨ ↓ (x1 x2 y1) (y ¯1 x3 y2) (y ¯2 x4 x5) ∨ ∨ ∧ ∨ ∨ ∧ ∨ ∨

Hard Problems 10 26 | Proving NP-hardness

If B is NP-complete and B reduces to C (in polynomial time), then C is NP-hard I Three common techniques: restriction, local replacement, component design I Restriction: 0-1 Linear Programming is NP-complete MINLP is NP-hard ⇒

Anders Skjäl: Complexity theory for the global optimizer Hard Problems 10 26 | Proving NP-hardness

If B is NP-complete and B reduces to C (in polynomial time), then C is NP-hard I Three common techniques: restriction, local replacement, component design I Restriction: 0-1 Linear Programming is NP-complete MINLP is NP-hard ⇒ I Local replacement: Satisfiability to 3-SAT . A local replacement of long clauses by clauses with three literals

x1 x2 x3 x4 x5 ∨ ∨ ∨ ∨ ↓ (x1 x2 y1) (y ¯1 x3 y2) (y ¯2 x4 x5) ∨ ∨ ∧ ∨ ∨ ∧ ∨ ∨

Anders Skjäl: Complexity theory for the global optimizer A recreational example: Lemmings is NP-complete (Cormode 2004) 8

Figure 5: Lemmings level encoding the formula (v ¯ v v¯ ) (v ¯ v v ) (v v¯ v¯ ) (v ¯ v¯ v¯ ) 1 ∨ 2 ∨ 3 ∧ 2 ∨ 3 ∨ 4 ∧ 1 ∨ 2 ∨ 4 ∧ 1 ∨ 3 ∨ 4 I Why didn’t we find one yet? I Often stated as “P versus NP”, one of the Millennium Prize problems

© Behnam Esfahbod / CC-BY-SA-3.0

Hard Problems 12 26 | Is P , NP? I Maybe all problems in NP have an efficient algorithm?

Anders Skjäl: Complexity theory for the global optimizer I Often stated as “P versus NP”, one of the Millennium Prize problems

© Behnam Esfahbod / CC-BY-SA-3.0

Hard Problems 12 26 | Is P , NP? I Maybe all problems in NP have an efficient algorithm? I Why didn’t we find one yet?

Anders Skjäl: Complexity theory for the global optimizer Hard Problems 12 26 | Is P , NP? I Maybe all problems in NP have an efficient algorithm? I Why didn’t we find one yet? I Often stated as “P versus NP”, one of the Millennium Prize problems

© Behnam Esfahbod / CC-BY-SA-3.0

Anders Skjäl: Complexity theory for the global optimizer I The optimization version is at least as hard as the decision version I In practice we rarely need the exact optimum I Is approximation any easier?

Optimization 13 26 | Decision Problems and Optimization

I Optimization problem: What is the minimum value of f(x),x X? ∈ I Related decision problem: Is there a solution x X with f(x) M? ∈ ≤

Anders Skjäl: Complexity theory for the global optimizer I Is approximation any easier?

Optimization 13 26 | Decision Problems and Optimization

I Optimization problem: What is the minimum value of f(x),x X? ∈ I Related decision problem: Is there a solution x X with f(x) M? ∈ ≤ I The optimization version is at least as hard as the decision version I In practice we rarely need the exact optimum

Anders Skjäl: Complexity theory for the global optimizer Optimization 13 26 | Decision Problems and Optimization

I Optimization problem: What is the minimum value of f(x),x X? ∈ I Related decision problem: Is there a solution x X with f(x) M? ∈ ≤ I The optimization version is at least as hard as the decision version I In practice we rarely need the exact optimum I Is approximation any easier?

Anders Skjäl: Complexity theory for the global optimizer ´ I The optimal number of bins N ∗ stuffi ´ ≥ d e I The first-fit solution NFF < 2 stuffi d e I NFF < 2N (an improved analysis gives NFF 1.7N + 2) ∗ ≤ ∗

Optimization 14 26 | Approximative Solutions A solution x to a minimization problem is ê-optimal if:

f(x) (1 + ê)f(x∗) ≤ I Example: Bin packing, first-fit

stuff5 stuff3 ... stuff2 stuff4 stuff1

Anders Skjäl: Complexity theory for the global optimizer I NFF < 2N (an improved analysis gives NFF 1.7N + 2) ∗ ≤ ∗

Optimization 14 26 | Approximative Solutions A solution x to a minimization problem is ê-optimal if:

f(x) (1 + ê)f(x∗) ≤ I Example: Bin packing, first-fit

stuff5 stuff3 ... stuff2 stuff4 stuff1 ´ I The optimal number of bins N ∗ stuffi ´ ≥ d e I The first-fit solution NFF < 2 stuffi d e

Anders Skjäl: Complexity theory for the global optimizer Optimization 14 26 | Approximative Solutions A solution x to a minimization problem is ê-optimal if:

f(x) (1 + ê)f(x∗) ≤ I Example: Bin packing, first-fit

stuff5 stuff3 ... stuff2 stuff4 stuff1 ´ I The optimal number of bins N ∗ stuffi ´ ≥ d e I The first-fit solution NFF < 2 stuffi d e I NFF < 2N (an improved analysis gives NFF 1.7N + 2) ∗ ≤ ∗

Anders Skjäl: Complexity theory for the global optimizer I A simple application shows that approximation of general Traveling Salesperson problems is NP-hard for any constant ê I Many hardness results followed on the PCP Theorem (Arora et al. 1990) . Probabilistically Checkable Proofs I A hierarchy of complexity classes emerges: APX PTAS FPTAS ⊃ ⊃ I The classes are not equal unless P = NP

Approximation 15 26 | Approximation Complexity

By reducing approximation problems to NP-complete decision problems they are shown to be hard I The gap technique . Objective range (0,a] [b,+ ) ⊂ ∪ ∞ . If it is NP-hard to decide if the minimum belongs to (0,a] . ... then approximation within ê = (b a)/a is NP-hard −

Anders Skjäl: Complexity theory for the global optimizer I A hierarchy of complexity classes emerges: APX PTAS FPTAS ⊃ ⊃ I The classes are not equal unless P = NP

Approximation 15 26 | Approximation Complexity

By reducing approximation problems to NP-complete decision problems they are shown to be hard I The gap technique . Objective range (0,a] [b,+ ) ⊂ ∪ ∞ . If it is NP-hard to decide if the minimum belongs to (0,a] . ... then approximation within ê = (b a)/a is NP-hard − I A simple application shows that approximation of general Traveling Salesperson problems is NP-hard for any constant ê I Many hardness results followed on the PCP Theorem (Arora et al. 1990) . Probabilistically Checkable Proofs

Anders Skjäl: Complexity theory for the global optimizer Approximation 15 26 | Approximation Complexity

By reducing approximation problems to NP-complete decision problems they are shown to be hard I The gap technique . Objective range (0,a] [b,+ ) ⊂ ∪ ∞ . If it is NP-hard to decide if the minimum belongs to (0,a] . ... then approximation within ê = (b a)/a is NP-hard − I A simple application shows that approximation of general Traveling Salesperson problems is NP-hard for any constant ê I Many hardness results followed on the PCP Theorem (Arora et al. 1990) . Probabilistically Checkable Proofs I A hierarchy of complexity classes emerges: APX PTAS FPTAS ⊃ ⊃ I The classes are not equal unless P = NP

Anders Skjäl: Complexity theory for the global optimizer I Metric Traveling Salesperson (symmetric distances, triangle inequality) 1 3 . Christofides’ algorithm, ê = 2 , O(n ) 1 . Approximation with ê < 219 is NP-hard (Papadimitriou and Vempala 2000) I < APX: Linear Integer Programming, general TSP, and Quadratic Assignment have no efficient approximation algorithms for any constant ê (unless P = NP)

Approximation 16 26 | APX - efficient approximation within constant ê

I Polynomial time approximation algorithms for some constant ê

Anders Skjäl: Complexity theory for the global optimizer I < APX: Linear Integer Programming, general TSP, and Quadratic Assignment have no efficient approximation algorithms for any constant ê (unless P = NP)

Approximation 16 26 | APX - efficient approximation within constant ê

I Polynomial time approximation algorithms for some constant ê I Metric Traveling Salesperson (symmetric distances, triangle inequality) 1 3 . Christofides’ algorithm, ê = 2 , O(n ) 1 . Approximation with ê < 219 is NP-hard (Papadimitriou and Vempala 2000)

Anders Skjäl: Complexity theory for the global optimizer Approximation 16 26 | APX - efficient approximation within constant ê

I Polynomial time approximation algorithms for some constant ê I Metric Traveling Salesperson (symmetric distances, triangle inequality) 1 3 . Christofides’ algorithm, ê = 2 , O(n ) 1 . Approximation with ê < 219 is NP-hard (Papadimitriou and Vempala 2000) I < APX: Linear Integer Programming, general TSP, and Quadratic Assignment have no efficient approximation algorithms for any constant ê (unless P = NP)

Anders Skjäl: Complexity theory for the global optimizer  d 1  I O nd+1(logn)(O(√d/ê)) − (Arora 1998)   I Two dimensions: O n3(logn)O(1/ê) I Grows exponentially with 1/ê

Approximation 17 26 | Polynomial Time Approximation Schemes

PTAS: problems with polynomial time approximation algorithms for any ê I Geometric Traveling Salesperson . Euclidean distances or lp ,p 1 norm ≥ . Dimension d

Anders Skjäl: Complexity theory for the global optimizer I Grows exponentially with 1/ê

Approximation 17 26 | Polynomial Time Approximation Schemes

PTAS: problems with polynomial time approximation algorithms for any ê I Geometric Traveling Salesperson . Euclidean distances or lp ,p 1 norm ≥ . Dimension d  d 1  I O nd+1(logn)(O(√d/ê)) − (Arora 1998)   I Two dimensions: O n3(logn)O(1/ê)

Anders Skjäl: Complexity theory for the global optimizer Approximation 17 26 | Polynomial Time Approximation Schemes

PTAS: problems with polynomial time approximation algorithms for any ê I Geometric Traveling Salesperson . Euclidean distances or lp ,p 1 norm ≥ . Dimension d  d 1  I O nd+1(logn)(O(√d/ê)) − (Arora 1998)   I Two dimensions: O n3(logn)O(1/ê) I Grows exponentially with 1/ê

Anders Skjäl: Complexity theory for the global optimizer I Knapsack Problem (see Vazirani 2001) . O(n3/ê) I . Compact polytope, t negative eigenvalues . L = complexity of solving a convex QP of the same size . (Vavasis 1992) l mt  O n(n + 1)/√ê L

. Fully polynomial if t is bounded

Approximation 18 26 | Fully Polynomial Time Approximation Schemes

FPTAS: problems with approximation schemes that are polynomial in both n and 1/ê

Anders Skjäl: Complexity theory for the global optimizer I Quadratic Programming . Compact polytope, t negative eigenvalues . L = complexity of solving a convex QP of the same size . (Vavasis 1992) l mt  O n(n + 1)/√ê L

. Fully polynomial if t is bounded

Approximation 18 26 | Fully Polynomial Time Approximation Schemes

FPTAS: problems with approximation schemes that are polynomial in both n and 1/ê I Knapsack Problem (see Vazirani 2001) . O(n3/ê)

Anders Skjäl: Complexity theory for the global optimizer Approximation 18 26 | Fully Polynomial Time Approximation Schemes

FPTAS: problems with approximation schemes that are polynomial in both n and 1/ê I Knapsack Problem (see Vazirani 2001) . O(n3/ê) I Quadratic Programming . Compact polytope, t negative eigenvalues . L = complexity of solving a convex QP of the same size . (Vavasis 1992) l mt  O n(n + 1)/√ê L

. Fully polynomial if t is bounded

Anders Skjäl: Complexity theory for the global optimizer I Continuous problems: “convex easy, nonconvex hard” . Polynomial-time interior-point methods for convex programming, Nesterov (1988) . Self-concordant barrier functions exist for all closed convex solids I Some exceptions p1 pn . Geometric programming: posynomials cx xn ,c > 0 1 ··· . Linear fractional programming: (pT x + Ó)/(qT x + Ô) . ...

Interlude 19 26 | So which Problems are Hard?

I Not always obvious for discrete problems I Plenty of references . Garey & Johnson: Computers and Intractability (1979) . Ausiello et al.: Complexity and Approximation (1999)

Anders Skjäl: Complexity theory for the global optimizer I Some exceptions p1 pn . Geometric programming: posynomials cx xn ,c > 0 1 ··· . Linear fractional programming: (pT x + Ó)/(qT x + Ô) . ...

Interlude 19 26 | So which Problems are Hard?

I Not always obvious for discrete problems I Plenty of references . Garey & Johnson: Computers and Intractability (1979) . Ausiello et al.: Complexity and Approximation (1999) I Continuous problems: “convex easy, nonconvex hard” . Polynomial-time interior-point methods for convex programming, Nesterov (1988) . Self-concordant barrier functions exist for all closed convex solids

Anders Skjäl: Complexity theory for the global optimizer Interlude 19 26 | So which Problems are Hard?

I Not always obvious for discrete problems I Plenty of references . Garey & Johnson: Computers and Intractability (1979) . Ausiello et al.: Complexity and Approximation (1999) I Continuous problems: “convex easy, nonconvex hard” . Polynomial-time interior-point methods for convex programming, Nesterov (1988) . Self-concordant barrier functions exist for all closed convex solids I Some exceptions p1 pn . Geometric programming: posynomials cx xn ,c > 0 1 ··· . Linear fractional programming: (pT x + Ó)/(qT x + Ô) . ...

Anders Skjäl: Complexity theory for the global optimizer I Genuine randomness versus pseudorandomness I ZPP - zero-error probabilistic polynomial-time . correct answers in polynomial time, but... . returns no answer with probability 1/2 ≤ I BPP - bounded-error probabilistic polynomial-time . wrong answer with probability 1/3 ≤ I RP - randomized polynomial-time . outputs ’no’ if the correct answer is ’no’ . outputs ’no’ if the correct answer is ’yes’ with probability 1/2 ≤

Randomized Algorithms 20 26 | Will Randomization Help?

Some tasks seem to benefit from using random numbers I Complexity theory provides frameworks for analysis

Anders Skjäl: Complexity theory for the global optimizer I ZPP - zero-error probabilistic polynomial-time . correct answers in polynomial time, but... . returns no answer with probability 1/2 ≤ I BPP - bounded-error probabilistic polynomial-time . wrong answer with probability 1/3 ≤ I RP - randomized polynomial-time . outputs ’no’ if the correct answer is ’no’ . outputs ’no’ if the correct answer is ’yes’ with probability 1/2 ≤

Randomized Algorithms 20 26 | Will Randomization Help?

Some tasks seem to benefit from using random numbers I Complexity theory provides frameworks for analysis I Genuine randomness versus pseudorandomness

Anders Skjäl: Complexity theory for the global optimizer . returns no answer with probability 1/2 ≤ I BPP - bounded-error probabilistic polynomial-time . wrong answer with probability 1/3 ≤ I RP - randomized polynomial-time . outputs ’no’ if the correct answer is ’no’ . outputs ’no’ if the correct answer is ’yes’ with probability 1/2 ≤

Randomized Algorithms 20 26 | Will Randomization Help?

Some tasks seem to benefit from using random numbers I Complexity theory provides frameworks for analysis I Genuine randomness versus pseudorandomness I ZPP - zero-error probabilistic polynomial-time . correct answers in polynomial time, but...

Anders Skjäl: Complexity theory for the global optimizer I BPP - bounded-error probabilistic polynomial-time . wrong answer with probability 1/3 ≤ I RP - randomized polynomial-time . outputs ’no’ if the correct answer is ’no’ . outputs ’no’ if the correct answer is ’yes’ with probability 1/2 ≤

Randomized Algorithms 20 26 | Will Randomization Help?

Some tasks seem to benefit from using random numbers I Complexity theory provides frameworks for analysis I Genuine randomness versus pseudorandomness I ZPP - zero-error probabilistic polynomial-time . correct answers in polynomial time, but... . returns no answer with probability 1/2 ≤

Anders Skjäl: Complexity theory for the global optimizer I RP - randomized polynomial-time . outputs ’no’ if the correct answer is ’no’ . outputs ’no’ if the correct answer is ’yes’ with probability 1/2 ≤

Randomized Algorithms 20 26 | Will Randomization Help?

Some tasks seem to benefit from using random numbers I Complexity theory provides frameworks for analysis I Genuine randomness versus pseudorandomness I ZPP - zero-error probabilistic polynomial-time . correct answers in polynomial time, but... . returns no answer with probability 1/2 ≤ I BPP - bounded-error probabilistic polynomial-time . wrong answer with probability 1/3 ≤

Anders Skjäl: Complexity theory for the global optimizer Randomized Algorithms 20 26 | Will Randomization Help?

Some tasks seem to benefit from using random numbers I Complexity theory provides frameworks for analysis I Genuine randomness versus pseudorandomness I ZPP - zero-error probabilistic polynomial-time . correct answers in polynomial time, but... . returns no answer with probability 1/2 ≤ I BPP - bounded-error probabilistic polynomial-time . wrong answer with probability 1/3 ≤ I RP - randomized polynomial-time . outputs ’no’ if the correct answer is ’no’ . outputs ’no’ if the correct answer is ’yes’ with probability 1/2 ≤

Anders Skjäl: Complexity theory for the global optimizer Randomized Algorithms 21 26 | Randomized Decision Classes

BPP

RP

ZPP

P

NP

Anders Skjäl: Complexity theory for the global optimizer I There is always an assignment satisfying 7 K clauses ≥ 8 I The fraction of such assignments is Ò(1/K) I Approximation within ê = 1/7 (r = 8/7 in CS texts) is in ZPP I Can be derandomized to give deterministic algorithm

Randomized Algorithms 22 26 | Example: MAX-3-SAT ^K (xk xk xk ) 1 ∨ 2 ∨ 3 k=1

I A clause (three different literals) is satisfied by a random assignment with probability

13 7 1 = − 2 8 I 7 The expected number of satisfied clauses is 8 K

Anders Skjäl: Complexity theory for the global optimizer I The fraction of such assignments is Ò(1/K) I Approximation within ê = 1/7 (r = 8/7 in CS texts) is in ZPP I Can be derandomized to give deterministic algorithm

Randomized Algorithms 22 26 | Example: MAX-3-SAT ^K (xk xk xk ) 1 ∨ 2 ∨ 3 k=1

I A clause (three different literals) is satisfied by a random assignment with probability

13 7 1 = − 2 8 I 7 The expected number of satisfied clauses is 8 K I There is always an assignment satisfying 7 K clauses ≥ 8

Anders Skjäl: Complexity theory for the global optimizer I Can be derandomized to give deterministic algorithm

Randomized Algorithms 22 26 | Example: MAX-3-SAT ^K (xk xk xk ) 1 ∨ 2 ∨ 3 k=1

I A clause (three different literals) is satisfied by a random assignment with probability

13 7 1 = − 2 8 I 7 The expected number of satisfied clauses is 8 K I There is always an assignment satisfying 7 K clauses ≥ 8 I The fraction of such assignments is Ò(1/K) I Approximation within ê = 1/7 (r = 8/7 in CS texts) is in ZPP

Anders Skjäl: Complexity theory for the global optimizer Randomized Algorithms 22 26 | Example: MAX-3-SAT ^K (xk xk xk ) 1 ∨ 2 ∨ 3 k=1

I A clause (three different literals) is satisfied by a random assignment with probability

13 7 1 = − 2 8 I 7 The expected number of satisfied clauses is 8 K I There is always an assignment satisfying 7 K clauses ≥ 8 I The fraction of such assignments is Ò(1/K) I Approximation within ê = 1/7 (r = 8/7 in CS texts) is in ZPP I Can be derandomized to give deterministic algorithm

Anders Skjäl: Complexity theory for the global optimizer I Miller-Rabin, O(D 2 logD loglogD) = O˜ (D 2) . Wrong answer for composite numbers with probability < 1/4 . Primality testing coRP ∈ I Elliptic Curve Primality Proving . Expected runtime O˜ (D 4) . Primality testing ZPP ∈ I Agrawal-Kayal-Saxena (2002), O˜ (D 6) . Deterministic . Primality testing P ∈

Randomized Algorithms 23 26 | Example: Primality Testing

Is n a prime number? I Let D = logn, the number of digits in n I Adleman-Pomerance-Rumely (Jacobi sums), O(D c loglogD ) . Deterministic, not polynomial-time

Anders Skjäl: Complexity theory for the global optimizer I Elliptic Curve Primality Proving . Expected runtime O˜ (D 4) . Primality testing ZPP ∈ I Agrawal-Kayal-Saxena (2002), O˜ (D 6) . Deterministic . Primality testing P ∈

Randomized Algorithms 23 26 | Example: Primality Testing

Is n a prime number? I Let D = logn, the number of digits in n I Adleman-Pomerance-Rumely (Jacobi sums), O(D c loglogD ) . Deterministic, not polynomial-time I Miller-Rabin, O(D 2 logD loglogD) = O˜ (D 2) . Wrong answer for composite numbers with probability < 1/4 . Primality testing coRP ∈

Anders Skjäl: Complexity theory for the global optimizer I Agrawal-Kayal-Saxena (2002), O˜ (D 6) . Deterministic . Primality testing P ∈

Randomized Algorithms 23 26 | Example: Primality Testing

Is n a prime number? I Let D = logn, the number of digits in n I Adleman-Pomerance-Rumely (Jacobi sums), O(D c loglogD ) . Deterministic, not polynomial-time I Miller-Rabin, O(D 2 logD loglogD) = O˜ (D 2) . Wrong answer for composite numbers with probability < 1/4 . Primality testing coRP ∈ I Elliptic Curve Primality Proving . Expected runtime O˜ (D 4) . Primality testing ZPP ∈

Anders Skjäl: Complexity theory for the global optimizer Randomized Algorithms 23 26 | Example: Primality Testing

Is n a prime number? I Let D = logn, the number of digits in n I Adleman-Pomerance-Rumely (Jacobi sums), O(D c loglogD ) . Deterministic, not polynomial-time I Miller-Rabin, O(D 2 logD loglogD) = O˜ (D 2) . Wrong answer for composite numbers with probability < 1/4 . Primality testing coRP ∈ I Elliptic Curve Primality Proving . Expected runtime O˜ (D 4) . Primality testing ZPP ∈ I Agrawal-Kayal-Saxena (2002), O˜ (D 6) . Deterministic . Primality testing P ∈

Anders Skjäl: Complexity theory for the global optimizer Randomized Algorithms 24 26 |

Primality testing was known to be in BPP, and now in P I It has been conjectured that P = BPP I Randomized algorithms might not be fundamentally stronger I But they may have advantages . Lower degree runtimes . Sometimes conceptually easier, faster to program I A probability of errors may be tolerable if it can be bounded . Example: “industrial strength primes”

Anders Skjäl: Complexity theory for the global optimizer 25 26 | References

G. Ausiello, P. Crescenzi, G. Gambosi, V. Kann, A. Marchetti-Spaccamala, and M. Protasi. Complexity and Approximation: Combinatorial Optimization Problems and Their Approximability Properties. Springer, 1999. G. Cormode. The hardness of the Lemmings game, or Oh no, more NP-completeness proofs. In Proceedings of Third International Conference on Fun with Algorithms, pages 65–76, 2004. Michael R. Garey and David S. Johnson. Computers and Intractability: A Guide to the Theory of NP-Completeness. W.H. Freeman and Company, 1979. Arnold Neumaier. Complete search in continuous global optimization and constraint satisfaction. Acta Numerica, 13:271–369, 5 2004. Vijay V. Vazirani. Approximation Algorithms. Springer, 2001.

Anders Skjäl: Complexity theory for the global optimizer Questions?

26 26 |

Thank you for listening!

Anders Skjäl: Complexity theory for the global optimizer 26 26 |

Thank you for listening!

Questions?

Anders Skjäl: Complexity theory for the global optimizer