* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Download Document
Steinitz's theorem wikipedia , lookup
History of geometry wikipedia , lookup
Topological quantum field theory wikipedia , lookup
System of polynomial equations wikipedia , lookup
Shapley–Folkman lemma wikipedia , lookup
Scale invariance wikipedia , lookup
Event symmetry wikipedia , lookup
Classical mathematics and new challenges Theorems and Algorithms László Lovász Microsoft Research One Microsoft Way, Redmond, WA 98052 [email protected] Algorithmic vs. structural mathematics ancient and classical algorithms Geometric constructions Euclidean algorithm Newton’s method Gaussian elimination An example: diophantine approximation and continued fractions Given , find rational approximation p / q such that | p / q | / q and q 1/ . m n | m n p || q p | q 1/ a0 a0 ? 1 1 a1 a0 a0 1 a0 a1 continued fraction expansion A mini-history of algorithms 30’s: Mathematical notion of algorithms recursive functions, Λ-calculus, Turing-machines Church, Turing, Post algorithmic and logical undecidability Church, Gödel 50’s, 60’s: Computers the significance of running time simple and complex problems sorting searching arithmetic … Travelling Salesman matching network flows factoring … late 60’s-80’s: Complexity theory Time, space, information complexity Nondeterminism, good characteriztion, completeness Polynomial hierarchy Classification of many real-life problems into P vs. NP-complete Randomization, parallelism P=NP? 90’s: Increasing sophistication upper and lower bounds on complexity algorithms negative results factoring volume computation semidefinite optimization topology algebraic geometry coding theory Higlights of the 90’s: Approximation algorithms positive and negative results Probabilistic algorithms Markov chains, high concentration, nibble methods, phase transitions Pseudorandom number generators from art to science: theory and constructions Approximation algorithms: The Max Cut Problem maximize NP-hard …Approximations? Easy with 50% error Erdős ~’65 ??? NP-hard with 6% error Arora-Lund-MotwaniSudan-Szegedy ’92 Hastad (Interactive proof systems, PCP) Polynomial with 12% error Goemans-Williamson ’93 (semidefinite optimization) Algorithms and probability Randomized algorithms (making coin flips): important applications (primality testing, integration, optimization, volume computation, simulation) difficult to analyze Algorithms with stochastic input: even more important applications even more difficult to analyze Difficulty: after a few iterations, complicated functions of the original random variables arise. New methods in probability: Strong concentration (Talagrand) Laws of Large Numbers: sums of independent random variables is strongly concentrated General strong concentration: very general “smooth” functions of independent random variables are strongly concentrated Nibble, martingales, rapidly mixing Markov chains,… Example O(q)? 3 a , a , a ,. .. G F ( q ) Want: Few vectors 1 2 3 such that: - any 3 linearly independent - every vector is a linear combination of 2 (was open for 30 years) Every finite projective plane of order q has a complete arc of size q polylog(q). Kim-Vu First idea: use algebraic construction (conics,…) gives only about q Second idea: choose a1 , a2 , a3 ,... at random ????? Solution: Rödl nibble + strong concentration results Driving forces for the next decade New areas of applications The study of very large structures More tools from classical areas in mathematics New areas of application: interaction between discrete and continuous Biology: genetic code population dynamics protein folding Physics: elementary particles, quarks, etc. (Feynman graphs) statistical mechanics (graph theory, discrete probability) Economics: indivisibilities (integer programming, game theory) Computing: algorithms, complexity, databases, networks, VLSI, ... Very large structures -internet -VLSI -databases -genetic code -brain -animal -ecosystem -economy -society How to model them? non-constant but stable partly random Very large structures: how to model them? Graph minors Robertson, Seymour, Thomas If a graph does not contain a given minor, then it is essentially up to a bounded number of additional nodes a 1-dimensional structure of essentially 2-dimensional pieces. embedable in a fixed surface tree-decomposition except for “fringes” of bounded depth Very large structures: how to model them? Regularity Lemma Szeméredi 74 The nodes of graph can be partitioned into a bounded number of essentially equal parts given >0 and k>1, # of parts is between k and f(k, ) so that difference at most 1 almost all bipartite graphs between 2 parts are essentially random (with different densities). with k2 exceptions for subsets X,Y of the two parts, # of edges between X and Y is p|X||Y| n2 Very large structures -internet -VLSI -databases -genetic code How to model them? How to handle them algorithmically? -brain heuristics/approximation algorithms -animal linear time algorithms -ecosystem sublinear time algorithms (sampling) -economy -society A complexity theory of linear time? More and more tools from classical math Example: Volume computation Given: K n , convex Want: volume of K by a membership oracle; B(0,1) K B(0, n2 ) with relative error ε in n Not possible in polynomial time, even if ε=ncn. Elekes, Bárány, Füredi Possible in randomized polynomial time, for arbitrarily small ε. Dyer, Frieze, Kannan Complexity: For self-reducible problems, counting sampling (Jerrum-Valiant-Vazirani) Enough to sample from convex bodies vol ( K ) | K S | vol ( B) |S| * * * * * * ** * must be exponential in n Complexity: For self-reducible problems, counting sampling (Jerrum-Valiant-Vazirani) K0 B Enough to sample from convex bodies vol( K 0 ) vol( K1 ) by sampling vol( K1 ) vol( K 2 ) by sampling … K1 B 21/ n K K2 B 22/ n K Complexity: For self-reducible problems, counting sampling (Jerrum-Valiant-Vazirani) Enough to sample from convex bodies Algorithmic results: Use rapidly mixing Markov chains (Broder; Jerrum-Sinclair) Enough to estimate the mixing rate of random walk on lattice in K Complexity: For self-reducible problems, counting sampling (Jerrum-Valiant-Vazirani) Enough to sample from convex bodies Algorithmic results: Use rapidly mixing Markov chains (Broder; Jerrum-Sinclair) Enough to estimate the mixing rate of random walk on lattice in K Probability: use eigenvalue gap Graph theory (expanders): use conductance to estimate eigenvalue gap Alon, Jerrum-Sinclair K” K’ F voln 1 ( F ) vol( K ) vol( K ') vol( K ") Complexity: For self-reducible problems, counting sampling (Jerrum-Valiant-Vazirani) Enough to sample from convex bodies Algorithmic results: Use rapidly mixing Markov chains (Broder; Jerrum-Sinclair) Enough to estimate the mixing rate of random walk on lattice in K Probability: use eigenvalue gap Graph theory (expanders): use conductance to estimate eigenvalue gap Alon, Jerrum-Sinclair Enough to prove isoperimetric inequality for subsets of K Differential geometry: Isoperimetric inequality Dyer Frieze Kannan 1989 O* (n27 ) Differential equations: bounds on Poincaré constant Paine-Weinberger bisection method, improved isoperimetric inequality LL-Simonovits 1990 O* (n16 ) Log-concave functions: reduction to integration Applegate-Kannan 1992 O* (n10 ) Convex geometry: Ball walk LL 1992 O* (n10 ) Statistics: Optimization: Functional analysis: isotropic position of convex bodies Better error handling Dyer-Frieze 1993 O* (n8 ) Better prepocessing LL-Simonovits 1995 O* (n7 ) O* (n5 ) achieving isotropic position Kannan-LL-Simonovits 1998 * Geometry: projective (Hilbert) distance O affine invariant isoperimetric inequality analysis of hit-and-run walk LL 1999 Differential equations: log-Sobolev inequality elimination of “start penalty” for lattice walk Frieze-Kannan 1999 log-Cheeger inequality elimination of “start penalty” for ball walk Kannan-LL 1999 Scientific computing: non-reversible chains mix better; lifting Diaconis-Holmes-Neal Feng-LL-Pak walk with inertia Aspnes-Kannan-LL ( n5 ) O* (n5 ) O* (n3 )?? More and more tools from classical math Linear algebra : eigenvalues semidefinite optimization higher incidence matrices homology theory Geometry : geometric representations convexity Analysis: generating functions Fourier analysis, quantum computing Number theory: cryptography Topology, group theory, algebraic geometry, special functions, differential equations,…