CS Cornell 40Th Anniversary Booklet

Total Page:16

File Type:pdf, Size:1020Kb

CS Cornell 40Th Anniversary Booklet Contents Welcome from the CS Chair .................................................................................3 Symposium program .............................................................................................4 Meet the speakers ..................................................................................................5 The Cornell environment The Cornell CS ambience ..............................................................................6 Faculty of Computing and Information Science ............................................8 Information Science Program ........................................................................9 College of Engineering ................................................................................10 College of Arts & Sciences ..........................................................................11 Selected articles Mission-critical distributed systems ............................................................ 12 Language-based security ............................................................................. 14 A grand challenge in computer networking .................................................16 Data mining, today and tomorrow ...............................................................18 Grid computing and Web services ...............................................................20 The science of networks .............................................................................. 22 Search engines that learn from experience ..................................................24 Sentiment analysis ....................................................................................... 26 Beauty is skin deep ......................................................................................28 Structural fi ngerprints of molecular evolution .............................................30 The impact of computing on medicine ........................................................32 Reasoning about knowledge ........................................................................34 Automated reasoning: The impossible made indispensable ........................36 Computational complexity ...........................................................................38 Sidelights The TRUST consortium ...............................................................................13 Programming methodology and program correctness .................................15 Computing in the arts ...................................................................................17 The marriage of structured and unstructured data .......................................19 The Cornell Theory Center ..........................................................................21 Algorithms at Cornell .................................................................................. 23 The father of information retrieval .............................................................. 25 CURIE comes to CS ....................................................................................27 Program of Computer Graphics ...................................................................29 Bridging the Rift ..........................................................................................31 The Game Design Initiative .........................................................................33 Forty years of numerical analysis and scientifi c computing ........................35 The scholarly publishing revolution ............................................................37 Computational complexity and the ice cube ................................................39 About the department Computer Science chairs ............................................................................. 40 Computer Science faculty ............................................................................40 Degrees granted ........................................................................................... 41 Selected faculty awards ............................................................................... 42 Grants to young faculty ............................................................................... 43 Books by the faculty ....................................................................................44 Selected invited addresses ........................................................................... 46 The 40th anniversary committee .........................................................................48 1 2 Welcome from the CS Chair There are many good departments of computer science but only a few great ones. What distinguishes the great from the merely good? I believe the difference comes down to three things. The fi rst hallmark of a great department is excellence in research and teaching. No Ten years ago, department can be a world leader in all areas of computer science, but a great department we realized that just as is a world leader in many of them. computer science The second hallmark of a great department is that it takes the lead in defi ning how the had revolutionized fi eld evolves and has the courage to invest heavily in that future. engineering in the 20th century, Finally, a great department must have character —it must possess some noteworthy characteristic, some idiosyncrasy, some texture that makes it unique among its peers. it would revolutionize all academic fi elds in the Looking through a draft of this brochure, I was reminded of what it is that makes the CS 21st century. department at Cornell one of the great CS departments in the world. For many years, we have had world-class research programs in areas such as algorithms, complexity theory, distributed systems, languages, and numerical analysis. These areas continue to fl ourish at Cornell, as you will see in these pages. Ten years ago, we realized that just as computer science had revolutionized engineering in the 20th century, it would revolutionize all academic fi elds in the 21st century. So, we worked with Cornell to create what is effectively a college of Computing and Information Sciences. In a very short time, this new structure has led to a university-wide explosion in interdisciplinary research organized around CS themes. Moreover, the timing of our This brochure presents the Cornell Department of Computer Science and the symposium that celebrates its 40th anni- decision to grow a fi rst-class group in AI and machine learning couldn’t have been better, versary. The information about the department is given in because this area is crucial to such interdisciplinary work. four ways: Finally, the unique texture of our department stems, I believe, from our collegiality. One • Description of the environment in which the Department of Computer Science operates, which has contributed in no of our enduring traditions is that the faculty gathers at noon every day to eat lunch while small way to our success. discussing the latest research results or barbecuing some unfortunate faculty candidate. A remarkable number of the collaborations between theoreticians and practitioners that you • Fourteen selected articles. Not all faculty are represented, will fi nd described in the pages of this brochure were sparked by lunchtime discussions at and only a portion of our research is described. Accom- panying these articles are sidelights that summarize other the Statler Club or a restaurant in Collegetown. aspects of the department. It is my privilege to chair this great department, and it is my pleasure to welcome you to • Information about the department: a list of faculty, books the symposium to celebrate the 40th anniversary of its founding. by the faculty, and so on. • A timeline of events, including when each tenure-track faculty member joined the department. 3 00 Welcome Robert L. Constable 9 Dean, CIS Introduced by Charles Van Loan Symposium Chair, Computer Science 15 When complexity was king program 9 and life thereafter Al Borodin Professor, Computer Science University of Toronto 45 From a bear to a lion Zvi Galil 9 Dean, Engineering & Applied Science Columbia University 10 15 Break 30 A cryptographer’s perspective on privacy-preserving 10 data mining and statistical disclosure control Cynthia Dwork Senior Researcher 40th Anniversary Microsoft Research Symposium 00 Certifying algorithms Kurt Mehlhorn 11 Director Max Planck Institut für Informatik, Saarbruecken Department of Vice President, Max Planck Society Computer Science, 30 Model checking: My 30-year quest to Cornell University 11 make verifi cation practical Ed Clarke FORE Systems Professor, Computer Science Carnegie Mellon University 1 October 2005 12 00 Lunch 30 Beyond mice and menus Barbara J. Grosz 1 Dean of Science, Radcliffe Inst. for Advanced Study Harvard University 00 Competition, cruelty, and compassion at Cornell 2 and the future of computer science Bobby Schnabel Vice Provost Colorado University 30 Gerry Salton’s information retrieval 2 reaches the masses Amit Singhal Distinguished Engineer Google 300 Break 30 Security in distributed systems: Where we are, how we 3 got there, and how Cornell is trying to save us Mike Reiter Professor, Computer Science Carnegie Mellon University 00 Databases aren’t dull 4 and other life lessons after Cornell Jennifer Widom Professor, Computer Science Stanford University 30 Evaluation + Design + 4 Implementation (Repeat) = Systems Randy Katz UMC Distinguished Professor, Computer Science Celebrating 40 years University of California Berkeley of leadership in 00 Banquet David
Recommended publications
  • Efficient Algorithms with Asymmetric Read and Write Costs
    Efficient Algorithms with Asymmetric Read and Write Costs Guy E. Blelloch1, Jeremy T. Fineman2, Phillip B. Gibbons1, Yan Gu1, and Julian Shun3 1 Carnegie Mellon University 2 Georgetown University 3 University of California, Berkeley Abstract In several emerging technologies for computer memory (main memory), the cost of reading is significantly cheaper than the cost of writing. Such asymmetry in memory costs poses a fun- damentally different model from the RAM for algorithm design. In this paper we study lower and upper bounds for various problems under such asymmetric read and write costs. We con- sider both the case in which all but O(1) memory has asymmetric cost, and the case of a small cache of symmetric memory. We model both cases using the (M, ω)-ARAM, in which there is a small (symmetric) memory of size M and a large unbounded (asymmetric) memory, both random access, and where reading from the large memory has unit cost, but writing has cost ω 1. For FFT and sorting networks we show a lower bound cost of Ω(ωn logωM n), which indicates that it is not possible to achieve asymptotic improvements with cheaper reads when ω is bounded by a polynomial in M. Moreover, there is an asymptotic gap (of min(ω, log n)/ log(ωM)) between the cost of sorting networks and comparison sorting in the model. This contrasts with the RAM, and most other models, in which the asymptotic costs are the same. We also show a lower bound for computations on an n × n diamond DAG of Ω(ωn2/M) cost, which indicates no asymptotic improvement is achievable with fast reads.
    [Show full text]
  • Tarjan Transcript Final with Timestamps
    A.M. Turing Award Oral History Interview with Robert (Bob) Endre Tarjan by Roy Levin San Mateo, California July 12, 2017 Levin: My name is Roy Levin. Today is July 12th, 2017, and I’m in San Mateo, California at the home of Robert Tarjan, where I’ll be interviewing him for the ACM Turing Award Winners project. Good afternoon, Bob, and thanks for spending the time to talk to me today. Tarjan: You’re welcome. Levin: I’d like to start by talking about your early technical interests and where they came from. When do you first recall being interested in what we might call technical things? Tarjan: Well, the first thing I would say in that direction is my mom took me to the public library in Pomona, where I grew up, which opened up a huge world to me. I started reading science fiction books and stories. Originally, I wanted to be the first person on Mars, that was what I was thinking, and I got interested in astronomy, started reading a lot of science stuff. I got to junior high school and I had an amazing math teacher. His name was Mr. Wall. I had him two years, in the eighth and ninth grade. He was teaching the New Math to us before there was such a thing as “New Math.” He taught us Peano’s axioms and things like that. It was a wonderful thing for a kid like me who was really excited about science and mathematics and so on. The other thing that happened was I discovered Scientific American in the public library and started reading Martin Gardner’s columns on mathematical games and was completely fascinated.
    [Show full text]
  • Benchmarks for IP Forwarding Tables
    Reviewers James Abello Richard Cleve Vassos Hadzilacos Dimitris Achilioptas James Clippinger Jim Hafner Micah Adler Anne Condon Torben Hagerup Oswin Aichholzer Stephen Cook Armin Haken William Aiello Tom Cormen Shai Halevi Donald Aingworth Dan Dooly Eric Hansen Susanne Albers Oliver Duschka Refael Hassin Eric Allender Martin Dyer Johan Hastad Rajeev Alur Ran El-Yaniv Lisa Hellerstein Andris Ambainis David Eppstein Monika Henzinger Amihood Amir Jeff Erickson Tom Henzinger Artur Andrzejak Kousha Etessami Jeremy Horwitz Boris Aronov Will Evans Russell Impagliazzo Sanjeev Arora Guy Even Piotr Indyk Amotz Barnoy Ron Fagin Sandra Irani Yair Bartal Michalis Faloutsos Ken Jackson Julien Basch Martin Farach-Colton David Johnson Saugata Basu Uri Feige John Jozwiak Bob Beals Joan Feigenbaum Bala Kalyandasundaram Paul Beame Stefan Felsner Ming-Yang Kao Steve Bellantoni Faith Fich Haim Kaplan Micahel Ben-Or Andy Fingerhut Bruce Kapron Josh Benaloh Paul Fischer Michael Kaufmann Charles Bennett Lance Fortnow Michael Kearns Marshall Bern Steve Fortune Sanjeev Khanna Nikolaj Bjorner Alan Frieze Samir Khuller Johannes Blomer Anna Gal Joe Kilian Avrim Blum Naveen Garg Valerie King Dan Boneh Bernd Gartner Philip Klein Andrei Broder Rosario Gennaro Spyros Kontogiannis Nader Bshouty Ashish Goel Gilad Koren Adam Buchsbaum Michel Goemans Dexter Kozen Lynn Burroughs Leslie Goldberg Dina Kravets Ran Canetti Paul Goldberg S. Ravi Kumar Pei Cao Oded Goldreich Eyal Kushilevitz Moses Charikar John Gray Stephen Kwek Chandra Chekuri Dan Greene Larry Larmore Yi-Jen Chiang
    [Show full text]
  • Variations on the Sensitivity Conjecture
    THEORY OF COMPUTING LIBRARY GRADUATE SURVEYS 4 (2011), pp. 1–27 www.theoryofcomputing.org Variations on the Sensitivity Conjecture Pooya Hatami Raghav Kulkarni Denis Pankratov Received: November 2, 2010; published: June 22, 2011. Abstract: The sensitivity of a Boolean function f of n Boolean variables is the maximum over all inputs x of the number of positions i such that flipping the i-th bit of x changes the value of f (x). Permitting to flip disjoint blocks of bits leads to the notion of block sensitivity, known to be polynomially related to a number of other complexity measures of f , including the decision-tree complexity, the polynomial degree, and the certificate complexity. A long-standing open question is whether sensitivity also belongs to this equivalence class. A positive answer to this question is known as the Sensitivity Conjecture. We present a selection of known as well as new variants of the Sensitivity Conjecture and point out some weaker versions that are also open. Among other things, we relate the problem to Communication Complexity via recent results by Sherstov (QIC 2010). We also indicate new connections to Fourier analysis. ACM Classification: F.0, F.1, F.2 AMS Classification: 68-02, 68Q05, 68Q25 Key words and phrases: sensitivity, block sensitivity, complexity measures of Boolean functions 1 The Sensitivity Conjecture For Boolean strings x;y 2 f0;1gn let x ⊕ y 2 f0;1gn denote the coordinate-wise exclusive or of x and y. n th Let ei 2 f0;1g denote an n-bit Boolean string whose i bit is 1 and the rest of the bits are 0: Let f : f0;1gn ! f0;1g be a Boolean function.
    [Show full text]
  • Four Results of Jon Kleinberg a Talk for St.Petersburg Mathematical Society
    Four Results of Jon Kleinberg A Talk for St.Petersburg Mathematical Society Yury Lifshits Steklov Institute of Mathematics at St.Petersburg May 2007 1 / 43 2 Hubs and Authorities 3 Nearest Neighbors: Faster Than Brute Force 4 Navigation in a Small World 5 Bursty Structure in Streams Outline 1 Nevanlinna Prize for Jon Kleinberg History of Nevanlinna Prize Who is Jon Kleinberg 2 / 43 3 Nearest Neighbors: Faster Than Brute Force 4 Navigation in a Small World 5 Bursty Structure in Streams Outline 1 Nevanlinna Prize for Jon Kleinberg History of Nevanlinna Prize Who is Jon Kleinberg 2 Hubs and Authorities 2 / 43 4 Navigation in a Small World 5 Bursty Structure in Streams Outline 1 Nevanlinna Prize for Jon Kleinberg History of Nevanlinna Prize Who is Jon Kleinberg 2 Hubs and Authorities 3 Nearest Neighbors: Faster Than Brute Force 2 / 43 5 Bursty Structure in Streams Outline 1 Nevanlinna Prize for Jon Kleinberg History of Nevanlinna Prize Who is Jon Kleinberg 2 Hubs and Authorities 3 Nearest Neighbors: Faster Than Brute Force 4 Navigation in a Small World 2 / 43 Outline 1 Nevanlinna Prize for Jon Kleinberg History of Nevanlinna Prize Who is Jon Kleinberg 2 Hubs and Authorities 3 Nearest Neighbors: Faster Than Brute Force 4 Navigation in a Small World 5 Bursty Structure in Streams 2 / 43 Part I History of Nevanlinna Prize Career of Jon Kleinberg 3 / 43 Nevanlinna Prize The Rolf Nevanlinna Prize is awarded once every 4 years at the International Congress of Mathematicians, for outstanding contributions in Mathematical Aspects of Information Sciences including: 1 All mathematical aspects of computer science, including complexity theory, logic of programming languages, analysis of algorithms, cryptography, computer vision, pattern recognition, information processing and modelling of intelligence.
    [Show full text]
  • Annual Report
    ANNUAL REPORT 2019FISCAL YEAR ACM, the Association for Computing Machinery, is an international scientific and educational organization dedicated to advancing the arts, sciences, and applications of information technology. Letter from the President It’s been quite an eventful year and challenges posed by evolving technology. for ACM. While this annual Education has always been at the foundation of exercise allows us a moment ACM, as reflected in two recent curriculum efforts. First, “ACM’s mission to celebrate some of the many the ACM Task Force on Data Science issued “Comput- hinges on successes and achievements ing Competencies for Undergraduate Data Science Cur- creating a the Association has realized ricula.” The guidelines lay out the computing-specific over the past year, it is also an competencies that should be included when other community that opportunity to focus on new academic departments offer programs in data science encompasses and innovative ways to ensure at the undergraduate level. Second, building on the all who work in ACM remains a vibrant global success of our recent guidelines for 4-year cybersecu- the computing resource for the computing community. rity curricula, the ACM Committee for Computing Edu- ACM’s mission hinges on creating a community cation in Community Colleges created a related cur- and technology that encompasses all who work in the computing and riculum targeted at two-year programs, “Cybersecurity arena” technology arena. This year, ACM established a new Di- Curricular Guidance for Associate-Degree Programs.” versity and Inclusion Council to identify ways to create The following pages offer a sampling of the many environments that are welcoming to new perspectives ACM events and accomplishments that occurred over and will attract an even broader membership from the past fiscal year, none of which would have been around the world.
    [Show full text]
  • Arxiv:2008.08417V2 [Cs.DS] 3 Nov 2020
    Modular Subset Sum, Dynamic Strings, and Zero-Sum Sets Jean Cardinal∗ John Iacono† Abstract a number of operations on a persistent collection strings, e modular subset sum problem consists of deciding, given notably, split and concatenate, as well as nding the longest a modulus m, a multiset S of n integers in 0::m − 1, and common prex (LCP) of two strings, all in at most loga- a target integer t, whether there exists a subset of S with rithmic time with updates being with high probability. We elements summing to t (mod m), and to report such a inform the reader of the independent work of Axiotis et + set if it exists. We give a simple O(m log m)-time with al., [ABB 20], contains much of the same results and ideas high probability (w.h.p.) algorithm for the modular sub- for modular subset sum as here has been accepted to ap- set sum problem. is builds on and improves on a pre- pear along with this paper in the proceedings of the SIAM vious O(m log7 m) w.h.p. algorithm from Axiotis, Backurs, Symposium on Simplicity in Algorithms (SOSA21). + Jin, Tzamos, and Wu (SODA 19). Our method utilizes the As the dynamic string data structure of [GKK 18] is ADT of the dynamic strings structure of Gawrychowski quite complex, we provide in §4 a new and far simpler alter- et al. (SODA 18). However, as this structure is rather com- native, which we call the Data Dependent Tree (DDT) struc- plicated we present a much simpler alternative which we ture.
    [Show full text]
  • Navigability of Small World Networks
    Navigability of Small World Networks Pierre Fraigniaud CNRS and University Paris Sud http://www.lri.fr/~pierre Introduction Interaction Networks • Communication networks – Internet – Ad hoc and sensor networks • Societal networks – The Web – P2P networks (the unstructured ones) • Social network – Acquaintance – Mail exchanges • Biology (Interactome network), linguistics, etc. Dec. 19, 2006 HiPC'06 3 Common statistical properties • Low density • “Small world” properties: – Average distance between two nodes is small, typically O(log n) – The probability p that two distinct neighbors u1 and u2 of a same node v are neighbors is large. p = clustering coefficient • “Scale free” properties: – Heavy tailed probability distributions (e.g., of the degrees) Dec. 19, 2006 HiPC'06 4 Gaussian vs. Heavy tail Example : human sizes Example : salaries µ Dec. 19, 2006 HiPC'06 5 Power law loglog ppk prob{prob{ X=kX=k }} ≈≈ kk-α loglog kk Dec. 19, 2006 HiPC'06 6 Random graphs vs. Interaction networks • Random graphs: prob{e exists} ≈ log(n)/n – low clustering coefficient – Gaussian distribution of the degrees • Interaction networks – High clustering coefficient – Heavy tailed distribution of the degrees Dec. 19, 2006 HiPC'06 7 New problematic • Why these networks share these properties? • What model for – Performance analysis of these networks – Algorithm design for these networks • Impact of the measures? • This lecture addresses navigability Dec. 19, 2006 HiPC'06 8 Navigability Milgram Experiment • Source person s (e.g., in Wichita) • Target person t (e.g., in Cambridge) – Name, professional occupation, city of living, etc. • Letter transmitted via a chain of individuals related on a personal basis • Result: “six degrees of separation” Dec.
    [Show full text]
  • The Best Nurturers in Computer Science Research
    The Best Nurturers in Computer Science Research Bharath Kumar M. Y. N. Srikant IISc-CSA-TR-2004-10 http://archive.csa.iisc.ernet.in/TR/2004/10/ Computer Science and Automation Indian Institute of Science, India October 2004 The Best Nurturers in Computer Science Research Bharath Kumar M.∗ Y. N. Srikant† Abstract The paper presents a heuristic for mining nurturers in temporally organized collaboration networks: people who facilitate the growth and success of the young ones. Specifically, this heuristic is applied to the computer science bibliographic data to find the best nurturers in computer science research. The measure of success is parameterized, and the paper demonstrates experiments and results with publication count and citations as success metrics. Rather than just the nurturer’s success, the heuristic captures the influence he has had in the indepen- dent success of the relatively young in the network. These results can hence be a useful resource to graduate students and post-doctoral can- didates. The heuristic is extended to accurately yield ranked nurturers inside a particular time period. Interestingly, there is a recognizable deviation between the rankings of the most successful researchers and the best nurturers, which although is obvious from a social perspective has not been statistically demonstrated. Keywords: Social Network Analysis, Bibliometrics, Temporal Data Mining. 1 Introduction Consider a student Arjun, who has finished his under-graduate degree in Computer Science, and is seeking a PhD degree followed by a successful career in Computer Science research. How does he choose his research advisor? He has the following options with him: 1. Look up the rankings of various universities [1], and apply to any “rea- sonably good” professor in any of the top universities.
    [Show full text]
  • A Short History of Computational Complexity
    The Computational Complexity Column by Lance FORTNOW NEC Laboratories America 4 Independence Way, Princeton, NJ 08540, USA [email protected] http://www.neci.nj.nec.com/homepages/fortnow/beatcs Every third year the Conference on Computational Complexity is held in Europe and this summer the University of Aarhus (Denmark) will host the meeting July 7-10. More details at the conference web page http://www.computationalcomplexity.org This month we present a historical view of computational complexity written by Steve Homer and myself. This is a preliminary version of a chapter to be included in an upcoming North-Holland Handbook of the History of Mathematical Logic edited by Dirk van Dalen, John Dawson and Aki Kanamori. A Short History of Computational Complexity Lance Fortnow1 Steve Homer2 NEC Research Institute Computer Science Department 4 Independence Way Boston University Princeton, NJ 08540 111 Cummington Street Boston, MA 02215 1 Introduction It all started with a machine. In 1936, Turing developed his theoretical com- putational model. He based his model on how he perceived mathematicians think. As digital computers were developed in the 40's and 50's, the Turing machine proved itself as the right theoretical model for computation. Quickly though we discovered that the basic Turing machine model fails to account for the amount of time or memory needed by a computer, a critical issue today but even more so in those early days of computing. The key idea to measure time and space as a function of the length of the input came in the early 1960's by Hartmanis and Stearns.
    [Show full text]
  • AN/N LOG N ALGORITHM for MINIMIZING STATES in Kf I N ITE AUTOMATON by JOHN HOPCROFT STAN-CS-71-190 January, 1971 COMPUTER SCIENC
    AN/N LOG N ALGORITHM FOR MINIMIZING STATES IN kF I N ITE AUTOMATON BY JOHN HOPCROFT STAN-CS-71-190 January, 1971 COMPUTER SCIENCE DEPARTMENT School of Humanities and Sciences STANFORD UN IVERS ITY AN N LOG N ALGORITHM FOR MINIMIZING STATES IN A FINITE AUTOMATON John Hopcroft Abstract An algorithm is given for minimizing the number of states in a finite automaton or for determining if two finite automata are equivalent. The asymptotic running time of the algorithm is bounded by knlogn where k is some constant and n is the number of states. The constant k depends linearly on the size of the input alphabet. This research was supported by the National Science Foundation under grant number NSF-GJ-96, and the Office of Naval Research under grant number N-00014-67-A-0112-0057 NR 044-402. Reproduction in whole or in part is permitted for any purpose of the United States Government. AN n log n ALGORITHM FOR MINIMIZING STATES IN A FINITE AUTOMATON John Hopcroft Stanford University Introduction Most basic texts on finite automata give algorithms for minimizing the number of states in a finite automaton [l, 21. However, a worst case analysis of these algorithms indicate that they are n2 processes where n is the number of states. For finite automata with large numbers of states, these algorithms are grossly inefficient. Thus in this paper we describe an algorithm for minimizing the states in which the asymptotic running time in a worst case analysis grows as n log n . The constant of proportionality depends linearly on the number of input symbols.
    [Show full text]
  • Download This PDF File
    News from New Zealand by C. S. Calude Department of Computer Science, University of Auckland Auckland, New Zealand [email protected] 1 Scientific and Community News The latest CDMTCS research reports are ( http://www.cs.auckland.ac.nz/ staff-cgi-bin/mjd/secondcgi.pl ): 403 U. Speidel. A Forward-Parsing Randomness Test Based on the Expected Codeword Length of T-codes 05 /2011 404. M.J. Dinneen, Y.-B. Kim and R. Nicolescu. An Adaptive Algorithm for P System Synchronization 05 /2011 405. A.A. Abbott, M. Bechmann, C.S. Calude, and A. Sebald. A Nuclear Mag- netic Resonance Implementation of a Classical Deutsch-Jozsa Algorithm 05 /2011 406. K. Tadaki. A Computational Complexity-Theoretic Elaboration of Weak Truth-Table Reducibility 07 /2011 29 BEATCS no 105 EATCS NEWS 2 A Dialogue with Juris Hartmanis about Com- plexity Professor Juris Hartmanis, http://en.wikipedia.org/wiki/Juris_ Hartmanis , a Turing Award Winner, is the Walter R. Read Professor of Engi- neering at Cornell University. He is a pioneer, founder and a major contributor to the area of computational complexity. Professor Hartmanis eminent career includes also a strong service compo- nent: he served in numerous important committees (Turing Award Committee, Gödel Prize Committee, Waterman Award Committee); he was director of NSF’s Directorate for Computer and Information Science and Engineering. Professor Hartmanis has been honoured with many awards and prizes. He was elected a member of the National Academy of Engineering and Latvian Academy of Sci- ences, and a fellow of the American Academy of Arts and Sciences, the Association for Computing Machinery, New York State Academy of Sciences, and the Ameri- can Association for the Advancement of Science.
    [Show full text]