Columbia Computer Science Faculty in the National Academies

Total Page:16

File Type:pdf, Size:1020Kb

Columbia Computer Science Faculty in the National Academies CUCS Resources You can subscribe to the CUCS news mailing list at CS@CU http://lists.cs.columbia.edu/mailman/listinfo/cucs-news/ NEWSLETTER OF THE DEPARTMENT OF COMPUTER SCIENCE AT COLUMBIA UNIVERSITY VOL.3 NO.1 FALL 2006 CUCS colloquium information is available at http://www.cs.columbia.edu/lectures Columbia Computer Visit the CUCS alumni portal at http://alum.cs.columbia.edu Science Faculty in the National Academies Alfred V. Aho Stephen M. Bellovin Columbia University in the City of New York Department of Computer Science NON-PROFIT ORG. 1214 Amsterdam Avenue U.S. POSTAGE Mailcode: 0401 PAID Zvi Galil Edward H. Shortliffe Joseph F. Traub Vladimir Vapnik New York, NY 10027-7003 NEW YORK, NY PERMIT 4332 At this year’s computer science department Alfred V. Aho is Lawrence Vice President, Director, ADDRESS SERVICE REQUESTED “Hello Meeting,” members of the faculty Gussman Professor of department head, and member of technical staff. This is the recalled the early days of the department Computer Science and Vice Chair for Undergraduate lab that invented UNIX, C and when the same meeting could be held on Education in the Computer C++. Al was also the General the first floor of Mudd, in the small area that Science Department. He Manager of the Information is now used for making sandwiches in the served as Chair of the Sciences and Technologies Carleton Lounge. Since then, the computer department from 1995 to Research Laboratory at 1997 and in the spring of Bellcore (now Telcordia). science department has grown both in size 2003. He was elected to the Professor Aho is the “A” in and prominence. The faculty now includes U.S. National Academy of AWK, a widely used pattern- six members of the National Academies: Engineering for contributions matching language (you can Al Aho, Steven M. Bellovin, Zvi Galil, Ted to the fields of algorithms think of AWK as the initial and programming tools. Shortliffe, Joe Traub, and Vladimir Vapnik. pure version of perl). “W” is Peter Weinberger and “K” Here is a brief overview of their achievements Professor Aho received his B.A.Sc in Engineering Physics is Brian Kernighan. Al also and contributions. from the University of Toronto wrote the initial versions of and a Ph.D. in Electrical the string pattern-matching Engineering/Computer Science programs egrep and fgrep from Princeton University. that first appeared on UNIX. Prior to his current position at His current research interests Columbia, Professor Aho served include quantum computing, in many capacities at the programming languages, Computing Sciences Research compilers, and algorithms. Center at Bell Labs, such as (continued on next page) CS@CU NEWSLETTER OF THE DEPARTMENT OF COMPUTER SCIENCE AT COLUMBIA UNIVERSITY WWW.CS.COLUMBIA.EDU CS@CU FALL 2006 1 Cover Story (continued) Professor Aho has won the and cybersecurity research journals and is the editor Care, while directing an active many countries. IBC studies National Academy of one hundred research papers. These six National IEEE John von Neumann needs. He was also a member of the book “Computational research program in clinical optimal algorithms and Engineering elected Professor His major achievements have Academy members Medal and has received of the information technology Algorithms on Strings.” information systems and computational complexity Vapnik as a new member in been the development of a honorary doctorates from subcommittee of an NRC Although being the Dean of decision support. He spear- for the continuous problems 2006 for “insights into the general theory for minimizing represent the growth the Universities of Helsinki study group on science versus Engineering and Applied headed the formation of a which arise in physical science, fundamental complexities of the expected risk of predictors and prominence of the and Waterloo. He is a terrorism. He was a member Science has taken up most of Stanford graduate degree pro- mathematical finance, learning and for inventing using empirical data, and a Columbia Computer Member of the American of the Internet Architecture his time for the last 12 years, he gram in biomedical informatics engineering and economics. practical and widely applied new type of learning machine Science Department since Academy of Arts and Board from 1996-2002, and reports, with great mirth, that and divided his time between A major focus of his current machine-learning algorithms”. (the Support Vector Machine) its modest beginnings. Sciences and a Fellow of was co-director of the Security he still tries to find the most clinical medicine and biomed- work is quantum computing. that possesses a high level of Professor Vapnik obtained his Undergraduates and the American Association Area of the IETF from 2002 efficient way to solve problems. ical informatics research. He has received numerous Masters Degree in Mathematics generalization ability. These for the Advancement of through 2004. He sits on Professor Shortliffe continues honors including election in 1958 at Uzbek State techniques have been used to graduate students alike Science, ACM, Bell Labs, the Department of Homeland solve many pattern recognition is to be closely involved with to the National Academy of University, Samarkand, USSR are fortunate to have and IEEE. He received Security’s Science and Edward H. Shortliffe and regression estimation Rolf A. Scholdager Professor biomedical informatics graduate Engineering in 1985, the and his Ph.D at the Institute such giant shoulders to the Great Teacher Award Technology Advisory Board. problems and have been and Chair of the Department training where he works to Emanuel R. Piore Medal from of Control Sciences in the for 2003 from the Society applied to the problems of stand on. of Biomedical Informatics create a new breed of health IEEE in 1991, selection by Academy of Sciences, Moscow, of Columbia Graduates. dependency estimation, fore- is the Dean of the at Columbia College of professional. His research the Accademia Nazionale dei in 1964. From 1964 to 1990 Zvi Galil casting, and constructing School of Engineering and Physicians and Surgeons. He interests include the broad Lincei in Rome to present he worked at the Institute, intelligent machines. Professor is a Applied Science. He is the is an elected member of the range of issues related to the 1993 Lezione Lincei, and where he became Head of the Steven M. Bellovin Vapnik’s research combines Professor of Computer Science. Julian Clarence Levi Professor Institute of Medicine of the integrated decision-support the 1999 Mayor’s Award for Computer Science Research mathematical analysis of the He was elected to the National of Mathematical Methods National Academy of Sciences. systems, their effective Excellence in Science and Department. He then joined problem of empirical inference Academy of Engineering and Computer Science, and implementation, and the role Technology in New York City. In AT&T Bell Laboratories, Holmdel, After receiving an A.B. in in complex (high dimensional) for contributions to network served as Chairman of the of the Internet in health care. 2001, he received an honorary NJ, as a Consultant, and Applied Mathematics from worlds with philosophical inter- applications and security. Department of Computer Doctorate of Science from the was appointed Professor of Harvard College in 1970, he pretation of such inferences Science from 1989 until July University of Central Florida. Computer Science and Statistics Professor Bellovin joined the moved to Stanford University Joseph F. Traub is the Edwin based on the imperative which 1995. In 2004, Dean Galil at Royal Holloway in 1995. faculty in 2005 after many where he could pursue Howard Armstrong Professor he formulated in 1995: “When was elected to the National years at Bell Labs and AT&T interests in medicine and of Computer Science. Vladimir Vapnik is Professor Professor Vapnik has taught solving a problem of interest, Academy of Engineering for Labs Research, where he was computer science, and was He came to Columbia in of Computer Science at and researched in computer do not solve a more general his contributions to the design an AT&T Fellow. He was first awarded a Ph.D. in Medical 1979 as founding chair Columbia and Fellow at NEC science and theoretical and problem as an intermediate and analysis of algorithms exposed to computers in 10th Information Sciences in of the Computer Science Labs. He is currently at the applied statistics for over thirty step. Try to get the answer and for leadership in computer grade at Stuyvesant High 1975 and an M.D. in 1976. Department. From 1971 Center for Computational years. He has published six that you really need but not a science and engineering. School and has been working Subsequently, he served as to 1979 he was Head of Learning Systems. The monographs and more than more general one.” in the field ever since. He Dean Galil began his studies in Professor of Medicine and of the Computer Science received a BA degree from applied mathematics; he soon Computer Science at Stanford Department at Carnegie Columbia University, and an turned to the emerging field University. In January 2000 Mellon University. He served MS and PhD in Computer of computer science, which he he assumed his new post at as founding Chairman of Science from the University saw as an extension of his Columbia University, where the Computer Science and of North Carolina at Chapel interest in mathematics. His he is also Deputy Vice Telecommunications Board Feature Article Hill. As a graduate student, main research pursuits are President (Columbia University (CSTB) of the National he helped create Netnews; in the design and analysis of Medical Center); Senior Academies from 1986 to 1992 for this, he and his fellow algorithms, computational Associate Dean for Strategic and is again serving as Chair. Professor Shree Nayar Named as 2006 Great Teacher contributors were awarded complexity, and cryptography.
Recommended publications
  • INSTITUTE for COMPUTATIONAL and BIOLOGICAL LEARNING (ICBL)
    ICBL (HANSON, GLUCK, VAPNIK) 1 August 20th 2003 Governor's Commission on Jobs & Economic Growth INSTITUTE FOR COMPUTATIONAL and BIOLOGICAL LEARNING (ICBL) PRIMARY PARTICIPANTS: Stephen José Hanson, Psych. Rutgers-Newark; Cog. Sci. Rutgers-NB, Info. Sci., NJIT; Advanced Imaging Center, UMDNJ (Principal Investigator and Co-Director) Mark A. Gluck, CMBN Rutgers-Newark (co-Principal Investigator and Co-Director) Vladimir Vapnik, NEC (co-Principal Investigator and Co-Director) We also include participants and contact personnel for other participating institutions including Rutgers NB-Piscataway, Rutgers-Camden, UMDNJ, NJIT, Princeton, NEC, ATT, Siemens Corporate Research, and Merck Pharmaceuticals (See attachments that follow this document.) PRIMARY OBJECTIVE: The Institute for Computational and Biological Learning (ICBL) will be an interdisciplinary center for research and training in the learning sciences, encompassing computational, biological,and psychological studies of systems that adapt, extract patterns from high complexity data and in general improve from experience. Core programs include learning theory, neuroinformatics, computational neuroscience of learning and memory, neural-network models, machine learning, human computer interaction and financial/economic modeling. The learning sciences is an emerging interdisciplinary technology, recently recognized by the National Science Foundation (NSF) as a national priority. The NSF announced this year a major multi-disciplinary initiative which will fund approximately five new national Learning Sciences Centers with budgets of $3M to $5M a year for up to 10 years, for a total commitment of up to $50,000,000 per center. As envisioned by NSF, these centers will be “large-scale, long-term Centers that will extend the frontiers of knowledge on learning and create the intellectual, organizational, and physical infrastructure needed for the long-term advancement of learning research.
    [Show full text]
  • Tarjan Transcript Final with Timestamps
    A.M. Turing Award Oral History Interview with Robert (Bob) Endre Tarjan by Roy Levin San Mateo, California July 12, 2017 Levin: My name is Roy Levin. Today is July 12th, 2017, and I’m in San Mateo, California at the home of Robert Tarjan, where I’ll be interviewing him for the ACM Turing Award Winners project. Good afternoon, Bob, and thanks for spending the time to talk to me today. Tarjan: You’re welcome. Levin: I’d like to start by talking about your early technical interests and where they came from. When do you first recall being interested in what we might call technical things? Tarjan: Well, the first thing I would say in that direction is my mom took me to the public library in Pomona, where I grew up, which opened up a huge world to me. I started reading science fiction books and stories. Originally, I wanted to be the first person on Mars, that was what I was thinking, and I got interested in astronomy, started reading a lot of science stuff. I got to junior high school and I had an amazing math teacher. His name was Mr. Wall. I had him two years, in the eighth and ninth grade. He was teaching the New Math to us before there was such a thing as “New Math.” He taught us Peano’s axioms and things like that. It was a wonderful thing for a kid like me who was really excited about science and mathematics and so on. The other thing that happened was I discovered Scientific American in the public library and started reading Martin Gardner’s columns on mathematical games and was completely fascinated.
    [Show full text]
  • A Priori Knowledge from Non-Examples
    A priori Knowledge from Non-Examples Diplomarbeit im Fach Bioinformatik vorgelegt am 29. März 2007 von Fabian Sinz Fakultät für Kognitions- und Informationswissenschaften Wilhelm-Schickard-Institut für Informatik der Universität Tübingen in Zusammenarbeit mit Max-Planck-Institut für biologische Kybernetik Tübingen Betreuer: Prof. Dr. Bernhard Schölkopf (MPI für biologische Kybernetik) Prof. Dr. Andreas Schilling (WSI für Informatik) Prof. Dr. Hanspeter Mallot (Biologische Fakultät) 1 Erklärung Die folgende Diplomarbeit wurde ausschließlich von mir (Fabian Sinz) erstellt. Es wurden keine weiteren Hilfmittel oder Quellen außer den Angegebenen benutzt. Tübingen, den 29. März 2007, Fabian Sinz Contents 1 Introduction 4 1.1 Foreword . 4 1.2 Notation . 5 1.3 Machine Learning . 7 1.3.1 Machine Learning . 7 1.3.2 Philosophy of Science in a Nutshell - About the Im- possibility to learn without Assumptions . 11 1.3.3 General Assumptions in Bayesian and Frequentist Learning . 14 1.3.4 Implicit prior Knowledge via additional Data . 17 1.4 Mathematical Tools . 18 1.4.1 Reproducing Kernel Hilbert Spaces . 18 1.5 Regularisation . 22 1.5.1 General regularisation . 22 1.5.2 Data-dependent Regularisation . 27 2 The Universum Algorithm 32 2.1 The Universum Idea . 32 2.1.1 VC Theory and Structural Risk Minimisation (SRM) in a Nutshell . 32 2.1.2 The original Universum Idea by Vapnik . 35 2.2 Implementing the Universum into SVMs . 37 2.2.1 Support Vector Machine Implementations of the Uni- versum . 38 2.2.1.1 Hinge Loss Universum . 38 2.2.1.2 Least Squares Universum . 46 2.2.1.3 Multiclass Universum .
    [Show full text]
  • The Best Nurturers in Computer Science Research
    The Best Nurturers in Computer Science Research Bharath Kumar M. Y. N. Srikant IISc-CSA-TR-2004-10 http://archive.csa.iisc.ernet.in/TR/2004/10/ Computer Science and Automation Indian Institute of Science, India October 2004 The Best Nurturers in Computer Science Research Bharath Kumar M.∗ Y. N. Srikant† Abstract The paper presents a heuristic for mining nurturers in temporally organized collaboration networks: people who facilitate the growth and success of the young ones. Specifically, this heuristic is applied to the computer science bibliographic data to find the best nurturers in computer science research. The measure of success is parameterized, and the paper demonstrates experiments and results with publication count and citations as success metrics. Rather than just the nurturer’s success, the heuristic captures the influence he has had in the indepen- dent success of the relatively young in the network. These results can hence be a useful resource to graduate students and post-doctoral can- didates. The heuristic is extended to accurately yield ranked nurturers inside a particular time period. Interestingly, there is a recognizable deviation between the rankings of the most successful researchers and the best nurturers, which although is obvious from a social perspective has not been statistically demonstrated. Keywords: Social Network Analysis, Bibliometrics, Temporal Data Mining. 1 Introduction Consider a student Arjun, who has finished his under-graduate degree in Computer Science, and is seeking a PhD degree followed by a successful career in Computer Science research. How does he choose his research advisor? He has the following options with him: 1. Look up the rankings of various universities [1], and apply to any “rea- sonably good” professor in any of the top universities.
    [Show full text]
  • Lecture Notes in Computer Science
    D. Dolev Z. Galil M. Rodeh (Eds.) Theory of Computing and Systems ISTCS '92, Israel Symposium Haifa, Israel, May 27-28, 1992 Proceedings Springer-Verlag Berlin Heidelberg NewYork London Paris Tokyo Hong Kong Barcelona Budapest Series Editors Gerhard Goos Juris Hartmanis Universit~it Karlsruhe Department of Computer Science Postfach 69 80 Cornell University Vincenz-Priessnitz-Stral3e 1 5149 Upson Hall W-7500 Karlsruhe, FRG Ithaca, NY 14853, USA Volume Editors Danny Dolev Hebrew University Givat Ram, Jerusalem, Israel Zvi Galil Columbia University, New York, NY 10027, USA and Tel Aviv University Ramat Aviv, Tel Aviv, Israel Michael Rodeh IBM Israel Ltd., Science and Technology, Technion City Haifa 32000, Israel CR Subject Classification (1991): F.1-4, B.3, C.2, G.1-2, 1.1, E.4, D.2-3 ISBN 3-540-55553-6 Springer-Verlag Berlin Heidelberg New York ISBN 0-387-55553-6 Springer-Verlag New York Berlin Heidelberg This work is subject to copyright. All rights are reserved, whether the whole or part of the material is concerned, specifically the rights of translation, reprinting, re-use of illustrations, recitation, broadcasting, reproduction on microfilms or in any other way, and storage in data banks. Duplication of this publication or parts thereof is permitted only under the provisions of the German Copyright Law of September 9, 1965, in its current version, and permission for use must always be obtained from Springer-Verlag. Violations are liable for prosecution under the German Copyright Law. Springer-Verlag Berlin Heidelberg 1992 Printed in Germany Typesetting: Camera ready by author/editor Printing and binding: Druckhaus Beltz, Hemsbach/Bergstr.
    [Show full text]
  • United States Patent 19 11 Patent Number: 5,640,492 Cortes Et Al
    US005640492A United States Patent 19 11 Patent Number: 5,640,492 Cortes et al. 45 Date of Patent: Jun. 17, 1997 54). SOFT MARGIN CLASSIFIER B.E. Boser, I. Goyon, and V.N. Vapnik, “A Training Algo rithm. For Optimal Margin Classifiers”. Proceedings Of The 75 Inventors: Corinna Cortes, New York, N.Y.: 4-th Workshop of Computational Learning Theory, vol. 4, Vladimir Vapnik, Middletown, N.J. San Mateo, CA, Morgan Kaufman, 1992. 73) Assignee: Lucent Technologies Inc., Murray Hill, Y. Le Cun, B. Boser, J. Denker, D. Henderson, R. Howard, N.J. W. Hubbard, and L. Jackel, "Handwritten Digit Recognition with a Back-Propagation Network', (D. Touretzky, Ed.), Advances in Neural Information Processing Systems, vol. 2, 21 Appl. No.: 268.361 Morgan Kaufman, 1990. 22 Filed: Jun. 30, 1994 A.N. Referes et al., “Stock Ranking: Neural Networks Vs. (51] Int. Cl. ................ G06E 1/00; G06E 3100 Multiple Linear Regression':, IEEE International Conf. on 52 U.S. Cl. .................................. 395/23 Neural Networks, vol. 3, San Francisco, CA, IEEE, pp. 58) Field of Search ............................................ 395/23 1419-1426, 1993. M. Rebollo et al., “AMixed Integer Programming Approach 56) References Cited to Multi-Spectral Image Classification”, Pattern Recogni U.S. PATENT DOCUMENTS tion, vol. 9, No. 1, Jan. 1977, pp. 47-51, 54-55, and 57. 4,120,049 10/1978 Thaler et al. ........................... 365/230 V. Uebele et al., "Extracting Fuzzy Rules From Pattern 4,122,443 10/1978 Thaler et al. ....... 340/.46.3 Classification Neural Networks”, Proceedings From 1993 5,040,214 8/1991 Grossberg et al. ....................... 381/43 International Conference on Systems, Man and Cybernetics, 5,214,716 5/1993 Refregier et al..
    [Show full text]
  • On the State of the Art in Machine Learning: a Personal Review
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by CiteSeerX University of Bristol DEPARTMENT OF COMPUTER SCIENCE On the state of the art in Machine Learning: a personal review Peter A. Flach December 2000 CSTR-00-018 On the state of the art in Machine Learning: a personal review Peter A. Flach December 15, 2000 Abstract This paper reviews a number of recent books related to current developments in machine learning. Some (anticipated) trends will be sketched. These include: a trend towards combining approaches that were hitherto regarded as distinct and were studied by separate research communities; a trend towards a more prominent role of representation; and a tighter integration of machine learning techniques with techniques from areas of application such as bioinformatics. The intended readership has some knowledge of what machine learning is about, but brief tuto- rial introductions to some of the more specialist research areas will also be given. 1 Introduction This paper reviews a number of books that appeared recently in my main area of exper- tise, machine learning. Following a request from the book review editors of Artificial Intelligence, I selected about a dozen books with which I either was already familiar, or which I would be happy to read in order to find out what’s new in that particular area of the field. The choice of books is therefore subjective rather than comprehensive (and also biased towards books I liked). However, it does reflect a personal view of what I think are exciting developments in machine learning, and as a secondary aim of this paper I will expand a bit on this personal view, sketching some of the current and anticipated future trends.
    [Show full text]
  • UNIVERSITY of CALIFORNIA Los Angeles Data-Driven Learning Of
    UNIVERSITY OF CALIFORNIA Los Angeles Data-Driven Learning of Invariants and Specifications A dissertation submitted in partial satisfaction of the requirements for the degree Doctor of Philosophy in Computer Science by Saswat Padhi 2020 © Copyright by Saswat Padhi 2020 ABSTRACT OF THE DISSERTATION Data-Driven Learning of Invariants and Specifications by Saswat Padhi Doctor of Philosophy in Computer Science University of California, Los Angeles, 2020 Professor Todd D. Millstein, Chair Although the program verification community has developed several techniques for analyzing software and formally proving their correctness, these techniques are too sophisticated for end users and require significant investment in terms of time and effort. In this dissertation, I present techniques that help programmers easily formalize the initial requirements for verifying their programs — specifications and inductive invariants. The proposed techniques leverage ideas from program synthesis and statistical learning to automatically generate these formal requirements from readily available program-related data, such as test cases, execution traces etc. I detail three of these data-driven learning techniques – FlashProfile and PIE for specification learning, and LoopInvGen for invariant learning. I conclude with some principles for building robust synthesis engines, which I learned while refining the aforementioned techniques. Since program synthesis is a form of function learning, it is perhaps unsurprising that some of the fundamental issues in program synthesis have also been explored in the machine learning community. I study one particular phenomenon — overfitting. I present a formalization of overfitting in program synthesis, and discuss two mitigation strategies, inspired by existing techniques. ii The dissertation of Saswat Padhi is approved. Adnan Darwiche Sumit Gulwani Miryung Kim Jens Palsberg Todd D.
    [Show full text]
  • Accelerating Text-As-Data Research in Computational Social Science
    Accelerating Text-as-Data Research in Computational Social Science Dallas Card August 2019 CMU-ML-19-109 Machine Learning Department School of Computer Science Carnegie Mellon University Pittsburgh, PA Thesis Committee: Noah A. Smith, Chair Artur Dubrawski Geoff Gordon Dan Jurafsky (Stanford University) Submitted in partial fulfillment of the requirements for the Degree of Doctor of Philosophy Copyright c 2019 Dallas Card This work was supported by a Natural Sciences and Engineering Research Council of Canada Postgraduate Scholarship, NSF grant IIS-1211277, an REU supplement to NSF grant IIS-1562364, a Bloomberg Data Science Research Grant, a University of Washington Innovation Award, and computing resources provided by XSEDE. Keywords: machine learning, natural language processing, computational social science, graphical models, interpretability, calibration, conformal methods Abstract Natural language corpora are phenomenally rich resources for learning about people and society, and have long been used as such by various disciplines such as history and political science. Recent advances in machine learning and natural language pro- cessing are creating remarkable new possibilities for how scholars might analyze such corpora, but working with textual data brings its own unique challenges, and much of the research in computer science may not align with the desiderata of social scien- tists. In this thesis, I present a line of work on developing methods for computational social science focused primarily on observational research using natural language text. Throughout, I take seriously the concerns and priorities of the social sciences, leading to a focus on aspects of machine learning which are otherwise sometimes secondary, including calibration, interpretability, and transparency. Two ideas which unify this work are the problems of exploration and measurement, and as a running example I consider the problem of analyzing how news sources frame contemporary political issues.
    [Show full text]
  • Sparsification–A Technique for Speeding up Dynamic Graph Algorithms
    Sparsification–A Technique for Speeding Up Dynamic Graph Algorithms DAVID EPPSTEIN University of California, Irvine, Irvine, California ZVI GALIL Columbia University, New York, New York GIUSEPPE F. ITALIANO Universita` “Ca’ Foscari” di Venezia, Venice, Italy AND AMNON NISSENZWEIG Tel-Aviv University, Tel-Aviv, Israel Abstract. We provide data structures that maintain a graph as edges are inserted and deleted, and keep track of the following properties with the following times: minimum spanning forests, graph connectivity, graph 2-edge connectivity, and bipartiteness in time O(n1/2) per change; 3-edge connectivity, in time O(n2/3) per change; 4-edge connectivity, in time O(na(n)) per change; k-edge connectivity for constant k, in time O(nlogn) per change; 2-vertex connectivity, and 3-vertex connectivity, in time O(n) per change; and 4-vertex connectivity, in time O(na(n)) per change. A preliminary version of this paper was presented at the 33rd Annual IEEE Symposium on Foundations of Computer Science (Pittsburgh, Pa.), 1992. D. Eppstein was supported in part by National Science Foundation (NSF) Grant CCR 92-58355. Z. Galil was supported in part by NSF Grants CCR 90-14605 and CDA 90-24735. G. F. Italiano was supported in part by the CEE Project No. 20244 (ALCOM-IT) and by a research grant from Universita` “Ca’ Foscari” di Venezia; most of his work was done while at IBM T. J. Watson Research Center. Authors’ present addresses: D. Eppstein, Department of Information and Computer Science, University of California, Irvine, CA 92717, e-mail: [email protected], internet: http://www.ics.
    [Show full text]
  • Teaching with IMPACT
    Teaching with IMPACT Carl Trimbach [email protected] Michael Littman [email protected] Abstract Like many problems in AI in their general form, supervised learning is computationally intractable. We hypothesize that an important reason humans can learn highly complex and varied concepts, in spite of the computational difficulty, is that they benefit tremendously from experienced and insightful teachers. This paper proposes a new learning framework that provides a role for a knowledgeable, benevolent teacher to guide the process of learning a target concept in a series of \curricular" phases or rounds. In each round, the teacher's role is to act as a moderator, exposing the learner to a subset of the available training data to move it closer to mastering the target concept. Via both theoretical and empirical evidence, we argue that this framework enables simple, efficient learners to acquire very complex concepts from examples. In particular, we provide multiple examples of concept classes that are known to be unlearnable in the standard PAC setting along with provably efficient algorithms for learning them in our extended setting. A key focus of our work is the ability to learn complex concepts on top of simpler, previously learned, concepts|a direction with the potential of creating more competent artificial agents. 1. Introduction In statistical learning theory, the PAC (probably approximately correct) model (Valiant, 1984) has been a valuable tool for framing and understanding the problem of learning from examples. Many interesting and challenging problems can be analyzed and attacked in the PAC setting, resulting in powerful algorithms. Nevertheless, there exist many central and relevant concept classes that are provably unlearnable this way.
    [Show full text]
  • Efficient Algorithms for Finding Maximum Matching in Graphs
    Efficient Algorithms for Finding Maximum Matching in Graphs ZVI GALIL Department of Computer Science, Columbia University, New York, N. Y., 10027 and Tel-Aviv University, Tel-Aviv, Israel This paper surveys the techniques used for designing the most efficient algorithms for finding a maximum cardinality or weighted matching in (general or bipartite) graphs. It also lists some open problems concerning possible improvements in existing algorithms and the existence of fast parallel algorithms for these problems. Categories and Subject Descriptors: F.2.2 [Analysis of Algorithms and Problem Complexity]: Nonnumerical Algorithms and Problems-computation on discrete structures; G.2.2 [Discrete Mathematics]: Graph Theory-graph algorithms General Terms: Algorithms, Theory, Verification Additional Key Words and Phrases: Algorithmic tools, the asexual case, assignment problem, augmenting path, blossoms, data structures, d-heap, duality, ET, generalized priority queue, Main Theorem of Botany, matching, monsters, moonlighting, polygamy, primal-dual method, shmathematics, shrink, warm-up INTRODUCTION tools are “divide and conquer” and dynamic There are no recipes for designing efficient programming [Aho et al. 19741. Alterna- algorithms. This is somewhat unfortunate tively, we may try to find a way to reduce from the point of view of applications: Any the number of steps in the original algo- time we have to design an algorithm, we rithm by finding a better way to organize may have to start (almost) from scratch. the information. However, it is fortunate from the point of (c) Data Structures. Sometimes we can view of researchers: It is unlikely that we speed up an algorithm by using an efficient are going to run out of problems or chal- data structure that supports the primitive lenges.
    [Show full text]