Probabilistic Supervised Learning

Total Page:16

File Type:pdf, Size:1020Kb

Probabilistic Supervised Learning Probabilistic supervised learning 1 † 1 ‡ 2 § 3 Frithjof Gressmann ∗ , Franz J. Király , Bilal Mateen , and Harald Oberhauser 1 Department of Statistical Science, University College London, Gower Street, London WC1E 6BT, United Kingdom 2Warwick Medical School, University of Warwick, Coventry CV4 7AL, United Kingdom 3Mathematical Institute, University of Oxford, Andrew Wiles Building, Oxford OX2 6GG, United Kingdom May 8, 2019 Abstract Predictive modelling and supervised learning are central to modern data science. With predictions from an ever-expanding number of supervised black-box strategies - e.g., kernel methods, random forests, deep learning aka neural networks - being employed as a basis for decision making processes, it is crucial to understand the statistical uncertainty associated with these predictions. As a general means to approach the issue, we present an overarching framework for black-box pre- diction strategies that not only predict the target but also their own predictions’ uncertainty. Moreover, the framework allows for fair assessment and comparison of disparate prediction strategies. For this, we formally consider strategies capable of predicting full distributions from feature variables (rather than just a class or number), so-called probabilistic supervised learning strategies. Our work draws from prior work including Bayesian statistics, information theory, and modern su- pervised machine learning, and in a novel synthesis leads to (a) new theoretical insights such as a prob- abilistic bias-variance decomposition and an entropic formulation of prediction, as well as to (b) new algorithms and meta-algorithms, such as composite prediction strategies, probabilistic boosting and bag- ging, and a probabilistic predictive independence test. Our black-box formulation also leads (c) to a new modular interface view on probabilistic supervised learning and a modelling workflow API design, which we have implemented in the newly released skpro machine learning toolbox, extending the familiar modelling interface and meta-modelling functionality of arXiv:1801.00753v3 [stat.ML] 7 May 2019 sklearn. The skpro package provides interfaces for construction, composition, and tuning of probabilistic supervised learning strategies, together with orchestration features for validation and comparison of any such strategy - be it frequentist, Bayesian, or other. [email protected] ∗ †[email protected][email protected] §[email protected] 1 Contents 1 Introduction 4 1.1 Main ideas . .6 1.2 Technical contributions . .7 1.3 Relation to prior art . .8 1.3.1 Model assessment of probabilistic models via predictive likelihood . .9 1.3.2 Connections of predictive model assessment to information theory . 10 1.3.3 Bayesian/frequentist synthesis . 10 1.3.4 Toolbox API designs for supervised learning and probabilistic learning. 10 2 The probabilistic prediction setting 13 2.1 Notation and Conventions . 13 2.2 Statistical setting . 14 2.3 Classical supervised learning . 14 2.4 Probabilistic supervised learning . 15 2.5 Probabilistic loss functionals . 15 2.6 Recovering the classical setting . 19 2.7 Probabilistic losses for mean-variance predictions . 20 2.8 Short note: why taking (posterior) expectations is a bad idea . 22 3 Model assessment and model selection 23 3.1 Out-of-sample estimation of the generalization loss . 23 3.2 Model comparison and model performance . 24 3.3 Uninformed baseline: label density estimation . 25 3.4 Short note: why permutation of features/labels is not the uninformed baseline . 28 3.5 Classical baseline: residual density estimation . 29 3.6 Bayes type information criteria and empirical risk asymptotics for the log-loss . 30 3.7 No counterexample when out-of-sample . 32 4 Mixed outcome distributions 33 4.1 Integrating a discrete loss applied to binary cut-offs . 33 4.2 Convolution of the target with a kernel . 35 4.3 Kernel discrepancy losses . 36 4.4 Realizing that all computing is discrete . 39 4.5 Decomposition into continuous and discrete part . 40 4.6 Mixed-to-continuous adaptors . 41 4.7 Non-mixed distributions . 42 5 Learning theory and model diagnostics 43 5.1 Variations on Jensen . 43 5.2 Bayesics on predictions and posteriors . 44 5.3 Bias and variance in the probabilistic setting . 46 5.4 Information, predictability, and independence . 49 5.5 Transformations and probabilistic residuals . 54 5.6 Visual diagnostics of probabilistic models . 57 6 Meta-algorithms for the probabilistic setting 59 6.1 Target re-normalization and outlier capping for the log-loss . 59 6.2 Target adaptors and target composite strategies . 61 6.2.1 Conversion to point prediction . 62 6.2.2 Adaptors for point prediction strategies . 63 6.2.3 Adaptors for empirical samples such as from the Bayesian predictive posterior . 64 6.2.4 Adaptors for mixed distributions . 65 6.2.5 Conversion to parametric distributions . 67 6.3 Bagging and model averaging for probabilistic prediction strategies . 67 2 6.4 Boosting of probabilistic predictors . 69 6.5 A multi-variate test for independence . 70 6.5.1 Constructing a probabilistic predictive independence test . 70 6.5.2 Workflow meta-interfaces in the probabilistic setting . 72 6.5.3 Relation to kernel discrepancy based testing procedures . 72 7 Probabilistic supervised learning algorithms 74 7.1 Bayesian models . 74 7.2 Probabilistic Classifiers . 74 7.3 Heteroscedastic regression and prediction intervals . 75 7.4 Conditional Density Estimation . 75 7.5 Gaussian Processes . 76 7.6 Mixture density estimators, and mixtures of experts . 77 7.7 Quantile Regression and Quantile Estimation . 77 8 An API for probabilistic supervised learning 79 8.1 Overview . 79 8.1.1 General design principles . 79 8.1.2 Main use cases . 79 8.1.3 API design requirements . 80 8.2 Pythonic skpro interface design . 82 8.3 Core API . 84 8.3.1 Probabilistic estimators and distributions . 86 8.3.2 Integration of third-party prediction strategies . 86 8.3.3 Bayesian and probabilistic programming interfaces . 86 8.3.4 Model selection and performance metrics . 86 8.3.5 Code example . 87 8.4 Prediction strategies . 87 8.4.1 Baselines . 88 8.4.2 PyMC-interface . 88 8.4.3 Composite parametric prediction . ..
Recommended publications
  • Piotr Fryzlewicz
    Piotr Fryzlewicz [email protected] https://stats.lse.ac.uk/fryzlewicz https://scholar.google.co.uk/citations?user=i846MWEAAAAJ Department of Statistics, London School of Economics Houghton Street London WC2A 2AE UK (no telephone number – please email to set up a Zoom call) 2011– Professor of Statistics, Department of Statistics, London School of Economics, UK. (Deputy Head of Department 2019–20, 2021–22.) 2009–2011 Reader in Statistics, Department of Statistics, London School of Economics, UK. 2005–2009 Lecturer, then part-time Lecturer, then part-time Senior Lecturer in Statistics, Depart- ment of Mathematics, University of Bristol, UK. 2003–2005 Chapman Research Fellow, Department of Mathematics, Imperial College London, UK. 2000–2003 PhD in Statistics, Department of Mathematics, University of Bristol, UK. Thesis title: Wavelet techniques for time series and Poisson data. Adviser: Prof. Guy Nason. 1995–2000 MSci in Mathematics, Faculty of Fundamental Problems of Technology, Wroclaw Uni- versity of Science and Technology, Poland. 1st class degree (with distinction). Associate (unremunerated) academic affiliations 2019– Associate Member, Department of Statistics, University of Oxford, UK. Professional accreditations 2018– Chartered Statistician; the Royal Statistical Society. Non-academic work experience 2008–2009 Researcher (on a 90% part-time basis), Winton Capital Management, London, UK. Awards and honours 2016 Distinguished Alumnus, Wroclaw University of Science and Technology, Poland. 2014–2019 Fellowship; Engineering and Physical Sciences Research Council. 2013 Guy Medal in Bronze; the Royal Statistical Society. 2007–2008 University Research Fellowship; University of Bristol. 2005–2007 Award to Newly Appointed Lecturers in Science, Engineering and Mathematics; Nuffield Foundation. 2000–2003 Overseas Research Student Award; Universities UK.
    [Show full text]
  • Department of Statistics Academic Program Review Spring 2015
    Department of Statistics Academic Program Review Spring 2015 Department of Statistics • Texas A&M University • 3143 TAMU • College Station, TX 77843-3143 (979) 845-3141 • www.stat.tamu.edu Table of Contents I. Introduction Brief History of the Department……………………………………………………. 4 Mission and Goals………………………………………………………………….. 8 Administrative Structure of the Department………………………………………... 11 Resources…………………………………………………………………………... 13 Analysis……………………………………………………………………………. 14 Undergraduate Program……………………………………………………………. 15 Graduate Program………………………………………………………………….. 16 Professional Education……………………………………………………………... 19 Challenges and Opportunities……………………………………………………… 20 Assessment………………………………………………………………………… 22 II. Student Report Graduate Program in Statistics……………………………………………………… 28 Master of Science Program…………………………………………………………. 29 Doctor of Philosophy Program……………………………………………………... 32 Internship Program…………………………………………………………………. 35 Undergraduate Course Offerings…………………………………………………… 36 Graduate Course Offerings………………………………………………………… 37 Scheduling Coursework…………………………………………………………….. 42 Graduate Program Admissions Criteria……………………………………………... 43 GRE Scores for First Time Students………………………………………………... 44 Graduate Students Applied, Admitted, and Enrolled ………………………………. 45 Students by GPR Range……………………………………………………………. 48 Institutional Support for Full Time Students……………………………………….. 50 Honors & Awards Received by Students…………………………………………… 51 Statistics Graduate Student Association…………………………………………….. 54 Student Publications………………………………………………………………..
    [Show full text]
  • Mathematisches Forschungsinstitut Oberwolfach Statistical Inference For
    Mathematisches Forschungsinstitut Oberwolfach Report No. 48/2013 DOI: 10.4171/OWR/2013/48 Statistical Inference for Complex Time Series Data Organised by Rainer Dahlhaus, Heidelberg Oliver Linton, Cambridge Wei-Biao Wu, Chicago Qiwei Yao, London 22 September – 28 September 2013 Abstract. During recent years the focus of scientific interest has turned from low dimensional stationary time series to nonstationary time series and high dimensional time series. In addition new methodological challenges are coming from high frequency finance where data are recorded and analyzed on a millisecond basis. The three topics “nonstationarity”, “high dimension- ality” and “high frequency” are on the forefront of present research in time series analysis. The topics also have some overlap in that there already ex- ists work on the intersection of these three topics, e.g. on locally stationary diffusion models, on high dimensional covariance matrices for high frequency data, or on multivariate dynamic factor models for nonstationary processes. The aim of the workshop was to bring together researchers from time se- ries analysis, nonparametric statistics, econometrics and empirical finance to work on these topics. This aim was successfully achieved and the workshops was very well attended. Mathematics Subject Classification (2010): 62M10. Introduction by the Organisers The workshop Statistical Inference for Complex Time Series Data, organised by Rainer Dahlhaus (Heidelberg), Oliver Linton (Cambridge), Wei-Biao Wu (Chicago) and Qiwei Yao (London), was held in 22-28 September 2013. The workshop was well attended with 51 participants with broad geographic representation from Europe, Australia, Canada and USA. The participants formed a nice blend of researchers with various backgrounds including statistics, probability, machine 2744 Oberwolfach Report 48/2013 learning and econometrics.
    [Show full text]
  • Meetings Inc
    Volume 47 • Issue 8 IMS Bulletin December 2018 International Prize in Statistics The International Prize in Statistics is awarded to Bradley Efron, Professor of Statistics CONTENTS and Biomedical Data Science at Stanford University, in recognition of the “bootstrap,” 1 Efron receives International a method he developed in 1977 for assess- Prize in Statistics ing the uncertainty of scientific results 2–3 Members’ news: Grace Wahba, that has had extraordinary and enormous Michael Jordan, Bin Yu, Steve impact across many scientific fields. Fienberg With the bootstrap, scientists were able to learn from limited data in a 4 President’s column: Xiao-Li Meng simple way that enabled them to assess the uncertainty of their findings. In 6 NSF Seeking Novelty in Data essence, it became possible to simulate Sciences Bradley Efron, a “statistical poet” a potentially infinite number of datasets 9 There’s fun in thinking just from an original dataset, and in looking at the differences, measure the uncertainty of one step more the result from the original data analysis. Made possible by computing, the bootstrap 10 Peter Hall Early Career Prize: powered a revolution that placed statistics at the center of scientific progress. It helped update and nominations to propel statistics beyond techniques that relied on complex mathematical calculations 11 Interview with Dietrich or unreliable approximations, and hence it enabled scientists to assess the uncertainty Stoyan of their results in more realistic and feasible ways. “Because the bootstrap is easy for a computer to calculate and is applicable in an 12 Recent papers: Statistical Science; Bernoulli exceptionally wide range of situations, the method has found use in many fields of sci- ence, technology, medicine, and public affairs,” says Sir David Cox, inaugural winner Radu Craiu’s “sexy statistics” 13 of the International Prize in Statistics.
    [Show full text]
  • Threshold Models in Time Series Analysis „ 30 Years On
    Statistics and Its Interface Volume 4 (2011) 107–118 Threshold models in time series analysis — 30 years on Howell Tong Re-visiting the past can lead to new discoveries. – Confucius (551 B.C.–479 B.C.) model that can generate periodic oscillations is Simple Har- monic Motion (SHM): This paper is a selective review of the development of the threshold model in time series analysis over the past d2x(t) = −ω2x(t), 30 years or so. First, the review re-visits the motivation of dt2 the model. Next, it describes the various expressions of the model, highlighting the underlying principle and the main where ω is a real constant. This is an elementary linear probabilistic and statistical properties. Finally, after listing differential equation. It admits the solution some of the recent offsprings of the threshold model, the x(t)=A cos(ωt)+B sin(ωt), review finishes with some on-going research in the context of threshold volatility. where the arbitrary constants A and B are fixed by the ini- Keywords and phrases: Bi-modality, Canadian lynx, tial condition, e.g. x(t)anddx(t)/dt at t = 0. However, not Chaos, Conditional least squares, Drift criterion, Het- so well-known is the fact that this highly idealized model is eroscedasticity, Ecology, Ergodicity, Indicator time series, not practically useful, because different periodic oscillations result from different initial values. Thus, SHM does not pro- Invertibility, Limit cycles, Markov chain, Markov switching, duce a period that is robust against small disturbances to Nonlinear oscillations, Piecewise linearity, River-flow, Se- the initial value.
    [Show full text]
  • An Adaptive Estimation of Dimension Reduction Space
    J. R. Statist. Soc. B (2002) 64, Part 3, pp. 363–410 An adaptive estimation of dimension reduction space Yingcun Xia, University of Cambridge, UK, and Jinan University, People’s Republic of China Howell Tong, University of Hong Kong, People’s Republic of China, and London School of Economics and Political Science, UK W. K. Li University of Hong Kong, People’s Republic of China and Li-Xing Zhu University of Hong Kong and Chinese Academy of Sciences, Beijing, People’s Republic of China [Read before The Royal Statistical Society at a meeting organized by the Research Section on Wednesday, February 13th, 2002, Professor D. Firth in the Chair] Summary. Searching for an effective dimension reduction space is an important problem in regression, especially for high dimensional data. We propose an adaptive approach based on semiparametric models, which we call the (conditional) minimum average variance estimation (MAVE) method, within quite a general setting. The MAVE method has the following advantages. Most existing methods must undersmooth the nonparametric link function estimator to achieve a faster rate of consistency for the estimator of the parameters (than for that of the nonparametric function). In contrast, a faster consistency rate can be achieved by the MAVE method even with- out undersmoothing the nonparametric link function estimator. The MAVE method is applicable to a wide range of models, with fewer restrictions on the distribution of the covariates, to the extent that even time series can be included. Because of the faster rate of consistency for the parameter estimators, it is possible for us to estimate the dimension of the space consistently.
    [Show full text]
  • New IMS President Welcomed
    Volume 49 • Issue 7 IMS Bulletin October/November 2020 New IMS President welcomed As you will have noticed, there was no physical IMS Annual Meeting this year, as CONTENTS it was due to be held at the Bernoulli–IMS World Congress in Seoul, which is now 1 New IMS President and postponed to next year. This meant that the IMS meetings that would normally have Council meet taken place there were held online instead, including the handover of the Presidency 2 Members’ news: David (traditionally done by passing the gavel, but this year with a virtual elbow bump!). Madigan, Rene Carmona, Susan Murphy [below left] handed the Presidency to Regina Liu [below right]. Michael Ludkovski, Philip Ernst, Nancy Reid 4 C.R. Rao at 100 6 COPSS Presidents’ Award interview; Call for COPSS nominations 8 Nominate for IMS Awards; Breakthrough Prizes 9 New COPSS Leadership Look out for an article from Regina in the next issue. Award; World Statistics Day; There was also a virtual IMS Council Meeting this year, also held via Zoom, which Saul Blumenthal included discussions about how the IMS can recruit more new members — and, 10 Obituaries: Hélène Massam, crucially, retain our existing members. If you have any thoughts on this, do share Xiangrong Yin them with Regina: [email protected]. There was also an update on the plans for 12 Radu’s Rides: The Stink of the 2022 IMS Annual Meeting, which will be held in London, just before the COLT Mathematical Righteousness meeting (http://www.learningtheory.org/), with a one-day workshop of interest to both communities held in between the meetings.
    [Show full text]
  • Short Curriculum Vitae of Professor Howell Tong
    Short Curriculum Vitae of Professor Howell Tong Born in Hong Kong Education B.Sc (1966), M.Sc (1969), Ph.D (1972) all from the University of Manchester. Employment 1967-1968: Lecturer in Mathematics, Polytechnic of North London, UK 1968-1970: Demonstrator in Mathematics, UMIST, UK 1970-1977: Lecturer in Statistics, UMIST, UK 1977-1982: Senior Lecturer in Statistics, UMIST, UK 1982-1985: Founding Chair of Statistics, Chinese University of Hong Kong, Hong Kong 1986-1999: Chair of Statistics, University of Kent at Canterbury, UK 1999-2004: Chair of Statistics and sometimes Pro-Vice Chancellor and Founding Dean of the Graduate School, University of Hong Kong 1999-: Chair of Statistics, London School of Economics, UK Honours 1983: Member of the International Statistical Institute 1993: Fellow of the Institute of Mathematical Statistics 1994: Special Plenary Lecturer at 15th Nordic Meeting in Mathematical Statistics 1999: Honorary Fellow, Institute of Actuaries, England 1999: Alan T. Craig Lecture, University of Iowa, USA 2000: Foreign member of the Norwegian Academy of Science and Letters 2000: National Natural Science Prize, China (Mathematics, Class 2) 2002: Distinguished Research Achievement Award, University of Hong Kong. 2007: Guy medal in Silver (Royal Statistical Society, UK) Editorships 1988: Guest editor, the Statistician 1991-1994: Associate editor, JRSS (B) 1994: Joint editor, Chaos and Prediction, Phil. Trans. Roy. Soc. (London) (with (now Lord) Robert May and Dr. Bryan Grenfell) 1991-1999: Associate editor, Statistica Sinica 1993-:
    [Show full text]
  • Statistical Inference for Dynamical Systems: a Review Kevin Mcgoff, Sayan Mukherjee, Natesh S
    Statistical Inference for Dynamical Systems: A review Kevin McGoff, Sayan Mukherjee, Natesh S. Pillai Abstract. The topic of statistical inference for dynamical systems has been studied extensively across several fields. In this survey we focus on the problem of parameter estimation for nonlinear dynamical systems. Our objective is to place results across distinct disciplines in a common setting and highlight opportunities for further research. 1. INTRODUCTION The problem of parameter estimation in dynamical systems appears in many areas of science and engineering. Often the form of the model can be derived from some knowledge about the process under investigation, but parameters of the model must be inferred from empirical observations in the form of time series data. As this problem has appeared in many different contexts, partial solutions to this problem have been proposed in a wide variety of disciplines, including nonlinear dynamics in physics, control theory in engineering, state space model- ing in statistics and econometrics, and ergodic theory and dynamical systems in mathematics. One purpose of this study is to present these various approaches in a common language, with the hope of unifying some ideas and pointing towards interesting avenues for further study. By a dynamical system we mean a stochastic process of the form (Xn;Yn)n, where Xn+1 depends only on Xn and possibly some noise, and Yn depends only on Xn and possibly some noise. We think of Xn as the true state of the system at time n and Yn as our observation of the system at time n. The case when no noise is present has been most often considered by mathematicians in the field of dynamical systems and ergodic theory.
    [Show full text]
  • Call for 2021 JSM Proposals
    Volume 49 • Issue 6 IMS Bulletin September 2020 Call for 2021 JSM Proposals CONTENTS IMS Calls for Invited Session Proposals 1 JSM 2021 call for Invited The 2020 Virtual JSM may have just Session proposals finished, but thoughts are turning to next 2 Members’ news: Rina Foygel year’s meeting. Barber, Daniela Witten Starting from the 2021 Joint Statistical Meetings, which will take place 3 Nominate for IMS Awards in Seattle from August 7–12, 2021, the 4 Members’ news: ASA Awards, IMS now requires half of our invited Emmanuel Candès; Nominate session allocation to be competitive, in IMS Special Lecturers odd-numbered years when the JSM is 5 ASA Fellows also the IMS Annual Meeting. 6 SSC Gold Medal: Paul our Gustafson; SSC lectures We need y 7 Schramm Lectures; Royal Society Awards; Ethel submission! Newbold Prize Invited sessions — which includes invited papers, panels, and posters — showcase some 8 Donors to IMS Funds of the most innovative, impactful, and cutting-edge work in our profession. Invited 10 Recent papers: Annals of paper sessions consist of 2–6 speakers and discussants reporting new discoveries or Probability, Annals of Applied advances in a common topic; invited panels include 3–6 panelists providing commen- Probability, ALEA, Probability tary, discussion, and engaging debate on a particular topic of contemporary interest; and Mathematical Statistics and invited posters consist of up to 40 electronic posters presented during the Opening 12 Obituaries: Willem van Zwet, Mixer. Anyone can propose a session, but someone from the program committee must Don Gaver, Yossi Yahav accept the proposal for it to be part of the invited program.
    [Show full text]
  • Historical Gregynog Statistical Conferences
    Historical Gregynog Statistical Conferences 1 Speakers by date Talk Edition # Speaker Talk Title Notes 1 1st Gregynog (1965) 1 Dr. John Aitchison (Cambridge) Prediction See ‘Research notes’ 2 2 Dr. Larry Brown (Cornell) Admissable Estimators of Location Parameters 3 3 Prof. Henry Daniels (Birmingham) Stock Market Problem See ‘Research notes’ 4 4 Dr. Frank Downton (Birmingham) Order Statistics See ‘Research notes’ 5 5 Dr. Toby Lewis (UCL) Factorisation of Characteristic Function 6 6 Mr. David Wishart (St Andrews) Bacterial Colonies: Stochastic Models See ‘Research notes’ 7 2nd Gregynog (1966) 1 Prof. David Cox (Birkbeck / Imperial) Analysis of Qualitative Data See ‘Research notes’ 8 2 Dr. Marvin Zelen (National Cancer Institute, Operational Methods in the Design of Experiments See ‘Research notes’ Bethesda, MD) 9 3 Dr. Alan Durrant (Aberystwyth) Analysis of Genetic Variation See ‘Research notes’ 10 4 Prof. B.L. Welch (Leeds) Comparisons between Confidence Point Procedures See ‘Research notes’ 11 5 Dr. Peter Bickel (Berkeley) Asymptotic Theory of Bayes Solutions See ‘Research notes’ 12 6 Mr. Jeff Harrison (ICI) Short-Term Sales Forecasting See ‘Research notes’ 13 7 Dr. Murray Rosenblatt (San Diego) Spectra and their Estimates See ‘Research notes’ 14 8 Dr. John Bather (Manchester) Continuous Control of a Simple Inventory on Dam See ‘Research notes’ 15 3rd Gregynog (1967) 1 Prof. James Coleman (John Hopkins) Two Alternative Methods for Attitude Change See ‘Research notes’ 16 2 Prof. Paul Meier (Chicago) Estimation from Incomplete Observations See ‘Research notes’ Continued on next page Talk Edition # Speaker Talk Title Notes 17 3 Prof. H.D. Brunk (Oregon State) Certain Generalzed Means and Associated Families of Distributions 18 4 Prof.
    [Show full text]
  • A Conversation with Howell Tong 3
    Statistical Science 2014, Vol. 29, No. 3, 425–438 DOI: 10.1214/13-STS464 c Institute of Mathematical Statistics, 2014 A Conversation with Howell Tong Kung-Sik Chan and Qiwei Yao Abstract. Howell Tong has been an Emeritus Professor of Statistics at the London School of Economics since October 1, 2009. He was appointed to a lectureship at the University of Manchester Institute of Science and Technology shortly after he started his Master program in 1968. He re- ceived his Ph.D. in 1972 under the supervision of Maurice Priestley, thus making him an academic grandson of Maurice Bartlett. He stayed at UMIST until 1982, when he took up the Founding Chair of Statistics at the Chinese University of Hong Kong. In 1986, he returned to the UK, as the first Chinese to hold a Chair of Statistics in the history of the UK, by accepting the Chair at the University of Kent at Canterbury. He stayed there until 1997, when he went to the University of Hong Kong, first as a Distinguished Visiting Professor, and then as the Chair Professor of Statistics. At the University of Hong Kong, he served as a Pro-Vice- Chancellor and was the Founding Dean of the Graduate School. He was appointed to his Chair at the London School of Economics in 1999. He is a pioneer in the field of nonlinear time series analysis and has been a sci- entific leader both in Hong Kong and in the UK. His work on threshold models has had lasting influence both on theory and applications.
    [Show full text]