Sparse Machine Learning Methods with Applications in Multivariate Signal Processing

Total Page:16

File Type:pdf, Size:1020Kb

Sparse Machine Learning Methods with Applications in Multivariate Signal Processing Sparse Machine Learning Methods with Applications in Multivariate Signal Processing Thomas Robert Diethe A dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy of the University of London. Department of Computer Science University College London 2010 2 I, Thomas Robert Diethe, confirm that the work presented in this thesis is my own. Where informa- tion has been derived from other sources, I confirm that this has been indicated in the thesis. Abstract This thesis details theoretical and empirical work that draws from two main subject areas: Machine Learning (ML) and Digital Signal Processing (DSP). A unified general framework is given for the appli- cation of sparse machine learning methods to multivariate signal processing. In particular, methods that enforce sparsity will be employed for reasons of computational efficiency, regularisation, and compress- ibility. The methods presented can be seen as modular building blocks that can be applied to a variety of applications. Application specific prior knowledge can be used in various ways, resulting in a flexible and powerful set of tools. The motivation for the methods is to be able to learn and generalise from a set of multivariate signals. In addition to testing on benchmark datasets, a series of empirical evaluations on real world datasets were carried out. These included: the classification of musical genre from polyphonic audio files; a study of how the sampling rate in a digital radar can be reduced through the use of Com- pressed Sensing (CS); analysis of human perception of different modulations of musical key from Electroencephalography (EEG) recordings; classification of genre of musical pieces to which a listener is attending from Magnetoencephalography (MEG) brain recordings. These applications demonstrate the efficacy of the framework and highlight interesting directions of future research. Acknowledgements To my parents, who have supported my education from start to finish, thank-you so much for giving me this opportunity. To my supervisor John Shawe-Taylor, whose breadth and depth of knowledge never ceases to amaze me, thank-you for your guidance. The research leading to the results presented here has received funding from the EPSRC grant agreement EP-D063612-1, “Learning the Structure of Music”. Contents List of Figures 9 List of Tables 11 1 Introduction 12 1.1 MachineLearning................................. ..... 12 1.2 SparsityinMachineLearning. ......... 12 1.3 MultivariateSignalProcessing . ........... 13 1.4 ApplicationAreas................................ ...... 14 1.4.1 LearningtheStructureofMusic . ....... 14 1.4.2 MusicInformationRetrieval . ....... 15 1.4.3 AutomaticanalysisofBrainSignals . ......... 15 1.4.4 AdditionalApplicationAreas . ....... 15 1.4.5 PublishedWorks ................................ .. 16 1.5 Structureofthisthesis . ........ 16 2 Background 18 2.1 MachineLearning................................. ..... 18 2.1.1 ReproducingKernelHilbertSpaces . ........ 19 2.1.2 Regression .................................... 20 2.1.3 Lossfunctionsforregression. ........ 20 2.1.4 Linearregressioninafeaturespace . ......... 21 2.1.5 StabilityofRegression . ...... 22 2.1.6 Regularisation ................................ ... 24 2.1.7 SparseRegression .............................. ... 25 2.1.8 Classification ................................. ... 27 2.1.9 Lossfunctionsforclassification . ......... 27 2.1.10 MaximumMarginclassification . ....... 31 Contents 6 2.1.11 Boosting ..................................... 32 2.1.12 SubspaceMethods ... ... .... .... .... ... .... .... ... 37 2.1.13 Multi-viewLearning . ..... 38 2.2 DigitalSignalProcessing(DSP) . .......... 39 2.2.1 Bases,Frames,DictionariesandTransforms . ............ 39 2.2.2 SparseandRedundantSignals . ...... 42 2.2.3 GreedyMethodsforSparseEstimation . ........ 43 2.2.4 CompressedSensing(CS) . .... 45 2.2.5 IncoherenceWithRandomMeasurements . ........ 46 2.2.6 MultivariateSignalProcessing . ......... 47 3 Sparse Machine Learning Framework for Multivariate Signal Processing 48 3.1 FrameworkOutline ................................ ..... 48 3.2 GreedymethodsforMachineLearning. .......... 51 3.2.1 Matching Pursuit Kernel Fisher Discriminant Analysis.............. 51 3.2.2 Nystr¨omLow-RankApproximations . ....... 53 3.3 KernelPolytopeFacesPursuit . ......... 63 3.3.1 Generalisationerrorbound . ....... 64 3.3.2 Experiments ................................... 68 3.3.3 BoundExperiments... ... .... .... .... ... .... .... ... 69 3.4 LearninginaNystr¨omApproximatedSubspace . ............. 70 3.4.1 Theory of Support Vector Machine (SVM) in Nystr¨om Subspace......... 72 3.4.2 Experiments:Classification . ....... 77 3.4.3 Experiments:Regression. ...... 78 3.5 Multi-ViewLearning .. .... ... .... .... .... ... .... .. ...... 81 3.5.1 Kernel Canonical Correlation Analysis with ProjectedNearestNeighbours . 83 3.5.2 Convex Multi-View Fisher Discriminant Analysis . ............. 84 3.6 ConclusionsandFurtherWork . ........ 96 4 Applications I 97 4.1 Introduction.................................... ..... 97 4.2 GenreClassification... .... ... .... .... .... ... .... ....... 98 4.2.1 MIREX....................................... 99 4.2.2 FeatureSelection .............................. 100 4.2.3 Framelevelfeatures .. ... .... .... .... ... .... .... 101 4.2.4 FeatureAggregation . 103 4.2.5 Algorithms .................................... 103 4.2.6 Multiclass Linear Programming Boosting (LPBoost) Formulation (LPMBoost ) . 104 4.2.7 Experiments ................................... 105 Contents 7 4.2.8 Results ....................................... 107 4.3 CompressedSensingforRadar . ........108 4.3.1 ReviewofCompressiveSampling . ......109 4.3.2 ApplicationofCSToRadar . 109 4.3.3 ExperimentalApproach . 110 4.3.4 ResultsAndAnalysis. 112 4.4 Conclusions..................................... 117 5 Applications II 118 5.1 Introduction.................................... 118 5.2 Experiment 1: Classification of tonality from EEG recordings...............119 5.2.1 Participants.................................. 120 5.2.2 Design .......................................121 5.2.3 EEGMeasurements ............................... 121 5.2.4 DataPreprocessing... ... .... .... .... ... .... .... 121 5.2.5 FeatureExtraction ... ... .... .... .... ... .... .... 122 5.2.6 Results ....................................... 124 5.2.7 Leave-one-outAnalysis . 126 5.3 Discussion...................................... 127 5.4 Experiment 2: Classification of genre from MEG recordings ...............128 5.4.1 Participants.................................. 130 5.4.2 Design .......................................130 5.4.3 Procedure..................................... 130 5.4.4 FeatureExtraction ... ... .... .... .... ... .... .... 131 5.4.5 Results ....................................... 132 5.4.6 Discussion .................................... 134 6 Conclusions 135 6.1 Conclusions..................................... 135 6.1.1 Greedymethods ................................. 135 6.1.2 Low-rankapproximationmethods . .......135 6.1.3 Multiviewmethods... ... .... .... .... ... .... .... 136 6.1.4 Experimentalapplications . .......136 6.2 FurtherWork ..................................... 139 6.2.1 Synthesis of greedy/Nystr¨om methods and Multi-View Learning (MVL) methods 139 6.2.2 Nonlinear Dynamics of Chaotic and Stochastic Systems .............140 6.3 One-classFisherDiscriminantAnalysis . .............141 6.4 SummaryandConclusions . ......142 A Mathematical Addenda 143 Contents 8 B Acronyms 144 Bibliography 147 List of Figures 2.1 Modularityofkernelmethods . ........ 19 2.2 StructuralRiskMinimisation . .......... 24 2.3 Minimisationontonormballs. ......... 26 2.4 Some examples of convex loss functions used in classification .............. 28 2.5 CommontasksinDigitalSignalProcessing . ............ 39 3.1 Diagrammatic view of the process of machine learning from multivariate signals . 50 3.2 Diagrammatic representation of the Nystr¨om method . ................ 54 3.3 Plot of generalisation error bound for different values of k usingRBFkernels . 62 3.4 Plot showing how the norm of the deflated kernel matrix and the test error vary with k .. 63 3.5 Generalisation error bound for the ‘Boston housing’ dataset ............... 69 ǫ/2 8 3.6 Plotof f(ǫ)=(1 ǫ/2) ln(1 + 1 ǫ ) ǫ/2 and f(ǫ)= ǫ for ǫ 0, 0.5 ...... 76 − − − ∈{ } 3.7 Error and run-time as a function of k on‘BreastCancer’forNFDA,KFDA . 79 3.8 Error and run-time as a function of k on‘FlareSolar’forNFDA,KFDA . 80 3.9 Error and run-time as a function of k on‘Bodyfat’byKRR,NRRandKRR . 81 3.10 Error and run-time as a function of k for‘Housing’byKRR,NRRandKRR . 82 3.11 Diagrammatic view of the process of a) MSL, b) MVL and c) MKL ........... 83 3.12 Plates diagram showing the hierarchical Bayesian interpretationofMFDA . 87 3.13 WeightsgivenbyMFDAandSMFDAonthetoydataset . ........... 94 3.14 Average precision recall curves for 3 VOC 2007 datasets forSMFDAandPicSOM . 95 4.1 Confusion Matrix of human performance on Anders Meng datasetd004 . 106 4.2 ThemodifiedreceiverchainforCSradar. ...........109 4.3 Fast-timesamplesofthestationarytarget. ...............112 4.4 Rangeprofilesofthestationarytarget. .............113 4.5 Fast-time samples constructed from largest three coefficients.. 113 4.6 Range profiles constructed from largest three coefficients..................114 4.7 The range-frequencysurfaces for the moving targets. .................114 LISTOFFIGURES 10 4.8 Range-frequencysurfacesfor
Recommended publications
  • CMPSCI 585 Programming Assignment 1 Spam Filtering Using Naive Bayes
    CMPSCI 585 Programming Assignment 1 Out: February 5, 2003 Due: February 17, 2003 Spam Filtering using Naive Bayes Naive Bayes is a simple, effective machine learning solution to the problem of document classification. For this assignment, you will implement a naive Bayes classifier to classify email messages as either spam (junk mail) or ham (legitimate messages). Data The data was collected from http://spamassassin.org/publiccorpus/. For this assignment, use the slightly modified version found at http://canberra.cs.umass.edu/~culotta/cs585/ass1-data.tgz (8.3M). The data consists of 4150 ham messages and 1897 spam messages, with original header information intact. Tasks Read data: Read in all messages and store them in some efficient manner. To do this you must decide how to tokenize each message – i.e. designate which characters indicate the start of a new “word.” See http://www.paulgrahm.com/spam.html for one way of doing this. (You will probably also assign each unique word an integer index you can use as an index into arrays of word counts and word proabilities.) Split Data: Randomly split the messages into a training set (70% of the messages) and a testing set (30%). Train Classifier: Using the training set only, estimate and store the prior class distributions P (spam) and P (ham), as well as the conditional probability distributions P (w|spam) and P (w|ham). Test Classifier: Classify each message in the testing set as spam or ham according to the Naive Bayes formulation. Note that you will most likely run out of floating-point resolution unless you do the product over words 1 Spam Ham Spam TP FP Ham FN TN Table 1: Confusion Matrix: TP = “true positive”, TN = “true negative”, FP = “false positive”, ‘FN = “false negative”.
    [Show full text]
  • Models of Statistical Self-Similarity for Signal and Image Synthesis Seungsin Lee
    Rochester Institute of Technology RIT Scholar Works Theses Thesis/Dissertation Collections 2004 Models of statistical self-similarity for signal and image synthesis Seungsin Lee Follow this and additional works at: http://scholarworks.rit.edu/theses Recommended Citation Lee, Seungsin, "Models of statistical self-similarity for signal and image synthesis" (2004). Thesis. Rochester Institute of Technology. Accessed from This Dissertation is brought to you for free and open access by the Thesis/Dissertation Collections at RIT Scholar Works. It has been accepted for inclusion in Theses by an authorized administrator of RIT Scholar Works. For more information, please contact [email protected]. Models of Statistical Self-Similarity for Signal and Image Synthesis by Seungsin Lee B.S., Electrical Engineering, Yonsei University, Seoul, Korea, 1994 M.S., Electrical Engineering, Rochester Institute of Technology, Rochester, NY, 2000 A dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in the Chester F. Carlson Center for Imaging Science Rochester Institute of Technology July, 2004 Signature of the Author ________Lee Seungsin--=- __________ _ Harvey E. Rhody Accepted by 2H~oov Coordinator, Ph.D. Degree Program Date CHESTER F. CARLSON CENTER FOR IMAGING SCIENCE ROCHESTER INSTITUTE OF TECHNOLOGY ROCHESTER, NEW YORK CERTIFICATE OF APPROVAL Ph.D. DEGREE DISSERTATION The Ph.D. Degree Dissertation of Seungsin Lee has been examined and approved by the dissertation committee as satisfactory for the dissertation required for the Ph.D. degree in Imaging Science RaghuveerRao Dr. Raghuveer Rao, dissertation Advisor Daniel Phillips Dr. Daniel B. Phillips John R. Schott Dr. John Schott Navalgund Rao Dr. N avalgund Rao ( Date ii DISSERTATION RELEASE PERMISSION ROCHESTER INSTITUTE OF TECHNOLOGY CHESTER F.
    [Show full text]
  • Using Machine Learning to Improve Dense and Sparse Matrix Multiplication Kernels
    Iowa State University Capstones, Theses and Graduate Theses and Dissertations Dissertations 2019 Using machine learning to improve dense and sparse matrix multiplication kernels Brandon Groth Iowa State University Follow this and additional works at: https://lib.dr.iastate.edu/etd Part of the Applied Mathematics Commons, and the Computer Sciences Commons Recommended Citation Groth, Brandon, "Using machine learning to improve dense and sparse matrix multiplication kernels" (2019). Graduate Theses and Dissertations. 17688. https://lib.dr.iastate.edu/etd/17688 This Dissertation is brought to you for free and open access by the Iowa State University Capstones, Theses and Dissertations at Iowa State University Digital Repository. It has been accepted for inclusion in Graduate Theses and Dissertations by an authorized administrator of Iowa State University Digital Repository. For more information, please contact [email protected]. Using machine learning to improve dense and sparse matrix multiplication kernels by Brandon Micheal Groth A dissertation submitted to the graduate faculty in partial fulfillment of the requirements for the degree of DOCTOR OF PHILOSOPHY Major: Applied Mathematics Program of Study Committee: Glenn R. Luecke, Major Professor James Rossmanith Zhijun Wu Jin Tian Kris De Brabanter The student author, whose presentation of the scholarship herein was approved by the program of study committee, is solely responsible for the content of this dissertation. The Graduate College will ensure this dissertation is globally accessible and will not permit alterations after a degree is conferred. Iowa State University Ames, Iowa 2019 Copyright c Brandon Micheal Groth, 2019. All rights reserved. ii DEDICATION I would like to dedicate this thesis to my wife Maria and our puppy Tiger.
    [Show full text]
  • Arxiv:1902.03680V3 [Cs.LG] 17 Jun 2019 1
    Learning From Noisy Labels By Regularized Estimation Of Annotator Confusion Ryutaro Tanno1 ∗ Ardavan Saeedi2 Swami Sankaranarayanan2 Daniel C. Alexander1 Nathan Silberman2 1University College London, UK 2Butterfly Network, New York, USA 1 2 fr.tanno, [email protected] fasaeedi,swamiviv,[email protected] Abstract tations is modest or the tasks are ambiguous. For example, The predictive performance of supervised learning algo- vision applications in medical image analysis [2] require an- rithms depends on the quality of labels. In a typical label notations from clinical experts, which incur high costs and collection process, multiple annotators provide subjective often suffer from high inter-reader variability [3, 4, 5, 6]. noisy estimates of the “truth” under the influence of their However, if the exact process by which each annotator varying skill-levels and biases. Blindly treating these noisy generates the labels was known, we could correct the an- labels as the ground truth limits the accuracy of learning al- notations accordingly and thus train our model on a cleaner gorithms in the presence of strong disagreement. This prob- set of data. Furthermore, this additional knowledge of the lem is critical for applications in domains such as medical annotators’ skills can be utilized to decide on which exam- imaging where both the annotation cost and inter-observer ples to be labeled by which annotators [7, 8, 9]. Therefore, variability are high. In this work, we present a method for methods that can accurately model the label noise of anno- simultaneously learning the individual annotator model and tators are useful for improving not only the accuracy of the the underlying true label distribution, using only noisy ob- trained model, but also the quality of labels in the future.
    [Show full text]
  • Performance Metric Elicitation from Pairwise Classifier Comparisons Arxiv:1806.01827V2 [Stat.ML] 18 Jan 2019
    Performance Metric Elicitation from Pairwise Classifier Comparisons Gaurush Hiranandani† Shant Boodaghians† Ruta Mehta† [email protected] [email protected] [email protected] Oluwasanmi Koyejo† [email protected] January 21, 2019 Abstract Given a binary prediction problem, which performance metric should the classifier optimize? We address this question by formalizing the problem of Metric Elicitation. The goal of metric elicitation is to discover the performance metric of a practitioner, which reflects her innate rewards (costs) for correct (incorrect) classification. In particular, we focus on eliciting binary classification performance metrics from pairwise feedback, where a practitioner is queried to provide relative preference between two classifiers. By exploiting key geometric properties of the space of confusion matrices, we obtain provably query efficient algorithms for eliciting linear and linear-fractional performance metrics. We further show that our method is robust to feedback and finite sample noise. 1 Introduction Selecting an appropriate performance metric is crucial to the real-world utility of predictive machine learning. Specialized teams of statisticians and economists are routinely hired in the arXiv:1806.01827v2 [stat.ML] 18 Jan 2019 industry to monitor many metrics { since optimizing the wrong metric directly translates into lost revenue [6]. Medical predictions are another important application, where ignoring cost sensitive trade-offs can directly impact lives [23]. Unfortunately, there is scant formal guidance
    [Show full text]
  • MASTER TP^ Ostrlbunon of THIS DOCUMENT IS UNL.M.TED Kaon Content of Three-Prong Decays of the Tan Lepton
    LBL--30035 DE91 007647 Kaon Content of Three-Prong Decays of the Tau Lepton* James Jackson Eastman Ph.D. Thesis "This work was supported by the Director, Office of Energy Research, Office of High Energy and Nuclear Physics, Division of High Energy Physics, of the U.S. Department of Energy under Contract No. DE-AC03-76SF00098. MASTER TP^ OSTRlBUnON OF THIS DOCUMENT IS UNL.M.TED Kaon Content of Three-Prong Decays of the Tan Lepton James Jackson Eastman Ph.D. Thesis Department of Physics University of California, Berkeley, CA 94720 and Lawrence Berkeley Laboratory, University of California, Berkeley, CA 94720 December 2, 1990 Abstract We present a series of measurements involving the production of charged kaons in three-prong hadronic decays of the T lepton. The data sample was obtained with the TPC/Two-Gamma detector facility at PEP. We set a limit on the branch­ ing fraction BR(T~ -> vrK~K°) < 0.26% at the 95% confidence level. The pro­ cess T~ —» v,K~Ka is related via SU(3) to the second-class current decay r~ —» vTit~t). We also present new measurements of the three-prong branching frac­ + + tions BR{T~ -» vTK~ir Tr- + neutrals) = 0.70 ±8:i?% and BR(T~ -> t/TK-K ir + neutrals) = 0.16 +g:J?%. This work is supported by the United States Department of Energy under Contract DE-AC03-76SF00098. ii FOR Ann. Acknowledgements HIGH-ENERGY PHYSICS EXPERIMENTS are great big projects, and I want to ex­ press my gratitude to the many scientists, engineers, technicians and programmers who built and ran the PEP4/9 experiment.
    [Show full text]
  • Designing Alternative Representations of Confusion Matrices to Support Non-Expert Public Understanding of Algorithm Performance
    153 Designing Alternative Representations of Confusion Matrices to Support Non-Expert Public Understanding of Algorithm Performance HONG SHEN, Carnegie Mellon University, USA HAOJIAN JIN, Carnegie Mellon University, USA ÁNGEL ALEXANDER CABRERA, Carnegie Mellon University, USA ADAM PERER, Carnegie Mellon University, USA HAIYI ZHU, Carnegie Mellon University, USA JASON I. HONG, Carnegie Mellon University, USA Ensuring effective public understanding of algorithmic decisions that are powered by machine learning techniques has become an urgent task with the increasing deployment of AI systems into our society. In this work, we present a concrete step toward this goal by redesigning confusion matrices for binary classification to support non-experts in understanding the performance of machine learning models. Through interviews (n=7) and a survey (n=102), we mapped out two major sets of challenges lay people have in understanding standard confusion matrices: the general terminologies and the matrix design. We further identified three sub-challenges regarding the matrix design, namely, confusion about the direction of reading the data, layered relations and quantities involved. We then conducted an online experiment with 483 participants to evaluate how effective a series of alternative representations target each of those challenges in the context of an algorithm for making recidivism predictions. We developed three levels of questions to evaluate users’ objective understanding. We assessed the effectiveness of our alternatives for accuracy in answering those questions, completion time, and subjective understanding. Our results suggest that (1) only by contextualizing terminologies can we significantly improve users’ understanding and (2) flow charts, which help point out the direction ofreading the data, were most useful in improving objective understanding.
    [Show full text]
  • A Generalized Hierarchical Approach for Data Labeling Zachary D. Blanks
    A Generalized Hierarchical Approach for Data Labeling by Zachary D. Blanks B.S. Operations Research, United States Air Force Academy Submitted to the Sloan School of Management in partial fulfillment of the requirements for the degree of Master of Science in Operations Research at the MASSACHUSETTS INSTITUTE OF TECHNOLOGY June, 2019 © Zachary D. Blanks, 2019. All rights reserved. The author hereby grants to MIT and DRAPER permission to reproduce and to distribute publicly paper and electronic copies of this thesis document in whole or in part in any medium now known or hereafter created. Author....................................................................................... Sloan School of Management May 17, 2019 Certifiedby.................................................................................. Dr. Troy M. Lau The Charles Stark Draper Laboratory Technical Supervisor Certifiedby.................................................................................. Prof. Rahul Mazumder Assistant Professor of Operations Research and Statistics Thesis Supervisor Acceptedby.................................................................................. Prof. Dimitris Bertsimas Boeing Professor of Operations Research Co-Director, Operations Research Center 1 THIS PAGE INTENTIONALLY LEFT BLANK 2 A Generalized Hierarchical Approach for Data Labeling by Zachary D. Blanks Submitted to the Sloan School of Management on May 17, 2019 in partial fulfillment of the requirements for the degree of Master of Science in Operations Research Abstract The goal of this thesis was to develop a data type agnostic classification algorithm best suited for problems where there are a large number of similar labels (e.g., classifying a port versus a shipyard). The most common approach to this issue is to simply ignore it, and attempt to fit a classifier against all targets at once (a “flat” classifier). The problem with this technique is that it tends to do poorly due to label similarity.
    [Show full text]
  • PREDICTING ACADEMIC PERFORMANCE of POTENTIAL ELECTRICAL ENGINEERING MAJORS a Thesis by SANJHANA SUNDARARAJ Submitted to the Offi
    PREDICTING ACADEMIC PERFORMANCE OF POTENTIAL ELECTRICAL ENGINEERING MAJORS A Thesis by SANJHANA SUNDARARAJ Submitted to the Office of Graduate and Professional Studies of Texas A&M University in partial fulfillment of the requirements for the degree of MASTER OF SCIENCE Chair of Committee, Tie Liu Co-Chair, Xiaoning Qian Ulisses Braga Neto Xia Hu Head of Department, Miroslav Begovic August 2017 Major Subject: Electrical Engineering Copyright 2017 Sanjhana Sundararaj ABSTRACT Data Analytics for education is fast growing into an important part of higher learning institutions, which helps to improve student success rate and decision-making with regards to teaching methods, course selection, and student retention. The undergraduate program at Texas A&M University requires students to take up a general engineering program during their freshman and sophomore years. During the course of this period, student academic performance, abilities and participation is assessed. As per the Entry-to-a-Major policy, departments place the students in the best possible ma- jor based on their displayed capacities and in alignment with their goals. Our focus is on the Electrical Engineering department and the success rate of students with aspirations and background in this major. An approach to improve student retention rate is to predict beforehand the performance of students in specific course disciplines based on the infor- mation that is mined from their previous records. Based on the outcome, decisions can be made in advance regarding their further enrollment in the area and need for specific attention in certain aspects to get students up to the benchmark. In this thesis, we put together a set attributes related to students in the general pro- gram and with an electrical engineering aligned background.
    [Show full text]
  • Jets + Missing Energy Signatures at the Large Hadron Collider
    Jets + Missing Energy Signatures At The Large Hadron Collider DISSERTATION Presented in Partial Fulfillment of the Requirements for the Degree Doctor of Philosophy in the Graduate School of The Ohio State University By Khalida S. Hendricks, M.S. Graduate Program in Physics The Ohio State University 2019 Dissertation Committee: Linda Carpenter, Advisor Amy Connolly Annika Peter Antonio Boveia c Copyright by Khalida S. Hendricks 2019 Abstract In this work we consider new ways to use jets plus missing energy signatures in searches at the Large Hadron Collider. We study the Higgs boson (h) decay to two light jets at the 14 TeV High-Luminosity- LHC (HL-LHC), where a light jet (j) represents any non-flavor tagged jet from the obser- vational point of view. We estimate the achievable bounds on the decay product branching fractions through the associated production V h (V = W ±;Z). As a reasonable estimation, we only focus on the boosted region of high pT (h) and the three leptonic decay channels of the vector boson. We find that with 3000 fb−1 data at the HL-LHC, we should expect approximately 1σ statistical significance on the SM V h(gg) signal in this channel. This cor- responds to a reachable upper bound BR(h jj) 4 BRSM (h gg) at 95% confidence ! ≤ ! level. A consistency fit also leads to an upper bound of BR(h cc) < 15 BRSM (h cc) ! ! at 95% confidence level. The estimated bound may be further strengthened by adopting multiple variable analyses, or adding other production channels. We then consider some simple machine learning techniques applied to the same channels.
    [Show full text]
  • Multiple Classifier Systems Incorporating Uncertainty
    Christian Thiel Multiple Classifier Systems Incorporating Uncertainty The title page and cover show the painting doctors by Michael S. Parker (www.michaelsparker.com), 2003. Used with friendly permission of the artist. Cover layout by Kreativfrosch Grafik + Design (www.kreativfrosch.eu), Wegscheid bei Passau, Germany. Universit¨atUlm | 89069 Ulm | Germany Dissertation zur Erlangung des Doktorgrades Doktor der Naturwissenschaften (Dr. rer. nat.) der Fakult¨at f¨urIngenieurwissenschaften und Informatik der Universit¨at Ulm vorgelegt von Christian Andr´eThiel Angefertigt am Institut f¨urNeuroinformatik Einreichung der Dissertation: Dezember 2009 Tag der Promotion: 10. Juni 2010 Gutachter: Prof. Dr. G¨unther Palm Gutachter: Prof. Dr. Heiko Neumann Gutachter: Prof. Dr. Barbara Hammer Amtierender Dekan der Fakult¨at: Prof. Dr.-Ing. Michael Weber www.ChristianThiel.com Abstract The inclusion of uncertain class information into multi classifier systems (MCS) is the central theme in this thesis. A multi classifier system relies on multiple base classifiers, each of which is trained on a separate view of the problem at hand. Combining their answers will lead to a more accurate final decision. An example would be emotion recognition, with decisions based on observations of the mouth, the eyes or the pitch of the voice. Traditionally in classification one sample is associated with exactly one class, for example anger. But, in practical applications, such a hard distinction is not warranted; instead a sample should have soft class memberships, thus being associated fuzzily with multiple classes at the same time. The inclusion of this uncertain information into various, but isolated building blocks of a MCS has been tackled by a great many researchers.
    [Show full text]
  • Models of Statistical Self-Similarity for Signal and Image Synthesis
    Rochester Institute of Technology RIT Scholar Works Theses 2004 Models of statistical self-similarity for signal and image synthesis Seungsin Lee Follow this and additional works at: https://scholarworks.rit.edu/theses Recommended Citation Lee, Seungsin, "Models of statistical self-similarity for signal and image synthesis" (2004). Thesis. Rochester Institute of Technology. Accessed from This Dissertation is brought to you for free and open access by RIT Scholar Works. It has been accepted for inclusion in Theses by an authorized administrator of RIT Scholar Works. For more information, please contact [email protected]. Models of Statistical Self-Similarity for Signal and Image Synthesis by Seungsin Lee B.S., Electrical Engineering, Yonsei University, Seoul, Korea, 1994 M.S., Electrical Engineering, Rochester Institute of Technology, Rochester, NY, 2000 A dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in the Chester F. Carlson Center for Imaging Science Rochester Institute of Technology July, 2004 Signature of the Author ________Lee Seungsin--=- __________ _ Harvey E. Rhody Accepted by 2H~oov Coordinator, Ph.D. Degree Program Date CHESTER F. CARLSON CENTER FOR IMAGING SCIENCE ROCHESTER INSTITUTE OF TECHNOLOGY ROCHESTER, NEW YORK CERTIFICATE OF APPROVAL Ph.D. DEGREE DISSERTATION The Ph.D. Degree Dissertation of Seungsin Lee has been examined and approved by the dissertation committee as satisfactory for the dissertation required for the Ph.D. degree in Imaging Science RaghuveerRao Dr. Raghuveer Rao, dissertation Advisor Daniel Phillips Dr. Daniel B. Phillips John R. Schott Dr. John Schott Navalgund Rao Dr. N avalgund Rao ( Date ii DISSERTATION RELEASE PERMISSION ROCHESTER INSTITUTE OF TECHNOLOGY CHESTER F.
    [Show full text]