A Provable Algorithm for Learning Interpretable Scoring Systems

A Provable Algorithm for Learning Interpretable Scoring Systems

A Provable Algorithm for Learning Interpretable Scoring Systems Nataliya Sokolovska Yann Chevaleyre Jean-Daniel Zucker University Paris 6, INSERM, France University Paris Dauphine, France IRD Bondy, INSERM, France Abstract Variable Thresholds Score Age <40 0 Score learning aims at taking advantage of su- 40–49 1 pervised learning to produce interpretable mod- 50 – 59 2 els which facilitate decision making. Scoring >60 3 systems are simple classification models that let Glycated hemoglobin <6.5 0 users quickly perform stratification. Ideally, a 6.5 – 6.9 2 scoring system is based on simple arithmetic op- 7 – 8.9 4 erations, is sparse, and can be easily explained by > 96 human experts. Insuline No 0 Yes 10 In this contribution, we introduce an origi- Other drugs No 0 nal methodology to simultaneously learn inter- Yes 3 pretable binning mapped to a class variable, and the weights associated with these bins contribut- Classify as Remission if sum of scores < 7 Classify as Non-remission if sum of scores 7 ing to the score. We develop and show the theo- ≥ retical guarantees for the proposed method. We demonstrate by numerical experiments on bench- Table 1: The DiaRem Score to assess the outcome of the mark data sets that our approach is competitive bariatric surgery [24] compared to the state-of-the-art methods. We il- lustrate by a real medical problem of type 2 dia- betes remission prediction that a scoring system Clinical scoring systems learned automatically purely from data is compa- are of particular interest since they rable to one manually constructed by clinicians. are expected to predict a state of a patient and to help physi- cians to provide accurate diagnostics. An example of such a score, shown in Table 1, is the DiaRem score [24] which is a preoperative method to predict remission of type 2 di- 1 Introduction abetes after a gastric bypass surgery. The DiaRem is based on four clinical variables and a few thresholds per variable. Scoring systems are simple linear classification models that Only one arithmetic operation is involved into the DiaRem are based on addition, subtraction, and multiplication of computation: the scores are added, and if the sum is < 7, a few small numbers. These models are applied to make then a patient is likely to benefit from the surgery and to get quick predictions, without use of a computer. Tradition- the diabetes remission. Some other widely used medical ally, a problem in supervised machine learning is cast as scores are SAPS I, II, and III [10, 21] and APACHE I, II, III a binary or multi-class classification where the goal is to to assess intensive care units mortality risks [14], CHADS2 learn real-valued weights of a model. However, although to assess the risk of stroke [9]; TIMI to estimate the risk the generalizing error is an important criterion, in some ap- of death of ischemic events [2]. Despite widespread use in plications, the interpretability of a model plays even a more clinical routines, there has been no principled approach to significant role. Most machine learning methods produce learn scores from observational data. Most of existing clin- highly complex models, not designed to provide explana- ical scores are built by a panel of experts, or by combining tions about predictions. multiple heuristics. In many applications, although continuous features are Proceedings of the 21st International Conference on Artificial available for a prediction task, it is often beneficial to use Intelligence and Statistics (AISTATS) 2018, Lanzarote, Spain. discretized features or categories. Predictors that use cate- PMLR: Volume 84. Copyright 2018 by the author(s). gorical variables need smaller memory footprint, are easier A Provable Algorithm for Learning Interpretable Scoring Systems to interpret, and can be applied directly by a human expert interest for diagnostic purposes. to make a new prediction. The difficulty to learn discrete In our work, we cast the problem of binning as a feature classifiers is well known (see, e.g., [4]): minimizing a con- selection task, where to add a bin, i.e. to add a thresh- vex loss function with discrete weights is NP-complete. old, is equivalent to add a feature into a model. It is In this paper, we propose a principled approach to learn known that feature selection and data categorization can discrete scoring systems. Our approach is unique since slightly degrade performance relative to a real-valued pre- it learns both the thresholds to discretize continuous vari- dictor, however, in domains such as medical diagnostics, an ables, and the weights for the corresponding bins. The interpretable model is preferred to a complex real-valued weights can be also discretized with a randomized rounding model which is the most accurate, if their performances after training. To our knowledge, this paper is the first at- are comparable. It was demonstrated [23] that it is pos- tempt to learn a discrete scoring system which relies on si- sible to estimate sparse predictors efficiently while com- multaneous learning of bins and their corresponding scores. promising on prediction accuracy. Binning or supervised discretization was reported to simplify the models, and not The algorithm we provide has the best of two worlds: ac- to degrade the generalizing performance. Usually, binning curacy and interpretability. It is fully optimised for feature is performed as a pre-processing step before learning (see, selection, and it converges to an optimal solution. e.g., [6, 19, 18, 3, 20, 11]). This paper is organised as follows. We discuss the related Very recently, [1] introduced a new penalization called bi- work in Section 2. In Section 3, we introduce the novel narsity which penalizes the weights of a model learned algorithm and show its theoretical properties. The results from grouped one-hot encodings. Their approach is an at- of the numerical experiments are discussed in Section 4. tempt to learn an interpretable model using a penalty term. Concluding remarks and perspectives close the paper. 3 Learning Scoring Systems 2 Related Work In this section, we introduce a novel algorithm called Fully Our contribution is related to the new methods for inter- Corrective Binning (FCB) which efficiently performs both pretable machine learning. The SLIM (Supersparse Linear binning and continuous weights learning. We also discuss Integer Models) [27] is formulated as an integer program- how to produce a discrete scoring system, i.e. a model with ming task and optimizes directly the accuracy, the 0-1 loss, discrete weights after the fully corrective binning proce- and the degree of sparsity. However, optimizing the 0-1 dure. loss is NP-hard even with continuous weights, and training of a SLIM model on a large data set can be challenging. 3.1 Preliminaries Another modern avenue of research are Bayesian-based In a supervised learning scenario, an algorithm has access approaches to learn scoring systems. So, [7] introduced to training data X ,Y N ( )N , and the goal is a Bayesian model where a prior favours fewer significant { i i}i=1 2 X⇥Y to find a rule to discriminate observations into two or more digits, and, therefore, the solution is sparse. A Bayesian classes as accurate as possible. The matrix of observations model is also developed in [29] to construct a falling rule X has N rows (samples), and p columns (variables), and list, which is a list of simple if-then rules containing a let X [ ⌦, ⌦]. decision-making process and which stratifies patients from ij 2 − Definition 1. (Encodings). For any X , we define the the highest at-risk group to the lowest at-risk group. A sim- 2X ilar idea, also based on Bayesian learning is considered by interval encoding [16, 30] where the main motivation is to construct simple rules which are interpretable by human experts and can be 1, if Xij ]l, u] , Zijlu = 2 (3.1) used by healthcare providers. (0, otherwise. Recently, [28] proposed to solve the score learning task Therefore, Z could be viewed as a matrix with N rows and with a cutting plane algorithm which is computationally ef- an extended number of d columns (where d p) indexed ficient, since it iteratively solves a surrogate problem with by the triplets j, l, u. The j-th column X j is thus replaced · a linear approximation of the loss function. in Z by dj columns containing only zeros and ones. The state-of-the-art methods [27, 7, 16, 30, 28] are reported We will show later that our problem can be cast as learning to be accurate, but an obvious drawback is that their out- a linear prediction model on Z. This linear model will be put, the learned scores, apply to real-valued data (if the in- represented by a parameter vector ✓ ⇥ d put data were real). Although medical data are often real 2 ⇢ indeed, a model which provides some interpretable dis- Without loss of generality, we consider a binary classifica- cretization or learns diagnostic thresholds, is of a bigger tion problem, where 1, 1 . Y2{− } Nataliya Sokolovska, Yann Chevaleyre, Jean-Daniel Zucker var 1 ( , 1.6] ( 1.6, + ) 1 − − 1 X 1.6 Z 1 0 1 − 1 X2 2.2 Z2 0 1 Table 2: A one-dimensional dataset composed of two samples (on the left), and the interval encoding of the dataset (on the right). The learning problem is defined as the minimization of a For any scoring system in its minimal form, we define loss function `(., ., .) as follows: ✓ = ✓ ✓ . (3.10) N k kfused | jlr − jru| 1 j=1,...,p;l,r,u [ ⌦;⌦] R(✓)=min `(Zi,Yi,✓). (3.2) X 2 − ✓ ⇥ N 2 i=1 X For example, a possible scoring model ✓ for the data set The sparsity of the vector ✓ is defined as a number of non- presented in Table 2 could be zero elements in ✓, and is defined as the L0 norm: ✓1 1.6 = 2,✓1 1.6+ =2 , (3.11) ✓ = i : ✓ =0 .

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    9 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us