<<

Index

˛, see significance level deep learning, 192 ˇ2, see Asimov dataset, 231 , see unnormalized asymmetric errors, 99, 110 1, see skewness combination of, 123 2, see excess asymptotic formulae for test , ", see efficiency 231 . see Gaussian average value or signal average value strength continuous case, 27 , see correlation coefficient discrete case, 12 , see or Gaussian standard in , 69 deviation , see lifetime ˚, see Gaussian cumulative distribution back propagation, neural network, 191 2 background distribution, 32 dependence of Feldman–Cousins upper method, 114 limits, 218, 220 binned case, 119 determination from control regions, in multiple dimensions, 132 129 , 32, 114, 120 fluctuation for significance level, 205 Baker–Cousins, 120 in convolution and unfolding, 160 Neyman’s, 119 modeling in extended likelihood, 107 Pearson’s, 119 modeling with Argus function, 43 , see sample space rejection in hypothesis test, 176 3 evidence, 207 treatment in iterative unfolding, 171 5 observation, 207 uncertainty in significance evaluation, 209 uncertainty in test , 227, 236 Baker–Cousins 2, 120 activation function, 191 , 73 adaptive boosting, 198 Bayes’ theorem, 59 AI, artificial intelligence, 195 learning process, 67 , 175 Bayesian Anderson–Darling test, 184 inference, 68 Argus function, 43 , 59, 64 artificial intelligence, 195 visual derivation, 60 artificial neural network, 181, 190 unfolding, 166

© Springer International Publishing AG 2017 251 L. Lista, Statistical Methods for Analysis in Particle Physics, Lecture Notes in Physics 941, DOI 10.1007/978-3-319-62840-0 252 Index

BDT, see boosted decision trees Clopper–Pearson binomial interval, 147 Bernoulli CLs method, 221 , 17 CNN, see convolutional neural network random process, 16 coefficient of determination R2, 117 random variable, 17 combination Bertrand’s paradox, 7 of measurements, 129 best linear unbiased estimator, 133 principle, 136, 140 conservative correlation assumption, 137 conditional intrinsic information weight, 136 distribution, 53 iterative application, 139 probability, 9 marginal information weight, 136 confidence negative weights, 135, 137 interval, 100, 109, 143 relative importance, 136 level, 100 , 83 conservative bias, 102 CLs method, 221, 223 in maximum likelihood estimators, 113 correlation assumption, BLUE method, 137 bifurcated Gaussian, 124 interval, 147 bimodal distribution, 14 limit, 217 bin migration, 158 consistency of an estimator, 102 binned Poissonian fit, 120 control binning, 118 region, 129 in convolution, 158 sample, 130 binomial convergence in probability, 22 coefficient, 18 ConvNet, see convolutional neural network interval, 147 convolution, 155 probability distribution, 18, 147 convolutional neural network, 194 Poissonian limit, 40 Fourier transform, 156 random process, 17 convolutional neural network, 193 random variable, 18 feature map, 194 BLUE, see best linear unbiased estimator local receptive fields, 194 boosted decision trees, 181 correlation coefficient, 14 adaptive boosting, 198 counting , 208, 212, 216, 227 boosting, 198 Cousins–Highlands method, 227 cross entropy, 196 , 14 decision forest, 197 matrix, 14 Gini index, 196 coverage, 100 leaf, 196 Cramér–Rao bound, 102 node, 196 Cramér–von Mises test, 184 boosting, boosted decision trees, 198 , 70 Box–Muller transformation, 89 cross entropy, decision tree, 196 Brazil plot, 225 Crystal Ball function, 44 breakdown point, robust estimator, 103 cumulative distribution, 28 Breit–Wigner cut, 176 non-relativistic distribution, 41 relativistic distribution, 42 data sample, 99 decision , 41 forest, 197 central tree, 196 interval, 70 deep learning, artificial neural network, 192 limit theorem, 46 degree of belief, 65 value, 99 dices, 4–6, 16, 21 chaotic regime, 82 differential probability, 26 classical probability, 4 discovery, 205, 207, 208 Index 253 distribution, see probability distribution flat (uniform) distribution, 6, 30 dogma, extreme Bayesian prior, 66 flip-flopping, 150 drand48 function from C standard library, 84 forest, boosted decision trees, 197 Fourier transform of PDF convolution, 156 efficiency frequentist hit-or-miss Monte Carlo, 90 inference, 100 of a detector, 10, 158 probability, 3, 22 estimate, 104 full width at half maximum, 31, 41 of an estimator, 102 fully asymmetric interval, 70 elementary event, 4, 6, 9 FWHM, full width at half maximum, , 4, 6, 25 31 ergodicity, 94 error of a measurement gamma function, 33, 71 Bayesian approach, 70 Gaussian frequentist approach, 99 average value, ,31 of the first kind, 177 bifurcated, 124 of the second kind, 177 contours in two dimensions, 55 propagation cumulative distribution, 31 Bayesian case, 79 distribution, 31 frequentist case, 121 in more dimensions, 54 simple cases, 121 intervals, 32, 58 estimate, 68, 97, 99, 100 , 108 estimator, 100 random number generator, 89 efficiency, 102 , 88 maximum likelihood, 105 standard deviation, ,31 properties, 101 generator, see pseudorandom number generator robust, 103 Gini index, decision tree, 196 Euler characteristic, 246 global significance level, 242 event, 2 goodness of fit, 33, 118, 120 counting experiment, 187, 206 gsl_rng_rand function from GSL library, elementary, 4, 6, 9 84 in physics, 105 in statistics, 2 independent, 10 Hastings ratio, 93 evidence , 119 3 significance level, 207 convolution, 158 Bayes factor, 73 in Asimov dataset, 231 excess, 15 PDF approximation, 182 exclusion, 211 hit-or-miss Monte Carlo, 90 . see average value homogeneous Markov chain, 93 , 34 Hui’s triangle, 18, 23 random number generator, 87 hybrid frequentist approach, 227 extended likelihood function, 106, 186 hypothesis test, 175 fast Fourier transform, 156 IID, independent identically distributed feature map, convolutional neural network, 194 random variables, 82 feedforward multilayer perceptron, 190 IIW, intrinsic information weight, BLUE Feldman–Cousins unified intervals, 152 method, 136 FFT. see fast Fourier transform importance , 91 , 75, 102, 136 improper prior distribution, 76 Fisher’s linear discriminant, 178 incomplete Gamma function, 41 254 Index independent , 115 and identically distributed random local variables, 82, 106 receptive fields, 194 events, 10 significance level, 210, 242 random variables, 50 log , 33 inference, 97 logistic map, 82 Bayesian, 68 look elsewhere effect, 210, 242 intersubjective probability, 75 in more dimensions, 246 intrinsic information weight, BLUE method, Lorentz distribution, 41 136 , 191 invariant prior, see Jeffreys’ prior lower limit, 70 iterative unfolding, 166 lrand48 function from C standard library, 84

Jeffreys’ prior, 75 machine learning, 188 joint probability distribution, 49 observation, 189 supervised, 188 kernel function, see response function unsupervised, 188 Kolmogorov distribution, 183 marginal Kolmogorov–Smirnov test, 182 distribution, 49 kurtosis, 15 information weight, BLUE method, 136 coefficient, 15 Markov chain, 93 homogeneous, 93 Monte Carlo, 69, 93 L’Ecuyer pseudorandom number generator, 84 maximum likelihood L-curve, 165 estimator, 105 Lüscher pseudorandom number generator, 84 bias, 113 , 46 properties, 112 large numbers, law of, 21 method, 69, 105 law uncertainty, 109 of large numbers, 21 MC, see Monte Carlo of total probability, 11 MCMC, , 93 leaf, decision tree, 196 , 14, 28, 103 learning Mersenne-Twistor pseudorandom number process in , 67 generator, 84 rate parameter, artificial neural network, Metropolis–Hastings 191 algorithm, 93 method, 114 proposal distribution, 93 lifetime, 35, 39 ratio, 95 Bayesian inference, 76 minimum , 77 2 method, see 2 method maximum likelihood estimate, 112 bound, 102 measurement combination, 140 MINUIT, 106, 110 likelihood misidentification probability, 176 function, 67, 105 MIW, marginal information weight, BLUE extended, 106, 186 method, 136 Gaussian, 108 , 14, 28 in Bayesian probability, 67 modified frequentist approach, 221 ratio , 6, 46, 69, 81 discriminant, 181 hit-or-miss, 90 in search for new signals, 185, 209 numerical integration, 92 projective discriminant, 182 sampling, 89 test statistic in Neyman–Pearson multilayer perceptron, 190 lemma, 181 multimodal distribution, 14, 28 Index 255 , 20 pooling, convolutional neural network, 194 multivariate analysis, 178 posterior MVA, multivariate analysis, 178 odds, 64, 73 probability, 60, 65, 67 prior negative weights, 135, 137 odds, 73 nested hypotheses, see Wilks’ theorem probability, 60, 65 neural network, see artificial neural network distribution, 67 Neyman distribution, improper, 76 confidence belt distribution, uniform, 71, 74 binomial case, 147 subjective choice, 74 construction, 144 uninformative, 69, 75 Feldman–Cousins, 152, 218 probability, 2 Gaussian case, 146 axiomatic definition, 8 inversion, 146 Bayesian, 3, 4 confidence intervals, 215 classical, 4 Neyman’s 2, 119 density, 25 Neyman–Pearson lemma, 181 dice rolls, 5 node, decision tree, 196 distribution, 9, 25 normal 2,32 distribution, see Gaussian distribution Bernoulli, 17 random variable, 31 beta, 83 normalization condition, 9, 26 bimodal, 14 nuisance parameter, 69, 98, 226, 227 binomial, 18 null hypothesis, 175 Breit–Wigner, non-relativistic, 41 Breit–Wigner, relativistic, 42 cumulative, 28 observation in machine learning, 189 exponential, 34 observation, 5 significance level, 207 Gaussian, 31 odds Gaussian, in more dimensions, 54 posterior, 64, 73 joint, 49 prior, 73 Landau, 46 ordering rule, 144 log normal, 33 outlier, 103 Lorentz, 41 overcoverage, 100, 144, 146, 147, 217 marginal, 49 multimodal, 14, 28 multinomial, 20 p-value, 118, 206 normal, 31 parameter Poissonian, 35 estimate, 100 standard normal, 31 Bayesian, 68 uniform, 6, 30 nuisance, 226, 227 distribution function, 26 of interest, 69, 98 in more dimensions, 49 Pascal’s triangle, 23 frequentist, 3, 22 PDF, see probability distribution function posterior, 60, 65, 67 Pearson’s 2, 119 prior (see ) , 28 theory, 2 period of a pseudorandom number generator, profile likelihood, 185, 228 84 projective likelihood ratio discriminant, 182 POI, parameter of interest, 69, 98 pseudorandom number, 81 , 35 generator, 82 Gaussian limit, 40 drand48 function, 84 Poissonian, see Poisson distribution gsl_rng_rand function, 84 random variable, 35 lrand48 function, 84 256 Index

exponential, 87 response from cumulative inversion, 86 function, 155 Gaussian, Box–Muller, 89 Gaussian case, 157 Gaussian, central limit theorem, 88 matrix, 158 L’Ecuyer, 84 RMS, see root square Lüscher, 84 robust estimator, 103 Mersenne-Twistor, 84 ROC curve, receiver operating characteristic, period, 84 177 RANLUX,84 ROOT, 106, 173 seed, 84 root mean square, 13 uniform, 84 ROOUNFOLD, 173 uniform on a sphere, 87 purity, 64, 196 sample space, 8 seed, pseudorandom number generator, 84 quantile, 28 selection, 176 efficiency, 176 shortest interval, 70 R2, coefficient of determination, 117 sigmoid function, 191 RANLUX pseudorandom number generator, 84 signal random exclusion, 211 number, see pseudorandom number region, 129 generator, see pseudorandom number strength, 185, 186 generator signal-injected expected limit, 225 process, 2 significance level, 177, 205, 207–210 variable, 4 simultaneous fit, 130 2,32 singular value decomposition, 171 Bernoulli, 17 skewness, 14 binomial, 18 unnormalized, 15, 126 exponential, 34 smearing, 156 Gaussian, 31 sources of systematic uncertainty, 100 independent, 50 standard log normal, 33 deviation, 13 normal, 31 continuous case, 27 Poissonian, 35 normal standard normal, 31 distribution, 31 uncorrelated, 14 random variable, 31 uniform, 6, 30 statistical uncertainty, 100 Random forest, 197 subjective probability, see Bayesian probability rate parameter, 35 supervised machine learning, 188 receiver operating characteristic, 177, 200 SVD, see singular value decomposition171 rectified linear units, convolutional neural symmetric interval, 70 network, 195 systematic uncertainty, 99, 226, 227 reference analysis, 76 sources, 100 reference prior, 76 regularization strength, 164 regularized unfolding, 163 Tartaglia’s triangle, 23 relative importance, BLUE method, 136 test religious belief, Bayesian extreme probability, sample, 189 66 statistic, 176 ReLU, rectified linear units, 195 for discovery, 230 repeatable experiment, 2 for Higgs boson search, 231 residuals, 114 for positive signal strength, 230 resonance, 41 for upper limits, 230 Index 257

Tikhonov regularization, 164 regularization strength, 164 total probability, law of, 11 regularized, 163 toy Monte Carlo, 120 response matrix inversion, 160 training, 188 singular value decomposition, 171 sample, 180, 182, 188 , 164 transformation of variables, 15, 29, 121 unified intervals, Feldman–Cousins, 152, 218 Bayesian posterior, 79 uniform trial factor, 243 distribution, 6, 30 trimmed average, 103 random number generator, 84 TUNFOLD, 173 uninformative prior, 69, 75 type-I error, 177 unknown parameter, 97 type-II error, 177 unnormalized skewness, 15, 126 unsupervised machine learning, 188 upcrossing, 244 uncertainty, 68, 70, 97, 99 upper limit, 70, 211, 215 interval, 99 Bayesian, 70 frequentist, 100 variance with maximum likelihood method, 109 continuous case, 27 uncorrelated random variables, 14 discrete case, 13 undercoverage, 100 underfluctuation, 211 unfolding, 155 weighted average, 131 L curve, 165 Wilks’ theorem, 120, 184 Bayesian, 166 nested hypotheses, 184 bin-to-bin correction factors, 163 in more dimensions, 173 iterative, 166 Z, see significance level