Basic Statistics and Monte-Carlo Method -2

Total Page:16

File Type:pdf, Size:1020Kb

Basic Statistics and Monte-Carlo Method -2 Applied Statistical Mechanics Lecture Note - 10 Basic Statistics and Monte-Carlo Method -2 고려대학교 화공생명공학과 강정원 Table of Contents 1. General Monte Carlo Method 2. Variance Reduction Techniques 3. Metropolis Monte Carlo Simulation 1.1 Introduction Monte Carlo Method Any method that uses random numbers Random sampling the population Application • Science and engineering • Management and finance For given subject, various techniques and error analysis will be presented Subject : evaluation of definite integral b I = ρ(x)dx a 1.1 Introduction Monte Carlo method can be used to compute integral of any dimension d (d-fold integrals) Error comparison of d-fold integrals Simpson’s rule,… E ∝ N −1/ d − Monte Carlo method E ∝ N 1/ 2 purely statistical, not rely on the dimension ! Monte Carlo method WINS, when d >> 3 1.2 Hit-or-Miss Method Evaluation of a definite integral b I = ρ(x)dx a h X X X ≥ ρ X h (x) for any x X Probability that a random point reside inside X O the area O O I N' O O r = ≈ O O (b − a)h N a b N : Total number of points N’ : points that reside inside the region N' I ≈ (b − a)h N 1.2 Hit-or-Miss Method Start Set N : large integer N’ = 0 h X X X X X Choose a point x in [a,b] = − + X Loop x (b a)u1 a O N times O O y = hu O O Choose a point y in [0,h] 2 O O a b if [x,y] reside inside then N’ = N’+1 I = (b-a) h (N’/N) End 1.2 Hit-or-Miss Method Error Analysis of the Hit-or-Miss Method It is important to know how accurate the result of simulations are The rule of 3σ’s Identifying Random Variable N = 1 X X n N n=1 From central mean theorem , X is normal variable in the limit of large N each X n X mean value : μ mean value : μ variance : σ 2 variance : σ 2 / N 1.2 Hit-or-Miss Method Sample Mean : estimation of actual mean value (μ) N μ = 1 ' xn N n=1 1 N N' μ = = = x01 ' xn r N = N n 1 p(x) r 1-r N = xn N' n=1 Accuracy of simulation the most probable error 0.6745σ σ = V (X ) N 1.2 Hit-or-Miss Method Estimation of error 0.6745σ σ = V (X ) N We do not know exact value of s , m We can also estimate the variance and the mean value from samples … 1 N σ '2 = (x − μ')2 − n N 1 n=1 1.2 Hit-or-Miss Method For present problem (evaluation of integral) exact answer (I) is known estimation of error is, V (X ) = E(X 2 ) − []E(X ) 2 E(X ) = xp(x) =1× r + 0×(1− r) = r = μ E(X 2 ) = x2 p(x) =1× r + 0×(1− r) = r V (X ) = r − r 2 = (1− r)r = σ 2 σ = (1− r)r I = r(a − b)h = μ(a − b)h 0.6745σ r(1− r) I((b − a)h − I) I HM = = 0.6745×(b − a)h = 0.6745 error N N N 1.3 Sample Mean Method ρ(x) is a continuous function in x and has a mean value ; b ρ (x)dx 1 b < ρ >= a = ρ(x)dx b − a dx b a a b ∴I = ρ(x)dx = (b − a) < ρ > a N < ρ >= 1 ρ (xn ) N n=1 = − + = xn (b a)un a un uniform random variable in [0,1] 1.3 Sample Mean Method Error Analysis of Sample Mean Method Identifying random variable N = 1 Y Yn N n=1 1 N 1 N μ =< ρ >≈ y = ρ(x ) Y n n N n=1 N n=1 Variance N < ρ 2 >= 1 ρ 2 (xn ) 2 2 V (ρ) =< ρ > −[]< ρ > N n=1 V (x) I SM = 0.6745σ = (b − a)×0.6745 error N 1.3 Sample Mean Method If we know the exact answer, b L ρ(x)2 dx − I 2 I SM = 0.6745 a error N 1.3 Sample Mean Method Start Set N : large integer s1 = 0, s2 = 0 Loop xn = (b-a) un + a N times ρ yn = (xn) 2 s1 = s1 + yn , s2 = s2 + yn Estimate mean μ’=s /N SM V ' 1 I = (b − a)×0.6745 Estimate variance V’ = s /N – μ’2 error 2 N End QUIZ Compare the error for the integral π / 2 π I = sin xdx = − cos x / 2 =1 0 0 using HM and SM method 0.6745σ r(1− r) I((b − a)h − I) I HM = = 0.6745×(b − a)h = 0.6745 error N N N b L ρ(x)2 dx − I 2 I SM = 0.6745 a error N Example : Comparison of HM and SM Evaluate the integral π / 2 π I = sin xdx = − cos x / 2 =1 0 0 I((b − a)h − I) 1((π / 2 − 0)×1−1) π / 2 −1 I HM = 0.6745 = 0.6745 = 0.6745 error N N N < ρ >= I /(b − a) = 2 /π 1 π / 2 < ρ 2 >= sin 2 xdx =1/ 2 b − a 0 1 2 V (ρ) = − 2 π 2 V ' (π 2 / 4)(1/ 2 − 4 /π 2 ) π 2 /8 −1 I SM = 0.6745(b − a) = 0.6745 = 0.6745 error N N N Example : Comparison of HM and SM Comparison of error π 2 SM −1 I error = 8 ≈ 0.64 SM method has 36 % less error than HM I HM π error −1 2 No of evaluation having the same error π 2 −1 SM = 8 HM ≈ HM SM method is more than twice faster than HM N π N 0.41N −1 2 2.1 Variance Reduction Technique - Introduction Monte Carlo Method and Sampling Distribution Monte Carlo Method : Take values from random sample From central limit theorem, 2 μ = μ σ = σ 2 / N 3s rule P(μ − 3σ ≤ X ≤ μ − 3σ ) ≈ 0.9973 Most probable error 0.6745σ Error ≈ ± N Important characteristics Error ∝1/ N Error ∝ σ 2.1 Variance Reduction Technique - Introduction Reducing error *100 samples reduces the error order of 10 Reducing variance Variance Reduction Technique The value of variance is closely related to how samples are taken Unbiased sampling Biased sampling • More points are taken in important parts of the population 2.2 Motivation of Variance Reduction Technique If we are using sample-mean Monte Carlo Method Variance depends very much on the behavior of ρ(x) ρ(x) varies little variance is small ρ(x) = const variance=0 Evaluation of a integral b − a N I'= (b − a)μ = ρ(x ) Y n N n=1 Near minimum points contribute less to the summation Near maximum points contribute more to the summation More points are sampled near the peak ”importance sampling strategy” 2.3 Variance Reduction using Rejection Technique Variance Reduction for Hit-or-Miss method In the domain [a,b] choose a comparison function w(x) ≥ ρ(x) w(x) x ρ(x) X = X W (x) w(x)dx X −∞ X b X = O A w(x)dx O a O O O − Au = w(x) x = W 1(Au) O O a b Points are generated on the area under w(x) function Random variable that follows distribution w(x) 2.3 Variance Reduction using Rejection Technique Points lying above ρ(x) is rejected N' I ≈ A N w(x) ρ(x) = X y w(x )u X n n n X X ≤ ρ X 1 if yn (xn ) O q = O n > ρ O 0 if yn (xn ) O O O O a b q10 = P(q) r 1-r r I / A 2.3 Variance Reduction using Rejection Technique Error Analysis E(Q) = r, V (Q) = r(1− r) I = Ar = AE(Q) − RJ ≈ I(A I) Hit or Miss method Ierror 0.67 N A = (b − a)h Error reduction A → I then Error → 0 2.3 Variance Reduction using Rejection Technique Start Set N : large integer w(x) ρ(x) X X X N’ = 0 X X O 1 O Loop Generate u1, x= W- (Au1) O N times O O O O Generate u2, y=u2 w(x) a b If y<= f(x) accept value N’ = N’+1 Else : reject value − RJ I'(A I') I = (b-a) h (N’/N) I ≈ 0.67 error N End 2.4 Importance Sampling Method Basic idea Put more points near maximum Put less points near minimum F(x) : transformation function (or weight function_ x F(x) = f (x)dx −∞ y = F(x) x = F −1(y) 2.4 Importance Sampling Method dy / dx = f (x) → dx = dy / f (x) b ρ(x) b ρ(x) I = dy = f (x)dx a a f (x) f (x) ρ(x) γ (x) = f (x) <η > = bη f (x) f (x)dx a if we choose f(x) = cρ(x) ,then variance will be small The magnitude of error depends on the choice of f(x) = b γ =< γ > I (x) f (x)dx f a 2.4 Importance Sampling Method Estimate of error N =< γ > ≈ 1 γ I f (xn ) N n=1 V (γ ) I = 0.67 f error N γ =< γ 2 > − < γ > 2 V f ( ) f ( f ) < γ 2 > −I 2 I IS = 0.67 f error N 2.4 Importance Sampling Method Start Set N : large integer s1 = 0 , s2 = 0 Loop Generate xn according to f(x) N times γ ρ n = (xn) / f ( xn) γ Add n to s1 γ Add n to s2 V ' 2 IS I ‘ =s1/N , V’=s2/N-I’ I = 0.67 error N End 3.
Recommended publications
  • Kalman and Particle Filtering
    Abstract: The Kalman and Particle filters are algorithms that recursively update an estimate of the state and find the innovations driving a stochastic process given a sequence of observations. The Kalman filter accomplishes this goal by linear projections, while the Particle filter does so by a sequential Monte Carlo method. With the state estimates, we can forecast and smooth the stochastic process. With the innovations, we can estimate the parameters of the model. The article discusses how to set a dynamic model in a state-space form, derives the Kalman and Particle filters, and explains how to use them for estimation. Kalman and Particle Filtering The Kalman and Particle filters are algorithms that recursively update an estimate of the state and find the innovations driving a stochastic process given a sequence of observations. The Kalman filter accomplishes this goal by linear projections, while the Particle filter does so by a sequential Monte Carlo method. Since both filters start with a state-space representation of the stochastic processes of interest, section 1 presents the state-space form of a dynamic model. Then, section 2 intro- duces the Kalman filter and section 3 develops the Particle filter. For extended expositions of this material, see Doucet, de Freitas, and Gordon (2001), Durbin and Koopman (2001), and Ljungqvist and Sargent (2004). 1. The state-space representation of a dynamic model A large class of dynamic models can be represented by a state-space form: Xt+1 = ϕ (Xt,Wt+1; γ) (1) Yt = g (Xt,Vt; γ) . (2) This representation handles a stochastic process by finding three objects: a vector that l describes the position of the system (a state, Xt X R ) and two functions, one mapping ∈ ⊂ 1 the state today into the state tomorrow (the transition equation, (1)) and one mapping the state into observables, Yt (the measurement equation, (2)).
    [Show full text]
  • A Fourier-Wavelet Monte Carlo Method for Fractal Random Fields
    JOURNAL OF COMPUTATIONAL PHYSICS 132, 384±408 (1997) ARTICLE NO. CP965647 A Fourier±Wavelet Monte Carlo Method for Fractal Random Fields Frank W. Elliott Jr., David J. Horntrop, and Andrew J. Majda Courant Institute of Mathematical Sciences, 251 Mercer Street, New York, New York 10012 Received August 2, 1996; revised December 23, 1996 2 2H k[v(x) 2 v(y)] l 5 CHux 2 yu , (1.1) A new hierarchical method for the Monte Carlo simulation of random ®elds called the Fourier±wavelet method is developed and where 0 , H , 1 is the Hurst exponent and k?l denotes applied to isotropic Gaussian random ®elds with power law spectral the expected value. density functions. This technique is based upon the orthogonal Here we develop a new Monte Carlo method based upon decomposition of the Fourier stochastic integral representation of the ®eld using wavelets. The Meyer wavelet is used here because a wavelet expansion of the Fourier space representation of its rapid decay properties allow for a very compact representation the fractal random ®elds in (1.1). This method is capable of the ®eld. The Fourier±wavelet method is shown to be straightfor- of generating a velocity ®eld with the Kolmogoroff spec- ward to implement, given the nature of the necessary precomputa- trum (H 5 Ad in (1.1)) over many (10 to 15) decades of tions and the run-time calculations, and yields comparable results scaling behavior comparable to the physical space multi- with scaling behavior over as many decades as the physical space multiwavelet methods developed recently by two of the authors.
    [Show full text]
  • Approaching Mean-Variance Efficiency for Large Portfolios
    Approaching Mean-Variance Efficiency for Large Portfolios ∗ Mengmeng Ao† Yingying Li‡ Xinghua Zheng§ First Draft: October 6, 2014 This Draft: November 11, 2017 Abstract This paper studies the large dimensional Markowitz optimization problem. Given any risk constraint level, we introduce a new approach for estimating the optimal portfolio, which is developed through a novel unconstrained regression representation of the mean-variance optimization problem, combined with high-dimensional sparse regression methods. Our estimated portfolio, under a mild sparsity assumption, asymptotically achieves mean-variance efficiency and meanwhile effectively controls the risk. To the best of our knowledge, this is the first time that these two goals can be simultaneously achieved for large portfolios. The superior properties of our approach are demonstrated via comprehensive simulation and empirical studies. Keywords: Markowitz optimization; Large portfolio selection; Unconstrained regression, LASSO; Sharpe ratio ∗Research partially supported by the RGC grants GRF16305315, GRF 16502014 and GRF 16518716 of the HKSAR, and The Fundamental Research Funds for the Central Universities 20720171073. †Wang Yanan Institute for Studies in Economics & Department of Finance, School of Economics, Xiamen University, China. [email protected] ‡Hong Kong University of Science and Technology, HKSAR. [email protected] §Hong Kong University of Science and Technology, HKSAR. [email protected] 1 1 INTRODUCTION 1.1 Markowitz Optimization Enigma The groundbreaking mean-variance portfolio theory proposed by Markowitz (1952) contin- ues to play significant roles in research and practice. The optimal mean-variance portfolio has a simple explicit expression1 that only depends on two population characteristics, the mean and the covariance matrix of asset returns. Under the ideal situation when the underlying mean and covariance matrix are known, mean-variance investors can easily compute the optimal portfolio weights based on their preferred level of risk.
    [Show full text]
  • A Family of Skew-Normal Distributions for Modeling Proportions and Rates with Zeros/Ones Excess
    S S symmetry Article A Family of Skew-Normal Distributions for Modeling Proportions and Rates with Zeros/Ones Excess Guillermo Martínez-Flórez 1, Víctor Leiva 2,* , Emilio Gómez-Déniz 3 and Carolina Marchant 4 1 Departamento de Matemáticas y Estadística, Facultad de Ciencias Básicas, Universidad de Córdoba, Montería 14014, Colombia; [email protected] 2 Escuela de Ingeniería Industrial, Pontificia Universidad Católica de Valparaíso, 2362807 Valparaíso, Chile 3 Facultad de Economía, Empresa y Turismo, Universidad de Las Palmas de Gran Canaria and TIDES Institute, 35001 Canarias, Spain; [email protected] 4 Facultad de Ciencias Básicas, Universidad Católica del Maule, 3466706 Talca, Chile; [email protected] * Correspondence: [email protected] or [email protected] Received: 30 June 2020; Accepted: 19 August 2020; Published: 1 September 2020 Abstract: In this paper, we consider skew-normal distributions for constructing new a distribution which allows us to model proportions and rates with zero/one inflation as an alternative to the inflated beta distributions. The new distribution is a mixture between a Bernoulli distribution for explaining the zero/one excess and a censored skew-normal distribution for the continuous variable. The maximum likelihood method is used for parameter estimation. Observed and expected Fisher information matrices are derived to conduct likelihood-based inference in this new type skew-normal distribution. Given the flexibility of the new distributions, we are able to show, in real data scenarios, the good performance of our proposal. Keywords: beta distribution; centered skew-normal distribution; maximum-likelihood methods; Monte Carlo simulations; proportions; R software; rates; zero/one inflated data 1.
    [Show full text]
  • Random Processes
    Chapter 6 Random Processes Random Process • A random process is a time-varying function that assigns the outcome of a random experiment to each time instant: X(t). • For a fixed (sample path): a random process is a time varying function, e.g., a signal. – For fixed t: a random process is a random variable. • If one scans all possible outcomes of the underlying random experiment, we shall get an ensemble of signals. • Random Process can be continuous or discrete • Real random process also called stochastic process – Example: Noise source (Noise can often be modeled as a Gaussian random process. An Ensemble of Signals Remember: RV maps Events à Constants RP maps Events à f(t) RP: Discrete and Continuous The set of all possible sample functions {v(t, E i)} is called the ensemble and defines the random process v(t) that describes the noise source. Sample functions of a binary random process. RP Characterization • Random variables x 1 , x 2 , . , x n represent amplitudes of sample functions at t 5 t 1 , t 2 , . , t n . – A random process can, therefore, be viewed as a collection of an infinite number of random variables: RP Characterization – First Order • CDF • PDF • Mean • Mean-Square Statistics of a Random Process RP Characterization – Second Order • The first order does not provide sufficient information as to how rapidly the RP is changing as a function of timeà We use second order estimation RP Characterization – Second Order • The first order does not provide sufficient information as to how rapidly the RP is changing as a function
    [Show full text]
  • Fourier, Wavelet and Monte Carlo Methods in Computational Finance
    Fourier, Wavelet and Monte Carlo Methods in Computational Finance Kees Oosterlee1;2 1CWI, Amsterdam 2Delft University of Technology, the Netherlands AANMPDE-9-16, 7/7/2016 Kees Oosterlee (CWI, TU Delft) Comp. Finance AANMPDE-9-16 1 / 51 Joint work with Fang Fang, Marjon Ruijter, Luis Ortiz, Shashi Jain, Alvaro Leitao, Fei Cong, Qian Feng Agenda Derivatives pricing, Feynman-Kac Theorem Fourier methods Basics of COS method; Basics of SWIFT method; Options with early-exercise features COS method for Bermudan options Monte Carlo method BSDEs, BCOS method (very briefly) Kees Oosterlee (CWI, TU Delft) Comp. Finance AANMPDE-9-16 1 / 51 Agenda Derivatives pricing, Feynman-Kac Theorem Fourier methods Basics of COS method; Basics of SWIFT method; Options with early-exercise features COS method for Bermudan options Monte Carlo method BSDEs, BCOS method (very briefly) Joint work with Fang Fang, Marjon Ruijter, Luis Ortiz, Shashi Jain, Alvaro Leitao, Fei Cong, Qian Feng Kees Oosterlee (CWI, TU Delft) Comp. Finance AANMPDE-9-16 1 / 51 Feynman-Kac theorem: Z T v(t; x) = E g(s; Xs )ds + h(XT ) ; t where Xs is the solution to the FSDE dXs = µ(Xs )ds + σ(Xs )d!s ; Xt = x: Feynman-Kac Theorem The linear partial differential equation: @v(t; x) + Lv(t; x) + g(t; x) = 0; v(T ; x) = h(x); @t with operator 1 Lv(t; x) = µ(x)Dv(t; x) + σ2(x)D2v(t; x): 2 Kees Oosterlee (CWI, TU Delft) Comp. Finance AANMPDE-9-16 2 / 51 Feynman-Kac Theorem The linear partial differential equation: @v(t; x) + Lv(t; x) + g(t; x) = 0; v(T ; x) = h(x); @t with operator 1 Lv(t; x) = µ(x)Dv(t; x) + σ2(x)D2v(t; x): 2 Feynman-Kac theorem: Z T v(t; x) = E g(s; Xs )ds + h(XT ) ; t where Xs is the solution to the FSDE dXs = µ(Xs )ds + σ(Xs )d!s ; Xt = x: Kees Oosterlee (CWI, TU Delft) Comp.
    [Show full text]
  • Efficient Monte Carlo Methods for Estimating Failure Probabilities
    Reliability Engineering and System Safety 165 (2017) 376–394 Contents lists available at ScienceDirect Reliability Engineering and System Safety journal homepage: www.elsevier.com/locate/ress ffi E cient Monte Carlo methods for estimating failure probabilities MARK ⁎ Andres Albana, Hardik A. Darjia,b, Atsuki Imamurac, Marvin K. Nakayamac, a Mathematical Sciences Dept., New Jersey Institute of Technology, Newark, NJ 07102, USA b Mechanical Engineering Dept., New Jersey Institute of Technology, Newark, NJ 07102, USA c Computer Science Dept., New Jersey Institute of Technology, Newark, NJ 07102, USA ARTICLE INFO ABSTRACT Keywords: We develop efficient Monte Carlo methods for estimating the failure probability of a system. An example of the Probabilistic safety assessment problem comes from an approach for probabilistic safety assessment of nuclear power plants known as risk- Risk analysis informed safety-margin characterization, but it also arises in other contexts, e.g., structural reliability, Structural reliability catastrophe modeling, and finance. We estimate the failure probability using different combinations of Uncertainty simulation methodologies, including stratified sampling (SS), (replicated) Latin hypercube sampling (LHS), Monte Carlo and conditional Monte Carlo (CMC). We prove theorems establishing that the combination SS+LHS (resp., SS Variance reduction Confidence intervals +CMC+LHS) has smaller asymptotic variance than SS (resp., SS+LHS). We also devise asymptotically valid (as Nuclear regulation the overall sample size grows large) upper confidence bounds for the failure probability for the methods Risk-informed safety-margin characterization considered. The confidence bounds may be employed to perform an asymptotically valid probabilistic safety assessment. We present numerical results demonstrating that the combination SS+CMC+LHS can result in substantial variance reductions compared to stratified sampling alone.
    [Show full text]
  • “Mean”? a Review of Interpreting and Calculating Different Types of Means and Standard Deviations
    pharmaceutics Review What Does It “Mean”? A Review of Interpreting and Calculating Different Types of Means and Standard Deviations Marilyn N. Martinez 1,* and Mary J. Bartholomew 2 1 Office of New Animal Drug Evaluation, Center for Veterinary Medicine, US FDA, Rockville, MD 20855, USA 2 Office of Surveillance and Compliance, Center for Veterinary Medicine, US FDA, Rockville, MD 20855, USA; [email protected] * Correspondence: [email protected]; Tel.: +1-240-3-402-0635 Academic Editors: Arlene McDowell and Neal Davies Received: 17 January 2017; Accepted: 5 April 2017; Published: 13 April 2017 Abstract: Typically, investigations are conducted with the goal of generating inferences about a population (humans or animal). Since it is not feasible to evaluate the entire population, the study is conducted using a randomly selected subset of that population. With the goal of using the results generated from that sample to provide inferences about the true population, it is important to consider the properties of the population distribution and how well they are represented by the sample (the subset of values). Consistent with that study objective, it is necessary to identify and use the most appropriate set of summary statistics to describe the study results. Inherent in that choice is the need to identify the specific question being asked and the assumptions associated with the data analysis. The estimate of a “mean” value is an example of a summary statistic that is sometimes reported without adequate consideration as to its implications or the underlying assumptions associated with the data being evaluated. When ignoring these critical considerations, the method of calculating the variance may be inconsistent with the type of mean being reported.
    [Show full text]
  • Monte Carlo Simulation by Marco Liu, Cqf
    OPTIMAL NUMBER OF TRIALS FOR MONTE CARLO SIMULATION BY MARCO LIU, CQF This article presents a way to estimate the number of trials required several hours. It would be beneficial to know what level of precision, for a desired confidence interval in the context of a Monte Carlo or confidence interval, we could achieve for a certain number of simulation. iterations in advance of running the program. Without the confidence interval, the time commitment for a trial-and-error process would Traditional valuation approaches such as Option Pricing Method be significant. In this article, we will present a simple method to (“OPM”) or Probability-Weighted Expected Return Method estimate the number of simulations needed for a desired level of (“PWERM”) may not be adequate in providing fair value estimation precision when running a Monte Carlo simulation. for financial instruments that require distribution assumptions on multiple input parameters. In such cases, a numerical method, Monte Carlo simulation for instance, is often used. The Monte Carlo 95% of simulation is a computerized algorithmic procedure that outputs a area wide range of values – typically unknown probability distribution – by simulating one or multiple input parameters via known probability distributions. Statistically, 95% of the area under a normal distribution curve is described as being plus or minus 1.96 standard This technique is often used to find fair value for financial deviations from the mean. For 90%, the z-statistic is 1.64. instruments for which probabilistic distributions are unknown. The simulation procedure is typically repeated multiple times and the average of the results is taken.
    [Show full text]
  • A Tutorial on Quantile Estimation Via Monte Carlo
    A Tutorial on Quantile Estimation via Monte Carlo Hui Dong and Marvin K. Nakayama Abstract Quantiles are frequently used to assess risk in a wide spectrum of applica- tion areas, such as finance, nuclear engineering, and service industries. This tutorial discusses Monte Carlo simulation methods for estimating a quantile, also known as a percentile or value-at-risk, where p of a distribution’s mass lies below its p-quantile. We describe a general approach that is often followed to construct quantile estimators, and show how it applies when employing naive Monte Carlo or variance-reduction techniques. We review some large-sample properties of quantile estimators. We also describe procedures for building a confidence interval for a quantile, which provides a measure of the sampling error. 1 Introduction Numerous application settings have adopted quantiles as a way of measuring risk. For a fixed constant 0 < p < 1, the p-quantile of a continuous random variable is a constant x such that p of the distribution’s mass lies below x. For example, the median is the 0:5-quantile. In finance, a quantile is called a value-at-risk, and risk managers commonly employ p-quantiles for p ≈ 1 (e.g., p = 0:99 or p = 0:999) to help determine capital levels needed to be able to cover future large losses with high probability; e.g., see [33]. Nuclear engineers use 0:95-quantiles in probabilistic safety assessments (PSAs) of nuclear power plants. PSAs are often performed with Monte Carlo, and the U.S. Nuclear Regulatory Commission (NRC) further requires that a PSA accounts for the Hui Dong Amazon.com Corporate LLC∗, Seattle, WA 98109, USA e-mail: [email protected] ∗This work is not related to Amazon, regardless of the affiliation.
    [Show full text]
  • Two-Dimensional Monte Carlo Filter for a Non-Gaussian Environment
    electronics Article Two-Dimensional Monte Carlo Filter for a Non-Gaussian Environment Xingzi Qiang 1 , Rui Xue 1,* and Yanbo Zhu 2 1 School of Electrical and Information Engineering, Beihang University, Beijing 100191, China; [email protected] 2 Aviation Data Communication Corporation, Beijing 100191, China; [email protected] * Correspondence: [email protected] Abstract: In a non-Gaussian environment, the accuracy of a Kalman filter might be reduced. In this paper, a two- dimensional Monte Carlo Filter is proposed to overcome the challenge of the non-Gaussian environment for filtering. The two-dimensional Monte Carlo (TMC) method is first proposed to improve the efficacy of the sampling. Then, the TMC filter (TMCF) algorithm is proposed to solve the non-Gaussian filter problem based on the TMC. In the TMCF, particles are deployed in the confidence interval uniformly in terms of the sampling interval, and their weights are calculated based on Bayesian inference. Then, the posterior distribution is described more accurately with less particles and their weights. Different from the PF, the TMCF completes the transfer of the distribution using a series of calculations of weights and uses particles to occupy the state space in the confidence interval. Numerical simulations demonstrated that, the accuracy of the TMCF approximates the Kalman filter (KF) (the error is about 10−6) in a two-dimensional linear/ Gaussian environment. In a two-dimensional linear/non-Gaussian system, the accuracy of the TMCF is improved by 0.01, and the computation time reduced to 0.067 s from 0.20 s, compared with the particle filter.
    [Show full text]
  • The Monte Carlo Method
    The Monte Carlo Method Introduction Let S denote a set of numbers or vectors, and let real-valued function g(x) be defined over S. There are essentially three fundamental computational problems associated with S and g. R Integral Evaluation Evaluate S g(x)dx. Note that if S is discrete, then integration can be replaced by summation. Optimization Find x0 2 S that minimizes g over S. In other words, find x0 2 S such that, for all x 2 S, g(x0) ≤ g(x). Mixed Optimization and Integration If g is defined over the space S × Θ, where Θ is a param- R eter space, then the problem is to determine θ 2 Θ for which S g(x; θ)dx is a minimum. The mathematical disciplines of real analysis (e.g. calculus), functional analysis, and combinatorial optimization offer deterministic algorithms for solving these problems (by solving we mean finding an exact solution or an approximate solution) for special instances of S and g. For example, if S is a closed interval of numbers [a; b] and g is a polynomial, then Z Z b g(x)dx = g(x)dx = G(b) − G(a); S a where G is an antiderivative of g. As another example, if G = (V; E; w) is a connected weighted simple graph, S is the set of all spanning trees of G, and g(x) denotes the sum of all the weights on the edges of spanning tree x, then Prim's algorithm may be used to find a spanning tree of G that minimizes g.
    [Show full text]