Bayesian Optimization Meets Bayesian Optimal Stopping

Total Page:16

File Type:pdf, Size:1020Kb

Bayesian Optimization Meets Bayesian Optimal Stopping Bayesian Optimization Meets Bayesian Optimal Stopping Zhongxiang Dai 1 Haibin Yu 1 Bryan Kian Hsiang Low 1 Patrick Jaillet 2 Abstract Silver et al., 2016). However, a major difficulty faced by ML Bayesian optimization (BO) is a popular paradigm practitioners is the choice of model hyperparameters which for optimizing the hyperparameters of machine significantly impacts the predictive performance. This calls learning (ML) models due to its sample efficiency. for the need to develop hyperparameter optimization algo- Many ML models require running an iterative rithms that have to be sample-efficient since the training training procedure (e.g., stochastic gradient de- of many modern ML models consumes massive computa- scent). This motivates the question whether in- tional resources. To this end, Bayesian optimization (BO) formation available during the training process is a popular paradigm due to its high sample efficiency and (e.g., validation accuracy after each epoch) can be strong theoretical performance guarantee (Shahriari et al., exploited for improving the epoch efficiency of 2016). In particular, the BO algorithm based on the Gaus- BO algorithms by early-stopping model training sian process-upper confidence bound (GP-UCB) acquisition under hyperparameter settings that will end up function has been shown to achieve no regret asymptotically under-performing and hence eliminating unneces- and perform competitively in practice (Srinivas et al., 2010). sary training epochs. This paper proposes to unify Many ML models require running an iterative training proce- BO (specifically, Gaussian process-upper confi- dure for some number of epochs such as stochastic gradient dence bound (GP-UCB)) with Bayesian optimal descent for neural networks (LeCun et al., 2015) and boost- stopping (BO-BOS) to boost the epoch efficiency ing procedure for gradient boosting machines (Friedman, of BO. To achieve this, while GP-UCB is sample- 2001). During BO, any query of a hyperparameter setting efficient in the number of function evaluations, usually involves training the ML model for a fixed number of BOS complements it with epoch efficiency for epochs. Information typically available during the training each function evaluation by providing a principled process (e.g., validation accuracy after each epoch) is rarely optimal stopping mechanism for early stopping. exploited for improving the epoch efficiency of BO algo- BO-BOS preserves the (asymptotic) no-regret per- rithms, specifically, by early-stopping model training under formance of GP-UCB using our specified choice hyperparameter settings that will end up under-performing, of BOS parameters that is amenable to an ele- hence eliminating unnecessary training epochs. Note that gant interpretation in terms of the exploration- this objective is different from that of standard early stop- exploitation trade-off. We empirically evaluate ping during the training of neural networks, which is used the performance of BO-BOS and demonstrate its to prevent overfitting. generality in hyperparameter optimization of ML models and two other interesting applications. To address this challenging issue, a number of works have been proposed to make BO more epoch-efficient: Freeze- 1. Introduction thaw BO (Swersky et al., 2014) explores a diverse collection of hyperparameter settings in the initial stage by training The state-of-the-art machine learning (ML) models have their ML models with a small number of epochs, and then recently reached an unprecedented level of predictive per- gradually focuses on (exploits) a small number of promising formance in several applications such as image recognition, settings. Despite its promising epoch efficiency, its perfor- complex board games, among others (LeCun et al., 2015; mance is not theoretically guaranteed and its computational 1 1 cost can be excessive. Multi-fidelity BO (Kandasamy et al., Department of Computer Science, National University of 2016; 2017), as well as its extensions using predictive en- Singapore, Republic of Singapore. 2Department of Electrical Engineering and Computer Science, Massachusetts Institute of tropy search (Zhang et al., 2017) and value of information Technology, USA. Correspondence to: Bryan Kian Hsiang Low for fidelity selection (Wu & Frazier, 2017), reduces the <[email protected]>. resource consumption of BO by utilizing low-fidelity func- Proceedings of the 36 th International Conference on Machine 1A closely related counterpart is multi-fidelity active learn- Learning, Long Beach, California, PMLR 97, 2019. Copyright ing (Zhang et al., 2016). 2019 by the author(s). Bayesian Optimization Meets Bayesian Optimal Stopping tions which can be obtained by using a subset of the training objectives of GP-UCB vs. BOS (respectively, objective func- data or by training the ML model for a small number of tion v.s. expected loss), BO-BOS can preserve the trademark epochs. However, in each BO iteration, since the fidelity (asymptotic) no-regret performance of GP-UCB with our (e.g., number of epochs) is determined before function eval- specified choice of BOS parameters that is amenable to an el- uation, it is not influenced by information that is typically egant interpretation in terms of the exploration-exploitation available during the training process (e.g., validation accu- trade-off (Section 4). Though the focus of our work here is racy after each epoch). In addition to BO, attempts have on epoch-efficient BO for hyperparameter tuning, we addi- also been made to improve the epoch efficiency of other tionally evaluate the performance of BO-BOS empirically hyperparameter optimization algorithms: Some heuristic in two other interesting applications to demonstrate its gen- methods (Baker et al., 2017; Domhan et al., 2015; Klein erality: policy search for reinforcement learning, and joint et al., 2017) predict the final training outcome based on hyperparameter tuning and feature selection (Section 5). partially trained learning curves in order to identify hyper- parameter settings that are predicted to under-perform and 2. Background and Notations early-stop their model training. Hyperband (Li et al., 2017), which dynamically allocates the computational resource 2.1. Bayesian Optimization (BO) and GP-UCB (e.g., training epochs) through random sampling and elimi- Consider the problem of sequentially maximizing an un- nates under-performing hyperparameter settings by succes- known objective function f : R representing the val- D! sive halving, has been proposed and shown to perform well idation accuracy over a compact input domain Rd of D✓ in practice. Both the learning curve prediction methods and different hyperparameter settings for training an ML model: Hyperband can be combined with BO to further improve In each iteration t =1,...,T, an input query zt , [xt,nt] the epoch efficiency (Domhan et al., 2015; Falkner et al., of a hyperparameter setting (comprising N <n N 0 t 2018; Klein et al., 2017), but their resulting performances training epochs and a vector xt of the other hyperparam- are not theoretically guaranteed. Despite these recent ad- eters) is selected for evaluating the validation accuracy f vances, we still lack an epoch-efficient algorithm that can of the ML model to yield a noisy observed output (valida- incorporate early stopping into BO (i.e., by exploiting infor- tion accuracy) yt , f(zt)+✏ with i.i.d. Gaussian noise mation available during the training process) and yet offer a ✏ (0,σ2) and noise variance σ2. Since every eval- ⇠N theoretical performance guarantee, the design of which is uation of f is costly (Section 1), our goal is to strategi- likely to require a principled decision-making mechanism cally select input queries to approach the global maximizer for determining the optimal stopping time. z⇤ , arg maxz f(z) as rapidly as possible. This can be 2D achieved by minimizing a standard BO objective such as Optimal stopping is a classic research topic in statistics and the simple regret ST , f(z⇤) maxt 1,...,T f(zt).A operations research regarding sequential decision-making − 2{ } BO algorithm is said to guarantee no regret asymptotically problems whose objective is to make the optimal stopping if it satisfies limT ST =0, thus implying that it will decision with a small number of observations (Ferguson, !1 eventually converge to the global maximum. 2006). In Bayesian optimal stopping (BOS) or Bayesian sequential design, the decision between stopping vs. continu- To guarantee no regret, the belief of f is modeled by a ing is made to maximize the expected utility or, equivalently, Gaussian process (GP). Let f(z) z denote a GP, that { } 2D minimize the expected loss (Powell & Ryzhov, 2012). BOS is, every finite subset of f(z) z follows a multivariate { } 2D has found success in application domains such as finance Gaussian distribution (Rasmussen & Williams, 2006). Then, (Longstaff & Schwartz, 2001), clinical design (Brockwell & a GP is fully specified by its prior mean µ(z) and covariance Kadane, 2003; Muller¨ et al., 2007; Wathen & Thall, 2008), k(z, z0) for all z, z0 , which are, respectively, assumed 2D and economics (Davis & Cairns, 2012). The capability of w.l.o.g. to be µ(z)=0and k(z, z0) 1 for notational BOS in providing a principled optimal stopping mechanism simplicity. Given a column vector yT , [yt]t>=1,...,T of makes it a prime candidate for introducing early stopping noisy outputs observed from evaluating f at the selected into BO in a theoretically sound and rigorous way. input queries z ,...,z after T iterations, the GP 1 T 2D posterior belief of f at some input z is a Gaussian with This paper proposes to unify Bayesian optimization (specifi- 2D the following posterior mean µ (z) and variance σ2 (z): cally, GP-UCB) with Bayesian optimal stopping (BO-BOS) T T to boost the epoch efficiency of BO (Section 3). Intu- 2 1 µT (z) , kT (z)>(KT + σ I)− yT , itively, GP-UCB is acclaimed for being sample-efficient 2 2 1 (1) σ (z) k(z, z) k (z)>(K + σ I)− k (z) in the number of function evaluations while BOS can reduce T , − T T T the required number of epochs for each function evalua- where KT , [k(zt, zt0 )]t,t0=1,...,T and kT (z) , tion.
Recommended publications
  • Deep Optimal Stopping
    Journal of Machine Learning Research 20 (2019) 1-25 Submitted 4/18; Revised 1/19; Published 4/19 Deep Optimal Stopping Sebastian Becker [email protected] Zenai AG, 8045 Zurich, Switzerland Patrick Cheridito [email protected] RiskLab, Department of Mathematics ETH Zurich, 8092 Zurich, Switzerland Arnulf Jentzen [email protected] SAM, Department of Mathematics ETH Zurich, 8092 Zurich, Switzerland Editor: Jon McAuliffe Abstract In this paper we develop a deep learning method for optimal stopping problems which directly learns the optimal stopping rule from Monte Carlo samples. As such, it is broadly applicable in situations where the underlying randomness can efficiently be simulated. We test the approach on three problems: the pricing of a Bermudan max-call option, the pricing of a callable multi barrier reverse convertible and the problem of optimally stopping a fractional Brownian motion. In all three cases it produces very accurate results in high- dimensional situations with short computing times. Keywords: optimal stopping, deep learning, Bermudan option, callable multi barrier reverse convertible, fractional Brownian motion 1. Introduction N We consider optimal stopping problems of the form supτ E g(τ; Xτ ), where X = (Xn)n=0 d is an R -valued discrete-time Markov process and the supremum is over all stopping times τ based on observations of X. Formally, this just covers situations where the stopping decision can only be made at finitely many times. But practically all relevant continuous- time stopping problems can be approximated with time-discretized versions. The Markov assumption means no loss of generality. We make it because it simplifies the presentation and many important problems already are in Markovian form.
    [Show full text]
  • Optimal Stopping Time and Its Applications to Economic Models
    OPTIMAL STOPPING TIME AND ITS APPLICATIONS TO ECONOMIC MODELS SIVAKORN SANGUANMOO Abstract. This paper gives an introduction to an optimal stopping problem by using the idea of an It^odiffusion. We then prove the existence and unique- ness theorems for optimal stopping, which will help us to explicitly solve opti- mal stopping problems. We then apply our optimal stopping theorems to the economic model of stock prices introduced by Samuelson [5] and the economic model of capital goods. Both economic models show that those related agents are risk-loving. Contents 1. Introduction 1 2. Preliminaries: It^odiffusion and boundary value problems 2 2.1. It^odiffusions and their properties 2 2.2. Harmonic function and the stochastic Dirichlet problem 4 3. Existence and uniqueness theorems for optimal stopping 5 4. Applications to an economic model: Stock prices 11 5. Applications to an economic model: Capital goods 14 Acknowledgments 18 References 19 1. Introduction Since there are several outside factors in the real world, many variables in classi- cal economic models are considered as stochastic processes. For example, the prices of stocks and oil do not just increase due to inflation over time but also fluctuate due to unpredictable situations. Optimal stopping problems are questions that ask when to stop stochastic pro- cesses in order to maximize utility. For example, when should one sell an asset in order to maximize his profit? When should one stop searching for the keyword to obtain the best result? Therefore, many recent economic models often include op- timal stopping problems. For example, by using optimal stopping, Choi and Smith [2] explored the effectiveness of the search engine, and Albrecht, Anderson, and Vroman [1] discovered how the search cost affects the search for job candidates.
    [Show full text]
  • The Duration of Optimal Stopping Problems
    arXiv: 1810.11557 THE DURATION OF OPTIMAL STOPPING PROBLEMS By Simon Demers Optimal stopping problems give rise to random distributions de- scribing how many applicants the decision-maker will sample or in- terview before choosing one, a quantity sometimes referred to as the search time or process duration. This research note surveys several variants of optimal stopping problems, extends earlier results in var- ious directions, and shows how many interviews are expected to be conducted in various settings. The focus is on problems that require a decision-maker to choose a candidate from a pool of sequential appli- cants with no recall, in the vein of previously studied Cayley-Moser, Secretary and Sultan's Dowry problems. 1. Introduction. Although optimal stopping problems have been stud- ied and refined extensively over time (Gilbert and Mosteller, 1966; Freeman, 1983; Ferguson, 1989a), relatively little is currently known about the statis- tical properties of the random distributions they give rise to. For example, it is not obvious how the expected number of observations (\applicants") to be considered (\interviewed") before one is ultimately chosen (\hired") will vary when the decision-maker (\employer") can observe the actual pay- off values (\talent") or when the horizon (\applicant pool") grows. This is unfortunate because it might be useful in some practical settings to antici- pate, for example, how many interviews the decision-maker should plausibly schedule or prepare for. This is what this research note attempts to elucidate. The focus is on the problem of choosing a candidate from a pool of ap- plicants with no recall and uniformly distributed talent.
    [Show full text]
  • Optimal Stopping Problems: Autonomous Trading Over an Infinite Time Horizon
    Imperial College London Department of Mathematics Optimal Stopping Problems: Autonomous Trading Over an Infinite Time Horizon Author: Viraj Shah (CID: 01191054) A thesis submitted for the degree of MSc in Mathematics and Finance, 2019-2020 Declaration The work contained in this thesis is my own work unless otherwise stated. i Acknowledgements I would like to take this opportunity to thank Dr. Alex Tse for being a wonderful supervisor throughout my thesis project. His understanding in this area of financial mathematics has been brilliant and he has pushed me to achieve my potential. I have thoroughly enjoyed working with Dr. Tse and will carry forward the lessons learned from this experience. Without his judgement and continual encouragement, this project would not have been possible. I am extremely grateful for the opportunity I have had to study at Imperial College London and would like to express my gratitude and appreciation to all those who have helped me to this point. I would like to recognise the support and companionship my friends have provided me with, as well as the guidance I have received from numerous academic staff. Lastly, I am indebted to my family and want to honour their unconditional support towards me. I am truly thankful for their consistent confidence in me and for everything that they have given me. ii Abstract Statistical arbitrage (StatArb) has taken off ever since its advent in the 1980's and is being increasingly used by hedge funds and investment banks to produce profitable, quantitative trading strategies. With financial firms pushing for a larger amount of automation, we are driven to inves- tigate trading methods that decide optimally on our behalf.
    [Show full text]
  • Optimal Stopping Problems
    Optimal stopping problems Part II. Applications A. N. Shiryaev M. V. Zhitlukhin Steklov Mathematical Institute, Moscow The University of Manchester, UK Bonn { May 2013 Outline The second part of the course contains some particular optimal stopping problems related to mathematical finance. The questions we consider concern problems of portfolio re-balancing and choosing optimal moments of time to sell (or buy) stock. Problems of this type play the central role in the technical analysis of financial markets; the field, which is much less theoretically developed in comparison with the two other analyses: the fundamental and the quantitative ones. Most of the methods in the technical analysis are based on empirical evidence of \rules of thumb", while we will try to present a mathematical foundation for them. 1/145 The contents of this part of the course 1. Overview of general facts from the optimal stopping theory 2. Stopping a Brownian motion at its maximum Peskir, Shiryaev. Optimal stopping and free-boundary problems, 2006; sec. 30 3. Trading rule \Buy and hold" Shiryaev, Xu, Zhou. Thou shalt buy and hold, 2008 4. Sequential hypothesis testing Shiryaev. Optimal stopping rules, 2007; ch. 4 5. Sequential parameters estimation C¸etin, Novikov, Shiryaev. LSE preprint, 2012 6. Quickest disorder detection Shiryaev. Optimal stopping rules, 2000; ch. IV Shiryaev. Quickest detection problems: 50 years later, 2010 2/145 1. General optimal stopping theory Formulation of an optimal stopping problem Let (Ω; F ; (Ft)t>0; P) be a filtered probability space and a G = (Gt)t>0 be a stochastic process on it, where Gt is interpreted as the gain if the observation is stopped at time t.
    [Show full text]
  • Time Pressure and Regret in Sequential Search
    A Service of Leibniz-Informationszentrum econstor Wirtschaft Leibniz Information Centre Make Your Publications Visible. zbw for Economics Klimm, Felix; Kocher, Martin G.; Optiz, Timm; Schudy, Simeon Andreas Dermot Working Paper Time Pressure and Regret in Sequential Search CESifo Working Paper, No. 9122 Provided in Cooperation with: Ifo Institute – Leibniz Institute for Economic Research at the University of Munich Suggested Citation: Klimm, Felix; Kocher, Martin G.; Optiz, Timm; Schudy, Simeon Andreas Dermot (2021) : Time Pressure and Regret in Sequential Search, CESifo Working Paper, No. 9122, Center for Economic Studies and Ifo Institute (CESifo), Munich This Version is available at: http://hdl.handle.net/10419/236664 Standard-Nutzungsbedingungen: Terms of use: Die Dokumente auf EconStor dürfen zu eigenen wissenschaftlichen Documents in EconStor may be saved and copied for your Zwecken und zum Privatgebrauch gespeichert und kopiert werden. personal and scholarly purposes. Sie dürfen die Dokumente nicht für öffentliche oder kommerzielle You are not to copy documents for public or commercial Zwecke vervielfältigen, öffentlich ausstellen, öffentlich zugänglich purposes, to exhibit the documents publicly, to make them machen, vertreiben oder anderweitig nutzen. publicly available on the internet, or to distribute or otherwise use the documents in public. Sofern die Verfasser die Dokumente unter Open-Content-Lizenzen (insbesondere CC-Lizenzen) zur Verfügung gestellt haben sollten, If the documents have been made available under an Open gelten abweichend von diesen Nutzungsbedingungen die in der dort Content Licence (especially Creative Commons Licences), you genannten Lizenz gewährten Nutzungsrechte. may exercise further usage rights as specified in the indicated licence. www.econstor.eu 9122 2021 May 2021 Time Pressure and Regret in Sequential Search Felix Klimm, Martin G.
    [Show full text]
  • Time Costs and Search Behavior
    Munich Personal RePEc Archive Time Costs and Search Behavior Hsiao, Yu-Chin and Kemp, Simon and Servátka, Maroš and Ward, Matt and Zhang, Le University of Canterbury, MGSM Experimental Economics Laboratory, Macquarie Business School, University of Economics in Bratislava 19 January 2021 Online at https://mpra.ub.uni-muenchen.de/105412/ MPRA Paper No. 105412, posted 02 Feb 2021 04:44 UTC Time Costs and Search Behavior Yu-Chin Hsiao University of Canterbury & MGSM Experimental Economics Laboratory, Macquarie Business School & Lingnan University Simon Kemp University of Canterbury Maroš Servátka MGSM Experimental Economics Laboratory, Macquarie Business School & University of Economics in Bratislava Matt Ward University of Canterbury Le Zhang MGSM Experimental Economics Laboratory, Macquarie Business School January 19, 2021 Abstract: Sequential search is often costly and time-consuming. The time cost is usually unknown ex ante and its presence and duration must be inferred as the search progresses. We disentangle the effect of time cost on search behavior from people’s (in)ability to perceive time delay between offers. We find that people are able to infer the existence of the time cost, but their inference is imperfect. We also compare the effect of time cost with the effect of monetary cost and find that the time cost reduces the amount of exerted search, but not as much as the monetary cost does. Discriminating between the effects is critical for increasing the empirical validity of search models and designing mechanisms capable of improving the quality of decisions, especially in unfamiliar or infrequently encountered situations. Keywords: Sequential Search, Time Cost, Search Cost, Experiment JEL Classification: C6, C91, D83 Acknowledgements: This paper is based on Yu-Chin Hsiao’s dissertation written jointly at the University of Canterbury and Macquarie Graduate School of Management (Hsiao, 2018).
    [Show full text]
  • J. Appl. Prob. 38(3) 647-658 Pricing Exotic Options
    J. Appl. Prob. 38, 647–658 (2001) Printed in Israel Applied Probability Trust 2001 SOME OPTIMAL STOPPING PROBLEMS WITH NONTRIVIAL BOUNDARIES FOR PRICING EXOTIC OPTIONS XIN GUO,∗ IBM LARRY SHEPP,∗∗ Rutgers University Abstract We solve the following three optimal stopping problems for different kinds of options, based on the Black–Scholes model of stock fluctuations. (i) The perpetual lookback American option for the running maximum of the stock price during the life of the option. This problem is more difficult than the closely related one for the Russian option, and we show that for a class of utility functions the free boundary is governed by a nonlinear ordinary differential equation. (ii) A new type of stock option, for a company, where the company provides a guaranteed minimum as an added incentive in case the market appreciation of the stock is low, thereby making the option more attractive to the employee. We show that the value of this option is given by solving a nonalgebraic equation. (iii) A new call option for the option buyer who is risk-averse and gets to choose, a priori, a fixed constant l as a ‘hedge’ on a possible downturn of the stock price, where the buyer gets the maximum of l and the price at any exercise time. We show that the optimal policy depends on the ratio of x/l, where x is the current stock price. Keywords: Lookback options; Black–Scholes model; optimal stopping; Bellman equa- tion; free boundary AMS 2000 Subject Classification: Primary 91B28; 60H30; 60G44 1. Problems and main results Let Xt be a geometric Brownian motion with drift µ and σ , that is, dXt = Xt µ dt + Xt σ dWt , (1) where µ, σ are constants and Wt is the standard Wiener process.
    [Show full text]
  • New Developments of the Odds Theorem
    NEW DEVELOPMENTS OF THE ODDS THEOREM REMI´ DENDIEVEL Universit´eLibre de Bruxelles Facult´edes Sciences D´epartement de Math´ematique Abstract. The odds theorem and the corresponding solution algo- rithm (odds algorithm) are tools to solve a wide range of optimal stop- ping problems. Its generality and tractability have caught much at- tention. (Google for instance “Bruss odds” to obtain a quick overview.) Many extensions and modifications of the this result have appeared since publication in 2000. This article reviews the important gnew develop- ments and applications in this field. The spectrum of application com- prises as different fields as secretary problems, more general stopping problems, robotic maintenance problems, compassionate use clinical tri- als and others. This review also includes a new contribution of our own. 1. The Original Odds Theorem The odds-theorem is a result in the theory of optimal stopping which can be applied for many interesting sequential decision problems. The original version of the Odds-algorithm is the work of Bruss (2000). He discovered it when he saw common features between quite different looking optimal stopping problems. The framework is as follows. There are n random variables which are ob- served sequentially. It is desired to stop online with maximum probability on a last specific event. No recall is permitted. Here “specific” is under- stood as being defined in terms of the interest of the decision maker. Such problems can be readily translated into the equivalent problem of stopping arXiv:1212.1391v1 [math.PR] 6 Dec 2012 on a sequence of independent indicators. The parameters of those indica- tors variables are supposed to be known.
    [Show full text]
  • Pricing the American Option Using Itô's Formula and Optimal Stopping
    U.U.D.M. Project Report 2014:3 Pricing the American Option Using Itô’s Formula and Optimal Stopping Theory Jonas Bergström Examensarbete i matematik, 15 hp Handledare och examinator: Erik Ekström Januari 2014 Department of Mathematics Uppsala University PRICING THE AMERICAN OPTION USING ITO'S^ FORMULA AND OPTIMAL STOPPING THEORY JONAS BERGSTROM¨ Abstract. In this thesis the goal is to arrive at results concerning the value of American options and a formula for the perpetual American put option. For the stochastic dynamics of the underlying asset I look at two cases. The first is the standard Black-Scholes model and the second allows for the asset to jump to zero i.e default. To achieve the goals stated above the first couple of sections introduces some basic concepts in probability such as processes and information. Before introducing It^o's formula this paper contains a not too rigorous introduction of stochastic differential equations and stochastic integration. Then the Black-Scholes model is introduced followed by a section about optimal stopping theory in order to arrive at the American option. Acknowledgements. I would like to thank my supervisor Erik Ek- str¨omfor his support. His input has been essential for the progress and structure of this thesis. 1. Introduction This section contains the definition of one of the most important building blocks in continuous probability namely Wiener process. This is followed by definitions of information and martingales. Definition 1.1. A Wiener process, W = fWt; t ≥ 0g, starting from W0 = 0 is a continuous time stochastic process taking values in R s.t • W has independent increments i.e Wv − Wu and Wt − Ws are inde- pendent whenever u ≤ v ≤ s ≤ t • Ws+t − Ws ∼ N(0; t) W Definition 1.2.
    [Show full text]
  • Optimal Stopping Under Drift Uncertainty
    U.U.D.M. Report 2015:1 Optimal Stopping under Drift Uncertainty Juozas Vaicenavicius Filosofie licentiatavhandling i matematik som framläggs för offentlig granskning den 13 maj 2015, kl 13.15, sal 2005, Ångströmlaboratoriet, Uppsala Department of Mathematics Uppsala University Preface Many real-world processes exhibit an inherent trend visible at a macroscopic level. A river flow has a direction and speed, various financial and economic indicators are noted to have near constant growth or decline rates when sufficiently long time series are taken. However, at a microscopic level, the macroscopic trends tend to be overshadowed by random fluctuations, be it a Brownian bombardment of a pollen grain by nearby water molecules, or intraday volatility in a stock market. How to make optimal decisions that highly depend on the macroscopic trend when one is only observing a development of the process at a microscopic scale is the main theme of this dissertation. The thesis consists of two essays: 1. `Bayesian sequential testing of the drift of a Brownian motion', 2. `Optimal liquidation of an asset under drift uncertainty'; both written together with my PhD supervisor Prof. Erik Ekstr¨om. In the first article, an optimal sequential statistical procedure for de- termining the sign of the drift of a Brownian motion is investigated in the Bayesian framework under a general prior distribution. The second paper studies a financial problem that concerns optimal timing for the sale of an indivisible asset under arbitrary prior beliefs about the drift of the price process. 1 Acknowledgements I want to express my wholehearted gratitude to my PhD supervisor Prof.
    [Show full text]
  • Risk Averse Stochastic Programming: Time Consistency and Optimal Stopping
    Risk averse stochastic programming: time consistency and optimal stopping Alois Pichler Fakult¨atf¨urMathematik Technische Universit¨atChemnitz D–09111 Chemnitz, Germany Rui Peng Liu and Alexander Shapiro∗ School of Industrial and Systems Engineering Georgia Institute of Technology Atlanta, GA 30332-0205 June 13, 2019 Abstract This paper addresses time consistency of risk averse optimal stopping in stochastic optimization. It is demonstrated that time consistent optimal stopping entails a specific structure of the functionals describing the transition between consecutive stages. The stopping risk measures capture this structural behavior and they allow natural dynamic equations for risk averse decision making over time. Consequently, associated optimal policies satisfy Bellman’s principle of optimality, which characterizes optimal policies for optimization by stating that a decision maker should not reconsider previous decisions retrospectively. We also discuss numerical approaches to solving such problems. Keywords: Stochastic programming, coherent risk measures, time consistency, dynamic equations, optimal stopping time, Snell envelope. arXiv:1808.10807v3 [math.OC] 12 Jun 2019 1 Introduction Optimal Stopping (OS) is a classical topic of research in statistics and operations research going back to the pioneering work of Wald(1947, 1949) on sequential analysis. For a thorough ∗Research of this author was partly supported by NSF grant 1633196, and DARPA EQUiPS program grant SNL 014150709. discussion of theoretical foundations of OS we can refer to Shiryaev(1978), for example. The classical formulation of OS assumes availability of the probability distribution of the considered data process. Of course in real life applications the ‘true distribution’ is never known exactly and in itself can be viewed as uncertain.
    [Show full text]