1 1. Data Transformations

Total Page:16

File Type:pdf, Size:1020Kb

1 1. Data Transformations 260 Archives ofDisease in Childhood 1993; 69: 260-264 STATISTICS FROM THE INSIDE Arch Dis Child: first published as 10.1136/adc.69.2.260 on 1 August 1993. Downloaded from 1 1. Data transformations M J R Healy Additive and multiplicative effects adding a constant quantity to the correspond- In the statistical analyses for comparing two ing before readings. Exactly the same is true of groups of continuous observations which I the unpaired situation, as when we compare have so far considered, certain assumptions independent samples of treated and control have been made about the data being analysed. patients. Here the assumption is that we can One of these is Normality of distribution; in derive the distribution ofpatient readings from both the paired and the unpaired situation, the that of control readings by shifting the latter mathematical theory underlying the signifi- bodily along the axis, and this again amounts cance probabilities attached to different values to adding a constant amount to each of the of t is based on the assumption that the obser- control variate values (fig 1). vations are drawn from Normal distributions. This is not the only way in which two groups In the unpaired situation, we make the further of readings can be related in practice. Suppose assumption that the distributions in the two I asked you to guess the size of the effect of groups which are being compared have equal some treatment for (say) increasing forced standard deviations - this assumption allows us expiratory volume in one second in asthmatic to simplify the analysis and to gain a certain children. Your reply might well be 'Oh, amount of power by utilising a single pooled perhaps +20%'. This implies a multiplicative estimate of variance. It is necessary to stress effect of the treatment. A child with an initial that the importance of these assumptions is level of 2-01 would be expected to increase this often exaggerated. Although they form part of by 0-41, one with an initial level of 3-01 would the mathematical framework, departures from be expected to increase this by 061. The dif- them have very little effect upon the outcome ference between the 'before' and 'after' read- of the analyses unless they are particularly ings varies systematica1ly with the level of marked. response. If data following this pattern were to http://adc.bmj.com/ A third assumption is a good deal less be analysed using the assumption of an addi- obvious and a good deal more important. This tive treatment effect, this kind of systematic is the assumption that the effect of the factor discrepancy between the (after-before) dif- which defines the two groups is additive, mean- ferences would be treated as if it was random ing by this that it produces (apart from error) a and interpreted as being due to error. constant difference between the readings in the two groups. Consider for example a paired situation, as when we compare readings before Logarithmic transformation of data on October 1, 2021 by guest. Protected copyright. and after a treatment on the same subjects. For A multiplicative treatment effect of this kind the purposes of analysis, we form the dif- can be converted to an additive effect by trans- ferences between the before and after readings forming all the original readings to logarithms for each subject, and the variability between before doing the analysis. Your memories of these differences is regarded as error. Put logarithms may be of a rather outdated aid to another way, we assume that we can derive the arithmetic, but their usefulness in statistical after readings on the subjects (error apart) by analysis goes far beyond this. For statistical purposes, the only thing you need to remember about logarithms is that the logarithm of the product of two numbers is the sum of the loga- rithms of the numbers, while the logarithm of the ratio of two numbers is the difference between the logarithms. In symbols log(axb) =log(a) +log(b) log(a/b) =log(a) -log(b) There are various kinds of logarithm, 23 Coleridge Court, depending upon what number is chosen to Milton Road, have a logarithm of 1I0. So-called 'common' Harpenden, Herts ALS 5LD I5I 2 logarithms are based on log(10)=1-0 (and 5 10 15 20 2E consequently log(100) =2-0, log(1000) =3-0, Correspondence to: Professor Healy. Figure 1 Two distributions differing by a constant log(01)=-1-0 and so on, using the rules No reprints available. amount. above). Mathematicians often use logarithms Data transformations 261 based on log(e)= 1-0, where e is a magic num- Ifinstead you analyse the differences ofthe logs ber equal to 2-7818 ... and rather presumpt- of the original readings, that is the logs of the uously call these 'natural' logarithms. The ratios, you will get t=0-104/0-0067=15-52, a numbers that describe the successive titres in a considerably larger value. Much more to the Arch Dis Child: first published as 10.1136/adc.69.2.260 on 1 August 1993. Downloaded from series of twofold dilutions are actually 'binary' point, the mean and standard error for the logarithms based on log(2)= 1-0. The numbers differences are +0-80 and 0-093, so that the 10, e, and 2 are called the bases of the loga- 95% confidence interval for the true mean rithms. Note that log(l)=0 0, no matter what difference (using the 5% significance value the base is (you can prove this ifyou like, using of t with 6 degrees of freedom) is the second of the above rules). It is usually 0-80 ±2-447 x 0-093 = + 0 57 to + 1 03. For the unimportant for statistical purposes what base logarithms of the ratios the mean and standard is used, since the different kinds of logarithm error are 0-104 and 0-0067, giving 95% differ by no more than a scale factor. The limits for the true mean difference of 'natural' logarithm of a number, for example, 0-104±2A447X0*0067=0-088 to 0-120. This is simply the 'common' logarithm multiplied interval relates to the differences between the by 2-3026. However, if you have to quote logarithms of the original data; taking the logarithmic values in the text or tables of a antilogs of these figures, the 95% confidence paper, it is essential to specify the base of the interval for the true ratio in the original data is logarithms so as to avoid ambiguity - you can from 1 22 to 1-32, a much more incisive result. write logl0, loge, or log2 as required. I person- ally use logarithmic values so often that I keep in my head a two figure table of common Transformation ofdata and its effects logarithms - In some circumstances an additive effect is rather implausible, so that it should be con- sidered whether transforming the data to loga- No 1 2 3 4 5 6 7 8 9 10 log 0-00 0-30 0-48 0-60 0-70 0-78 0-85 0-90 0-95 1 -00 rithms would be advantageous. One such situation arises when the data values are neces- sarily positive (as with, for example, the You may like to satisfy yourself that, for concentrations of some chemical in the blood) example, log(24) = 1-38, log(0 15) = -0-82. but are close to a definite zero point, in the The treatment effect of +20% that we spoke sense that zero is within one or two standard of amounts to assuming that (apart from error) deviations ofthe mean so that the coefficient of the treatment multiplies all the 'before' read- variation (the ratio of the standard deviation ings by the constant factor 1-20. When the over the mean) is 50% or more. Suppose that observations are transformed to logarithms, this describes the distribution of the control the treatment effect correspondingly increases values in a trial; then with an additive treat- all the 'before' log observations by the constant ment effect, the treated values would have to amount 0 079, which is the logarithm (to the start abruptly at some non-zero value, an http://adc.bmj.com/ base 10) of 1-20. unlikely state of affairs. A multiplicative effect A simple example, using paired data and as shown in fig 2 is a more realistic model for logs to the base 10, is shown in the table. Here data of this kind. A great many measurements, you can see quite a clear tendency for the dif- both physiological and biochemical, exhibit ference within each pair of original readings to this behaviour. increase with the average level. The ratios But it is worth noticing that this situation, show no such systematic tendency, suggesting marked by a large ratio of standard deviation that the data are closer to showing a constant to mean, is also unlikely to meet the other two on October 1, 2021 by guest. Protected copyright. ratio rather than a constant difference, in other assumptions I have mentioned, of Normality words, that the effect of treatment is closer to and equal variability. With a mean close to being multiplicative than additive. zero in the above sense, a Normal distribution If you calculate the paired t value from the would necessarily imply an appreciable prob- differences, you will get t=0-80/0-093=8-60. ability for the impossible negative values. Figure 2 also shows that, with a fixed Logaithmic transformation terminus at zero for both distributions, the Before After Difference Ratio Original data 2-0 2-4 +0-4 1-20 2-2 2-9 +0-7 1-32 2-5 3-2 +0-7 1-28 2-8 3-6 +0-8 1-29 3-2 4.3 +1-1 1-34 3.7 4.5 +0-8 1-22 4.3 5-4 +1-1 1-26 Mean (SE) +0-80 (0-093) Logged data 0-301 0-380 +0-079 0-342 0 462 +0-120 0-398 0-505 +0-107 0-447 0-556 +0 109 0-505 0-633 +0- 128 0-568 0-653 +0-085 0-633 0-732 +0-099 Mean (SE) +0-104 (0-0067) Figure 2 Two distributions differing by a multiplicative factor.
Recommended publications
  • Frequency Response and Bode Plots
    1 Frequency Response and Bode Plots 1.1 Preliminaries The steady-state sinusoidal frequency-response of a circuit is described by the phasor transfer function Hj( ) . A Bode plot is a graph of the magnitude (in dB) or phase of the transfer function versus frequency. Of course we can easily program the transfer function into a computer to make such plots, and for very complicated transfer functions this may be our only recourse. But in many cases the key features of the plot can be quickly sketched by hand using some simple rules that identify the impact of the poles and zeroes in shaping the frequency response. The advantage of this approach is the insight it provides on how the circuit elements influence the frequency response. This is especially important in the design of frequency-selective circuits. We will first consider how to generate Bode plots for simple poles, and then discuss how to handle the general second-order response. Before doing this, however, it may be helpful to review some properties of transfer functions, the decibel scale, and properties of the log function. Poles, Zeroes, and Stability The s-domain transfer function is always a rational polynomial function of the form Ns() smm as12 a s m asa Hs() K K mm12 10 (1.1) nn12 n Ds() s bsnn12 b s bsb 10 As we have seen already, the polynomials in the numerator and denominator are factored to find the poles and zeroes; these are the values of s that make the numerator or denominator zero. If we write the zeroes as zz123,, zetc., and similarly write the poles as pp123,, p , then Hs( ) can be written in factored form as ()()()s zsz sz Hs() K 12 m (1.2) ()()()s psp12 sp n 1 © Bob York 2009 2 Frequency Response and Bode Plots The pole and zero locations can be real or complex.
    [Show full text]
  • Computational Entropy and Information Leakage∗
    Computational Entropy and Information Leakage∗ Benjamin Fuller Leonid Reyzin Boston University fbfuller,[email protected] February 10, 2011 Abstract We investigate how information leakage reduces computational entropy of a random variable X. Recall that HILL and metric computational entropy are parameterized by quality (how distinguishable is X from a variable Z that has true entropy) and quantity (how much true entropy is there in Z). We prove an intuitively natural result: conditioning on an event of probability p reduces the quality of metric entropy by a factor of p and the quantity of metric entropy by log2 1=p (note that this means that the reduction in quantity and quality is the same, because the quantity of entropy is measured on logarithmic scale). Our result improves previous bounds of Dziembowski and Pietrzak (FOCS 2008), where the loss in the quantity of entropy was related to its original quality. The use of metric entropy simplifies the analogous the result of Reingold et. al. (FOCS 2008) for HILL entropy. Further, we simplify dealing with information leakage by investigating conditional metric entropy. We show that, conditioned on leakage of λ bits, metric entropy gets reduced by a factor 2λ in quality and λ in quantity. ∗Most of the results of this paper have been incorporated into [FOR12a] (conference version in [FOR12b]), where they are applied to the problem of building deterministic encryption. This paper contains a more focused exposition of the results on computational entropy, including some results that do not appear in [FOR12a]: namely, Theorem 3.6, Theorem 3.10, proof of Theorem 3.2, and results in Appendix A.
    [Show full text]
  • A Weakly Informative Default Prior Distribution for Logistic and Other
    The Annals of Applied Statistics 2008, Vol. 2, No. 4, 1360–1383 DOI: 10.1214/08-AOAS191 c Institute of Mathematical Statistics, 2008 A WEAKLY INFORMATIVE DEFAULT PRIOR DISTRIBUTION FOR LOGISTIC AND OTHER REGRESSION MODELS By Andrew Gelman, Aleks Jakulin, Maria Grazia Pittau and Yu-Sung Su Columbia University, Columbia University, University of Rome, and City University of New York We propose a new prior distribution for classical (nonhierarchi- cal) logistic regression models, constructed by first scaling all nonbi- nary variables to have mean 0 and standard deviation 0.5, and then placing independent Student-t prior distributions on the coefficients. As a default choice, we recommend the Cauchy distribution with cen- ter 0 and scale 2.5, which in the simplest setting is a longer-tailed version of the distribution attained by assuming one-half additional success and one-half additional failure in a logistic regression. Cross- validation on a corpus of datasets shows the Cauchy class of prior dis- tributions to outperform existing implementations of Gaussian and Laplace priors. We recommend this prior distribution as a default choice for rou- tine applied use. It has the advantage of always giving answers, even when there is complete separation in logistic regression (a common problem, even when the sample size is large and the number of pre- dictors is small), and also automatically applying more shrinkage to higher-order interactions. This can be useful in routine data analy- sis as well as in automated procedures such as chained equations for missing-data imputation. We implement a procedure to fit generalized linear models in R with the Student-t prior distribution by incorporating an approxi- mate EM algorithm into the usual iteratively weighted least squares.
    [Show full text]
  • C-17Bpages Pdf It
    SPECIFICATION DEFINITIONS FOR LOGARITHMIC AMPLIFIERS This application note is presented to engineers who may use logarithmic amplifiers in a variety of system applica- tions. It is intended to help engineers understand logarithmic amplifiers, how to specify them, and how the loga- rithmic amplifiers perform in a system environment. A similar paper addressing the accuracy and error contributing elements in logarithmic amplifiers will follow this application note. INTRODUCTION The need to process high-density pulses with narrow pulse widths and large amplitude variations necessitates the use of logarithmic amplifiers in modern receiving systems. In general, the purpose of this class of amplifier is to condense a large input dynamic range into a much smaller, manageable one through a logarithmic transfer func- tion. As a result of this transfer function, the output voltage swing of a logarithmic amplifier is proportional to the input signal power range in dB. In most cases, logarithmic amplifiers are used as amplitude detectors. Since output voltage (in mV) is proportion- al to the input signal power (in dB), the amplitude information is displayed in a much more usable format than accomplished by so-called linear detectors. LOGARITHMIC TRANSFER FUNCTION INPUT RF POWER VS. DETECTED OUTPUT VOLTAGE 0V 50 MILLIVOLTS/DIV. 250 MILLIVOLTS/DIV. 0V 50.0 ns/DIV. 50.0 ns/DIV. There are three basic types of logarithmic amplifiers. These are: • Detector Log Video Amplifiers • Successive Detection Logarithmic Amplifiers • True Log Amplifiers DETECTOR LOG VIDEO AMPLIFIER (DLVA) is a type of logarithmic amplifier in which the envelope of the input RF signal is detected with a standard "linear" diode detector.
    [Show full text]
  • Linear and Logarithmic Scales
    Linear and logarithmic scales. Defining Scale You may have thought of a scale as something to weigh yourself with or the outer layer on the bodies of fish and reptiles. For this lesson, we're using a different definition of a scale. A scale, in this sense, is a leveled range of values/numbers from lowest to highest that measures something at regular intervals. A great example is the classic number line that has numbers lined up at consistent intervals along a line. What Is a Linear Scale? A linear scale is much like the number line described above. They key to this type of scale is that the value between two consecutive points on the line does not change no matter how high or low you are on it. For instance, on the number line, the distance between the numbers 0 and 1 is 1 unit. The same distance of one unit is between the numbers 100 and 101, or -100 and -101. However you look at it, the distance between the points is constant (unchanging) regardless of the location on the line. A great way to visualize this is by looking at one of those old school intro to Geometry or mid- level Algebra examples of how to graph a line. One of the properties of a line is that it is the shortest distance between two points. Another is that it is has a constant slope. Having a constant slope means that the change in x and y from one point to another point on the line doesn't change.
    [Show full text]
  • The Indefinite Logarithm, Logarithmic Units, and the Nature of Entropy
    The Indefinite Logarithm, Logarithmic Units, and the Nature of Entropy Michael P. Frank FAMU-FSU College of Engineering Dept. of Electrical & Computer Engineering 2525 Pottsdamer St., Rm. 341 Tallahassee, FL 32310 [email protected] August 20, 2017 Abstract We define the indefinite logarithm [log x] of a real number x> 0 to be a mathematical object representing the abstract concept of the logarithm of x with an indeterminate base (i.e., not specifically e, 10, 2, or any fixed number). The resulting indefinite logarithmic quantities naturally play a mathematical role that is closely analogous to that of dimensional physi- cal quantities (such as length) in that, although these quantities have no definite interpretation as ordinary numbers, nevertheless the ratio of two of these entities is naturally well-defined as a specific, ordinary number, just like the ratio of two lengths. As a result, indefinite logarithm objects can serve as the basis for logarithmic spaces, which are natural systems of logarithmic units suitable for measuring any quantity defined on a log- arithmic scale. We illustrate how logarithmic units provide a convenient language for explaining the complete conceptual unification of the dis- parate systems of units that are presently used for a variety of quantities that are conventionally considered distinct, such as, in particular, physical entropy and information-theoretic entropy. 1 Introduction The goal of this paper is to help clear up what is perceived to be a widespread arXiv:physics/0506128v1 [physics.gen-ph] 15 Jun 2005 confusion that can found in many popular sources (websites, popular books, etc.) regarding the proper mathematical status of a variety of physical quantities that are conventionally defined on logarithmic scales.
    [Show full text]
  • Logarithmic Amplifier for Ultrasonic Sensor Signal Conditioning
    Application Report SLDA053–March 2020 Logarithmic Amplifier for Ultrasonic Sensor Signal Conditioning Akeem Whitehead, Kemal Demirci ABSTRACT This document explains the basic operation of a logarithmic amplifier, how a logarithmic amplifier works in an ultrasonic front end, the advantages and disadvantages of using a logarithmic amplifier versus a linear time-varying gain amplifier, and compares the performance of a logarithmic amplifier versus a linear amplifier. These topics apply to the TUSS44x0 device family (TUSS4440 and TUSS4470), which includes Texas Instrument’s latest ultrasonic driver and logarithmic amplifier-based receiver front end integrated circuits. The receive signal path of the TUSS44x0 devices includes a low-noise linear amplifier, a band pass filter, followed by a logarithmic amplifier for input-level dependent amplification. The logarithmic amplifier allows for high sensitivity for weak echo signals and offers a wide input dynamic range over full range of reflected echoes. Contents 1 Introduction to Logarithmic Amplifier (Log Amp)......................................................................... 2 2 Logarithmic Amplifier in Ultrasonic Sensing.............................................................................. 8 3 Logarithmic versus Time Varying Linear Amplifier..................................................................... 10 4 Performance of Log Amp in Ultrasonic Systems....................................................................... 11 List of Figures 1 Linear Input and Logarithmic
    [Show full text]
  • Arxiv:1712.03037V1 [Cs.CV] 8 Dec 2017 Community
    A Frequency Domain Neural Network for Fast Image Super-resolution Junxuan Li1∗ Shaodi You1;2 Antonio Robles-Kelly1;2 1College of Eng. and Comp. Sci., Australian National University, Canberra, ACT 2601, Australia 2Datat61-CSIRO, Black Mountain Laboratories, Acton, ACT 2601, Australia Abstract by aggregating multiple frames with complementary spa- tial information or by relating the higher-resolved image to In this paper, we present a frequency domain neural net- the lower resolution one by a sparse linear system. For in- work for image super-resolution. The network employs the stance, Baker and Kanade [1] formulated the problem in a convolution theorem so as to cast convolutions in the spatial regularization setting where the examples are constructed domain as products in the frequency domain. Moreover, the using a pyramid approach. Protter et al. [22] used block non-linearity in deep nets, often achieved by a rectifier unit, matching to estimate a motion model and use exemplars to is here cast as a convolution in the frequency domain. This recover super-resolved videos. Yang et al. [31] used sparse not only yields a network which is very computationally ef- coding to perform super-resolution by learning a dictionary ficient at testing but also one whose parameters can all be that can then be used to produce the output image, by lin- learnt accordingly. The network can be trained using back early combining learned exemplars. propagation and is devoid of complex numbers due to the Note that, the idea of super-solution “by example” can use of the Hartley transform as an alternative to the Fourier be viewed as hinging on the idea of learning functions so transform.
    [Show full text]
  • Lecture 1-2: Sound Pressure Level Scale
    Lecture 1-2: Sound Pressure Level Scale Overview 1. Psychology of Sound. We can define sound objectively in terms of its physical form as pressure variations in a medium such as air, or we can define it subjectively in terms of the sensations generated by our hearing mechanism. Of course the sensation we perceive from a sound is linked to the physical properties of the sound wave through our hearing mechanism, whereby some aspects of the pressure variations are converted into firings in the auditory nerve. The study of how the objective physical character of sound gives rise to subjective auditory sensation is called psychoacoustics. The three dimensions of our psychological sensation of sound are called Loudness , Pitch and Timbre . 2. How can we measure the quantity of sound? The amplitude of a sound is a measure of the average variation in pressure, measured in pascals (Pa). Since the mean amplitude of a sound is zero, we need to measure amplitude either as the distance between the most negative peak and the most positive peak or in terms of the root-mean-square ( rms ) amplitude (i.e. the square root of the average squared amplitude). The power of a sound source is a measure of how much energy it radiates per second in watts (W). It turns out that the power in a sound is proportional to the square of its amplitude. The intensity of a sound is a measure of how much power can be collected per unit area in watts per square metre (Wm -2). For a sound source of constant power radiating equally in all directions, the sound intensity must fall with the square of the distance to the source (since the radiated energy is spread out through the surface of a sphere centred on the source, and the surface area of the sphere increases with the square of its radius).
    [Show full text]
  • Logarithmic System of Units
    Appendix A Logarithmic System of Units A.1 Introduction This appendix deals with the concepts and fundamentals of the logarithmic units and their applications in the engineering of radiowaves propagation. Use of these units, expressing the formulas, and illustrating figures/graphs in the logarithmic systems are common and frequently referred in the book. A.2 Definition According to the definition, logarithm of the ratio of two similar quantities in decimal base is called “bel,” and ten times of it is called “decibel.” = p bel log10 pr =[ ]= p decibel dB 10 log10 pr As the above formula suggests, it is a logarithmic scale, and due to the persistent usage of decimal base in the logarithmic units, it is normally omitted from the related symbol for simplicity. Since values in the logarithmic scale are the logarithm of the ratio of two similar quantities, hence they are dimensionless and measured in the same unit. In other words, the unit of denominator, pr, is the base of comparison and includes the required unit. To indicate the unit of the base in the logarithmic system of units, one or few characters are used according to the following examples: dBw: Decibel unit compared to 1 W power A. Ghasemi et al., Propagation Engineering in Radio Links Design, 523 DOI 10.1007/978-1-4614-5314-7, © Springer Science+Business Media New York 2013 524 A Logarithmic System of Units dBm: Decibel unit compared to 1 mW power dBkw: Decibel unit compared to 1 kW power Example A.1. Output power of a transmitter is 2 W, state its power in dBm,dBw, and dBkw.
    [Show full text]
  • Logistic Regression
    CHAPTER 5 Logistic regression Logistic regression is the standard way to model binary outcomes (that is, data yi that take on the values 0 or 1). Section 5.1 introduces logistic regression in a simple example with one predictor, then for most of the rest of the chapter we work through an extended example with multiple predictors and interactions. 5.1 Logistic regression with a single predictor Example: modeling political preference given income Conservative parties generally receive more support among voters with higher in- comes. We illustrate classical logistic regression with a simple analysis of this pat- tern from the National Election Study in 1992. For each respondent i in this poll, we label yi =1ifheorshepreferredGeorgeBush(theRepublicancandidatefor president) or 0 if he or she preferred Bill Clinton (the Democratic candidate), for now excluding respondents who preferred Ross Perot or other candidates, or had no opinion. We predict preferences given the respondent’s income level, which is characterized on a five-point scale.1 The data are shown as (jittered) dots in Figure 5.1, along with the fitted logistic regression line, a curve that is constrained to lie between 0 and 1. We interpret the line as the probability that y =1givenx—in mathematical notation, Pr(y =1x). We fit and display the logistic regression using the following R function calls:| fit.1 <- glm (vote ~ income, family=binomial(link="logit")) Rcode display (fit.1) to yield coef.est coef.se Routput (Intercept) -1.40 0.19 income 0.33 0.06 n=1179,k=2 residual deviance = 1556.9, null deviance = 1591.2 (difference = 34.3) 1 The fitted model is Pr(yi =1)=logit− ( 1.40 + 0.33 income).
    [Show full text]
  • Measurement Invariance, Entropy, and Probability
    Entropy 2010, 12, 289-303; doi:10.3390/e12030289 OPEN ACCESS entropy ISSN 1099-4300 www.mdpi.com/journal/entropy Article Measurement Invariance, Entropy, and Probability Steven A. Frank1;2;? and D. Eric Smith2 1 Department of Ecology and Evolutionary Biology, University of California, Irvine, CA 92697–2525, USA 2 Santa Fe Institute, 1399 Hyde Park Road, Santa Fe, NM 87501, USA ? Author to whom correspondence should be addressed; E-mail: [email protected]. Received: 18 January 2010; in revised form: 16 February 2010 / Accepted: 23 February 2010 / Published: 26 February 2010 Abstract: We show that the natural scaling of measurement for a particular problem defines the most likely probability distribution of observations taken from that measurement scale. Our approach extends the method of maximum entropy to use measurement scale as a type of information constraint. We argue that a very common measurement scale is linear at small magnitudes grading into logarithmic at large magnitudes, leading to observations that often follow Student’s probability distribution which has a Gaussian shape for small fluctuations from the mean and a power law shape for large fluctuations from the mean. An inverse scaling often arises in which measures naturally grade from logarithmic to linear as one moves from small to large magnitudes, leading to observations that often follow a gamma probability distribution. A gamma distribution has a power law shape for small magnitudes and an exponential shape for large magnitudes. The two measurement scales are natural inverses connected by the Laplace integral transform. This inversion connects the two major scaling patterns commonly found in nature.
    [Show full text]