Calibration Methods – Nomenclature and Classification

Total Page:16

File Type:pdf, Size:1020Kb

Calibration Methods – Nomenclature and Classification CHAPTER 8 CALIBRATION METHODS – NOMENCLATURE AND CLASSIFICATION Paweł Kościelniak Institute of Analytical Chemistry, Faculty of Chemistry, Jagiellonian University, R. Ingardena 3, 30-060 Kraków, Poland ABSTRACT Reviewing the analytical literature, including academic textbooks, one can notice that in fact there is no precise and clear terminology dealing with the analytical calibration. Especially a great confusion exists in nomenclature related to the calibration methods: not only different names are used with reference to a given method, but they do not express the principles and the nature of different methods properly (e.g. "the set of standard method" or "the internal standard method"). The problem mentioned above is of great importance. A lack of good terminology can be a source of misunderstandings and, consequently, can be even a reason of carrying out an analytical treatment against the rules. Finally, the aspect of rather psychological nature is worth to be stressed, namely just an analyst is (or should be at least) especially sensitive to such terms as "order" and "purity" irrespectively of what analytical area is considered. Chapter 8 1 INTRODUCTION Reading the professional literature, one is bound to arrive at the conclusion that in analytical chemistry there is a lack of clearly defined, current nomenclature relating to the problems of analytical calibration. It is characteristic that, among other things, in spite of the inevitable necessity of carrying out calibration in instrumental analysis and the common usage of the term ‘analytical calibration’ itself, it is not defined even in texts on nomenclature problems in chemistry [1,2], or otherwise the definitions are not connected with analytical practice [3]. Another surprising fact is that in the face of the huge significance of calibration problems, they are treated too cursorily and quite inconsistently, even in textbooks on chemical analytics [4-6]. Most importantly, there is a lack of a universal classification of calibration methods which would be clear and logical enough so that, for example, the terms ‘calibration method’ and ‘analytical method’ are explicitly separate and relate to different fields of problems in analytical chemistry. The above-mentioned problem is undoubtedly of considerable significance. Lack of nomenclature ordering in the domain of calibration can become a source of numerous misunderstandings and obscurities, which, as a result, can cause improper or totally fallacious analytical proceeding. The situation is also unfavourable for didactic purposes, since it is difficult to credibly convey analytical knowledge on the basis of a particular textbook or academic script when using a language which is not only felt as incorrect, but which is also distinct from the one that can be found in other generally available sources. Finally, one ought to remember the psychological aspect: it is no-one else but the analytical chemist who is (or at least should be) particularly sensitive to ‘neatness’ and ‘order’, regardless of which fields of analytical chemistry the terms refer to. These problems were the chief reason that gave rise to this paper. The suggestions of definitions, terminology and classification of calibration terms made in this paper are the author’s own. They, however, met with understanding in the author’s closest academic society, and are being gradually introduced to the syllabus of analytical chemistry at the Faculty of Chemistry of the Jagiellonian University. Presentation of these suggestions to a large circle of analytical chemists has as its aim the initiation of a discussion which may shed some new light on the problems in question and contribute to the establishment of a generally accepted standpoint in this field. 2 ANALYTICAL CALIBRATION The basic analytical task consists in determining the concentration of a given substance (known as analyte) in a sample of the material being tested. It can only be done by using a chosen analytical instrument 1 and on the basis of the measurment data (the analytical signals) provided by the instrument for the sample. 1 Quantitative analysis of a sample is always carried out by means of a specific measuring instrument, although the devices used for this purpose are sometimes (in comparison with the commonly used nowadays, fully automated and computer-controlled apparatus of complex structure) so simple that their ‘instrumental’ character is no longer perceived and appreciated. These are, most importantly, the analytical balance and the burette, which are used for conducting respectively gravimetric and volumetrical analyses. From this point of view, there is no point in the commonly made division of chemical analysis into classical analysis (i.e. gravimetric and volumetrical) and instrumental analysis. 111 Chapter 8 Regardless of the instrument used in a particular case, the measured analytical signal (R) depends not only on the concentration of the analyte (ca), but also on a number of other factors. These include the parameters of the instrument (pa1, pa2, …, pam) and the factors determining the physicochemical conditions of the sample being analysed (pc1, pc2, …, pck), as well as the concentrations of other substances present in the sample (cd1, cd2, …, cdn), which are either natural components of the sample or which are added to it in the course of the analytical procedure. Thus, the general relation between the analytical signal and the aforementioned factors can be represented by the formula 2: R = f (ca; cd1, cd2, …, cdn; pa1, pa2, …, pam; pc1, pc2, …, pck) (1) Before the analytical signal is measured, the measuring instrument and the sample are adjusted to each other by setting the parameters determining the instrumental and physicochemical conditions at fixed, optimal levels 3. Thus, at the stage of making measurements, relation (1) takes a simplified form: R = f (K; ca; cd1, cd2, …, cdn) (2) where K is the parameter determining the general conditions of the analysis. The fundamental analytical problem is to determine the types and concentrations of the components (cd1, cd2, …, cdn) which influence the analytical signal that indicates the presence and concentration of the analyte in the sample. This influence is termed the interference effect, and the substances that cause the effect are called interferents. Although theoretically the interference effect may not occur or be negligible in a given case, it must always be reckoned with, at least because of the possibility of inserting the interferents into the sample when it is being prepared for the measurement. Accurate specification of the parameters determining the analytical conditions and determination of the types and concentrations of the interferents in the sample (cd1, cd2, …, cdn) on the basis of theoretical or even semi-empirical deliberations are always extremely difficult, and most often impossible. In consequence, direct determination of the concentration of the analyte in the sample on the basis of the analytical signal and by means of a precisely specified formula (2) is not possible either. Therefore, a purely empirical approach is used for this purpose, which is the domain of analytical calibration. The term ‘analytical calibration’ ought to be understood as denoting a process which consists in representing the actual (real, theoretical) dependence of the analytical signal on the concentration of the analyte (which in this context is called calibration function) in an empirical form (calibration plot4), and then using the plot to determine the 2 In some cases the analytical signal is not measured directly for the analyte, but for some other component which is added to the sample in a known quantity and which binds the analyte in a reaction of a known stoichiometry. Then, the signal depends additionally on the concentration of this component. However, since the presence of such a component in the sample is only of auxiliary character and serves the purpose of determining the concentration of the analyte, it has not been taken into account in the formula. 3 These activities when carried out in relation to the instrument are often also called ‘calibration’, which must not, however, be confused or identified with the concept of ‘analytical calibration’. 4 ‘Calibration function’ is more accurate, but due to the commonly accepted habit of representing calibration data in a graphical form, the term ‘calibration plot’ has been decided upon. 112 Chapter 8 concentration of the analyte in the sample under examination (i.e. obtaining analytical result). The calibration plot is constructed under specified, constant analytical conditions (both instrumental and physicochemical), with the use of one or several standard solutions 5, i.e. solutions of known, precisely determined concentrations of at least one component (usually analyte). Thus, analytical calibration consists of three stages: laboratory (i.e. the preparation of the standard solutions), measurement (i.e. the construction of the calibration plot) and mathematical (i.e. the calculation of the analytical result). A detailed mode of performing respective stages constitutes the calibration procedure. Each calibration procedure ought to be followed according to the strictly specified rules which determine a more general way of proceeding, leading to the attainment of (besides the main calibration objective) certain additional analytical goals. Such a specific mode of calibration procedure can be called a calibration method. 3 NOMENCLATURE OF CALIBRATION METHODS There
Recommended publications
  • Pressure Measuring Instruments
    testo-312-2-3-4-P01 21.08.2012 08:49 Seite 1 We measure it. Pressure measuring instruments For gas and water installers testo 312-2 HPA testo 312-3 testo 312-4 BAR °C www.testo.com testo-312-2-3-4-P02 23.11.2011 14:37 Seite 2 testo 312-2 / testo 312-3 We measure it. Pressure meters for gas and water fitters Use the testo 312-2 fine pressure measuring instrument to testo 312-2 check flue gas draught, differential pressure in the combustion chamber compared with ambient pressure testo 312-2, fine pressure measuring or gas flow pressure with high instrument up to 40/200 hPa, DVGW approval, incl. alarm display, battery and resolution. Fine pressures with a resolution of 0.01 hPa can calibration protocol be measured in the range from 0 to 40 hPa. Part no. 0632 0313 DVGW approval according to TRGI for pressure settings and pressure tests on a gas boiler. • Switchable precision range with a high resolution • Alarm display when user-defined limit values are • Compensation of measurement fluctuations caused by exceeded temperature • Clear display with time The versatile pressure measuring instrument testo 312-3 testo 312-3 supports load and gas-rightness tests on gas and water pipelines up to 6000 hPa (6 bar) quickly and reliably. testo 312-3 versatile pressure meter up to Everything you need to inspect gas and water pipe 300/600 hPa, DVGW approval, incl. alarm display, battery and calibration protocol installations: with the electronic pressure measuring instrument testo 312-3, pressure- and gas-tightness can be tested.
    [Show full text]
  • A Measuring Instrument for Multipoint Soil Temperature Underground
    A MEASURING INSTRUMENT FOR MULTIPOINT SOIL TEMPERATURE UNDERGROUND Cheng Wang, Chunjiang Zhao * , Xiaojun Qiao, Zhilong Xu National Engineering Research Center for Information Technology in Agriculture, Beijing, P. R. China, 100097 * Corresponding author, Address: Shuguang Huayuan Middle Road 11#, Beijing, 100097, P. R. China, Tel: +86-10-51503411, Fax: +86-10-51503449, Email: [email protected] Abstract: A new measuring instrument for 10 points soil temperatures in 0–50 centimeters depth underground was designed. System was based on Silicon Laboratories’ MCU C8051F310, single chip digital temperature sensor DS18B20, and other peripheral circuits. It was simultaneously able to measure, memory and display, and also convey data to computer via a standard RS232 interface. Keywords: Multi-point Soil Temperature; Portable; DS18B20; C8051F310 1. INTRODUCTION The temperature of soil is a vital environmental factor, which directly influences the activity of microorganisms and the decomposition of organic substances. It can affect roots absorbing water and mineral elements. It also plays an important role in the growth rate and range of roots. Statistically, roots of most plants are within 50 centimeters underground, so it becomes very significant to measure the soil temperature of different depth in this level. The Soil Temperature Measuring Instruments used nowadays mainly fall into three types, the first type is the measure temperature by making use of the relationship between the soil temperature and the temperature-sensitive resistor. Before using this sort of instruments, the system parameters need to Wang, C., Zhao, C., Qiao, X. and Xu, Z., 2008, in IFIP International Federation for Information Processing, Volume 259; Computer and Computing Technologies in Agriculture, Vol.
    [Show full text]
  • Made to Measure. Practical Guide to Electrical Measurements in Low Voltage Switchboards V
    Contact us A 250 500 200 150 V (b) 100 (a) 50 0 t Made to measure. Practical guide to electrical measurements in low voltage switchboards A 250 500 ABB SACE The data and illustrations are not binding. We reserve 200 the right to modify the contents of this document on the 150 Una divisione di ABB S.p.A. basis of technical development of the products, 100 Apparecchi Modulari without prior notice. 50 0 Viale dell’Industria, 18 Copyright 2010 ABB. All rights reserved. - 1.500 - CAL. 20010 Vittuone (MI) Tel.: 02 9034 1 Fax: 02 9034 7609 bol.it.abb.com www.abb.com V 80 V 60 2CSC445012D0201 - 12/2010 (f) 40 50 Hz 20 0 t Made to measure. Practical guide to electrical measurements in low voltage switchboards table of Made to measure. Practical guide to electrical measurements contents in low voltage switchboards 1 Electrical measurements 5.3.2 Current transformers ......................................................... 37 5.3.3 Voltage transformers ......................................................... 38 1.1 Why is it important to measure? .......................................... 3 5.3.4 Shunts for direct current .................................................... 38 1.2 Applicational contexts .......................................................... 4 1.3 Problems connected with energy networks ......................... 4 6 The measurements 1.4 Reducing consumption ........................................................ 7 1.5 Table of charges .................................................................. 8 6.1 TRMS Measurements
    [Show full text]
  • IQ Sensornet Nitravis 701 & 705 IQ Sensors User Manual
    OPERATIONS MANUAL ba76078e03 05/2017 NitraVis 701 IQ NitraVis 705 IQ OPTICAL SENSOR FOR NITRATE NitraVis 70x IQ Contact YSI 1725 Brannum Lane Yellow Springs, OH 45387 USA Tel: +1 937-767-7241 800-765-4974 Email: [email protected] Internet: www.ysi.com Copyright © 2017 Xylem Inc. 2 ba76078e03 05/2017 NitraVis 70x IQ Contents Contents 1 Overview . 5 1.1 How to use this component operating manual . 5 1.2 Field of application . 6 1.3 Measuring principle of the sensor NitraVis 70x IQ . 6 1.4 Structure of the sensor NitraVis 70x IQ . 7 2 Safety . 8 2.1 Safety information . 8 2.1.1 Safety information in the operating manual . 8 2.1.2 Safety signs on the product . 8 2.1.3 Further documents providing safety information . 8 2.2 Safe operation . 9 2.2.1 Authorized use . 9 2.2.2 Requirements for safe operation . 9 2.2.3 Unauthorized use . 9 3 Commissioning . 10 3.1 IQ SENSORNET system requirements . 10 3.2 Scope of delivery of the NitraVis 70x IQ . 10 3.3 Installation . 11 3.3.1 Mounting the sensor . 11 3.3.2 Mounting the shock protectors . 13 3.3.3 Connecting the sensor to the IQ SENSORNET . 14 3.4 Initial commissioning . 16 3.4.1 General information . 16 3.4.2 Settings . 17 4 Measurement / Operation . 21 4.1 Determination of measured values . 21 4.2 Measurement operation . 22 4.3 Calibration . 22 4.3.1 Overview . 22 4.3.2 User calibration . 25 4.3.3 Sensor check/Zero adjustment .
    [Show full text]
  • Bluemeter SIGMA
    WYLER AG Tel. 0041 (0) 52 233 66 66 Im Hölderli Fax. 0041 (0) 52 233 20 53 CH-8405 WINTERTHUR Switzerland Homepage: http://www.wylerag.com E-Mail: [email protected] Manual BlueMETER SIGMA INDEX Subject page 1 BASICS / INTRODUCTION 6 1.1 DESCRIPTION OF THE BLUEMETER SIGMA 6 1.2 PREPARATION AND START-UP OF THE BLUEMETER SIGMA 6 1.2.1 BATTERIES 6 1.2.2 POSSIBLE CONFIGURATIONS 8 2 INITIAL STARTUP OF THE BLUEMETER SIGMA AND THE MEASURING INSTRUMENTS/SENSORS 10 2.1 CONNECTING THE INSTRUMENTS / CONNECTING OPTIONS ON THE BLUEMETER SIGMA 11 2.2 START UP 12 2.2.1 OPERATING ELEMENTS/SHORT OVERVIEW 12 2.2.1.1 OVERVIEW KEYS AND DISPLAY 12 2.2.1.2 SWITCHING THE INSTRUMENT ON AND OFF 13 2.2.1.3 KEYS / FUNCTIONS / SHORT DESCRIPTIONS OF EACH SINGLE KEY 14 2.3 DISPLAY 16 2.3.1 SCALING OF THE DISPLAY 16 2.3.2 DISPLAY TYPES 16 2.3.3 BACKGROUND COLOUR 19 2.3.4 BRIGHTNESS OF THE DISPLAY 20 2.3.5 SHORT DESCRIPTION OF THE INDIVIDUAL DISPLAY AREAS 21 3 OPERATING INSTRUCTIONS BLUEMETER SIGMA 22 3.1 FUNCTIONS ON THE BLUEMETER SIGMA / OVERVIEW KEYS AND DISPLAY 22 3.2 STARTING THE BLUEMETER SIGMA 24 3.2.1 START WITH UNCHANGED CONFIGURAATION 24 3.2.2 START WITH A CHANGED CONFIGURATION 25 3.3 REFRESH 26 3.4 SENSOR 26 3.5 ZERO-SETTING / ABSOLUTE ZERO 28 3.5.1 SET ABSOLUTE ZERO (WITH A REVERSAL MEASUREMENT) 28 3.6 SELECTION OF THE MEASURING UNIT / UNIT 30 3.6.1 STANDARD-UNITS 30 3.6.2 UNITS WITH RELATIVE BASE LENGTH 30 3.7 FUNCTION HOLD 31 3.8 FUNCTION SEND (PRINT FUNCTION) 32 3.9 SELECTION OF THE FILTER UNDER DIFFERENT MEASURING CONDITIONS / FILTER 33 3.10 ABSOLUTE
    [Show full text]
  • Quick Guide to Precision Measuring Instruments
    E4329 Quick Guide to Precision Measuring Instruments Coordinate Measuring Machines Vision Measuring Systems Form Measurement Optical Measuring Sensor Systems Test Equipment and Seismometers Digital Scale and DRO Systems Small Tool Instruments and Data Management Quick Guide to Precision Measuring Instruments Quick Guide to Precision Measuring Instruments 2 CONTENTS Meaning of Symbols 4 Conformance to CE Marking 5 Micrometers 6 Micrometer Heads 10 Internal Micrometers 14 Calipers 16 Height Gages 18 Dial Indicators/Dial Test Indicators 20 Gauge Blocks 24 Laser Scan Micrometers and Laser Indicators 26 Linear Gages 28 Linear Scales 30 Profile Projectors 32 Microscopes 34 Vision Measuring Machines 36 Surftest (Surface Roughness Testers) 38 Contracer (Contour Measuring Instruments) 40 Roundtest (Roundness Measuring Instruments) 42 Hardness Testing Machines 44 Vibration Measuring Instruments 46 Seismic Observation Equipment 48 Coordinate Measuring Machines 50 3 Quick Guide to Precision Measuring Instruments Quick Guide to Precision Measuring Instruments Meaning of Symbols ABSOLUTE Linear Encoder Mitutoyo's technology has realized the absolute position method (absolute method). With this method, you do not have to reset the system to zero after turning it off and then turning it on. The position information recorded on the scale is read every time. The following three types of absolute encoders are available: electrostatic capacitance model, electromagnetic induction model and model combining the electrostatic capacitance and optical methods. These encoders are widely used in a variety of measuring instruments as the length measuring system that can generate highly reliable measurement data. Advantages: 1. No count error occurs even if you move the slider or spindle extremely rapidly. 2. You do not have to reset the system to zero when turning on the system after turning it off*1.
    [Show full text]
  • I Introduction to Measurement and Data Analysis
    I Introduction to Measurement and Data Analysis Introduction to Measurement In physics lab the activity in which you will most frequently be engaged is measuring things. Using a wide variety of measuring instruments you will measure times, temperatures, masses, forces, speeds, frequencies, energies, and many more physical quantities. Your tools will span a range of technologies from the simple (such as a ruler) to the complex (perhaps a digital computer). Certainly it would be worthwhile to devote a little time and thought to some of the details of \measuring things" that may have not yet occurred to you. True Value - How Tall? At first thought you might suppose that the goal of measurement is a very straightforward one: find the true value of the thing being measured. Alas, things are seldom as simple as we would like. Consider the following \case study." Suppose you wished to measure how your lab partner's height. One way might be to simply look at him or her and estimate, \Oh, about five-nine," meaning five feet, nine inches tall. Of course you couldn't be sure that five-eight or five-ten, or even five-eleven might be a better estimate. In other words, your measurement (estimate) is uncertain by some amount, perhaps an inch or two either way. The \true value" lies somewhere within a range of uncertainty and one way to express this notion is to say that your partner's height is five feet, nine inches plus or minus two inches or 69 ± 2 inches. It should begin to be clear that at least one of the goals of measurement is to reduce the uncertainty to as small an amount as is feasible and useful.
    [Show full text]
  • Experiment Zero (Lab Intro) Spring 2021 PHYS221 Introduction to Measurement & Error Analysis Data Sheet
    Experiment zero (Lab intro) Spring 2021 PHYS221 Introduction to measurement & error analysis data sheet http://relativity.phy.olemiss.edu/~thomas/ NAME________________________ Section_____ Date_________ In this lab you will measure various quantities using either an analog device (which does not have a digital display, e.g., a ruler or triple beam balance) or a device with a digital display (e.g., a digital caliper). The readability (or precision) of any type of device is defined as the smallest quantity that can be read from (i.e., measured by) that device. The uncertainty of an analog device is generally (but not always) defined as ½ of the readability of the scale. The uncertainty of a digital device is generally defined as the readability of the digital display. Part one- Readability and uncertainty of a measuring instrument Instrument Readability Uncertainty (include units) (include units) ruler +/- protractor +/- triple beam balance +/- electronic (digital) balance +/- Vernier caliper Use readability +/- Part two-Measurement of dimensions of metal block. Please note that dimensions (lengths) of the sides of the blocks are denoted as long, medium and short. All measurements must have units when applicable and all uncertainties should be written using one significant figure. For example (51.5 +/- 0.5) cm. This can also be written as 51.5 +/- 0.5 cm Length ± Absolute uncertainty of length Dimension (length) Ruler Caliper measured (include units) (include units) long±δlong medium±δmedium short±δshort 25.40 +/- mm How to calculate fractional of percent error. Say you are given a measurement with an absolute uncertainty of (51.5 +/- 0.5) cm. To convert to fractional (percent) uncertainty you would do the following calculation: Experiment zero (Lab intro) Spring 2021 PHYS221 Introduction to measurement & error analysis data sheet http://relativity.phy.olemiss.edu/~thomas/ 0.5 �� × 100 (51.5 ± 0.5)�� → 51.5 �� ± → 51.5 �� ± 0.9709 %.
    [Show full text]
  • Measuring Instruments for Food
    Measuring instruments for food. 2019 Contents Contact temperature measurement Self-adhesive temperature foils testo mini thermometer Mini temperature measuring instruments testo temperature foils Folding thermometer testo 103 Folding thermometer testo 104 One-hand temperature measuring instrument testo 105 Core temperature thermometer testo 106 Temperature measuring instrument testo 108 Temperature measuring instrument (1-channel) testo 110 Temperature measuring instrument (1-channel) testo 112 Temperature measuring instrument (1-channel) testo 926 Temperature measuring instrument (3-channel) testo 735 Temperature measuring instrument (1-channel) testo 720 Non-contact temperature measurement IR temperature measuring instrument testo 805 Multi-purpose infrared and penetration thermometer testo 104-IR IR temperature measuring instrument testo 826-T2 IR temperature measuring instrument testo 826-T4 IR temperature measuring instrument testo 831 Temperature data loggers Mini data logger temperature testo 174T Temperature data logger testo 175 T1 Temperature data logger testo 175 T2 Temperature data logger testo 175 T3 Temperature data logger testo 176 T1 Temperature data logger testo 176 T2 USB data logger temperature testo 184 T1 USB data logger temperature testo 184 T2 USB data logger temperature testo 184 T3 WiFi data logger system testo Saveris 2 Humidity measuring instruments and data logggers Thermal hygrometer testo 608 Mini data logger for temperature and humidity testo 174H USB data logger for humidity and temperature testo 184 H1
    [Show full text]
  • An Overview of Particulate Matter Measurement Instruments
    Atmosphere 2015, 6, 1327-1345; doi:10.3390/atmos6091327 OPEN ACCESS atmosphere ISSN 2073-4433 www.mdpi.com/journal/atmosphere Review An Overview of Particulate Matter Measurement Instruments Simone Simões Amaral 1, João Andrade de Carvalho Jr. 1,*, Maria Angélica Martins Costa 2 and Cleverson Pinheiro 3 1 Department of Energy, UNESP—Univ Estadual Paulista, Guaratinguetá , SP 12516-410, Brazil; E-Mail: [email protected] 2 Department of Biochemistry and Chemical Technology, Institute of Chemistry, UNESP—Univ Estadual Paulista, Araraquara, SP 14800-060, Brazil; E-Mail: [email protected] 3 Department of Logistic, Federal Institute of Education, Science and Technology of São Paulo, Jacareí, SP 12322-030, Brazil; E-Mail: [email protected] * Author to whom correspondence should be addressed; E-Mail: [email protected]; Tel.: +55-12-3123-2838; Fax: +55-12-3123-2835. Academic Editor: Pasquale Avino Received: 27 June 2015 / Accepted: 1 September 2015 / Published: 9 September 2015 Abstract: This review article presents an overview of instruments available on the market for measurement of particulate matter. The main instruments and methods of measuring concentration (gravimetric, optical, and microbalance) and size distribution Scanning Mobility Particle Sizer (SMPS), Electrical Low Pressure Impactor (ELPI), and others were described and compared. The aim of this work was to help researchers choose the most suitable equipment to measure particulate matter. When choosing a measuring instrument, a researcher must clearly define the purpose of the study and determine whether it meets the main specifications of the equipment. ELPI and SMPS are the suitable devices for measuring fine particles; the ELPI works in real time.
    [Show full text]
  • Quantum Dynamics of a Small Symmetry Breaking Measurement
    Quantum dynamics of a small symmetry breaking measurement device H. C. Donkera, H. De Raedtb, M. I. Katsnelsona,∗ aRadboud University, Institute for Molecules and Materials, Heyendaalseweg 135, NL-6525AJ Nijmegen, The Netherlands bZernike Institute for Advanced Materials, University of Groningen, Nijenborgh 4, NL-9747 AG Groningen, The Netherlands Abstract A quantum measuring instrument is constructed that utilises symmetry break- ing to enhance a microscopic signal. The entire quantum system consists of a system-apparatus-environment triad that is composed of a small set of spin-1/2 particles. The apparatus is a ferromagnet that measures the z-component of a single spin. A full quantum many-body calculation allows for a careful ex- amination of the loss of phase coherence, the formation and amplification of system-apparatus correlations, the irreversibility of registration, the fault toler- ance, and the bias of the device. 1. Introduction Bohr, in his discussions with Einstein [1], repeatedly emphasised that each peculiar feature or seemingly paradoxical phenomenon in quantum mechanics, is always to be viewed in the light of the experimental arrangement that is used to interrogate the quantal test object (e.g., an electron or atom). But if one takes seriously the idea that the laws of quantum mechanics are universally valid, then, in turn, the measurement instrument itself must obey the laws of quantum mechanics. The first attempt to consolidate the internal consistency of measurement and quantum theory was by von Neumann [2]. By detailing measurement as a non-unitary disruption of the density operator to diagonal arXiv:1804.08151v1 [quant-ph] 22 Apr 2018 form (in the basis determined by the measurement), von Neumann noted that the same result can be accomplished by considering an enlarged quantum sys- tem in which the Hilbert space is partitioned in two AB = A B [2].
    [Show full text]
  • Pre AP Physics 1: Measurement Lab
    Pre AP Physics 1: Measurement Lab Introduction: Proper measurement is the key to successful work in the physics laboratory. The devices you will use most frequently for measuring length are the meter stick, the caliper, and the micrometer. The most common devices for measuring mass are the triple beam balance and the spring scale. In this exercise we will learn the correct way of using these devices and learn something about their accuracy and precision. All physical measurements are uncertain to some degree. The degree of uncertainty depends upon both the measuring instrument and the person making the measurement. These uncertainties are the result of several unavoidable factors: the limitations in the construction of the measuring instruments, the conditions under which the measurements are made, and the ability of the person using the instrument, to name a few. Repeated measurements can help to reduce uncertainty, but it can never be completely eliminated. Objective: After completing this experiment, you should be able to make proper measurements with a meter stick, caliper, micrometer, triple beam balance, and spring scale. You should also understand that some uncertainty in measurements is unavoidable. Apparatus: meter stick or metric ruler vernier caliper micrometer wooden block triple beam balance spring scale ring stand clamp Procedure: After each group has obtained the apparatus, your instructor will demonstrate the proper use of the caliper and micrometer and the reading of their scales. In making measurements with a meter stick, keep the following suggestions in mind. (a) Since the end of the meter stick may be worn, start the measurement at some intermediate mark like 10 cm.
    [Show full text]