Sequence-To-Point Learning with Neural Networks for Non-Intrusive Load Monitoring

Total Page:16

File Type:pdf, Size:1020Kb

Sequence-To-Point Learning with Neural Networks for Non-Intrusive Load Monitoring Sequence-to-point learning with neural networks for non-intrusive load monitoring Chaoyun Zhang1, Mingjun Zhong2, Zongzuo Wang1, Nigel Goddard1, and Charles Sutton1 1School of Informatics, University of Edinburgh, United Kingdom [email protected], fngoddard,[email protected] 2School of Computer Science, University of Lincoln, United Kingdom [email protected] Abstract usage for every appliance in a given household, multiple devices exhibiting similar power consumption, and simul- Energy disaggregation (a.k.a nonintrusive load monitoring, taneous switching on/off of multiple devices. Therefore en- NILM), a single-channel blind source separation problem, ergy disaggregation has been an active area for the appli- aims to decompose the mains which records the whole house electricity consumption into appliance-wise readings. This cation of artificial intelligence and machine learning tech- problem is difficult because it is inherently unidentifiable. niques. Popular approaches have been based on factorial Recent approaches have shown that the identifiability prob- hidden Markov models (FHMM) (Kolter and Jaakkola 2012; lem could be reduced by introducing domain knowledge into Parson et al. 2012; Zhong, Goddard, and Sutton 2013; 2014; the model. Deep neural networks have been shown to be a 2015; Lange and Berges´ 2016) and signal processing meth- promising approach for these problems, but sliding windows ods (Pattem 2012; Zhao, Stankovic, and Stankovic 2015; are necessary to handle the long sequences which arise in sig- 2016; Batra, Singh, and Whitehouse 2016; Tabatabaei, Dick, nal processing problems, which raises issues about how to and Xu 2017). combine predictions from different sliding windows. In this paper, we propose sequence-to-point learning, where the in- Recently, it has been shown that single-channel BSS put is a window of the mains and the output is a single point of could be modelled by using sequence-to-sequence (seq2seq) the target appliance. We use convolutional neural networks to learning with neural networks (Grais, Sen, and Erdogan train the model. Interestingly, we systematically show that the 2014; Huang et al. 2014; Du et al. 2016). In particular, it has convolutional neural networks can inherently learn the signa- been applied to energy disaggregation (Kelly and Knotten- tures of the target appliances, which are automatically added belt 2015a) —both convolutional (CNN) and recurrent neu- into the model to reduce the identifiability problem. We ap- ral networks (RNN) were employed. The idea of sequence- plied the proposed neural network approaches to real-world to-sequence learning is to train a deep network to map be- household energy data, and show that the methods achieve tween an input sequence, such as the mains power readings state-of-the-art performance, improving two standard error in the NILM problem, and an output sequence, such as the measures by 84% and 92%. power readings of a single appliance. Energy disaggregation (Hart 1992) is a single-channel blind A difficulty immediately arises when applying seq2seq in source separation (BSS) problem that aims to decompose signal processing applications such as BSS. In these applica- the whole energy consumption of a dwelling into the en- tions, the input and output sequences can be long, for exam- ergy usage of individual appliances. The purpose is to help ple, in one of our data sets, the input and output sequences households to reduce their energy consumption by helping are 14,400 time steps. Such long sequences can make train- them to understand what is causing them to use energy, and ing both computationally difficult, both because of memory it has been shown that disaggregated information can help limitations in current graphics processing units (GPUs) and, householders to reduce energy consumption by as much as with RNNs, because of the vanishing gradient problem. A arXiv:1612.09106v3 [stat.AP] 18 Sep 2017 5 − 15% (Fischer 2008). However, current electricity me- common way to avoid these problems is a sliding window ters can only report the whole-home consumption data. This approach, that is, training the network to map a window of triggers the demand of machine-learning tools to infer the the input signal to the corresponding window of the out- appliance-specific consumption. put signal. However, this approach has several difficulties, Energy disaggregation is unidentifiable and thus a diffi- in that each element of the output signal is predicted many cult prediction problem because it is a single-channel BSS times, once for each sliding window; an average of multiple problem; we want to extract more than one source from a predictions is naturally used, which consequently smooths single observation. Additionally, there are a large number the edges. Further, we expect that some of the sliding win- of sources of uncertainty in the prediction problem, includ- dows will provide a better prediction of a single element ing noise in the data, lack of knowledge of the true power than others; particularly, those windows where the element is near the midpoint of the window rather than the edges, so Copyright c 2018, Association for the Advancement of Artificial that the network can make use of all nearby regions of the Intelligence (www.aaai.org). All rights reserved. input signal, past and future. But a simple sliding window approach cannot exploit this information. 2014). Various inference algorithms could then be employed In this paper, we propose a different architecture called to infer the appliance signals fXig (Kolter and Jaakkola sequence-to-point learning (seq2point) for single-channel 2012; Zhong, Goddard, and Sutton 2014; Shaloudegi et al. BSS. This uses a sliding window approach, but given a win- 2016). It is well-known that the problem is still uniden- dow of the input sequence, the network is trained to predict tifiable. To tackle the identifiability problem, various ap- the output signal only at the midpoint of the window. This proaches have been proposed by incorporating domain has the effect of making the prediction problem easier on the knowledge into the model. For example, local informa- network, as rather than needing to predict in total W (T −W ) tion, e.g., appliance power levels, ON-OFF state changes, outputs as in the seq2seq method, where T is the length of and durations, has been incorporated into the model (Kolter the input signal and W the size of the sliding window, the and Jaakkola 2012; Parson et al. 2012; Pattem 2012; Zhao, seq2point method predicts only T outputs. This allows the Stankovic, and Stankovic 2015; Tabatabaei, Dick, and Xu neural network to focus its representational power on the 2017); others have incorporated global information, e.g., to- midpoint of the window, rather than on the more difficult tal number of cycles and total energy consumption (Zhong, outputs on the edges, hopefully yielding more accurate pre- Goddard, and Sutton 2014; 2015; Batra, Singh, and White- dictions. house 2016). However, the domain knowledge required by We provide both an analytical and empirical analysis of these methods needs to be extracted manually, which makes the methods, showing that seq2point has a tighter approxi- the methods more difficult to use. As was previously noted, mation to the target distribution than seq2seq learning. On all these approaches require handcrafted features based on two different real-world NILM data sets (UK-DALE (Kelly the observation data. Instead, we propose to employ neu- and Knottenbelt 2015b) and REDD (Kolter and Johnson ral networks to extract those features automatically during 2011)), we find that sequence-to-point learning performs learning. dramatically better than previous work, with as much as 83% reduction in error. Sequence-to-sequence learning Finally, to have confidence in the models, it is vital to in- Kelly and Knottenbelt [2015a] have applied deep learning terpret the model predictions and understand what informa- methods to NILM. The neural networks learns a nonlin- tion the neural networks for NILM are relying on to make ear regression between a sequence of the mains readings their predictions. By visualizing the feature maps learned and a sequence of appliance readings with the same time by our networks, we found that our networks automati- stamps. We will refer to this as a sequence-to-sequence ap- cally extract useful features of the input signal, such as proach. Although RNN architectures are most commonly change points, and typical usage durations and power lev- used in sequence-to-sequence learning for text (Sutskever, els of appliances. Interestingly, these signatures have been Vinyals, and Le 2014), for NILM Kelly and Knottenbelt commonly incorporated into handcrafted features and archi- [2015a] employ both CNNs and RNNs. Similar sequence- tectures for the NILM problem (Kolter and Jaakkola 2012; to-sequence neural network approaches have been applied Parson et al. 2012; Pattem 2012; Zhao, Stankovic, and to single-channel BSS problems in audio and speech (Grais, Stankovic 2015; Zhong, Goddard, and Sutton 2014; 2015; Sen, and Erdogan 2014; Huang et al. 2014; Du et al. 2016). Batra, Singh, and Whitehouse 2016; Tabatabaei, Dick, and Sequence-to-sequence architectures define a neural net- Xu 2017), but in our work these features are learned auto- work Fs that maps sliding windows Yt:t+W −1 of the input matically. mains power to corresponding windows Xt:t+W −1 of the output appliance power, that is, they model Xt:t+W −1 = Energy disaggregation Fs(Yt:t+W −1)+, where is W -dimensional Gaussian ran- The goal of energy disaggregation is to recover the energy dom noise. Then, to train the network on a pair (X; Y ) of consumption of individual appliances from the mains sig- full sequences, the loss function is nal, which measures the total electricity consumption. Sup- T −W +1 X pose we have observed the mains Y which indicates the total Ls = log p(Xt:t+W −1jYt:t+W −1; θs); (1) power in Watts in a household, where Y = (y1; y2; :::; yT ) t=1 and yt 2 R+.
Recommended publications
  • Completeness in Quasi-Pseudometric Spaces—A Survey
    mathematics Article Completeness in Quasi-Pseudometric Spaces—A Survey ¸StefanCobzas Faculty of Mathematics and Computer Science, Babe¸s-BolyaiUniversity, 400 084 Cluj-Napoca, Romania; [email protected] Received:17 June 2020; Accepted: 24 July 2020; Published: 3 August 2020 Abstract: The aim of this paper is to discuss the relations between various notions of sequential completeness and the corresponding notions of completeness by nets or by filters in the setting of quasi-metric spaces. We propose a new definition of right K-Cauchy net in a quasi-metric space for which the corresponding completeness is equivalent to the sequential completeness. In this way we complete some results of R. A. Stoltenberg, Proc. London Math. Soc. 17 (1967), 226–240, and V. Gregori and J. Ferrer, Proc. Lond. Math. Soc., III Ser., 49 (1984), 36. A discussion on nets defined over ordered or pre-ordered directed sets is also included. Keywords: metric space; quasi-metric space; uniform space; quasi-uniform space; Cauchy sequence; Cauchy net; Cauchy filter; completeness MSC: 54E15; 54E25; 54E50; 46S99 1. Introduction It is well known that completeness is an essential tool in the study of metric spaces, particularly for fixed points results in such spaces. The study of completeness in quasi-metric spaces is considerably more involved, due to the lack of symmetry of the distance—there are several notions of completeness all agreing with the usual one in the metric case (see [1] or [2]). Again these notions are essential in proving fixed point results in quasi-metric spaces as it is shown by some papers on this topic as, for instance, [3–6] (see also the book [7]).
    [Show full text]
  • "Mathematics for Computer Science" (MCS)
    “mcs” — 2016/9/28 — 23:07 — page i — #1 Mathematics for Computer Science revised Wednesday 28th September, 2016, 23:07 Eric Lehman Google Inc. F Thomson Leighton Department of Mathematics and the Computer Science and AI Laboratory, Massachussetts Institute of Technology; Akamai Technologies Albert R Meyer Department of Electrical Engineering and Computer Science and the Computer Science and AI Laboratory, Massachussetts Institute of Technology 2016, Eric Lehman, F Tom Leighton, Albert R Meyer. This work is available under the terms of the Creative Commons Attribution-ShareAlike 3.0 license. “mcs” — 2016/9/28 — 23:07 — page ii — #2 “mcs” — 2016/9/28 — 23:07 — page iii — #3 Contents I Proofs Introduction 3 0.1 References4 1 What is a Proof? 5 1.1 Propositions5 1.2 Predicates8 1.3 The Axiomatic Method8 1.4 Our Axioms9 1.5 Proving an Implication 11 1.6 Proving an “If and Only If” 13 1.7 Proof by Cases 15 1.8 Proof by Contradiction 16 1.9 Good Proofs in Practice 17 1.10 References 19 2 The Well Ordering Principle 29 2.1 Well Ordering Proofs 29 2.2 Template for Well Ordering Proofs 30 2.3 Factoring into Primes 32 2.4 Well Ordered Sets 33 3 Logical Formulas 47 3.1 Propositions from Propositions 48 3.2 Propositional Logic in Computer Programs 51 3.3 Equivalence and Validity 54 3.4 The Algebra of Propositions 56 3.5 The SAT Problem 61 3.6 Predicate Formulas 62 3.7 References 67 4 Mathematical Data Types 93 4.1 Sets 93 4.2 Sequences 98 4.3 Functions 99 4.4 Binary Relations 101 4.5 Finite Cardinality 105 “mcs” — 2016/9/28 — 23:07 — page iv — #4 Contentsiv 5 Induction 125 5.1 Ordinary Induction 125 5.2 Strong Induction 134 5.3 Strong Induction vs.
    [Show full text]
  • MTH 304: General Topology Semester 2, 2017-2018
    MTH 304: General Topology Semester 2, 2017-2018 Dr. Prahlad Vaidyanathan Contents I. Continuous Functions3 1. First Definitions................................3 2. Open Sets...................................4 3. Continuity by Open Sets...........................6 II. Topological Spaces8 1. Definition and Examples...........................8 2. Metric Spaces................................. 11 3. Basis for a topology.............................. 16 4. The Product Topology on X × Y ...................... 18 Q 5. The Product Topology on Xα ....................... 20 6. Closed Sets.................................. 22 7. Continuous Functions............................. 27 8. The Quotient Topology............................ 30 III.Properties of Topological Spaces 36 1. The Hausdorff property............................ 36 2. Connectedness................................. 37 3. Path Connectedness............................. 41 4. Local Connectedness............................. 44 5. Compactness................................. 46 6. Compact Subsets of Rn ............................ 50 7. Continuous Functions on Compact Sets................... 52 8. Compactness in Metric Spaces........................ 56 9. Local Compactness.............................. 59 IV.Separation Axioms 62 1. Regular Spaces................................ 62 2. Normal Spaces................................ 64 3. Tietze's extension Theorem......................... 67 4. Urysohn Metrization Theorem........................ 71 5. Imbedding of Manifolds..........................
    [Show full text]
  • General Topology
    General Topology Tom Leinster 2014{15 Contents A Topological spaces2 A1 Review of metric spaces.......................2 A2 The definition of topological space.................8 A3 Metrics versus topologies....................... 13 A4 Continuous maps........................... 17 A5 When are two spaces homeomorphic?................ 22 A6 Topological properties........................ 26 A7 Bases................................. 28 A8 Closure and interior......................... 31 A9 Subspaces (new spaces from old, 1)................. 35 A10 Products (new spaces from old, 2)................. 39 A11 Quotients (new spaces from old, 3)................. 43 A12 Review of ChapterA......................... 48 B Compactness 51 B1 The definition of compactness.................... 51 B2 Closed bounded intervals are compact............... 55 B3 Compactness and subspaces..................... 56 B4 Compactness and products..................... 58 B5 The compact subsets of Rn ..................... 59 B6 Compactness and quotients (and images)............. 61 B7 Compact metric spaces........................ 64 C Connectedness 68 C1 The definition of connectedness................... 68 C2 Connected subsets of the real line.................. 72 C3 Path-connectedness.......................... 76 C4 Connected-components and path-components........... 80 1 Chapter A Topological spaces A1 Review of metric spaces For the lecture of Thursday, 18 September 2014 Almost everything in this section should have been covered in Honours Analysis, with the possible exception of some of the examples. For that reason, this lecture is longer than usual. Definition A1.1 Let X be a set. A metric on X is a function d: X × X ! [0; 1) with the following three properties: • d(x; y) = 0 () x = y, for x; y 2 X; • d(x; y) + d(y; z) ≥ d(x; z) for all x; y; z 2 X (triangle inequality); • d(x; y) = d(y; x) for all x; y 2 X (symmetry).
    [Show full text]
  • Continuity Properties and Sensitivity Analysis of Parameterized Fixed
    Continuity properties and sensitivity analysis of parameterized fixed points and approximate fixed points Zachary Feinsteina Washington University in St. Louis August 2, 2016 Abstract In this paper we consider continuity of the set of fixed points and approximate fixed points for parameterized set-valued mappings. Continuity properties are provided for the fixed points of general multivalued mappings, with additional results shown for contraction mappings. Further analysis is provided for the continuity of the approxi- mate fixed points of set-valued functions. Additional results are provided on sensitivity of the fixed points via set-valued derivatives related to tangent cones. Key words: fixed point problems; approximate fixed point problems; set-valued continuity; data dependence; generalized differentiation MSC: 54H25; 54C60; 26E25; 47H04; 47H14 1 Introduction Fixed points are utilized in many applications. Often the parameters of such models are estimated from data. As such it is important to understand how the set of fixed points aZachary Feinstein, ESE, Washington University, St. Louis, MO 63130, USA, [email protected]. 1 changes with respect to the parameters. We motivate our general approach by consider applications from economics and finance. In particular, the solution set of a parameterized game (i.e., Nash equilibria) fall under this setting. Thus if the parameters of the individuals playing the game are not perfectly known, sensitivity analysis can be done in a general way even without unique equilibrium. For the author, the immediate motivation was from financial systemic risk models such as those in [10, 8, 4, 12] where methodology to estimate system parameters are studied in, e.g., [15].
    [Show full text]
  • Inspector's Handbook: Subbase Construction
    INSPECTOR'S HANDBOOK SUBBASE CONSTRUCTION IOWA STATE HIGHWAY COMMISSION AMES, IOWA 1969 r 17:_:H53 s~sl4 1969 \------ -- "- SUBBASE CONSTRUCTION Walter Schneider Jerry Rodibaugh INTRODUCTION This handbook is an inspector's aid. It was written by two inspectors to bring together a 11 of the most­ of ten-needed information involved in their work. Much care has been taken to detai I each phase of construction, with particular attention to the r.equire­ ments and limitations of specifications. All applicable specification interpretations in Instructions to Resident Engineers have been included. The beginning inspector should look to the hand­ book as a reference for standards of good practice. The Standard.Specifications and Special Provisions should not, however, be overlooked as the basic sources of information on requirements and restrictions concerning workmanship and materials. I . CONTENTS SUBBASE CONSTRUCTION Page 1 Definition 1 Types 1 Figure 1 - Typical Section 1 INSPECTION 1 Plans, Proposals, and Specifications · 1 Preconstruction Detai Is 2 I l. STAKING 2 Purpose 2 Techniques 2 -- SUBGRADE CORRECTION 3 Application 3 Profile and Cross-Section Requirements 4 Figure 2 - Flexible Base Construction 5 INSPECTING DIFFERENT TYPES OF ROLLERS USED IN SUBBASE CONSTRUCTION 5 Tamping Rollers . s· Self-Propelled, Steel-Tired Rollers 5 Self-Propelled and Pull-Type Pneumatic Tire Roi lers 6 Figure 3 - Pneumatic-Tired Roller 6 SPREAD RA TES AND PLAN QUANTITIES 6 Spread Chart 6 Fig.ure 4 - Typical Spr~ad chart 7 Figure 5 - scale Ti ck 8 et -- Granular
    [Show full text]
  • Sufficient Generalities About Topological Vector Spaces
    (November 28, 2016) Topological vector spaces Paul Garrett [email protected] http:=/www.math.umn.edu/egarrett/ [This document is http://www.math.umn.edu/~garrett/m/fun/notes 2016-17/tvss.pdf] 1. Banach spaces Ck[a; b] 2. Non-Banach limit C1[a; b] of Banach spaces Ck[a; b] 3. Sufficient notion of topological vector space 4. Unique vectorspace topology on Cn 5. Non-Fr´echet colimit C1 of Cn, quasi-completeness 6. Seminorms and locally convex topologies 7. Quasi-completeness theorem 1. Banach spaces Ck[a; b] We give the vector space Ck[a; b] of k-times continuously differentiable functions on an interval [a; b] a metric which makes it complete. Mere pointwise limits of continuous functions easily fail to be continuous. First recall the standard [1.1] Claim: The set Co(K) of complex-valued continuous functions on a compact set K is complete with o the metric jf − gjCo , with the C -norm jfjCo = supx2K jf(x)j. Proof: This is a typical three-epsilon argument. To show that a Cauchy sequence ffig of continuous functions has a pointwise limit which is a continuous function, first argue that fi has a pointwise limit at every x 2 K. Given " > 0, choose N large enough such that jfi − fjj < " for all i; j ≥ N. Then jfi(x) − fj(x)j < " for any x in K. Thus, the sequence of values fi(x) is a Cauchy sequence of complex numbers, so has a limit 0 0 f(x). Further, given " > 0 choose j ≥ N sufficiently large such that jfj(x) − f(x)j < " .
    [Show full text]
  • Seqgan: Sequence Generative Adversarial Nets with Policy Gradient
    Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence (AAAI-17) SeqGAN: Sequence Generative Adversarial Nets with Policy Gradient Lantao Yu,† Weinan Zhang,†∗ Jun Wang,‡ Yong Yu† †Shanghai Jiao Tong University, ‡University College London {yulantao,wnzhang,yyu}@apex.sjtu.edu.cn, [email protected] Abstract of sequence increases. To address this problem, (Bengio et al. 2015) proposed a training strategy called scheduled sampling As a new way of training generative models, Generative Ad- (SS), where the generative model is partially fed with its own versarial Net (GAN) that uses a discriminative model to guide synthetic data as prefix (observed tokens) rather than the true the training of the generative model has enjoyed considerable success in generating real-valued data. However, it has limi- data when deciding the next token in the training stage. Nev- tations when the goal is for generating sequences of discrete ertheless, (Huszar´ 2015) showed that SS is an inconsistent tokens. A major reason lies in that the discrete outputs from the training strategy and fails to address the problem fundamen- generative model make it difficult to pass the gradient update tally. Another possible solution of the training/inference dis- from the discriminative model to the generative model. Also, crepancy problem is to build the loss function on the entire the discriminative model can only assess a complete sequence, generated sequence instead of each transition. For instance, while for a partially generated sequence, it is non-trivial to in the application of machine translation, a task specific se- balance its current score and the future one once the entire quence score/loss, bilingual evaluation understudy (BLEU) sequence has been generated.
    [Show full text]
  • Topological Vector Spaces
    An introduction to some aspects of functional analysis, 3: Topological vector spaces Stephen Semmes Rice University Abstract In these notes, we give an overview of some aspects of topological vector spaces, including the use of nets and filters. Contents 1 Basic notions 3 2 Translations and dilations 4 3 Separation conditions 4 4 Bounded sets 6 5 Norms 7 6 Lp Spaces 8 7 Balanced sets 10 8 The absorbing property 11 9 Seminorms 11 10 An example 13 11 Local convexity 13 12 Metrizability 14 13 Continuous linear functionals 16 14 The Hahn–Banach theorem 17 15 Weak topologies 18 1 16 The weak∗ topology 19 17 Weak∗ compactness 21 18 Uniform boundedness 22 19 Totally bounded sets 23 20 Another example 25 21 Variations 27 22 Continuous functions 28 23 Nets 29 24 Cauchy nets 30 25 Completeness 32 26 Filters 34 27 Cauchy filters 35 28 Compactness 37 29 Ultrafilters 38 30 A technical point 40 31 Dual spaces 42 32 Bounded linear mappings 43 33 Bounded linear functionals 44 34 The dual norm 45 35 The second dual 46 36 Bounded sequences 47 37 Continuous extensions 48 38 Sublinear functions 49 39 Hahn–Banach, revisited 50 40 Convex cones 51 2 References 52 1 Basic notions Let V be a vector space over the real or complex numbers, and suppose that V is also equipped with a topological structure. In order for V to be a topological vector space, we ask that the topological and vector spaces structures on V be compatible with each other, in the sense that the vector space operations be continuous mappings.
    [Show full text]
  • NETS There Are a Number of Footnotes, Which Can Safely Be Ignored
    NETS There are a number of footnotes, which can safely be ignored, especially on the first reading. 1. Nets, convergence, continuity Nets are generalization of sequences needed to deal with convergence in general topological spaces, where convergent sequences are not enough to describe the topology. The idea is to replace the set of natural numbers, which is used to index elements of sequences, by arbitrary sets such that it makes sense to talk about \large" elements. Recall that a partial order on a set I is a binary relation ≺ such that (a) i ≺ i; (b) if i ≺ j and j ≺ i; then i = j; (c) if i ≺ j and j ≺ k; then i ≺ k: Definition 1.1. A set I with a partial order1 ≺ is called directed if for any i; j 2 I there exists k 2 I such that i ≺ k and j ≺ k. The following example will be our main source of directed sets. Example 1.2. Let X be a set and I be a collection of subsets of X closed under finite intersections, so if A; B 2 I then A \ B 2 I. Define a partial order on I by A ≺ B iff B ⊂ A: Then I is a directed set: for any A; B 2 I, we have A ≺ A \ B and B ≺ A \ B. Definition 1.3. A net in a set X is a collection (xi)i2I of elements of X indexed by a nonempty directed set I, or in other words, it is a map I ! X. Definition 1.4.
    [Show full text]
  • Well-Order As a Construction Principle for Physical Theories
    Eur. Phys. J. C (2019) 79:937 https://doi.org/10.1140/epjc/s10052-019-7456-2 Regular Article - Theoretical Physics Well-order as a construction principle for physical theories Peter Schusta Sebastian-Bauer-Str. 37a, 81737 Munich, Germany Received: 19 August 2019 / Accepted: 4 November 2019 / Published online: 18 November 2019 © The Author(s) 2019 Abstract Physics has up to now missed to express in math- together with a binary order relation R that is itself a subset ematical terms the fundamental idea of events of a path in of the set of all ordered pairs A × A of A;see[1, p 10, p 17], time and space uniquely succeeding one another. An appro- [2, p 64, p 80]. priate mathematical concept that reflects this idea is a well- A densely ordered set is an ordered set in which between ordered set. In such a set every subset has a least element. any two elements of the set exists another element of the set; Thus every element of a well-ordered set has as its definite see [2, p 205]. Well known examples for this are the rational successor the least element of the subset of all elements larger numbers or the real numbers together with the natural order than itself. This is apparently contradictory to the densely of numbers; see [2, p 213]. By employing the real number ordered real number lines which conventionally constitute line as coordinate axes this is transferred into the resulting the coordinate axes in any representation of time and space Rn. The open balls as the base of the Euclidean topology are and in which between any two numbers exists always another a reflection of this.
    [Show full text]
  • Nets, the Axiom of Choice, and the Tychonoff Theorem
    Dean Katsaros 12/10/15 Nets, The Axiom of Choice, and the Tychonoff Theorem 1 A Brief History of the Tychonoff Theorem and The Axiom of Choice It turns out that the Axiom of Choice has direct implications for point-set topology. This might not be that interesting at first mention, nor should it be surprising given the intimate role that set theoretical methods play in proving the many results of point-set topology. Also un-surprising is that the following form of Axiom of Choice: Theorem 1.1 (Axiom of Choice). For every family (Xα)α2J of non-empty sets Xα, the product set Q α2J Xα is non-empty. Can be proven using the Tychonoff Theorem: Theorem 1.2 (Tychonoff Theorem). Let fXαgα2J such that each Xα is compact. Then, X = Q α2J Xα is compact in the product topology. This shouldn't be surprising because of the relative similarity of the statements - no matter one's mathematical background, the similarly of the statements is immediate. What is interesting is the role played by certain topologically motivated objects called nets in proving that the Axiom of Choice is in fact equivalent to the Tychonoff Theorem. The Tychonoff Theorem was stated in full generality in 1930 by Andrey Nikolayevich Tikhonov (Ty- chonoff), and was proven in 1937 by Eduard Cech. The Axiom of Choice and its equivalent statements became conjectures around the same time. The history of the AoC takes us back to Cantor, who spent 1 much time during his mathematical career considering the following two conjectures, the first of which he originally thought to be self evident (i.e., that it didn't need to be proven).
    [Show full text]