Quantifying the Consonance of Complex Tones with Missing Fundamentals

Total Page:16

File Type:pdf, Size:1020Kb

Quantifying the Consonance of Complex Tones with Missing Fundamentals QUANTIFYING THE CONSONANCE OF COMPLEX TONES WITH MISSING FUNDAMENTALS A THESIS SUBMITTED TO THE DEPARTMENT OF ELECTRICAL ENGINEERING AND THE COMMITTEE ON GRADUATE STUDIES OF STANFORD UNIVERSITY IN PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR THE DEGREE OF ENGINEER Song Hui Chon June 2008 c Copyright by Song Hui Chon 2008 All Rights Reserved ii Approved for the department. (Julius O. Smith, III) Advisor Approved for the Stanford University Committee on Graduate Studies. iii Abstract Consonance is one of the more fundamental perceptual criteria in music and sound on which Western music theory is built. It is closely related to the musical intervals, explaining why certain intervals sound better than others. Recently, the concept of tonal consonance has been distinguished from that of mu- sical consonance. Tonal consonance is about consonance or “pleasantness” of tones when sounded together, while musical consonance is about intervals. Tonal conso- nance appears to be a superset of musical consonance, in that musically consonant intervals are tonally consonant. The “missing fundamental” (or “virtual pitch”) is an interesting psychoacoustical phenomenon. When we hear a set of tones whose frequencies are integer multiples of a fundamental that is missing in the sound, we identify it as a complex tone whose pitch is the missing fundamental frequency. This phenomenon enables producing decent bass sound out of mini-sized speakers. Then the question is, why do we hear something that is not there? This thesis deals with the intersection of tonal consonance and missing funda- mental. While trying to explain some data from a psychoacoustical experiment, I had stumbled onto the concept tonal consonance. This work is built upon the earlier work, with the addition of missing fundamental analysis. The work covered in this thesis finds that the consonance of most sound stimuli stayed pretty constant regardless of the loudness level at which they were presented. iv It also supports a previous conclusion that each type of stimulus has its own intrinsic consonance value. A new discovery here is that the consonance values for the stimuli considered seem to be grouped by the size of bandwidth. v Acknowledgments I would first like to thank my advisor, Julius O. Smith, for his constant support and guidance. It was his advice that first prompted me to embark on this exciting journey of applying engineering techniques to the applications of music. I would also like to thank the CCRMA community including the faculty there – Jonathan Berger, Chris Chafe, Jonathan Abel, Marina Bosi, Malcolm Slaney and David Berners. I had the honor of being the teaching assistant for Drs. Abel, Berners and Bosi, to whom I owe much of my audio compression and effect knowledge. They were teachers, mentors and friends for the past four years of my life. There are my friends and colleagues at CCRMA and Electrical Engineering who have been there for me when I needed a guidance – Woon Seung Yeo, Kyogu Lee, Sook Young Won, Juhan Nam, Minjong Kim, Greg Sell, Gautham Mysore, Ed Berdahl and David Yeh. They would teach, help and discuss with me and that eventually led me here. I would especially like to thank Daniel Steele, with whom I worked on a research project and published a paper together. In doing the research we collaborated, I found the topic of consonance quantification and its underlying problems, which eventually became this thesis. I owe a big thank to my church community at Calvary Chapel Mountain View, including Inma and David Robinson, Lisa Erickson, Regina and Kirill Buryak, Brid- get Ingham and others. They have been unvaryingly understanding, supportive and vi loving during my many ups and downs in the last few years, with prayer and encour- agement. My life at Stanford would have been a lot more challenging without their support. My special thanks go to Doctor Ik Keun Hwang of Chonbuk National University in Korea. Without his medical cares, I would not be here now studying what I am curious about. I am indebted to him eternally for his thorough care and the miracle he produced. The biggest thanks is for my family for their love and support. They have showed a continuous and unconditional love throughout my entire life. I feel very blessed to belong to my family. It was my parents’ love and encouragements that enabled me to pursue a degree at Stanford. My caring sister was always there for me with arms stretched. My brother, who is also studying psychoacoustics at Seoul National University in Korea, was a friend and colleague with whom I had numerous research conversations. Last, but certainly not the least, I thank God for my past four years at Stanford. He brought me here and opened the doors for me to study music and engineering together. My life often took an unexpected turn, but no matter what I was going through, God always was faithful to His Words. I hope that this thesis is a testimony to His glory. vii Contents Abstract iv Acknowledgments vi 1 Introduction 1 1.1 IntroductionandMotivation . 1 1.2 Consonance ................................ 2 1.2.1 TonalConsonance......................... 3 1.3 MissingFundamental........................... 4 2 Quantifying Consonance 6 2.1 Introduction................................ 6 2.2 TheProposedAlgorithm . 10 2.2.1 HowYINWorks ......................... 14 2.3 Experiment ................................ 16 2.3.1 A Simple Counterexample for Plomp and Levelt . 16 3 Annoyance Perception Experiment 21 3.1 Stimuli................................... 21 3.2 Procedure ................................. 23 3.3 TheMatterofWeighting. .. .. 24 viii 3.4 Results................................... 28 4 Results and Discussion 30 4.1 QTC Values of Annoyance Stimuli . 30 4.2 CorrelationCoefficients. 32 4.3 Fundamental Estimation using YIN . 34 4.3.1 Comparison of Estimated Fundamentals with Human Perception 38 5 Conclusions and Future Works 40 A A Perceptual Study of Sound Annoyance 42 B Sound Annoyance as a Function of Consonance 49 Bibliography 58 ix List of Tables 3.1 Theoretically calculated perceived loudness values using dB(A), dB(B), dB(C)anddB(ITU-R468) . 27 4.1 The QTC values calculated with the proposed algorithm . ..... 30 4.2 The correlation coefficients of QTC values and loudness levels per stim- ulustype.................................. 32 4.3 The correlation coefficients of QTC values and annoyance order, per loudnesslevel ............................... 32 x List of Figures 2.1 The original QTC algorithm (from Chon et al. [5]). .... 7 2.2 Illustration of consonant and dissonant parts in a pair of pure tones (fromChonetal.[5]). .......................... 8 2.3 ThemodifiedQTCalgorithm(from[6]). 9 2.4 Perceived consonance and dissonance of a pair of pure tones as defined by Plomp and Levelt in [17] (reproduced from [6]). 10 2.5 TheproposedQTCalgorithm.. 12 2.6 Magnitude spectrum of a) the complex tone, b) the pure tone and c) themixture................................. 17 2.7 Spectrogram of a) the complex tone, b) the pure tone and c) the mixture. 18 2.8 Time-domain waveforms of a) the complex tone, b) the pure tone and c)themixture. .............................. 19 3.1 The magnitude spectrum of the six stimuli . .. 22 3.2 Weighting standards: A-, B-, C-weighting and ITU-R 468 . ...... 24 3.3 Weighted decibel levels of six stimuli using A- (solid red line), B- (blue dash dotted line), C- (black dotted line) and ITU-R 468 (green dashed line)weighting............................... 25 3.4 The result from annoyance perception experiment [20] . ....... 28 4.1 QTC values of twenty-four stimuli . 31 xi 4.2 Magnitude spectrum of the mixture from the Simple Test in section 2.3.1(fromfigure2.6)........................... 34 4.3 Magnitude spectrum of the mixture from the Simple Test in section 2.3.1 and the estimated fundamentals using YIN . 35 4.4 Fundamental estimation procedure in the Peaks and Fundamentals blockinfigure2.5............................. 36 xii Chapter 1 Introduction 1.1 Introduction and Motivation Consonance is one of the oldest concepts in music and sound. It is an intuitive concept that produces ambiguities in attempting a formal definition. During my annoyance perception studies [6] [20] I became curious about this concept of consonance and started to wonder if I could somehow quantify the consonance of sound stimuli and whether it could in turn explain the annoyance experiment data. It was relatively successful, because it could explain most cases. While studying papers on consonance, I realized that the literature was sparse. And I began to wonder if there should have been more questions in this field, now that we have achieved a significant development in psychoacoustics and also in com- putational processing capabilities. Thus it is time to make advances. I have decided to pursue the issue of consonance quantification in the case of missing fundamentals, described in section 1.3. My last paper [6] used a classic algorithm by Plomp and Levelt from [17] to quantify disso- nance between two tones, which regarded two pure tones (i.e., sinusoids) completely consonant when they are over 1.2 times a critical bandwidth apart. This will be the 1 CHAPTER 1. INTRODUCTION 2 right judgment for cases with only two sinusoids considered. But in real life, most audio stimuli have complex harmonic structures in them; therefore Plomp and Lev- elt’s theory may not hold up well. In fact, a simple listening test reveals that there is a perceived beating created between a pure tone and a set of complex tones with a missing fundamental, where the missing fundamental and the pure tone are within the same critical bandwidth, while the lowest existing harmonic and the pure tone are more than a critical bandwidth apart. We can see using MATLAB that the time- domain waveforms exhibit a physical modulation whose rate corresponds to the rate of beats in this case. This suggests not only that there is a perceptual phenomenon that can be explained by nonlinear processing in ear, but also a physical phenomenon of beating even though the fundamental is not there, and that beating is translated as dissonance in psychoacoustics.
Recommended publications
  • L Atdment OFFICF QO9ENT ROOM 36
    *;JiQYL~dW~llbk~ieira - - ~-- -, - ., · LAtDMENT OFFICF QO9ENT ROOM 36 ?ESEARC L0ORATORY OF RL C.f:'__ . /j16baV"BLI:1!S INSTITUTE 0i'7Cn' / PERCEPTION OF MUSICAL INTERVALS: EVIDENCE FOR THE CENTRAL ORIGIN OF THE PITCH OF COMPLEX TONES ADRIANUS J. M. HOUTSMA JULIUS L. GOLDSTEIN LO N OPY TECHNICAL REPORT 484 OCTOBER I, 1971 MASSACHUSETTS INSTITUTE OF TECHNOLOGY RESEARCH LABORATORY OF ELECTRONICS CAMBRIDGE, MASSACHUSETTS 02139 The Research Laboratory of Electronics is an interdepartmental laboratory in which faculty members and graduate students from numerous academic departments conduct research. The research reported in this document was made possible in part by support extended the Massachusetts Institute of Tech- nology, Research Laboratory of Electronics, by the JOINT SER- VICES ELECTRONICS PROGRAMS (U. S. Army, U. S. Navy, and U.S. Air Force) under Contract No. DAAB07-71-C-0300, and by the National Institutes of Health (Grant 5 PO1 GM14940-05). Requestors having DOD contracts or grants should apply for copies of technical reports to the Defense Documentation Center, Cameron Station, Alexandria, Virginia 22314; all others should apply to the Clearinghouse for Federal Scientific and Technical Information, Sills Building, 5285 Port Royal Road, Springfield, Virginia 22151. THIS DOCUMENT HAS BEEN APPROVED FOR PUBLIC RELEASE AND SALE; ITS DISTRIBUTION IS UNLIMITED. MASSACHUSETTS INSTITUTE OF TECHNOLOGY RESEARCH LABORATORY OF ELECTRONICS Technical Report 484 October 1, 1971 PERCEPTION OF MUSICAL INTERVALS: EVIDENCE FOR THE CENTRAL ORIGIN OF COMPLEX TONES Adrianus J. M. Houtsma and Julius L. Goldstein This report is based on a thesis by A. J. M. Houtsma submitted to the Department of Electrical Engineering at the Massachusetts Institute of Technology, June 1971, in partial fulfillment of the requirements for the degree of Doctor of Philosophy.
    [Show full text]
  • Contribution of the Conditioning Stage to the Total Harmonic Distortion in the Parametric Array Loudspeaker
    Universidad EAFIT Contribution of the conditioning stage to the Total Harmonic Distortion in the Parametric Array Loudspeaker Andrés Yarce Botero Thesis to apply for the title of Master of Science in Applied Physics Advisor Olga Lucia Quintero. Ph.D. Master of Science in Applied Physics Science school Universidad EAFIT Medellín - Colombia 2017 1 Contents 1 Problem Statement 7 1.1 On sound artistic installations . 8 1.2 Objectives . 12 1.2.1 General Objective . 12 1.2.2 Specific Objectives . 12 1.3 Theoretical background . 13 1.3.1 Physics behind the Parametric Array Loudspeaker . 13 1.3.2 Maths behind of Parametric Array Loudspeakers . 19 1.3.3 About piezoelectric ultrasound transducers . 21 1.3.4 About the health and safety uses of the Parametric Array Loudspeaker Technology . 24 2 Acquisition of Sound from self-demodulation of Ultrasound 26 2.1 Acoustics . 26 2.1.1 Directionality of Sound . 28 2.2 On the non linearity of sound . 30 2.3 On the linearity of sound from ultrasound . 33 3 Signal distortion and modulation schemes 38 3.1 Introduction . 38 3.2 On Total Harmonic Distortion . 40 3.3 Effects on total harmonic distortion: Modulation techniques . 42 3.4 On Pulse Wave Modulation . 46 4 Loudspeaker Modelling by statistical design of experiments. 49 4.1 Characterization Parametric Array Loudspeaker . 51 4.2 Experimental setup . 52 4.2.1 Results of PAL radiation pattern . 53 4.3 Design of experiments . 56 4.3.1 Placket Burmann method . 59 4.3.2 Box Behnken methodology . 62 5 Digital filtering techniques and signal distortion analysis.
    [Show full text]
  • Large Scale Sound Installation Design: Psychoacoustic Stimulation
    LARGE SCALE SOUND INSTALLATION DESIGN: PSYCHOACOUSTIC STIMULATION An Interactive Qualifying Project Report submitted to the Faculty of the WORCESTER POLYTECHNIC INSTITUTE in partial fulfillment of the requirements for the Degree of Bachelor of Science by Taylor H. Andrews, CS 2012 Mark E. Hayden, ECE 2012 Date: 16 December 2010 Professor Frederick W. Bianchi, Advisor Abstract The brain performs a vast amount of processing to translate the raw frequency content of incoming acoustic stimuli into the perceptual equivalent. Psychoacoustic processing can result in pitches and beats being “heard” that do not physically exist in the medium. These psychoac- oustic effects were researched and then applied in a large scale sound design. The constructed installations and acoustic stimuli were designed specifically to combat sensory atrophy by exer- cising and reinforcing the listeners’ perceptual skills. i Table of Contents Abstract ............................................................................................................................................ i Table of Contents ............................................................................................................................ ii Table of Figures ............................................................................................................................. iii Table of Tables .............................................................................................................................. iv Chapter 1: Introduction .................................................................................................................
    [Show full text]
  • An Infrasonic Missing Fundamental Rises at 18.5Hz Christopher D
    Papers & Publications: Interdisciplinary Journal of Undergraduate Research Volume 2 Article 11 2013 An Infrasonic Missing Fundamental Rises at 18.5Hz Christopher D. Lacomba University of North Georgia Steven A. Lloyd University of North Georgia Ryan A. Shanks University of North Georgia Follow this and additional works at: http://digitalcommons.northgeorgia.edu/papersandpubs Part of the Behavior and Behavior Mechanisms Commons, Biological and Chemical Physics Commons, Biological Psychology Commons, Cognition and Perception Commons, Cognitive Neuroscience Commons, Cognitive Psychology Commons, Computational Neuroscience Commons, Dynamic Systems Commons, Investigative Techniques Commons, Mental Disorders Commons, Non-linear Dynamics Commons, Numerical Analysis and Computation Commons, Psychological Phenomena and Processes Commons, Speech and Hearing Science Commons, and the Systems Neuroscience Commons Recommended Citation Lacomba, Christopher D.; Lloyd, Steven A.; and Shanks, Ryan A. (2013) "An Infrasonic Missing Fundamental Rises at 18.5Hz," Papers & Publications: Interdisciplinary Journal of Undergraduate Research: Vol. 2 , Article 11. Available at: http://digitalcommons.northgeorgia.edu/papersandpubs/vol2/iss1/11 This Article is brought to you for free and open access by the Center for Undergraduate Research and Creative Activities (CURCA) at Nighthawks Open Institutional Repository. It has been accepted for inclusion in Papers & Publications: Interdisciplinary Journal of Undergraduate Research by an authorized editor of Nighthawks Open Institutional Repository. The Missing Fundamental (MF) phenomenon creates perceptible tones, which do not actually exist in the audible range, from harmonic series. Sounds that follow a harmonic series are perceptually significant since they stand out from random background noise and frequently represent sounds of biological origin and importance. For example, both animal vocalizations and human speech produce harmonic series of sounds.
    [Show full text]
  • 11 – Music Temperament and Pitch
    11 – Music temperament and pitch Music sounds Every music instrument, including the voice, uses a more or less well defined sequence of music sounds, the notes, to produce music. The notes have particular frequencies that are called „pitch“ by musicians. The frequencies or pitches of the notes stand in a particular relation to each other. There are and were different ways to build a system of music sounds in different geographical regions during different historical periods. We will consider only the currently dominating system (used in classical and pop music) that was originated in ancient Greece and further developed in Europe. Pitch vs interval Only a small part of the people with normal hearing are able to percieve the absolute frequency of a sound. These people, who are said to have the „absolute pitch“, can tell what key has been hit on a piano without looking at it. The human ear and brain is much more sensitive to the relations between the frequencies of two or more sounds, the music intervals. The intervals carry emotional content that makes music an art. For instance, the major third sounds joyful and affirmative whereas the minor third sounds sad. The system of frequency relations between the sounds used in music production is called music temperament . 1 Music intervals and overtone series Perfect music intervals (that cannot be fully achieved practically, see below) are based on the overtone series studied before in this course. A typical music sound (except of a pure sinusiodal one) consists of a fundamental frequency f1 and its overtones: = = fn nf 1, n 3,2,1 ,..
    [Show full text]
  • Impairment of the Missing Fundamental Phenomenon in Individuals with Alzheimer’S Disease: a Neuropsychological and Voxel-Based Morphometric Study
    EXTRA Dement Geriatr Cogn Disord Extra 2018;8:23–32 DOI: 10.1159/000486331 © 2018 The Author(s) Received: November 17, 2017 Published by S. Karger AG, Basel Accepted: December 7, 2017 www.karger.com/dee Published online: February 1, 2018 This article is licensed under the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 Interna- tional License (CC BY-NC-ND) (http://www.karger.com/Services/OpenAccessLicense). Usage and distribu- tion for commercial purposes as well as any distribution of modified material requires written permission. Original Research Article Impairment of the Missing Fundamental Phenomenon in Individuals with Alzheimer’s Disease: A Neuropsychological and Voxel-Based Morphometric Study a a, b a a Makiko Abe Ken-ichi Tabei Masayuki Satoh Mari Fukuda c d a Hironobu Daikuhara Mariko Shiga Hirotaka Kida a, b Hidekazu Tomimoto a Department of Dementia Prevention and Therapeutics, Graduate School of Medicine, b Mie University, Mie, Japan; Department of Neurology, Graduate School of Medicine, c Mie University, Mie, Japan; Graduate School of Medicine, Mie University, Mie, Japan; d Mie Prefectural Dementia-Related Disease Medical Center, Mie, Japan Keywords Alzheimer’s disease · Dementia · Pitch perception · Missing fundamental phenomenon · Magnetic resonance imaging · Voxel-based morphometry Abstract Background/Aims: The missing fundamental phenomenon (MFP) is a universal pitch percep- tion illusion that occurs in animals and humans. In this study, we aimed to determine wheth- er the MFP is impaired in patients with Alzheimer’s disease (AD) using an auditory pitch per- ception experiment. We further examined anatomical correlates of the MFP in patients with AD by measuring gray matter volume (GMV) on magnetic resonance images via voxel-based morphometric analysis.
    [Show full text]
  • Locomotion-Induced Sounds and Sonations: Mechanisms, Communication Function, and Relationship with Behavior
    Chapter 4 Locomotion-Induced Sounds and Sonations: Mechanisms, Communication Function, and Relationship with Behavior Christopher James Clark Abstract Motion-induced sound is an intrinsic byproduct of essentially all animal behavior. Locomotion-induced sounds that have evolved specialization for commu- nication are termed sonations. The null hypothesis is that locomotion-induced sounds are noncommunicative (adventitious), produced by nonspecialized mor- phology, and are involuntary. A sound is a sonation if it is produced by specialized morphology, or is produced voluntarily. The production of locomotion-induced sound can be examined at two levels: the animal motions (kinematics) that lead to sound production and the physical acoustic mechanism(s) that generate(s) the sound itself. The physical acoustics of locomotion induced sound are diverse, with both aerodynamic and structural mechanisms, including aeroelastic flutter, percussion, stridulation, and presumably many other undescribed mechanisms. There is a direct sound–motion correspondence between aspects of an animal’s motions and ensuing locomotion-induced sounds, especially in the time domain. This correspondence has two implications. One is experimental: sound recordings are a useful and perhaps underutilized source of data about animal locomotion. The second is behavioral: locomotion-induced sounds intrinsically contain information about an animal’s motions (such as wingbeat frequencies) that may be of interest to other animals. Therefore, locomotion-induced sounds are intrinsically suited to be mechanisms by which animals directly evaluate the locomotor performance of other animals, such as during courtship. The sound–motion correspondence is also a constraint. Sonations seem less acoustically diverse than vocalizations. Because they require discrete behaviors to be produced, animals also have somewhat fewer opportunities to produce sonations strategically, and few sonations are frequency-­ modulated.
    [Show full text]
  • Ghost Stochastic Resonance with Distributed Inputs in Pulse-Coupled Electronic Neurons
    PHYSICAL REVIEW E 73, 021101 ͑2006͒ Ghost stochastic resonance with distributed inputs in pulse-coupled electronic neurons Abel Lopera,1 Javier M. Buldú,1,* M. C. Torrent,1 Dante R. Chialvo,2 and Jordi García-Ojalvo1,† 1Departamento de Física i Enginyeria Nuclear, Universitat Politècnica de Catalunya, Colom 11, E-08222 Terrassa, Spain 2Department of Physiology, Northwestern University, Chicago, Illinois 60611, USA ͑Received 29 September 2005; published 8 February 2006͒ We study experimentally the phenomenon of ghost stochastic resonance in pulse-coupled excitable systems, for input signals distributed among different elements. Specifically, two excitable electronic circuits are driven by different sinusoidal signals that produce periodic spikes at distinct frequencies. Their outputs are sent to a third circuit that processes these spiking signals and is additionally perturbed by noise. When the input signals are harmonics of a certain fundamental ͑that is not present in the inputs͒ the processing circuit exhibits, for an optimal amount of noise, a resonant response at the frequency of the missing fundamental ͑ghost frequency͒. In contrast with the standard case in which the signals being directly integrated are sinusoidal, this behavior relies here on a coincidence-detection mechanism. When the input signals are homogeneously shifted in frequency, the processing circuit responds with pulse packages composed of spikes at a frequency that depends linearly on the frequency shift. Expressions for the dependence of the package period and duration on the frequency shift and spike width, respectively, are obtained. These results provide an experimental verification of a recently proposed mechanism of binaural pitch perception. DOI: 10.1103/PhysRevE.73.021101 PACS number͑s͒: 05.40.Ϫa, 05.45.Ϫa I.
    [Show full text]
  • The Sounds of Music: Science of Musical Scales∗ 1
    SERIES ARTICLE The Sounds of Music: Science of Musical Scales∗ 1. Human Perception of Sound Sushan Konar Both, human appreciation of music and musical genres tran- scend time and space. The universality of musical genres and associated musical scales is intimately linked to the physics of sound, and the special characteristics of human acoustic sensitivity. In this series of articles, we examine the science underlying the development of the heptatonic scale, one of the most prevalent scales of the modern musical genres, both western and Indian. Sushan Konar works on stellar compact objects. She Introduction also writes popular science articles and maintains a Fossil records indicate that the appreciation of music goes back weekly astrophysics-related blog called Monday Musings. to the dawn of human sentience, and some of the musical scales in use today could also be as ancient. This universality of musi- cal scales likely owes its existence to an amazing circularity (or periodicity) inherent in human sensitivity to sound frequencies. Most musical scales are specific to a particular genre of music, and there exists quite a number of them. However, the ‘hepta- 1 1 tonic’ scale happens to have a dominating presence in the world Having seven base notes. music scene today. It is interesting to see how this has more to do with the physics of sound and the physiology of human auditory perception than history. We shall devote this first article in the se- ries to understand the specialities of human response to acoustic frequencies. Human ear is a remarkable organ in many ways. The range of hearing spans three orders of magnitude in frequency, extending Keywords from ∼20 Hz to ∼20,000 Hz (Figure 1) even though the sensitivity String vibration, beat frequencies, consonance-dissonance, pitch, tone.
    [Show full text]
  • Part D: Characterizing Sound
    Part D : Characterizing Sound Part D: Characterizing Sound Chapter 36. Parameters and Perception Simple Harmonic Motion of an object is itself not sound, but a vibrating object can be a source of sound. Each of the physical parameters of SHM vibration is closely related to a perceptual characteristic of the sound produced. • The duration of a vibration, how long it lasts, is mainly determines the duration of the resulting sound. • Vibrations with larger amplitudes create sounds that are louder, all other things being equal. • Vibrations with higher frequencies cause sounds that are higher in pitch. • Since period is the reciprocal of frequency, it also relates directly to pitch. A longer period corresponds to a lower pitch. • The initial phase of a vibration has no bearing on the perception of the sound. Since the value of initial phase depends on a choice made by an observer (specifically, the choice of the time origin), it can’t have real-world consequences. Recall from Chapter 21 that phase differences can be physically meaningful, but those will only arise when multiple sounds combine. These relationships are not exclusive, as each of these perceptions (duration, loudness, and pitch) can be influenced by all of the physical parameters except initial phase. For instance, perceived loudness depends slightly on frequency. Nevertheless, the connections above are the principal ones. The characteristics of a sound can vary over time. Over the course of seconds, like notes in a song, you might consider that separate sounds, or one changing sound. On a more rapid scale, the exact variations over the first few milliseconds of a sound, called the attack in some contexts, can be very important for recognizing the source.
    [Show full text]
  • Tu.P3.1 II - 1739 Fundamental Frequencies Were Mistuned by ∆F Hz the Normalized ACF Calculated by In-Phase and Maintaining a Harmonic Relationship
    Multiple Missing Fundamental Phenomenon as a Monaural-Beat Perception Ryota Shimokura a and Yoichi Ando b Graduate School of Science and Technology, Kobe University Now at: a Department of Energy, Nuclear and Environmental Control Engineering, Bologna University [email protected] Now at: bDepartment of Architecture and Civil Engineering, Kumamoto University [email protected] Abstract phenomenon of periodicity pitch or residue pitch has played roles to explain the mechanism of pitch It is found that a monaural beat induced by two missing perception [1][2]. The fundamental frequency is fundamentals of complex tones (A and B) may be absent in the spectrum, however the pitch of it can be perceived, even in the condition without the envelope perceived. Schouten mixed a pure tone 206 Hz on a component of beat. This was realized by mixing the complex tone with fundamental frequency 200 Hz and complex tone A with the missing fundamental frequency indicated that the beat corresponding to 6 Hz did not f0a and the complex tone B with the missing occur [2]. In this experiment, we shall examine the fundamental frequency f0b. When we listen together, missing fundamental phenomenon perceived as a beat, then an additional missing fundamental frequency ∆f (= which was caused by two fundamental frequencies. Two f0b - f0a) may be perceived as the beat. The ∆f was different fundamental frequencies (f0a and f0b) by two changed to 2, 4, 8 and 16 Hz. The total tone mixed with different complex tones may induce an additional two complex tones has the envelope fluctuation fundamental frequency: ∆f (= f0a - f0b).
    [Show full text]
  • A. Acoustic Theory and Modeling of the Vocal Tract
    A. Acoustic Theory and Modeling of the Vocal Tract by H.W. Strube, Drittes Physikalisches Institut, Universität Göttingen A.l Introduction This appendix is intended for those readers who want to inform themselves about the mathematical treatment of the vocal-tract acoustics and about its modeling in the time and frequency domains. Apart from providing a funda­ mental understanding, this is required for all applications and investigations concerned with the relationship between geometric and acoustic properties of the vocal tract, such as articulatory synthesis, determination of the tract shape from acoustic quantities, inverse filtering, etc. Historically, the formants of speech were conjectured to be resonances of cavities in the vocal tract. In the case of a narrow constriction at or near the lips, such as for the vowel [uJ, the volume of the tract can be considered a Helmholtz resonator (the glottis is assumed almost closed). However, this can only explain the first formant. Also, the constriction - if any - is usu­ ally situated farther back. Then the tract may be roughly approximated as a cascade of two resonators, accounting for two formants. But all these approx­ imations by discrete cavities proved unrealistic. Thus researchers have have now adopted a more reasonable description of the vocal tract as a nonuni­ form acoustical transmission line. This can explain an infinite number of res­ onances, of which, however, only the first 2 to 4 are of phonetic importance. Depending on the kind of sound, the tube system has different topology: • for vowel-like sounds, pharynx and mouth form one tube; • for nasalized vowels, the tube is branched, with transmission from pharynx through mouth and nose; • for nasal consonants, transmission is through pharynx and nose, with the closed mouth tract as a "shunt" line.
    [Show full text]