Model-Based Sound Synthesis

Total Page:16

File Type:pdf, Size:1020Kb

Model-Based Sound Synthesis EURASIP Journal on Applied Signal Processing Model-Based Sound Synthesis Guest Editors: Vesa Välimäki, Augusto Sarti, Matti Karjalainen, Rudolf Rabenstein, and Lauri Savioja EURASIP Journal on Applied Signal Processing Model-Based Sound Synthesis EURASIP Journal on Applied Signal Processing Model-Based Sound Synthesis Guest Editors: Vesa Välimäki, Augusto Sarti, Matti Karjalainen, Rudolf Rabenstein, and Lauri Savioja Copyright © 2004 Hindawi Publishing Corporation. All rights reserved. This is a special issue published in volume 2004 of “EURASIP Journal on Applied Signal Processing.” All articles are open access articles distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. Editor-in-Chief Marc Moonen, Belgium Senior Advisory Editor K. J. Ray Liu, College Park, USA Associate Editors Kiyoharu Aizawa, Japan A. Gorokhov, The Netherlands Antonio Ortega, USA Gonzalo Arce, USA Peter Handel, Sweden Montse Pardas, Spain Jaakko Astola, Finland Ulrich Heute, Germany Ioannis Pitas, Greece Kenneth Barner, USA John Homer, Australia Phillip Regalia, France Mauro Barni, Italy Jiri Jan, Czech Markus Rupp, Austria Sankar Basu, USA Søren Holdt Jensen, Denmark Hideaki Sakai, Japan Jacob Benesty, Canada Mark Kahrs, USA Bill Sandham, UK Helmut Bölcskei, Switzerland Thomas Kaiser, Germany Wan-Chi Siu, Hong Kong Chong-Yung Chi, Taiwan Moon Gi Kang, Korea Dirk Slock, France M. Reha Civanlar, Turkey Aggelos Katsaggelos, USA Piet Sommen, The Netherlands Tony Constantinides, UK Mos Kaveh, USA John Sorensen, Denmark Luciano Costa, Brazil C.-C. Jay Kuo, USA Michael G. Strintzis, Greece Satya Dharanipragada, USA Chin-Hui Lee, USA Sergios Theodoridis, Greece Petar M. Djurić, USA Kyoung Mu Lee, Korea Jacques Verly, Belgium Jean-Luc Dugelay, France Sang Uk Lee, Korea Xiaodong Wang, USA Touradj Ebrahimi, Switzerland Y. Geoffrey Li, USA Douglas Williams, USA Sadaoki Furui, Japan Mark Liao, Taiwan An-Yen (Andy) Wu, Taiwan Moncef Gabbouj, Finland Bernie Mulgrew, UK Xiang-Gen Xia, USA Sharon Gannot, Israel King N. Ngan, Hong Kong Fulvio Gini, Italy Douglas O’Shaughnessy, Canada Contents Editorial, Vesa Välimäki, Augusto Sarti, Matti Karjalainen, Rudolf Rabenstein, and Lauri Savioja Volume 2004 (2004), Issue 7, Pages 923-925 Physical Modeling of the Piano, N. Giordano and M. Jiang Volume 2004 (2004), Issue 7, Pages 926-933 Sound Synthesis of the Harpsichord Using a Computationally Efficient Physical Model, Vesa Välimäki, Henri Penttinen, Jonte Knif, Mikael Laurson, and Cumhur Erkut Volume 2004 (2004), Issue 7, Pages 934-948 Multirate Simulations of String Vibrations Including Nonlinear Fret-String Interactions Using the Functional Transformation Method, L. Trautmann and R. Rabenstein Volume 2004 (2004), Issue 7, Pages 949-963 Physically Inspired Models for the Synthesis of Stiff Strings with Dispersive Waveguides, I. Testa, G. Evangelista, and S. Cavaliere Volume 2004 (2004), Issue 7, Pages 964-977 Digital Waveguides versus Finite Difference Structures: Equivalence and Mixed Modeling, Matti Karjalainen and Cumhur Erkut Volume 2004 (2004), Issue 7, Pages 978-989 A Digital Synthesis Model of Double-Reed Wind Instruments, Ph. Guillemain Volume 2004 (2004), Issue 7, Pages 990-1000 Real-Time Gesture-Controlled Physical Modelling Music Synthesis with Tactile Feedback, David M. Howard and Stuart Rimell Volume 2004 (2004), Issue 7, Pages 1001-1006 Vibrato in Singing Voice: The Link between Source-Filter and Sinusoidal Models, Ixone Arroabarren and Alfonso Carlosena Volume 2004 (2004), Issue 7, Pages 1007-1020 A Hybrid Resynthesis Model for Hammer-String Interaction of Piano Tones, Julien Bensa, Kristoffer Jensen, and Richard Kronland-Martinet Volume 2004 (2004), Issue 7, Pages 1021-1035 Warped Linear Prediction of Physical Model Excitations with Applications in Audio Compression and Instrument Synthesis, Alexis Glass and Kimitoshi Fukudome Volume 2004 (2004), Issue 7, Pages 1036-1044 EURASIP Journal on Applied Signal Processing 2004:7, 923–925 c 2004 Hindawi Publishing Corporation Editorial Vesa Valim¨ aki¨ Laboratory of Acoustics and Audio Signal Processing, Helsinki University of Technology, P.O. Box 3000, 02015 Espoo, Finland Email: vesa.valimaki@hut.fi Augusto Sarti Dipartimento di Elettronica e Informazione, Politecnico di Milano, piazza Leonardo da Vinci 32, 20133 Milan, Italy Email: [email protected] Matti Karjalainen Laboratory of Acoustics and Audio Signal Processing, Helsinki University of Technology, P.O. Box 3000, 02015 Espoo, Finland Email: matti.karjalainen@hut.fi Rudolf Rabenstein Multimedia Communications and Signal Processing, University Erlangen-Nuremberg, 91058 Erlangen, Germany Email: [email protected] Lauri Savioja Laboratory of Telecommunications Software and Multimedia, Helsinki University of Technology, P.O. Box 5400, 02015 Espoo, Finland Email: lauri.savioja@hut.fi Model-based sound synthesis has become one of the most called ALMA (Algorithms for the Modelling of Acous- active research topics in musical signal processing and in tic Interactions, IST-2001-33059, see http://www-dsp.elet. musical acoustics. The earliest attempts in generating mu- polimi.it/alma/) where the guest editors and their research sical sound with a physical model were made over three teams collaborated in the period from 2001 to 2004. The decades ago. The first commercial products were seen only goal of the ALMA project was to develop an elegant, gen- some twenty years later. Recently, many refinements to pre- eral, and unifying strategy for a blockwise design of physi- vious signal processing algorithms and several new ones have cal models for sound synthesis. A “divide-and-conquer” ap- been introduced. We have learned that new signal processing proach was taken, in which the elements of the structure methods can still be devised or old ones modified to advance are individually modeled and discretized, while their inter- the field. action topology is separately designed and implemented in a Today there exist efficient model-based synthesis algo- dynamical and physically sound fashion. As a result, several rithms for many sound sources, while there are still some high-quality demonstrations of virtual musical instruments for which we do not have a good model. Certain issues, such played in a virtual environment were developed. During the as parameter estimation and real-time control, require fur- ALMA project, the guest editors realized that this special is- ther work for many model-based approaches. Finally, the ca- sue could be created, since the field was very active but there pabilities of human listeners to perceive details in synthetic had not been a special issue devoted to it for a long time. sound should be accounted for in a way similar as in percep- This EURASIP JASP special issue presents ten examples tual audio coding in order to optimize the algorithms. The of recent research in model-based sound synthesis. The first success and future of the model-based approach depends on two papers are related to keyboard instruments. First Gior- researchers and the results of their work. dano and Jiang discuss physical modeling synthesis of the pi- The roots of this special issue are in a European project ano using the finite-difference approach. Then Valim¨ aki¨ et al. 924 EURASIP Journal on Applied Signal Processing show how to synthesize the sound of the harpsichord based Westminster, London, UK. During the academic year 2001-2002 on measurements of a real instrument. An efficient imple- he was Professor of signal processing at the Pori School of Tech- nology and Economics, Tampere University of Technology (TUT), mentation using a visual software synthesis package is given Pori, Finland. In August 2002 he returned to HUT, where he for real-time synthesis. is currently Professor of audio signal processing. He was ap- In the third paper, Trautmann and Rabenstein present a pointed Docent in signal processing at the Pori School of Tech- multirate implementation of a vibrating string model that is nology and Economics, TUT, in 2003. His research interests are based on the functional transformation method. In the next in the application of digital signal processing to audio and mu- paper, Testa et al. investigate the modeling of stiff string be- sic. Dr. Valim¨ aki¨ is a Senior Member of the IEEE Signal Process- havior. The dispersive wave phenomenon, perceivable as in- ing Society and is a Member of the Audio Engineering Society, harmonicity in many string instrument sounds, is studied by the Acoustical Society of Finland, and the Finnish Musicological deriving different physically inspired models. Society. In the fourth paper, Karjalainen and Erkut propose a very interesting and general solution to the problem of how to Augusto Sarti, born in 1963, received the build composite models from digital waveguides and finite- “Laurea” degree (1988, cum laude) and the difference time-domain blocks. The next contribution is Ph.D. (1993) in electrical engineering, from from Guillemain, who proposes a real-time synthesis model the University of Padua, Italy, with research of double-reed wind instruments based on a nonlinear phys- on nonlinear communication systems. He completed his graduate studies at the Uni- ical model. versity of California at Berkeley, where he The paper by Howard and Rimell provides a viewpoint spent two years doing research on nonlinear ff quite di erent from the others in this special issue. It deals system control and on motion planning of with
Recommended publications
  • Minimoog Model D Manual
    3 IMPORTANT SAFETY INSTRUCTIONS WARNING - WHEN USING ELECTRIC PRODUCTS, THESE BASIC PRECAUTIONS SHOULD ALWAYS BE FOLLOWED. 1. Read all the instructions before using the product. 2. Do not use this product near water - for example, near a bathtub, washbowl, kitchen sink, in a wet basement, or near a swimming pool or the like. 3. This product, in combination with an amplifier and headphones or speakers, may be capable of producing sound levels that could cause permanent hearing loss. Do not operate for a long period of time at a high volume level or at a level that is uncomfortable. 4. The product should be located so that its location does not interfere with its proper ventilation. 5. The product should be located away from heat sources such as radiators, heat registers, or other products that produce heat. No naked flame sources (such as candles, lighters, etc.) should be placed near this product. Do not operate in direct sunlight. 6. The product should be connected to a power supply only of the type described in the operating instructions or as marked on the product. 7. The power supply cord of the product should be unplugged from the outlet when left unused for a long period of time or during lightning storms. 8. Care should be taken so that objects do not fall and liquids are not spilled into the enclosure through openings. There are no user serviceable parts inside. Refer all servicing to qualified personnel only. NOTE: This equipment has been tested and found to comply with the limits for a class B digital device, pursuant to part 15 of the FCC rules.
    [Show full text]
  • Additive Synthesis, Amplitude Modulation and Frequency Modulation
    Additive Synthesis, Amplitude Modulation and Frequency Modulation Prof Eduardo R Miranda Varèse-Gastprofessor [email protected] Electronic Music Studio TU Berlin Institute of Communications Research http://www.kgw.tu-berlin.de/ Topics: Additive Synthesis Amplitude Modulation (and Ring Modulation) Frequency Modulation Additive Synthesis • The technique assumes that any periodic waveform can be modelled as a sum sinusoids at various amplitude envelopes and time-varying frequencies. • Works by summing up individually generated sinusoids in order to form a specific sound. Additive Synthesis eg21 Additive Synthesis eg24 • A very powerful and flexible technique. • But it is difficult to control manually and is computationally expensive. • Musical timbres: composed of dozens of time-varying partials. • It requires dozens of oscillators, noise generators and envelopes to obtain convincing simulations of acoustic sounds. • The specification and control of the parameter values for these components are difficult and time consuming. • Alternative approach: tools to obtain the synthesis parameters automatically from the analysis of the spectrum of sampled sounds. Amplitude Modulation • Modulation occurs when some aspect of an audio signal (carrier) varies according to the behaviour of another signal (modulator). • AM = when a modulator drives the amplitude of a carrier. • Simple AM: uses only 2 sinewave oscillators. eg23 • Complex AM: may involve more than 2 signals; or signals other than sinewaves may be employed as carriers and/or modulators. • Two types of AM: a) Classic AM b) Ring Modulation Classic AM • The output from the modulator is added to an offset amplitude value. • If there is no modulation, then the amplitude of the carrier will be equal to the offset.
    [Show full text]
  • Real-Time Timbre Transfer and Sound Synthesis Using DDSP
    REAL-TIME TIMBRE TRANSFER AND SOUND SYNTHESIS USING DDSP Francesco Ganis, Erik Frej Knudesn, Søren V. K. Lyster, Robin Otterbein, David Sudholt¨ and Cumhur Erkut Department of Architecture, Design, and Media Technology Aalborg University Copenhagen, Denmark https://www.smc.aau.dk/ March 15, 2021 ABSTRACT Neural audio synthesis is an actively researched topic, having yielded a wide range of techniques that leverages machine learning architectures. Google Magenta elaborated a novel approach called Differ- ential Digital Signal Processing (DDSP) that incorporates deep neural networks with preconditioned digital signal processing techniques, reaching state-of-the-art results especially in timbre transfer applications. However, most of these techniques, including the DDSP, are generally not applicable in real-time constraints, making them ineligible in a musical workflow. In this paper, we present a real-time implementation of the DDSP library embedded in a virtual synthesizer as a plug-in that can be used in a Digital Audio Workstation. We focused on timbre transfer from learned representations of real instruments to arbitrary sound inputs as well as controlling these models by MIDI. Furthermore, we developed a GUI for intuitive high-level controls which can be used for post-processing and manipulating the parameters estimated by the neural network. We have conducted a user experience test with seven participants online. The results indicated that our users found the interface appealing, easy to understand, and worth exploring further. At the same time, we have identified issues in the timbre transfer quality, in some components we did not implement, and in installation and distribution of our plugin. The next iteration of our design will address these issues.
    [Show full text]
  • Presented at ^Ud,O the 99Th Convention 1995October 6-9
    Tunable Bandpass Filters in Music Synthesis 4098 (L-2) Robert C. Maher University of Nebraska-Lincoln Lincoln, NE 68588-0511, USA Presented at ^ uD,o the 99th Convention 1995 October 6-9 NewYork Thispreprinthas been reproducedfrom the author'sadvance manuscript,withoutediting,correctionsor considerationby the ReviewBoard. TheAES takesno responsibilityforthe contents. Additionalpreprintsmay be obtainedby sendingrequestand remittanceto theAudioEngineeringSocietY,60 East42nd St., New York,New York10165-2520, USA. All rightsreserved.Reproductionof thispreprint,or anyportion thereof,isnot permitted withoutdirectpermissionfromthe Journalof theAudio EngineeringSociety. AN AUDIO ENGINEERING SOCIETY PREPRINT TUNABLE BANDPASS FILTERS IN MUSIC SYNTHESIS ROBERT C. MAHER DEPARTMENT OF ELECTRICAL ENGINEERING AND CENTERFORCOMMUNICATION AND INFORMATION SCIENCE UNIVERSITY OF NEBRASKA-LINCOLN 209N WSEC, LINCOLN, NE 68588-05II USA VOICE: (402)472-2081 FAX: (402)472-4732 INTERNET: [email protected] Abst/act: Subtractive synthesis, or source-filter synthesis, is a well known topic in electronic and computer music. In this paper a description is given of a flexible subtractive synthesis scheme utilizing a set of tunable digital bandpass filters. Specific examples and applications are presented for realtime subtractive synthesis of singing and other musical signals. 0. INTRODUCTION Subtractive (or source-filter) synthesis is used widely in electronic and computer music applications. Subtractive synthesis general!y involves a source signal with a broad spectrum that is passed through a filter. The properties of the filter largely define the shape of the output spectrum by attenuating specific frequency ranges, hence the name subtractive synthesis [1]. The subtractive synthesis model is appropriate for the wide class of physical systems in which an input source drives a passive acoustical or mechanical system.
    [Show full text]
  • Computationally Efficient Music Synthesis
    HELSINKI UNIVERSITY OF TECHNOLOGY Department of Electrical and Communications Engineering Laboratory of Acoustics and Audio Signal Processing Jussi Pekonen Computationally Efficient Music Synthesis – Methods and Sound Design Master’s Thesis submitted in partial fulfillment of the requirements for the degree of Master of Science in Technology. Espoo, June 1, 2007 Supervisor: Professor Vesa Välimäki Instructor: Professor Vesa Välimäki HELSINKI UNIVERSITY ABSTRACT OF THE OF TECHNOLOGY MASTER’S THESIS Author: Jussi Pekonen Name of the thesis: Computationally Efficient Music Synthesis – Methods and Sound Design Date: June 1, 2007 Number of pages: 80+xi Department: Electrical and Communications Engineering Professorship: S-89 Supervisor: Professor Vesa Välimäki Instructor: Professor Vesa Välimäki In this thesis, the design of a music synthesizer for systems suffering from limitations in computing power and memory capacity is presented. First, different possible syn- thesis techniques are reviewed and their applicability in computationally efficient music synthesis is discussed. In practice, the applicable techniques are limited to additive and source-filter synthesis, and, in special cases, to frequency modulation, wavetable and sampling synthesis. Next, the design of the structures of the applicable techniques are presented in detail, and properties and design issues of these structures are discussed. A major implemen- tation problem is raised in digital source-filter synthesis, where the use of classic wave- forms, such as sawtooth wave, as the source signal is challenging due to aliasing caused by waveform discontinuities. Methods for existing bandlimited waveform synthesis are reviewed, and a new approach using polynomial bandlimited step function is pre- sented in detail with design rules for the applicable polynomials.
    [Show full text]
  • Frequency-Domain Additive Synthesis with an Oversampled Weighted Overlap-Add Filterbank for a Portable Low-Power MIDI Synthesizer
    Audio Engineering Society Convention Paper 6202 Presented at the 117th Convention 2004 October 28–31 San Francisco, CA, USA This convention paper has been reproduced from the author's advance manuscript, without editing, corrections, or consideration by the Review Board. The AES takes no responsibility for the contents. Additional papers may be obtained by sending request and remittance to Audio Engineering Society, 60 East 42nd Street, New York, New York 10165-2520, USA; also see www.aes.org. All rights reserved. Reproduction of this paper, or any portion thereof, is not permitted without direct permission from the Journal of the Audio Engineering Society. Frequency-Domain Additive Synthesis With An Oversampled Weighted Overlap-Add Filterbank For A Portable Low-Power MIDI Synthesizer King Tam1 1 Dspfactory Ltd., Waterloo, Ontario, N2V 1K8, Canada [email protected] ABSTRACT This paper discusses a hybrid audio synthesis method employing both additive synthesis and DPCM audio playback, and the implementation of a miniature synthesizer system that accepts MIDI as an input format. Additive synthesis is performed in the frequency domain using a weighted overlap-add filterbank, providing efficiency gains compared to previously known methods. The synthesizer system is implemented on an ultra-miniature, low-power, reconfigurable application specific digital signal processing platform. This low-resource MIDI synthesizer is suitable for portable, low-power devices such as mobile telephones and other portable communication devices. Several issues related to the additive synthesis method, DPCM codec design, and system tradeoffs are discussed. implementation using the Fourier transform and inverse Fourier transform. 1. INTRODUCTION While several other synthesis methods have been Additive synthesis in musical applications has been developed, interest in additive synthesis has continued.
    [Show full text]
  • 11C Software 1034-1187
    Section11c PHOTO - VIDEO - PRO AUDIO Computer Software Ableton.........................................1036-1038 Arturia ...................................................1039 Antares .........................................1040-1044 Arkaos ....................................................1045 Bias ...............................................1046-1051 Bitheadz .......................................1052-1059 Bomb Factory ..............................1060-1063 Celemony ..............................................1064 Chicken Systems...................................1065 Eastwest/Quantum Leap ............1066-1069 IK Multimedia .............................1070-1078 Mackie/UA ...................................1079-1081 McDSP ..........................................1082-1085 Metric Halo..................................1086-1088 Native Instruments .....................1089-1103 Propellerhead ..............................1104-1108 Prosoniq .......................................1109-1111 Serato............................................1112-1113 Sonic Foundry .............................1114-1127 Spectrasonics ...............................1128-1130 Syntrillium ............................................1131 Tascam..........................................1132-1147 TC Works .....................................1148-1157 Ultimate Soundbank ..................1158-1159 Universal Audio ..........................1160-1161 Wave Mechanics..........................1162-1165 Waves ...........................................1166-1185
    [Show full text]
  • THE COMPLETE SYNTHESIZER: a Comprehensive Guide by David Crombie (1984)
    THE COMPLETE SYNTHESIZER: A Comprehensive Guide By David Crombie (1984) Digitized by Neuronick (2001) TABLE OF CONTENTS TABLE OF CONTENTS...........................................................................................................................................2 PREFACE.................................................................................................................................................................5 INTRODUCTION ......................................................................................................................................................5 "WHAT IS A SYNTHESIZER?".............................................................................................................................5 CHAPTER 1: UNDERSTANDING SOUND .............................................................................................................6 WHAT IS SOUND? ...............................................................................................................................................7 THE THREE ELEMENTS OF SOUND .................................................................................................................7 PITCH ...................................................................................................................................................................8 STANDARD TUNING............................................................................................................................................8 THE RESPONSE OF THE HUMAN
    [Show full text]
  • Enhancing Digital Signal Processing Education with Audio Signal Processing and Music Synthesis
    AC 2008-1613: ENHANCING DIGITAL SIGNAL PROCESSING EDUCATION WITH AUDIO SIGNAL PROCESSING AND MUSIC SYNTHESIS Ed Doering, Rose-Hulman Institute of Technology Edward Doering received his Ph.D. in electrical engineering from Iowa State University in 1992, and has been a member the ECE faculty at Rose-Hulman Institute of Technology since 1994. He teaches courses in digital systems, circuits, image processing, and electronic music synthesis, and his research interests include technology-enabled education, image processing, and FPGA-based signal processing. Sam Shearman, National Instruments Sam Shearman is a Senior Product Manager for Signal Processing and Communications at National Instruments (Austin, TX). Working for the firm since 2000, he has served in roles involving product management and R&D related to signal processing, communications, and measurement. Prior to working with NI, he worked as a technical trade press editor and as a research engineer. As a trade press editor for "Personal Engineering & Instrumentation News," he covered PC-based test and analysis markets. His research engineering work involved embedding microstructures in high-volume plastic coatings for non-imaging optics applications. He received a BS (1993) in electrical engineering from the Georgia Institute of Technology (Atlanta, GA). Erik Luther, National Instruments Erik Luther, Textbook Program Manager, works closely with professors, lead users, and authors to improve the quality of Engineering education utilizing National Instruments technology. During his last 5 years at National Instruments, Luther has held positions as an academic resource engineer, academic field engineer, an applications engineer, and applications engineering intern. Throughout his career, Luther, has focused on improving education at all levels including volunteering weekly to teach 4th graders to enjoy science, math, and engineering by building Lego Mindstorm robots.
    [Show full text]
  • New Possibilities in Sound Analysis and Synthesis Xavier Rodet, Philippe Depalle, Guillermo Garcia
    New Possibilities in Sound Analysis and Synthesis Xavier Rodet, Philippe Depalle, Guillermo Garcia To cite this version: Xavier Rodet, Philippe Depalle, Guillermo Garcia. New Possibilities in Sound Analysis and Synthesis. ISMA: International Symposium of Music Acoustics, 1995, Dourdan, France. pp.1-1. hal-01157137 HAL Id: hal-01157137 https://hal.archives-ouvertes.fr/hal-01157137 Submitted on 27 May 2015 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. New Possibilities in Sound Analysis and Synthesis Xavier Rodet, Philippe Depalle, Guillermo Garcia ISMA 95, Dourdan (France), 1995 Copyright © ISMA 1995 Abstract In this presentation we exemplify the emergence of new possibilities in sound analysis and synthesis with three novel developments that have been done in the Analysis/Synthesis team at IRCAM. These examples address three main activities in our domain, and have reached a large public making or simply listening to music. The first example concerns synthesis using physical models. We have determined the behavior of a class of models, in terms of stability, oscillation, periodicity, and finally chaos, leading to a better control of these models in truly innovative musical applications. The second example concerns additive synthesis essentially based on the analysis of natural sounds.
    [Show full text]
  • Additive Synthesis
    Additive synthesis Additive synthesis is a sound synthesis technique that creates timbre by adding sine waves together.[1][2] Additive synthesis example The timbre of musical instruments can be considered in the light of Fourier theory to consist of multiple harmonic or 0:00 inharmonic partials or overtones. Each partial is a sine wave of different frequency and amplitude that swells and decays over A bell-like sound generated by time due to modulation from an ADSR envelope or low frequency oscillator. additive synthesis of 21 inharmonic partials Additive synthesis most directly generates sound by adding the output of multiple sine wave generators. Alternative implementations may use pre-computedwavetables or the inverse Fast Fourier transform. Problems playing this file? See media help. Contents 1 Explanation 2 Definitions 2.1 Harmonic form 2.2 Time-dependent amplitudes 2.3 Inharmonic form 2.4 Time-dependent frequencies 2.5 Broader definitions 3 Implementation methods 3.1 Oscillator bank synthesis 3.2 Wavetable synthesis 3.2.1 Group additive synthesis 3.3 Inverse FFT synthesis 4 Additive analysis/resynthesis 4.1 Products 5 Applications 5.1 Musical instruments 5.2 Speech synthesis 6 History 6.1 Timeline 7 Discrete-time equations 8 See also 9 References 10 External links Explanation The sounds that are heard in everyday life are not characterized by a single frequency. Instead, they consist of a sum of pure sine frequencies, each one at a different amplitude. When humans hear these frequencies simultaneously, we can recognize the sound. This is true for both "non-musical" sounds (e.g. water splashing, leaves rustling, etc) and for "musical sounds" (e.g.
    [Show full text]
  • Physically-Based Parametric Sound Synthesis and Control
    Physically-Based Parametric Sound Synthesis and Control Perry R. Cook Princeton Computer Science (also Music) Course Introduction Parametric Synthesis and Control of Real-World Sounds for virtual reality games production auditory display interactive art interaction design Course #2 Sound - 1 Schedule 0:00 Welcome, Overview 0:05 Views of Sound 0:15 Spectra, Spectral Models 0:30 Subtractive and Modal Models 1:00 Physical Models: Waveguides and variants 1:20 Particle Models 1:40 Friction and Turbulence 1:45 Control Demos, Animation Examples 1:55 Wrap Up Views of Sound • Sound is a recorded waveform PCM playback is all we need for interactions, movies, games, etc. (Not true!!) • Time Domain x( t ) (from physics) • Frequency Domain X( f ) (from math) • Production what caused it • Perception our image of it Course #2 Sound - 2 Views of Sound Time Domain is most closely related to Production Frequency Domain is most closely related to Perception we will see that many hybrids abound Views of Sound: Time Domain Sound is produced/modeled by physics, described by quantities of • Force force = mass * acceleration • Position x(t) actually < x(t), y(t), z(t) > • Velocity Rate of change of position dx/dt • Acceleration Rate of change of velocity dv/dt Examples: Mass+Spring+Damper Wave Equation Course #2 Sound - 3 Mass/Spring/Damper F = ma = - ky - rv - mg F = ma = - ky - rv (if gravity negligible) d 2 y r dy k + + y = 0 dt2 m dt m ( ) D2 + Dr / m + k / m = 0 2nd Order Linear Diff Eq. Solution 1) Underdamped: -t/τ ω y(t) = Y0 e cos( t ) exp.
    [Show full text]