Ambisonics Plug-In Suite for Production and Performance Usage

Total Page:16

File Type:pdf, Size:1020Kb

Ambisonics Plug-In Suite for Production and Performance Usage Ambisonics plug-in suite for production and performance usage Matthias KRONLACHNER Institute of Electronic Music and Acoustics, Graz (student) M.K. Ciurlionioˇ g. 5-15 LT - 03104 Vilnius [email protected] Abstract only) plug-in suite offers 3D encoders for 1st and nd th Ambisonics is a technique for the spatialization of 2 order as well as 2D encoders for 5 order. sound sources within a circular or spherical loud- For the plug-ins created in this work, speaker arrangement. This paper presents a suite of the C++ cross-platform programming library Ambisonics processors, compatible with most stan- JUCE5 [Storer, 2012] has been used to develop 1 dard DAW plug-in formats on Linux, Mac OS X audio plug-ins compatible to most DAWs on and Windows. Some considerations about usabil- Linux, Mac OS X and Windows. JUCE is be- ity did result in features of the user interface and automation possibilities, not available in other sur- ing developed and maintained by Julian Storer, round panning plug-ins. The encoder plug-in may be who used it as the base of the DAW Track- connected to a central program for visualisation and tion. JUCE is released under the GNU Public remote control purposes, displaying the current posi- Licence. A commercial license may be acquired tion and audio level of every single track in the DAW. for closed source projects. It is possible to build To enable monitoring of Ambisonics content without JUCE audio processors as LADSPA, VST, AU, an extensive loudspeaker setup, binaural decoders RTAS and AAX plug-ins or as Jack standalone for headphone playback have been implemented. applications. LV2 (LADSPA version 2) support Keywords currently has to be added manually from the separate project DISTRHO6 [Coelho and Ro- Ambisonics, Plug-ins, Digital Audio Workstations, drigues, 2012]. Binaural, Ardour Ambisonics suffers from different existing 1 Introduction standards concerning channel order and normal- ization. To overcome this problem, a conver- It turns out to be difficult finding platform inde- sion tool is included in the plug-in suite. En- pendent audio plug-ins for encoding and decod- coders, rotators and decoders from the authors ing Ambisonics. Following section should give a suite are designed for the ACN channel order brief overview of still maintained plug-ins. and SN3D normalization, proposed by [Nach- Fons Adriaensen's AMB-plugins2 offer en- rd bar et al., 2011]. Conversion between standards coders and rotators (yaw axis only) until 3 of the input and/or output format can be done order. LADSPA can be used with the DAW Ar- by the conversion plug-in. dour under Linux and Mac OS X. There is no Apart from platform compatibility, some con- Windows host supporting LADSPA. For decod- siderations about usability did result in features ing Ambisonics signals into loudspeaker feeds, of the user interface, not available in other sur- Adriaensen's Jack client application AmbDec round panning plug-ins. Continuously rotat- [Adriaensen, 2005] is often used. 3 ing a sound source results in a discontinuity of Bruce Wiggins offers his WigWare plug-ins the angular representation between −180◦ and in VST format for Windows and MacOS X. +180◦. This jump is also reflected when draw- These processors include 2D and 3D encoders rd st ing automation curves resulting in a mismatch until 3 order as well as 1 order decoders for between the visual representation and auditory several fixed loudspeaker arrangements. 4 perceived movement of a sound source. A solu- Daniel Courville's Audio Unit (Mac OS X tion allowing to define absolute starting points 1Digital Audio Workstation and angular velocities for relative movements is 2http://kokkinizita.linuxaudio.org 3http://www.brucewiggins.co.uk 5http://www.rawmaterialsoftware.com 4http://www.radio.uqam.ca/ambisonic/ 6http://distrho.sourceforge.net proposed. 2.1 Automation parameters For headphone monitoring several binaural Most current DAWs are limited to represent au- decoders have been implemented simulating the tomation parameters between 0:0 and 1:0 along Ambisonics half sphere of the medium sized a time line. Panning plug-ins usually map this IEM Cube with 24 speakers and the concert range for azimuth and elevation between −180◦ 7 hall Mumuth with 29 speakers in a elliptical and +180◦. A full circle rotation results in stretched half sphere. a mismatch between the visual representation For visualization and external control pur- (Fig. 2) of the automation curve and the per- poses, a bidirectional Open Sound Control ceived continuous rotation of the source. Ad- (OSC) communication layer has been imple- ditionally the plug-in host may interpolate be- mented in the encoder plug-in. An external pro- tween a jump from 1.0 to 0.0, resulting in a very gram is able to display the current position and fast audible jump. audio level of every track in the DAW. The visu- alization program may also take control over the sources. This can be very useful in performance situations while having a multitrack playback coming from the DAW and a central display to control the position of the individual tracks. Currently no audio plug-in format can handle dynamic input/output channel counts. There- fore all plug-ins may be compiled for fixed Am- bisonics orders. channel order/scaling L' alternative output Ambisonics Converter format B-Format Converter Figure 2: Automation curve for full circle rota- calibration Loudspeaker Loudspeaker- configuration head Ambisonics Head Tracking MicEnc impulse responses tion using Ardour and AMB plug-ins microphone left Mono sound Encoder L N source L Binaural Rotation Decoder filter Mono sound Encoder right To overcome this problem, automation pa- source azimuth, Loudspeaker Headphone elevation, signals signals rameters (Fig. 3) have been added for setting size Independent from Loudspeaker configuration start points (SetAzimuth, SetAzimuthRel) and M… Ambisonics Order L… Number of Ambisonics Channels angular velocities (tgl-rot-azimuth). The maxi- N… Number of Loudspeakers mum speed of the angular velocity may be ad- justed between 0 and 360deg=sec by an addi- Figure 1: Ambisonics production and playback tional parameter (max-speed). This guarantees chain a wide range of adjustment and at the same time accuracy for the rotation speed. 2 Encoder The encoder plug-in distributes a mono source signal into Ambisonics channels (derived from spherical harmonic functions), according to az- imuth and elevation settings, representing the coordinates on a spherical surface. An addi- tional parameter called size [0; 1] may be used to adjust the spatial directivity. Adjusting the size parameter from zero towards one results in a scaling of the higher order components. For a size setting of 1, all Ambisonics channels will be zero except the 0th order (also known as W Figure 3: Encoder automation parameters channel), resulting in an equally distributed sig- nal over all loudspeakers. 2.2 Remote control and visualization 7http://www.kug.ac.at/en/ Keeping track of all sound source positions studies-further-education/studies/ within a DAW may be a difficult task for the infrastructure/the-mumuth.html mixing engineer. To allow a better overview and control of the spatial scenery, a coopera- movements during binaural playback. The ro- tive visualization and control unit has been im- tation plug-in listens to an adjustable UDP port plemented. All encoder plug-ins are equipped for incoming OSC messages. This allows to by- with a bidirectional OSC layer (Fig. 4), send- pass host automation and controlling the rota- ing and receiving control and status parameters. tion directly from the head tracking software Currently a functional prototype has been im- (Sec. 4.1). plemented in Pd/GEM (Fig. 5) displaying all tracks (encoders) on a sphere, including visual- ization of their audio levels. Currently the audio level is represented by the variable length of the cylinder representing a source signal. This con- cept may be extended to a more sophisticated implementation. individual OSC/UDP connection from visualization to encoders azimuth, elevation, size Encoder central OSC/UDP Visualization Figure 6: Ambisonics Rotator Encoder azimuth, elevation, and size, audio level Remote Control ... 4 Binaural decoder Encoder The binaural decoder computes virtual loud- digital audio workstation speaker feeds as a linear combination of the Am- bisonics signals according to a given decoding Figure 4: Encoder OSC communication with re- matrix. These virtual loudspeaker signals are mote visualization and control convolved with their individual stereo impulse responses, modeling the transfer path from the loudspeaker position to the left and right ear of the listener (Fig. 1). 4.1 Head tracking Head tracking is a significant feature for virtual reality scenes and headphone playback. Small head movements change the relative position of a sound source in aspect to the listeners ears, making localization more easy and removing ambiguity. The KinectTM sensor as add-on for the gam- ing console XBox 360TM by Microsoft offers a low budget depth sensor. [Fanelli et al., 2011] developed a software to gather head orienta- Figure 5: Visualization and remote control of tion angles and the head position relative to the several encoders with Pd/GEM Kinect sensor (Fig. 7). The author extended this head pose estimation software about send- 3 Rotator ing OSC data to the Ambisonics rotator plug- The rotation plug-in (Fig. 6) may be used to in8. The Ambisonics sound field has to be ro- manipulate the orientation in the Ambisonics tated in opposite direction to keep the sound domain, as described in [Musil et al., 2003]. source positions fixed and suppress the rotation An optimized way to calculate rotation matri- with the listeners head. ces can be found in [Rumori, 2009]. Therefore ◦ yaw, pitch and roll can be adjusted by ±180 . 8http://github.com/kronihias/ This is very useful for the incorporation of head head-pose-estimation versatile loudspeaker setup that can be changed within a minute.
Recommended publications
  • Embedded Real-Time Audio Signal Processing with Faust Romain Michon, Yann Orlarey, Stéphane Letz, Dominique Fober, Dirk Roosenburg
    Embedded Real-Time Audio Signal Processing With Faust Romain Michon, Yann Orlarey, Stéphane Letz, Dominique Fober, Dirk Roosenburg To cite this version: Romain Michon, Yann Orlarey, Stéphane Letz, Dominique Fober, Dirk Roosenburg. Embedded Real- Time Audio Signal Processing With Faust. International Faust Conference (IFC-20), Dec 2020, Paris, France. hal-03124896 HAL Id: hal-03124896 https://hal.archives-ouvertes.fr/hal-03124896 Submitted on 29 Jan 2021 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Proceedings of the 2nd International Faust Conference (IFC-20), Maison des Sciences de l’Homme Paris Nord, France, December 1-2, 2020 EMBEDDED REAL-TIME AUDIO SIGNAL PROCESSING WITH FAUST Romain Michon,a;b Yann Orlarey,a Stéphane Letz,a Dominique Fober,a and Dirk Roosenburgd;e aGRAME – Centre National de Création Musicale, Lyon, France bCenter for Computer Research in Music and Acoustics, Stanford University, USA dTIMARA, Oberlin Conservatory of Music, USA eDepartment of Physics, Oberlin College, USA [email protected] ABSTRACT multi-channel audio, etc.). These products work in conjunction with specialized Linux distributions where audio processing tasks FAUST has been targeting an increasing number of embedded plat- forms for real-time audio signal processing applications in recent are carried out outside of the operating system, allowing for the years.
    [Show full text]
  • Extending the Faust VST Architecture with Polyphony, Portamento and Pitch Bend Yan Michalevsky Julius O
    Extending the Faust VST Architecture with Polyphony, Portamento and Pitch Bend Yan Michalevsky Julius O. Smith Andrew Best Department of Electrical Center for Computer Research in Blamsoft, Inc. Engineering, Music and Acoustics (CCRMA), [email protected] Stanford University Stanford University [email protected] AES Fellow [email protected] Abstract VST (Virtual Studio Technology) plugin stan- We introduce the vsti-poly.cpp architecture for dard was released by Steinberg GmbH (famous the Faust programming language. It provides sev- for Cubase and other music and sound produc- eral features that are important for practical use of tion products) in 1996, and was followed by the Faust-generated VSTi synthesizers. We focus on widespread version 2.0 in 1999 [8]. It is a partic- the VST architecture as one that has been used tra- ularly common format supported by many older ditionally and is supported by many popular tools, and newer tools. and add several important features: polyphony, note Some of the features expected from a VST history and pitch-bend support. These features take plugin can be found in the VST SDK code.2 Faust-generated VST instruments a step forward in Examining the list of MIDI events [1] can also terms of generating plugins that could be used in Digital Audio Workstations (DAW) for real-world hint at what capabilities are expected to be im- music production. plemented by instrument plugins. We also draw from our experience with MIDI instruments and Keywords commercial VST plugins in order to formulate sound feature requirements. Faust, VST, Plugin, DAW In order for Faust to be a practical tool for generating such plugins, it should support most 1 Introduction of the features expected, such as the following: Faust [5] is a popular music/audio signal pro- • Responding to MIDI keyboard events cessing language developed by Yann Orlarey et al.
    [Show full text]
  • Practises in Listener Envelopment with Controllable Spatial Audio
    DIGITAL AUDIO SYSTEMS BRUNO MARION WRITTEN REVIEW 2 440566533 Practises in Listener Envelopment with Controllable Spatial Audio Introduction The envelopment of the listener in the sound field can be created in a number of ways. The two most import factors to Binaural Quality Index in acoustic space, be it real or virtual; is the Auditory Source Width and the Listener Envelopment Factor (Cardenas, et al., 2012). Multiple solutions exist for the increase in ASW and LEV, those discussed here will be Natural Sounding Artificial Reverberation, Recirculating Delays, Higher-Order Ambisonics and finally Higher-Order Speakers. Reverberation A computationally effective artificial reverberation can be easily achieved by using an exponentially decaying impulse and convolving this with the original audio signal. Inversing a duplication of this signal can also give a stereo effect but when compared to reality, this single-impulse method does not come close to creating the complex interactions of plane waves in a given space as each time a wave reflects from a surface or another wave, it is properties are affected as a function of frequency, phase and amplitude. The two main problems cited by Manfred Schroeder with artificial reverberation are: 1. A non-flat Amplitude Frequency Response 2. Low Echo Density (Schroeder, 1961) Shroeder offers solutions to these problems by use of an all-pass filter. An allpass filter is a method of filtering an audio signal without manipulation of other properties and can be constructed by combining a Finite Impuse Respose (FIR) feed forward and feedback. Figure 1 The allpass filter where fc is equal to the cut off frequency, fcl is equal to the lower cu toff frequency and fch is equal to the higher cut off frequency.
    [Show full text]
  • Franz Zotter Matthias Frank a Practical 3D Audio Theory for Recording
    Springer Topics in Signal Processing Franz Zotter Matthias Frank Ambisonics A Practical 3D Audio Theory for Recording, Studio Production, Sound Reinforcement, and Virtual Reality Springer Topics in Signal Processing Volume 19 Series Editors Jacob Benesty, INRS-EMT, University of Quebec, Montreal, QC, Canada Walter Kellermann, Erlangen-Nürnberg, Friedrich-Alexander-Universität, Erlangen, Germany The aim of the Springer Topics in Signal Processing series is to publish very high quality theoretical works, new developments, and advances in the field of signal processing research. Important applications of signal processing will be covered as well. Within the scope of the series are textbooks, monographs, and edited books. More information about this series at http://www.springer.com/series/8109 Franz Zotter • Matthias Frank Ambisonics A Practical 3D Audio Theory for Recording, Studio Production, Sound Reinforcement, and Virtual Reality Franz Zotter Matthias Frank Institute of Electronic Music and Acoustics Institute of Electronic Music and Acoustics University of Music and Performing Arts University of Music and Performing Arts Graz, Austria Graz, Austria ISSN 1866-2609 ISSN 1866-2617 (electronic) Springer Topics in Signal Processing ISBN 978-3-030-17206-0 ISBN 978-3-030-17207-7 (eBook) https://doi.org/10.1007/978-3-030-17207-7 © The Editor(s) (if applicable) and The Author(s) 2019. This book is an open access publication. Open Access This book is licensed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing, adap- tation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license and indicate if changes were made.
    [Show full text]
  • Implementing a Parametric EQ Plug-In in C++ Using the Multi-Platform VST Specification
    2003:044 C EXTENDED ESSAY Implementing a parametric EQ plug-in in C++ using the multi-platform VST specification JONAS EKEROOT SCHOOL OF MUSIC Audio Technology Supervisor: Jan Berg 2003:044 • ISSN: 1402 – 1773 • ISRN: LTU - CUPP - - 03/44 - - SE Implementing a parametric EQ plug-in in C++ using the multi-platform VST specification Jonas Ekeroot Division of Sound Recording School of Music in Pite˚a Lule˚aUniversity of Technology April 23, 2003 Abstract As the processing power of desktop computer systems increase by every year, more and more real-time audio signal processing is per- formed on such systems. What used to be done in external effects units, e.g. adding reverb, can now be accomplished within the com- puter system using signal processing code modules – plug-ins. This thesis describes the development of a peak/notch parametric EQ VST plug-in. First a prototype was made in the graphical audio program- ming environment Max/MSP on MacOS, and then a C++ implemen- tation was made using the VST Software Development Kit. The C++ source code was compiled on both Windows and MacOS, resulting in versions of the plug-in that can be used in any VST host application on Windows and MacOS respectively. Writing a plug-in relieves the programmer of the burden to deal directly with audio interface details and graphical user interface specifics, since this is taken care of by the host application. It can thus be an interesting way to start developing audio DSP algorithms, since the host application also provides the op- portunity to listen to and measure the performance of the implemented plug-in algorithm.
    [Show full text]
  • Audio Signal Processing in Faust
    Audio Signal Processing in Faust Julius O. Smith III Center for Computer Research in Music and Acoustics (CCRMA) Department of Music, Stanford University, Stanford, California 94305 USA jos at ccrma.stanford.edu Abstract Faust is a high-level programming language for digital signal processing, with special sup- port for real-time audio applications and plugins on various software platforms including Linux, Mac-OS-X, iOS, Android, Windows, and embedded computing environments. Audio plugin formats supported include VST, lv2, AU, Pd, Max/MSP, SuperCollider, and more. This tuto- rial provides an introduction focusing on a simple example of white noise filtered by a variable resonator. Contents 1 Introduction 3 1.1 Installing Faust ...................................... 4 1.2 Faust Examples ...................................... 5 2 Primer on the Faust Language 5 2.1 Basic Signal Processing Blocks (Elementary Operators onSignals) .......... 7 2.2 BlockDiagramOperators . ...... 7 2.3 Examples ........................................ 8 2.4 InfixNotationRewriting. ....... 8 2.5 Encoding Block Diagrams in the Faust Language ................... 9 2.6 Statements ...................................... ... 9 2.7 FunctionDefinition............................... ...... 9 2.8 PartialFunctionApplication . ......... 10 2.9 FunctionalNotationforOperators . .......... 11 2.10Examples ....................................... 11 1 2.11 Summary of Faust NotationStyles ........................... 11 2.12UnaryMinus ..................................... 12 2.13 Fixing
    [Show full text]
  • How to Create Music with GNU/Linux
    How to create music with GNU/Linux Emmanuel Saracco [email protected] How to create music with GNU/Linux by Emmanuel Saracco Copyright © 2005-2009 Emmanuel Saracco How to create music with GNU/Linux Warning WORK IN PROGRESS Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or any later version published by the Free Software Foundation; with no Invariant Sections, no Front-Cover Texts, and no Back-Cover Texts. A copy of the license is available on the World Wide Web at http://www.gnu.org/licenses/fdl.html. Revision History Revision 0.0 2009-01-30 Revised by: es Not yet versioned: It is still a work in progress. Dedication This howto is dedicated to all GNU/Linux users that refuse to use proprietary software to work with audio. Many thanks to all Free developers and Free composers that help us day-by-day to make this possible. Table of Contents Forword................................................................................................................................................... vii 1. System settings and tuning....................................................................................................................1 1.1. My Studio....................................................................................................................................1 1.2. File system..................................................................................................................................1 1.3. Linux Kernel...............................................................................................................................2
    [Show full text]
  • Scene-Based Audio and Higher Order Ambisonics
    Scene -Based Audio and Higher Order Ambisonics: A technology overview and application to Next-Generation Audio, VR and 360° Video NOVEMBER 2019 Ferdinando Olivieri, Nils Peters, Deep Sen, Qualcomm Technologies Inc., San Diego, California, USA The main purpose of an EBU Technical Review is to critically examine new technologies or developments in media production or distribution. All Technical Reviews are reviewed by 1 (or more) technical experts at the EBU or externally and by the EBU Technical Editions Manager. Responsibility for the views expressed in this article rests solely with the author(s). To access the full collection of our Technical Reviews, please see: https://tech.ebu.ch/publications . If you are interested in submitting a topic for an EBU Technical Review, please contact: [email protected] EBU Technology & Innovation | Technical Review | NOVEMBER 2019 2 1. Introduction Scene Based Audio is a set of technologies for 3D audio that is based on Higher Order Ambisonics. HOA is a technology that allows for accurate capturing, efficient delivery, and compelling reproduction of 3D audio sound fields on any device, such as headphones, arbitrary loudspeaker configurations, or soundbars. We introduce SBA and we describe the workflows for production, transport and reproduction of 3D audio using HOA. The efficient transport of HOA is made possible by state-of-the-art compression technologies contained in the MPEG-H Audio standard. We discuss how SBA and HOA can be used to successfully implement Next Generation Audio systems, and to deliver any combination of TV, VR, and 360° video experiences using a single audio workflow. 1.1 List of abbreviations & acronyms CBA Channel-Based Audio SBA Scene-Based Audio HOA Higher Order Ambisonics OBA Object-Based Audio HMD Head-Mounted Display MPEG Motion Picture Experts Group (also the name of various compression formats) ITU International Telecommunications Union ETSI European Telecommunications Standards Institute 2.
    [Show full text]
  • Radium: a Music Editor Inspired by the Music Tracker
    Radium: A Music Editor Inspired by the Music Tracker Kjetil Matheussen Norwegian Center for Technology in Music and the Arts. (NOTAM) Sandakerveien 24D, Bygg F3 N-0473 Oslo Norway [email protected] Abstract Musical events are defined with pure text. Radium is a new type of music editor inspired by The event C#3 5-32-000 plays the note C the music tracker. Radium's interface differs from sharp at octave 3 using instrument number 5 at the classical music tracker interface by using graphi- volume 32. The last three zeroes can be used cal elements instead of text and by allowing musical for various types of sound effects, or to set new events anywhere within a tracker line. tempo. Chapter 1: The classical music tracker interface The tables are called patterns, and a song and how Radium differs from it. Chapter 2: Ra- usually contains several patterns. To control dium Features: a) The Editor; b) The Modular the order patterns are playbed back, we use a Mixer; c) Instruments and Audio Effects; d) In- strument Configuration; e) Common Music Nota- playlist. For example, if we have three patterns, tion. Chapter 3: Implementation details: a) Paint- a typical song could have a playlist like this: ing the Editor; b) Smooth Scrolling; c) Embed- 1, 2, 1, 2, 3, 1, 2. ding Pure Data; d) Collecting Memory Garbage in C and C++. Chapter 4: Related software. 1.1 How Radium Differs from the Classical Tracker Interface Keywords Radium4 differs from the music tracker inter- Radium, Music Tracker, GUI, Pure Data, Graphics face by using graphical elements instead of text Programming.
    [Show full text]
  • Proyecto Fin De Grado
    ESCUELA TÉCNICA SUPERIOR DE INGENIERÍA Y SISTEMAS DE TELECOMUNICACIÓN PROYECTO FIN DE GRADO TÍTULO: Implementación de un sintetizador en un kit de desarrollo de bajo coste AUTOR: Arturo Fernández TITULACIÓN: Grado en Ingeniería de Sonido e Imagen TUTOR: Antonio Mínguez Olivares DEPARTAMENTO: Teoría de la señal y Comunicación VºBº Miembros del Tribunal Calificador: PRESIDENTE: Eduardo Nogueira Díaz TUTOR: Antonio Mínguez Olivares SECRETARIO: Francisco Javier Tabernero Gil Fecha de lectura: Calificación: El Secretario, Agradecimientos − A mis padres y mi hermana, por apoyarme en todos los aspectos (con discrepancias inevitables) a lo largo de esta carrera de fondo. − A mis amigos y la novia, por no verme el pelo durante largos periodos de tiempo. − A los profesores que han conseguido despertar mi interés y dedicarme horas en tutorías. − A mi tutor por aceptar mi propuesta y tutelar el proyecto sin pensárselo. − A mi cuñado Rafa, que, a pesar de estar eternamente abarrotado, sacaba un hueco para ayudarme en los momentos más difíciles, motivarme y guiarme a lo largo de todos estos años. Sin su ayuda, esto no habría sido posible. − Finalmente, darme las gracias a mí mismo, por aguantar este ritmo de vida, tener la ambición de seguir y afrontar nuevos retos. Como me dice mi padre: “Tú hasta que no llegues a la Luna, no vas a parar, ¿no?”. Implementación de un sintetizador en un kit de desarrollo de bajo coste Resumen La realización de este proyecto consiste en el diseño de un sintetizador ‘Virtual Analógico’ basado en sintetizadores de los años 60 y 70. Para ello, se estudia la teoría que conforma la topología básica de los sintetizadores sustractivos, así como, los elementos que intervienen en la generación y modificación del sonido.
    [Show full text]
  • Spatialized Audio Rendering for Immersive Virtual Environments
    Spatialized Audio Rendering for Immersive Virtual Environments Martin Naef1 Oliver Staadt2 Markus Gross1 1Computer Graphics Laboratory 2Computer Science Department ETH Zurich, Switzerland University of California, Davis, USA +41-1-632-7114 +1-530-752-4821 {naef, grossm}@inf.ethz.ch [email protected] ABSTRACT We are currently developing a novel networked SID, the blue-c, which combines immersive projection with multi-stream video We present a spatialized audio rendering system for the use in acquisition and advanced multimedia communication. Multiple immersive virtual environments. The system is optimized for ren- interconnected portals will allow remotely located users to meet, to dering a sufficient number of dynamically moving sound sources communicate, and to collaborate in a shared virtual environment. in multi-speaker environments using off-the-shelf audio hardware. We have identified spatial audio as an important component for Based on simplified physics-based models, we achieve a good this system. trade-off between audio quality, spatial precision, and perfor- In this paper, we present an audio rendering pipeline and system mance. Convincing acoustic room simulation is accomplished by suitable for spatially immersive displays. The pipeline is targeted integrating standard hardware reverberation devices as used in the at systems which require a high degree of flexibility while using professional audio and broadcast community. We elaborate on inexpensive consumer-level audio hardware. important design principles for audio rendering as well as on prac- For successful deployment in immersive virtual environments, tical implementation issues. Moreover, we describe the integration the audio rendering pipeline has to meet several important require- of the audio rendering pipeline into a scene graph-based virtual ments: reality toolkit.
    [Show full text]
  • Localization of 3D Ambisonic Recordings and Ambisonic Virtual
    Localization of 3D Ambisonic Recordings and Ambisonic Virtual Sources Sebastian Braun and Matthias Frank UniversitÄatfÄurMusik und darstellende Kunst Graz, Austria Institut fÄurElektronische Musik und Akustik, Email: [email protected] Abstract The accurate recording and reproduction of real 3D sound environments is still a challenging task. In this paper, the two most established 3D microphones are evaluated: the Sound¯eld SPS200 and MH Acoustics' Eigenmike EM32, according to 1st and 4th order Ambisonics. They are compared to virtual encoded sound sources of the same orders in a localization test. For the reproduction, an Ambisonics system with 24 (12+8+4) hemispherically arranged loudspeakers is used. In order to compare to existing results from the literature, this paper focuses on sound sources in the horizontal plane. As expected, the 4th order sources yield better localization as the 1st order sources. Within each order, the real recordings and the virtual encoded sources show a good correspondence. Introduction Coordinate system. Fig. 1 shows the coordinate sys- tem used in this paper. The directions θ are vectors Surround sound systems have entered movie theaters and of unit length that depend on the azimuth angle ' and our living rooms during the last decades. But these zenith angle # (# = 90± in the horizontal plane) systems are still limited to horizontal sound reproduc- 0 1 tion. Recently, further developments, such as 22.2 [1] cos(') sin(#) and AURO 3D [2], expanded to the third dimension by θ('; #) = @sin(') sin(#)A : (1) adding height channels. cos(#) The third dimension envolves new challenges in panning and recording [3]. VBAP [4] is the simplest method for z panning and at the same time the most flexible one, as it works with arbitrary loudspeaker arrangements.
    [Show full text]