Implementing Real-Time Granular Synthesis Ross Bencina Draft of 31St August 2001
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Navigating the Landscape of Computer Aided Algorithmic Composition Systems: a Definition, Seven Descriptors, and a Lexicon of Systems and Research
NAVIGATING THE LANDSCAPE OF COMPUTER AIDED ALGORITHMIC COMPOSITION SYSTEMS: A DEFINITION, SEVEN DESCRIPTORS, AND A LEXICON OF SYSTEMS AND RESEARCH Christopher Ariza New York University Graduate School of Arts and Sciences New York, New York [email protected] ABSTRACT comparison and analysis, this article proposes seven system descriptors. Despite Lejaren Hiller’s well-known Towards developing methods of software comparison claim that “computer-assisted composition is difficult to and analysis, this article proposes a definition of a define, difficult to limit, and difficult to systematize” computer aided algorithmic composition (CAAC) (Hiller 1981, p. 75), a definition is proposed. system and offers seven system descriptors: scale, process-time, idiom-affinity, extensibility, event A CAAC system is software that facilitates the production, sound source, and user environment. The generation of new music by means other than the public internet resource algorithmic.net is introduced, manipulation of a direct music representation. Here, providing a lexicon of systems and research in computer “new music” does not designate style or genre; rather, aided algorithmic composition. the output of a CAAC system must be, in some manner, a unique musical variant. An output, compared to the 1. DEFINITION OF A COMPUTER-AIDED user’s representation or related outputs, must not be a ALGORITHMIC COMPOSITION SYSTEM “copy,” accepting that the distinction between a copy and a unique variant may be vague and contextually determined. This output may be in the form of any Labels such as algorithmic composition, automatic sound or sound parameter data, from a sequence of composition, composition pre-processing, computer- samples to the notation of a complete composition. -
Granular Synthesis and Physical Modeling with This Manual Is Released Under the Creative Commons Attribution 3.0 Unported License
Granular synthesis and physical modeling with This manual is released under the Creative Commons Attribution 3.0 Unported License. You may copy, distribute, transmit and adapt it, for any purpose, provided you include the following attribution: Kaivo and the Kaivo manual by Madrona Labs. http://madronalabs.com. Version 1.3, December 2016. Written by George Cochrane and Randy Jones. Illustrated by David Chandler. Typeset in Adobe Minion using the TEX document processing system. Any trademarks mentioned are the sole property of their respective owners. Such mention does not imply any endorsement of or associa- tion with Madrona Labs. Introduction “The future evolution of virtual devices is less constrained than that of real devices.” –Julius O. Smith Kaivo is a software instrument combining two powerful synthesis techniques (physical modeling and granular synthesis) in an easy-to- use semi-modular package. It’s laid out a bit like an acoustic instru- For more information on the theory be- ment; the GRANULATOR module acts like the player’s touch, exciting hind Kaivo, see Chapter 1, “Physi-who? Granu-what?” one or more tuned objects (here, the RESONATOR module, based on physical models of resonant objects) that come together in a central resonating body (the BODY module, also physics-based). This allows for a natural (or uncanny) sense of space, pleasing in- teractions between voices, and a ton of expressive potential—all traits in short supply in digital synthesis. The acoustic comparison begins to pale when you realize that your “touch” is really a granular sam- ple player with scads of options, and that the physical properties of the resonating modules are widely variable, and in real time. -
Syllabus EMBODIED SONIC MEDITATION
MUS CS 105 Embodied Sonic Meditation Spring 2017 Jiayue Cecilia Wu Syllabus EMBODIED SONIC MEDITATION: A CREATIVE SOUND EDUCATION Instructor: Jiayue Cecilia Wu Contact: Email: [email protected]; Cell/Text: 650-713-9655 Class schedule: CCS Bldg 494, Room 143; Mon/Wed 1:30 pm – 2:50 pm, Office hour: Phelps room 3314; Thu 1: 45 pm–3:45 pm, or by appointment Introduction What is the difference between hearing and listening? What are the connections between our listening, bodily activities, and the lucid mind? This interdisciplinary course on acoustic ecology, sound art, and music technology will give you the answers. Through deep listening (Pauline Oliveros), compassionate listening (Thich Nhat Hanh), soundwalking (Westerkamp), and interactive music controlled by motion capture, the unifying theme of this course is an engagement with sonic awareness, environment, and self-exploration. Particularly we will look into how we generate, interpret and interact with sound; and how imbalances in the soundscape may have adverse effects on our perceptions and the nature. These practices are combined with readings and multimedia demonstrations from sound theory, audio engineering, to aesthetics and philosophy. Students will engage in hands-on audio recording and mixing, human-computer interaction (HCI), group improvisation, environmental film streaming, and soundscape field trip. These experiences lead to a final project in computer-assisted music composition and creative live performance. By the end of the course, students will be able to use computer software to create sound, design their own musical gears, as well as connect with themselves, others and the world around them in an authentic level through musical expressions. -
62 Years and Counting: MUSIC N and the Modular Revolution
62 Years and Counting: MUSIC N and the Modular Revolution By Brian Lindgren MUSC 7660X - History of Electronic and Computer Music Fall 2019 24 December 2019 © Copyright 2020 Brian Lindgren Abstract. MUSIC N by Max Mathews had two profound impacts in the world of music synthesis. The first was the implementation of modularity to ensure a flexibility as a tool for the user; with the introduction of the unit generator, the instrument and the compiler, composers had the building blocks to create an unlimited range of sounds. The second was the impact of this implementation in the modular analog synthesizers developed a few years later. While Jean-Claude Risset, a well known Mathews associate, asserts this, Mathews actually denies it. They both are correct in their perspectives. Introduction Over 76 years have passed since the invention of the first electronic general purpose computer,1 the ENIAC. Today, we carry computers in our pockets that can perform millions of times more calculations per second.2 With the amazing rate of change in computer technology, it's hard to imagine that any development of yesteryear could maintain a semblance of relevance today. However, in the world of music synthesis, the foundations that were laid six decades ago not only spawned a breadth of multifaceted innovation but continue to function as the bedrock of important digital applications used around the world today. Not only did a new modular approach implemented by its creator, Max Mathews, ensure that the MUSIC N lineage would continue to be useful in today’s world (in one of its descendents, Csound) but this approach also likely inspired the analog synthesizer engineers of the day, impacting their designs. -
The Sonification Handbook Chapter 9 Sound Synthesis for Auditory Display
The Sonification Handbook Edited by Thomas Hermann, Andy Hunt, John G. Neuhoff Logos Publishing House, Berlin, Germany ISBN 978-3-8325-2819-5 2011, 586 pages Online: http://sonification.de/handbook Order: http://www.logos-verlag.com Reference: Hermann, T., Hunt, A., Neuhoff, J. G., editors (2011). The Sonification Handbook. Logos Publishing House, Berlin, Germany. Chapter 9 Sound Synthesis for Auditory Display Perry R. Cook This chapter covers most means for synthesizing sounds, with an emphasis on describing the parameters available for each technique, especially as they might be useful for data sonification. The techniques are covered in progression, from the least parametric (the fewest means of modifying the resulting sound from data or controllers), to the most parametric (most flexible for manipulation). Some examples are provided of using various synthesis techniques to sonify body position, desktop GUI interactions, stock data, etc. Reference: Cook, P. R. (2011). Sound synthesis for auditory display. In Hermann, T., Hunt, A., Neuhoff, J. G., editors, The Sonification Handbook, chapter 9, pages 197–235. Logos Publishing House, Berlin, Germany. Media examples: http://sonification.de/handbook/chapters/chapter9 18 Chapter 9 Sound Synthesis for Auditory Display Perry R. Cook 9.1 Introduction and Chapter Overview Applications and research in auditory display require sound synthesis and manipulation algorithms that afford careful control over the sonic results. The long legacy of research in speech, computer music, acoustics, and human audio perception has yielded a wide variety of sound analysis/processing/synthesis algorithms that the auditory display designer may use. This chapter surveys algorithms and techniques for digital sound synthesis as related to auditory display. -
Combining Granular Synthesis with Frequency Modulation
Combining granular synthesis with frequency modulation. Kim ERVIK Øyvind BRANDSEGG Department of music Department of music University of Science and Technology University of Science and Technology Norway Norway [email protected] [email protected] Abstract acoustical events. Sound as particles has been Both granular synthesis and frequency used in applications like independent time and modulation are well-established synthesis pitch scaling, formant modification, analog synth techniques that are very flexible. This paper modeling, clouds of sound, granular delays and will investigate different ways of combining reverbs, etc [3]. Examples of granular synthesis the two techniques. It will describe the rules parameters are density, grain pitch, grain duration, of spectra that emerge when combining, compare it to similar synthesis techniques and grain envelope, the global arrangement of the suggest some aesthetic perspectives on the grains, and of course the content of the grains matter. (which can be a synthetic waveform or sampled sound). Granular synthesis gives the musician vast Keywords expressive possibilities [10]. Granular synthesis, frequency modulation, 1.3 FM synthesis partikkel, csound, sound synthesis Frequency modulation has been a known method of coding audio into radio signal since the beginning of commercial radio. In 1964 that John Chowning discovered the implication of frequency modulation of audio working on synthesis of brass instrument at Stanford [2]. He discovered that modulating the frequency resulted in sideband emerging from the carrier frequency. Fig 1:Grain Pitch modulation Yamaha later implemented the technique in the hugely successful DX7. 1 Introduction If we consider FM synthesis using sinusoidal carrier and modulation oscillator, the spectral 1.1 Method components present in a FM sound can be mathematically stated as in figure 2. -
The Path to Half Life
The Path to Half-life Curtis Roads A composer does not often pause to explain a musical path. When that path is a personal break- through, however, it may be worthwhile to reflect on it, especially when it connects to more gen- eral trends in today’s musical scene. My electronic music composition Half-life por- trays a virtual world in which sounds are born and die in an instant or emerge in slow motion. As emerging sounds unfold, they remain stable or mutate before expiring. Interactions between different sounds sug- gest causalities, as if one sound spawned, trig- gered, crashed into, bonded with, or dis- solved into another sound. Thus the introduc- tion of every new sound contributes to the unfolding of a musical narrative. Most of the sound material of Half-life was produced by the synthesis of brief acoustic sound particles or grains. The interactions of acoustical particles can be likened to photo- graphs of bubble chamber experiments, which were designed to visualize atomic inter- actions. These strikingly beautiful images por- tray intricate causalities as particles enter a chamber at high speed, leading to collisions in which some particles break apart or veer off in strange directions, indicating the pres- ence of hidden forces (figure 1). Figure 1. Bubble chamber image (© CERN, Geneva). Composed years ago, in 1998 and 1999, Half-life is not my newest composition, nor does it incor- porate my most recent techniques. The equipment and software used to make it was (with exception of some custom programs) quite standard. Nonetheless this piece is deeply significant to me as a turning point in a long path of composition. -
Eastman Computer Music Center (ECMC)
Upcoming ECMC25 Concerts Thursday, March 22 Music of Mario Davidovsky, JoAnn Kuchera-Morin, Allan Schindler, and ECMC composers 8:00 pm, Memorial Art Gallery, 500 University Avenue Saturday, April 14 Contemporary Organ Music Festival with the Eastman Organ Department & College Music Department Steve Everett, Ron Nagorcka, and René Uijlenhoet, guest composers 5:00 p.m. + 7:15 p.m., Interfaith Chapel, University of Rochester Eastman Computer Wednesday, May 2 Music Cente r (ECMC) New carillon works by David Wessel and Stephen Rush th with the College Music Department 25 Anniversa ry Series 12:00 pm, Eastman Quadrangle (outdoor venue), University of Rochester admission to all concerts is free Curtis Roads & Craig Harris, ecmc.rochester.edu guest composers B rian O’Reilly, video artist Thursday, March 8, 2007 Kilbourn Hall fire exits are located along the right A fully accessible restroom is located on the main and left sides, and at the back of the hall. Eastman floor of the Eastman School of Music. Our ushers 8:00 p.m. Theatre fire exits are located throughout the will be happy to direct you to this facility. Theatre along the right and left sides, and at the Kilbourn Hall back of the orchestra, mezzanine, and balcony Supporting the Eastman School of Music: levels. In the event of an emergency, you will be We at the Eastman School of Music are grateful for notified by the stage manager. the generous contributions made by friends, If notified, please move in a calm and orderly parents, and alumni, as well as local and national fashion to the nearest exit. -
Neural Granular Sound Synthesis
Neural Granular Sound Synthesis Adrien Bitton Philippe Esling Tatsuya Harada IRCAM, CNRS UMR9912 IRCAM, CNRS UMR9912 The University of Tokyo, RIKEN Paris, France Paris, France Tokyo, Japan [email protected] [email protected] [email protected] ABSTRACT Granular sound synthesis is a popular audio generation technique based on rearranging sequences of small wave- resynthesis generated sequence match decode form windows. In order to control the synthesis, all grains grains in a given corpus are analyzed through a set of acoustic continuous + f(↵) descriptors. This provides a representation reflecting some discrete + condition + form of local similarities across the grains. However, the grain space + quality of this grain space is bound by that of the descrip- + + + grain tors. Its traversal is not continuously invertible to signal grain + + latent space sample and does not render any structured temporality. library dz acoustic R We demonstrate that generative neural networks can im- analysis encode plement granular synthesis while alleviating most of its target signal shortcomings. We efficiently replace its audio descriptor input basis by a probabilistic latent space learned with a Vari- signal ational Auto-Encoder. In this setting the learned grain space is invertible, meaning that we can continuously syn- thesize sound when traversing its dimensions. It also im- Figure 1. Left: A grain library is analysed and scattered (+) into the acoustic dimensions. A target is defined, by analysing an other signal (o) plies that original grains are not stored for synthesis. An- or as a free trajectory, and matched to the library through the acoustic other major advantage of our approach is to learn struc- descriptors. -
43558913.Pdf
! ! ! Generative Music Composition Software Systems Using Biologically Inspired Algorithms: A Systematic Literature Review ! ! ! Master of Science Thesis in the Master Degree Programme ! !Software Engineering and Management! ! ! KEREM PARLAKGÜMÜŞ ! ! ! University of Gothenburg Chalmers University of Technology Department of Computer Science and Engineering Göteborg, Sweden, January 2014 The author grants Chalmers University of Technology and University of Gothenburg the non-exclusive right to publish the work electronically and in a non-commercial purpose and to make it accessible on the Internet. The author warrants that he/she is the author of the work, and warrants that the work does not contain texts, pictures or other material that !violates copyright laws. The author shall, when transferring the rights of the work to a third party (like a publisher or a company), acknowledge the third party about this agreement. If the author has signed a copyright agreement with a third party regarding the work, the author warrants hereby that he/she has obtained any necessary permission from this third party to let Chalmers University of Technology and University of Gothenburg store the work electronically and make it accessible on the Internet. ! ! Generative Music Composition Software Systems Using Biologically Inspired Algorithms: A !Systematic Literature Review ! !KEREM PARLAKGÜMÜŞ" ! !© KEREM PARLAKGÜMÜŞ, January 2014." Examiner: LARS PARETO, MIROSLAW STARON" !Supervisor: PALLE DAHLSTEDT" University of Gothenburg" Chalmers University of Technology" Department of Computer Science and Engineering" SE-412 96 Göteborg" Sweden" !Telephone + 46 (0)31-772 1000" ! ! ! Department of Computer Science and Engineering" !Göteborg, Sweden, January 2014 Abstract My original contribution to knowledge is to examine existing work for methods and approaches used, main functionalities, benefits and limitations of 30 Genera- tive Music Composition Software Systems (GMCSS) by performing a systematic literature review. -
Concatenative Sound Synthesis: the Early Years
CONCATENATIVE SOUND SYNTHESIS: THE EARLY YEARS Diemo Schwarz Ircam – Centre Pompidou 1, place Igor-Stravinsky, 75003 Paris, France http://www.ircam.fr/anasyn/schwarz http://concatenative.net [email protected] ABSTRACT Concatenative sound synthesis (CSS) methods use a large database of source sounds, segmented into units, Concatenative sound synthesis is a promising method and a unit selection algorithm that finds the sequence of of musical sound synthesis with a steady stream of work units that match best the sound or phrase to be synthe- and publications for over five years now. This article of- sised, called the target. The selection is performed ac- fers a comparative survey and taxonomy of the many dif- cording to the descriptors of the units, which are charac- ferent approaches to concatenative synthesis throughout teristics extracted from the source sounds, or higher level the history of electronic music, starting in the 1950s, even descriptors attributed to them. The selected units can then if they weren't known as such at their time, up to the recent be transformed to fully match the target specification, and surge of contemporary methods. Concatenative sound are concatenated. However, if the database is sufficiently synthesis methods use a large database of source sounds, large, the probability is high that a matching unit will be segmented into units, and a unit selection algorithm that found, so the need to apply transformations, which always finds the units that match best the sound or musical phrase degrade sound quality, is reduced. The units can be non- to be synthesised, called the target. The selection is per- uniform (heterogeneous), i.e. -
Chroma Palette: Chromatic Maps of Sound As Granular Synthesis
Proceedings of the 2007 Conference on New Interfaces for Musical Expression (NIME07), New York, NY, USA Chroma Palette: Chromatic Maps of Sound As Granular Synthesis Interface Justin Donaldson Ian Knopke Chris Raphael Indiana University School of Indiana University School of Indiana University School of Informatics Informatics Informatics 1900 E. 10th Street, Room 931 1900 E. 10th Street, Room 932 1900 E. 10th Street, Room 933 Bloomington, IN 47406 Bloomington, IN 47406 Bloomington, IN 47406 [email protected] [email protected] [email protected] ABSTRACT There are many different categories of granular synthesis, Chroma based representations of acoustic phenomenon are such as synchronous, quasi-synchronous, and asynchronous representations of sound as pitched acoustic energy. A frame- forms, referring to the regularity with which grains are wise chroma distribution over an entire musical piece is a useful reassembled. Grains are usually windowed, both to aid resynthesis and straightforward representation of its musical pitch over time. and to avoid audible clicks. However, the choice of window This paper examines a method of condensing the block-wise function can also have a pronounced effect on the resulting timbre chroma information of a musical piece into a two dimensional and is an important component of the synthesis process. Granular embedding. Such an embedding is a representation or map of the synthesis has similarities to other common analysis/resynthesis different pitched energies in a song, and how these energies relate methodologies such as the short-term Fourier transform and to each other in the context of the song. The paper presents an wavelet-based techniques. Figure 1 shows an example of an interactive version of this representation as an exploratory envelope windowing and overlap arrangement for four different analytical tool or instrument for granular synthesis.