Chuck: a Concurrent, On-The-Fly, Audio Programming Language

Total Page:16

File Type:pdf, Size:1020Kb

Chuck: a Concurrent, On-The-Fly, Audio Programming Language In Proceedings of the 2003 International Computer Music Conference. ChucK: A Concurrent, On-the-fly, Audio Programming Language Ge Wang and †Perry R. Cook Computer Science Department, †(also Music) Princeton University {gewang,prc}@cs.princeton.edu Abstract shreds and concurrency in ChucK, and demonstrate multiple and simultaneous control paths and control ChucK is a new audio programming language for rates. Section 4 presents an overview of the real-time synthesis, composition, and performance, integration of many types of input/output related which runs on commodity operating systems. ChucK devices and operations into the language. Section 5 natively supports concurrency, multiple, presents a special aspect of ChucK: its ability to be simultaneous, dynamic control rates, and the ability programmed on-the-fly during run-time. ChucK is to add, remove, and modify code, on-the-fly, while the implemented as a virtual machine running with a program is running, without stopping or restarting. special run-time compiler with low-level audio It offers composers and performers a powerful and engine. Section 6 takes a step back and reasons about flexible programming tool for building and the performance benefits/drawbacks of ChucK. experimenting with complex audio synthesis programs, and real-time interactive control. 2. The ChucK Operator 1. Ideas in ChucK ChucK is a strongly-typed, imperative programming language. Its syntax and semantics are In this paper we present four key ideas that form governed by a flexible type system. ChucK includes the foundation of ChucK. The goal is to design a standard features (arithmetic, bit-wise, memory natural audio programming language (1) to operations, etc…) and control flow mechanisms ( , concurrently and accurately represent complex audio if synthesis, (2) to enable fine-grain, flexible control for, while, switch, goto, break, continue, etc…) over time, (3) to provide the capability to operate on common to most modern imperative programming multiple, dynamic and simultaneous control rates, languages. But the heart of ChucK's syntax is based and (4) to make possible an on-the-fly style of around the ChucK operator (written as =>). programming. ChucK runs on commodity operating st systems (Linux, Windows, Solaris, MacOS). 1 rule of ChucK: make use of the left-to-right ChucK operator. => originates from the slang term Four major ideas form the basis of ChucK. "chuck", meaning to throw an entity into or at another entity. The language uses this notion to help express • A unifying, massively overloaded operator. sequential operations and data flow. • A precise timing model that is capable of true concurrency and arbitrarily fine granularity. The (440 => osc) => env => filt => audio[0]; language semantic supports multiple, The above code fragment constructs a simple simultaneous, and dynamic control rates, and synthesis instrument using a series of unit generators naturally amortizes operations over time. (Mathews 1969) (their declarations are omitted for • Native language support of arbitrarily many the moment): an oscillator, an envelope, a filter, and, input/output sources, MIDI, network, serial, finally, audio channel 0. Notice that the single line USB, graphics, and any input/output device you captures the flow of the signals from left to right - the can connect to the computer. same order as we read and type. • On-the-fly programming enables dynamically modifiable programs for performance and A ChucK statement can be composed of any experimentation. appropriate types of objects (including user-defined), unit generators, operations, values, and variables. We present ChucK with respect to these four The semantic of the statement depends on the types major facets. We provide some informal 'rules' of of the objects, and the overloading of the ChucK ChucK, each of which embodies some key aspect of operator on those types. the programming language. In Section 2, we look in detail at the ChucK operator. In Section 3, we present The main ideas behind the ChucK operator are (1) the ChucK timing model, introduce the concept of it locally captures the left-to-right nature of program 1 In Proceedings of the 2003 International Computer Music Conference. flow (with the exception of nested ChucK 5 => int x; statements), representing the order of sequential operations more faithfully, (2) it can be chained to With the ChucK operator, we represent any any arbitrary length, and (3) it reacts, or behaves, sequence of operations in a left-to-right manner. As differently depending on the types of the objects shown below, a piece of code is chucked over TCP to being chucked. =>’s behavior is defined/altered a remote host (also running ChucK), waits for a through overloading using the ChucK type system. message, and prints out the result. code_seg => tcp(140.180.141.103:8888) The basic ChucK operation takes place between => local_receiver => stdout; two entities, the ChucKer and the ChucKee: The ChucK operator can also naturally x => y synchronize on events and objects. Synchronization primitives such as semaphores, condition variables, The operation performed depends on the types of and monitors (Lampson 1980) can be inserted into x (ChucKer) and of y (ChucKee), the combination of ChucK chains as clear points of synchronization. which maps to a particular action that is overloaded on the ChucK operator. For example, it might assign code_seg => mutex => machine; the value to a variable, or connect/disconnect one unit generator to/from another, or anything depending on button[0] => play_note(); the particular overloading. For example, the notion of behavioral abstraction in Nyquist (Dannenberg In the next sections, we will look at the issues of 1997) can be realized through this type of timing and concurrency, input/output, and on-the-fly overloading =>. programming, each of which employs some aspect of the ChucK operator. The following are some sample ChucK statements, where the identifiers x, y etc. can be any 3. Time, Shreds, Rates well-typed entities. The notions of time, duration, synchronization, Binary ChucK: the simplest ChucK statement, control rates, and simultaneity are captured chucking object/value x to object/value y. automatically by ChucK’s timing mechanism. ChucK allows the programmer, composer, and x => y; performer to write truly concurrent code using the framework of the timing semantic. There is no fixed Chain ChucK: ChucK statements can be a sequence notion of control rate – the control rate is a natural of ChucK operations of arbitrary length. The product of using the timing constructs. The timing operations are performed left to right, in exactly the semantic, along with concurrency leads to the ability same order as written. to dynamically change control rate, as well as have many different control rates. w => x => y => z; 3.1 Time and Duration Nested ChucK: evaluation of ChucK expressions gives local precedence to parenthesis. In the The now keyword (of type time) is defined to always hold the exact, current ChucK time. The statement below, the operation v=>w is evaluated, keyword dur refers to the duration between points in then x=>y, and the result of v=>w is chucked to the time. These allow the programmer to do precise result of x=>y. Finally, that result is chucked to z. arithmetic with time and duration as we see in the v => w => ( x => y ) => z; following examples. Cross ChucK: cross-chucking allows one or more Each duration value has a unit attached to it and is ChucKer objects/values to be chucked to more than declared in the following way: one ChucKee. In the statement below, the object/value w is chucked to x, then y, and then z. 0.01:second => dur midi_rate; (This can also be achieved by a more verbose In the above example, we chuck a value of sequence of three ChucK statements.) 0.01:second to a newly declared variable of type w => ( x, y, z ); dur, called midi_rate. The "built-in" units are samp (duration of one sample), ms, second, minute, The ChucK programming language has no hour, day, week. We can also use any variable of assignment operator (=), but unifies this operation type dur as a unit to inductively construct other under the ChucK operator, as seen here: durations: 2 In Proceedings of the 2003 International Computer Music Conference. 0.7:second => dur quarter; responsibility for deciding when to "step out" of 4:<quarter> => dur whole; suspended animation and advance time. He/she can do it in one of two ways. In the first method (chuck- The above statement constructs a duration value to-now) the programmer can allow time to advance of 4:<quarter> (quarter defined in the previous by explicitly chucking a duration value or a time statement) and chucks a relationship (symbolized by value to now, as shown above. This allows for a the angle brackets) to a new variable called whole. In natural programming approach that embeds the this case, chucking a new dur value to quarter will timing control directly in the language, giving the automatically change whole. With this system, we programmer the ability to perform computations at can easily define and use any duration greater than or arbitrary points in time, and to "move forward" in equal to the duration of a single sample. ChucK time in a precise manner. Time is defined in terms of existing time values. The second method to advance time in ChucK is We can perform arithmetic using time and duration to by waiting on some event(s). These could be obtain new time values. Here we calculate the time synchronization mechanisms (such as mutexes, value for 10:second after now, and store this value semaphores, and/or condition variables), input in a new variable called later. devices, or any asynchronous event(s) such as MIDI input or packets arriving over TCP or UDP. 10:second after now => time later; Execution will resume when the synchronization condition is fulfilled.
Recommended publications
  • Synchronous Programming in Audio Processing Karim Barkati, Pierre Jouvelot
    Synchronous programming in audio processing Karim Barkati, Pierre Jouvelot To cite this version: Karim Barkati, Pierre Jouvelot. Synchronous programming in audio processing. ACM Computing Surveys, Association for Computing Machinery, 2013, 46 (2), pp.24. 10.1145/2543581.2543591. hal- 01540047 HAL Id: hal-01540047 https://hal-mines-paristech.archives-ouvertes.fr/hal-01540047 Submitted on 15 Jun 2017 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. A Synchronous Programming in Audio Processing: A Lookup Table Oscillator Case Study KARIM BARKATI and PIERRE JOUVELOT, CRI, Mathématiques et systèmes, MINES ParisTech, France The adequacy of a programming language to a given software project or application domain is often con- sidered a key factor of success in software development and engineering, even though little theoretical or practical information is readily available to help make an informed decision. In this paper, we address a particular version of this issue by comparing the adequacy of general-purpose synchronous programming languages to more domain-specific
    [Show full text]
  • Peter Blasser CV
    Peter Blasser – [email protected] - 410 362 8364 Experience Ten years running a synthesizer business, ciat-lonbarde, with a focus on touch, gesture, and spatial expression into audio. All the while, documenting inventions and creations in digital video, audio, and still image. Disseminating this information via HTML web page design and YouTube. Leading workshops at various skill levels, through manual labor exploring how synthesizers work hand and hand with acoustics, culminating in montage of participants’ pieces. Performance as touring musician, conceptual lecturer, or anything in between. As an undergraduate, served as apprentice to guild pipe organ builders. Experience as racquetball coach. Low brass wind instrumentalist. Fluent in Java, Max/MSP, Supercollider, CSound, ProTools, C++, Sketchup, Osmond PCB, Dreamweaver, and Javascript. Education/Awards • 2002 Oberlin College, BA in Chinese, BM in TIMARA (Technology in Music and Related Arts), minors in Computer Science and Classics. • 2004 Fondation Daniel Langlois, Art and Technology Grant for the project “Shinths” • 2007 Baltimore City Grant for Artists, Craft Category • 2008 Baltimore City Grant for Community Arts Projects, Urban Gardening List of Appearances "Visiting Professor, TIMARA dep't, Environmental Studies dep't", Oberlin College, Oberlin, Ohio, Spring 2011 “Babier, piece for Dancer, Elasticity Transducer, and Max/MSP”, High Zero Festival of Experimental Improvised Music, Theatre Project, Baltimore, September 2010. "Sejayno:Cezanno (Opera)", CEZANNE FAST FORWARD. Baltimore Museum of Art, May 21, 2010. “Deerhorn Tapestry Installation”, Curators Incubator, 2009. MAP Maryland Art Place, September 15 – October 24, 2009. Curated by Shelly Blake-Pock, teachpaperless.blogspot.com “Deerhorn Micro-Cottage and Radionic Fish Drier”, Electro-Music Gathering, New Jersey, October 28-29, 2009.
    [Show full text]
  • Chuck: a Strongly Timed Computer Music Language
    Ge Wang,∗ Perry R. Cook,† ChucK: A Strongly Timed and Spencer Salazar∗ ∗Center for Computer Research in Music Computer Music Language and Acoustics (CCRMA) Stanford University 660 Lomita Drive, Stanford, California 94306, USA {ge, spencer}@ccrma.stanford.edu †Department of Computer Science Princeton University 35 Olden Street, Princeton, New Jersey 08540, USA [email protected] Abstract: ChucK is a programming language designed for computer music. It aims to be expressive and straightforward to read and write with respect to time and concurrency, and to provide a platform for precise audio synthesis and analysis and for rapid experimentation in computer music. In particular, ChucK defines the notion of a strongly timed audio programming language, comprising a versatile time-based programming model that allows programmers to flexibly and precisely control the flow of time in code and use the keyword now as a time-aware control construct, and gives programmers the ability to use the timing mechanism to realize sample-accurate concurrent programming. Several case studies are presented that illustrate the workings, properties, and personality of the language. We also discuss applications of ChucK in laptop orchestras, computer music pedagogy, and mobile music instruments. Properties and affordances of the language and its future directions are outlined. What Is ChucK? form the notion of a strongly timed computer music programming language. ChucK (Wang 2008) is a computer music program- ming language. First released in 2003, it is designed to support a wide array of real-time and interactive Two Observations about Audio Programming tasks such as sound synthesis, physical modeling, gesture mapping, algorithmic composition, sonifi- Time is intimately connected with sound and is cation, audio analysis, and live performance.
    [Show full text]
  • Implementing Stochastic Synthesis for Supercollider and Iphone
    Implementing stochastic synthesis for SuperCollider and iPhone Nick Collins Department of Informatics, University of Sussex, UK N [dot] Collins ]at[ sussex [dot] ac [dot] uk - http://www.cogs.susx.ac.uk/users/nc81/index.html Proceedings of the Xenakis International Symposium Southbank Centre, London, 1-3 April 2011 - www.gold.ac.uk/ccmc/xenakis-international-symposium This article reflects on Xenakis' contribution to sound synthesis, and explores practical tools for music making touched by his ideas on stochastic waveform generation. Implementations of the GENDYN algorithm for the SuperCollider audio programming language and in an iPhone app will be discussed. Some technical specifics will be reported without overburdening the exposition, including original directions in computer music research inspired by his ideas. The mass exposure of the iGendyn iPhone app in particular has provided a chance to reach a wider audience. Stochastic construction in music can apply at many timescales, and Xenakis was intrigued by the possibility of compositional unification through simultaneous engagement at multiple levels. In General Dynamic Stochastic Synthesis Xenakis found a potent way to extend stochastic music to the sample level in digital sound synthesis (Xenakis 1992, Serra 1993, Roads 1996, Hoffmann 2000, Harley 2004, Brown 2005, Luque 2006, Collins 2008, Luque 2009). In the central algorithm, samples are specified as a result of breakpoint interpolation synthesis (Roads 1996), where breakpoint positions in time and amplitude are subject to probabilistic perturbation. Random walks (up to second order) are followed with respect to various probability distributions for perturbation size. Figure 1 illustrates this for a single breakpoint; a full GENDYN implementation would allow a set of breakpoints, with each breakpoint in the set updated by individual perturbations each cycle.
    [Show full text]
  • Isupercolliderkit: a Toolkit for Ios Using an Internal Supercollider Server As a Sound Engine
    ICMC 2015 – Sept. 25 - Oct. 1, 2015 – CEMI, University of North Texas iSuperColliderKit: A Toolkit for iOS Using an Internal SuperCollider Server as a Sound Engine Akinori Ito Kengo Watanabe Genki Kuroda Ken’ichiro Ito Tokyo University of Tech- Watanabe-DENKI Inc. Tokyo University of Tech- Tokyo University of Tech- nology [email protected] nology nology [email protected]. [email protected]. [email protected]. ac.jp ac.jp ac.jp The editor client sends the OSC code-fragments to its server. ABSTRACT Due to adopting the model, SuperCollider has feature that a iSuperColliderKit is a toolkit for iOS using an internal programmer can dynamically change some musical ele- SuperCollider Server as a sound engine. Through this re- ments, phrases, rhythms, scales and so on. The real-time search, we have adapted the exiting SuperCollider source interactivity is effectively utilized mainly in the live-coding code for iOS to the latest environment. Further we attempt- field. If the iOS developers would make their application ed to detach the UI from the sound engine so that the native adopting the “sound-server” model, using SuperCollider iOS visual objects built by objective-C or Swift, send to the seems to be a reasonable choice. However, the porting situa- internal SuperCollider server with any user interaction tion is not so good. SonicPi[5] is the one of a musical pro- events. As a result, iSuperColliderKit makes it possible to gramming environment that has SuperCollider server inter- utilize the vast resources of dynamic real-time changing nally. However, it is only for Raspberry Pi, Windows and musical elements or algorithmic composition on SuperCol- OSX.
    [Show full text]
  • The Viability of the Web Browser As a Computer Music Platform
    Lonce Wyse and Srikumar Subramanian The Viability of the Web Communications and New Media Department National University of Singapore Blk AS6, #03-41 Browser as a Computer 11 Computing Drive Singapore 117416 Music Platform [email protected] [email protected] Abstract: The computer music community has historically pushed the boundaries of technologies for music-making, using and developing cutting-edge computing, communication, and interfaces in a wide variety of creative practices to meet exacting standards of quality. Several separate systems and protocols have been developed to serve this community, such as Max/MSP and Pd for synthesis and teaching, JackTrip for networked audio, MIDI/OSC for communication, as well as Max/MSP and TouchOSC for interface design, to name a few. With the still-nascent Web Audio API standard and related technologies, we are now, more than ever, seeing an increase in these capabilities and their integration in a single ubiquitous platform: the Web browser. In this article, we examine the suitability of the Web browser as a computer music platform in critical aspects of audio synthesis, timing, I/O, and communication. We focus on the new Web Audio API and situate it in the context of associated technologies to understand how well they together can be expected to meet the musical, computational, and development needs of the computer music community. We identify timing and extensibility as two key areas that still need work in order to meet those needs. To date, despite the work of a few intrepid musical Why would musicians care about working in explorers, the Web browser platform has not been the browser, a platform not specifically designed widely considered as a viable platform for the de- for computer music? Max/MSP is an example of a velopment of computer music.
    [Show full text]
  • Rock Around Sonic Pi Sam Aaron, Il Live Coding E L’Educazione Musicale
    Rock around Sonic Pi Sam Aaron, il live coding e l’educazione musicale Roberto Agostini IC9 Bologna, Servizio Marconi TSI (USR-ER) 1. Formazione “Hands-On” con Sam Aaron “Imagine if the only allowed use of reading and writing was to make legal documents. Would that be a nice world? Now, today’s programmers are all like that”1. In questa immagine enigmatica e un po’ provocatoria proposta da Sam Aaron nelle prime fasi di una sua recente conferenza è racchiusa un’originale visione della programmazione ricca di implicazioni, e non solo per l’educazione musicale. Ma andiamo con ordine: il 14 febbraio 2020, all’Opificio Golinelli di Bologna, Sam Aaron, ​ ​ creatore del software per live coding musicale Sonic Pi, è stato protagonista di un denso ​ ​ pomeriggio dedicato al pensiero computazionale nella musica intitolato “Rock around Sonic Pi”. L’iniziativa rientrava nell’ambito della “Formazione Hands-On” promossa dal Future Lab dell’Istituto Comprensivo di Ozzano dell’Emilia in collaborazione con il Servizio Marconi TSI ​ dell’USR-ER. Oltre alla menzionata conferenza, aperta a tutti, Sam Aaron ha tenuto un laboratorio di tre ore a numero chiuso per i docenti delle scuole2. ● Presentazione e programma dettagliato della giornata “Rock Around Sonic Pi” (presentazione a cura ​ di Rosa Maria Caffio). ● Ripresa integrale della conferenza di Sam Aaron (20.2.2020) (registrazione a cura dello staff ​ dell’Opificio Golinelli). 1 “Immaginate che l’unico modo consentito di usare la lettura e la scrittura fosse quello di produrre documenti legali. Sarebbe un bel mondo? Ora, i programmatori di oggi sono nella stessa situazione”.
    [Show full text]
  • A Supercollider Environment for Synthesis-Oriented Live Coding
    Colliding: a SuperCollider environment for synthesis-oriented live coding Gerard Roma CVSSP, University of Surrey Guildford, United Kingdom [email protected] Abstract. One of the motivations for live coding is the freedom and flexibility that a programming language puts in the hands of the performer. At the same time, systems with explicit constraints facilitate learning and often boost creativity in unexpected ways. Some simplified languages and environments for music live coding have been developed during the last few years, most often focusing on musical events, patterns and sequences. This paper describes a constrained environment aimed at exploring the creation and modification of sound synthesis and processing networks in real time, using a subset of the SuperCollider programming language. The system has been used in educational and concert settings, two common applications of live coding that benefit from the lower cognitive load. Keywords: Live coding, sound synthesis, live interfaces Introduction The world of specialized music creation programming languages has been generally dominated by the Music-N paradigm pioneered by Max Mathews (Mathews 1963). Programming languages like CSound or SuperCollider embrace the division of the music production task in two separate levels: a signal processing level, which is used to define instruments as networks of unit generators, and a compositional level that is used to assemble and control the instruments. An exception to this is Faust, which is exclusively concerned with signal processing. Max and Pure Data use a different type of connection for signals and events, although under a similar interaction paradigm. Similarly, Chuck has specific features for connecting unit generators and controlling them along time.
    [Show full text]
  • Final Presentation
    MBET Senior Project: Final Presentation Jonah Pfluger What was the project? And how did I spend my time? ● The core of this project centralized around two main areas of focus: ○ The designing and building of custom audio software in Max/MSP and SuperCollider ○ The organization, promotion, and performance of a concert event that utilized these custom audio tools in a free improvised setting alongside other instruments ● Regarding “MBET”: ○ The former part of the project addresses “Technology” ○ The latter part (most directly) addresses “Business” ■ Also addresses “Technology” as I undertook various technical tasks in the setup process such as: setting up a Quadraphonic speaker system for synthesizer playback, setting up a separate stereo system for vocal playback/monitoring, and other tasks standard to “live event production” Max/MSP vs SuperCollider ● To develop the various audio tools/software instruments I heavily utilized Max/MSP and SuperCollider. ● Max is a visual programming language whereas SuperCollider uses a language derivative of C++ called “sclang” ● All Max projects eventually were built out into graphic user interfaces, 1/2 of SC projects were (with the other ½ having a “live code” style workflow) ● Ultimately, both were extremely useful but are simply different and, therefore, better at different tasks The instruments… (part 1) ● Max instrument: Generative Ambient Console ● Expanded class project based on generative ambient "drone", good for Eno-esque textures, or a ghostly electronic Tanpura-esque sound. The instruments…
    [Show full text]
  • Applications in Heart Rate Variability
    Data Analysis through Auditory Display: Applications in Heart Rate Variability Mark Ballora Faculty of Music McGill University, Montréal May, 2000 A thesis submitted to the Faculty of Graduate Studies and Research in partial fulfillment of the requirements of the degree of Doctor of Philosophy in Music © Mark Ballora, May 2000 Table of Contents Abstract......................................................................................... v Résumé ........................................................................................ vi Acknowledgements ..........................................................................vii 1. Introduction 1.1 Purpose of Study.................................................................... 1 1.2 Auditory Display.................................................................... 2 1.3 Types of Auditory Display ........................................................ 3 1.4 Heart Rate Variability.............................................................. 4 1.5 Design of the Thesis................................................................ 6 2. Survey of Related Literature 2.1 Data in Music 2.1.1 Data Music—Making Art from Information............................ 8 2.1.2 Biofeedback Music........................................................14 2.1.3 Nonlinear Dynamics in Music ...........................................16 2.1.3.1 Fractal Music...................................................16 2.1.3.2 Mapping Chaotic (and other) Data ..........................18 2.1.4 Concluding Thoughts
    [Show full text]
  • A Tabletop Waveform Editor for Live Performance
    Proceedings of the 2008 Conference on New Interfaces for Musical Expression (NIME08), Genova, Italy A tabletop waveform editor for live performance Gerard Roma Anna Xambo´ Music Technology Group Music Technology Group Universitat Pompeu Fabra Universitat Pompeu Fabra [email protected] [email protected] ABSTRACT We present an audio waveform editor that can be oper- ated in real time through a tabletop interface. The system combines multi-touch and tangible interaction techniques in order to implement the metaphor of a toolkit that allows di- rect manipulation of a sound sample. The resulting instru- ment is well suited for live performance based on evolving loops. Keywords tangible interface, tabletop interface, musical performance, interaction techniques 1. INTRODUCTION The user interface of audio editors has changed relatively little over time. The standard interaction model is centered on the waveform display, allowing the user to select portions Figure 1: Close view of the waveTable prototype. of the waveform along the horizontal axis and execute com- mands that operate on those selections. This model is not very different to that of the word processor, and its basics Barcelona 2005) which inspired this project. The use of a are usually understood by computer users even with little standard audio editor on a laptop computer as a sophisti- or no experience in specialized audio tools. As a graphical cated looper served the performers’ minimalist glitch aes- representation of sound, the waveform is already familiar thetics. Perhaps more importantly, the projected waveform for many people approaching computers for audio and mu- provided a visual cue that helped the audience follow the sic composition.
    [Show full text]
  • A Comparison of Solenoid-Based Strategies for Robotic Drumming
    A COMPARISON OF SOLENOID-BASED STRATEGIES FOR ROBOTIC DRUMMING Ajay Kapur 1,2,3 Trimpin Eric Singer2 Afzal Suleman1 George Tzanetakis1 Music Intelligence and Sound Technology Interdisciplinary Centre (MISTIC) 1 University of Victoria, Victoria, British Columbia, Canada League of Electronic Urban Music Robots (LEMUR) 2 Brooklyn, New York, USA KarmetiK Technology (A Division of KarmetiK LLC) 3 Reno, Nevada, USA ABSTRACT drums gives the students a more realistic paradigm for concentrated rehearsal. Solenoids are important components of robotic A number of different drumming robots have been drumming systems and several designs have been designed in the academic and artistic communities. proposed. In this paper, we compare different designs Researchers at Harvard University struggled to create an in terms of speed and dynamic range and discuss the accurate robotic drum roll [1], while next door tradeoffs involved. The evaluation is performed in the researchers at MIT developed Cog to control the context of MahaDeviBot, a custom built 12-armed MIDI number of times a stick can bounce [2]. Gil Weinberg controlled mechanical device that performs a variety of developed Haile to explore human to robot interaction Indian folk instruments, including frame-drums, bells, [3]. Mitsuo Kawato continues to develop hydraulic and shakers. To measure speed and dynamic range a systems for humanoid drumming [4]. Many artists have haptic robotic feedback system using piezo sensors was presented a number of different pieces including built. The evaluation methods presented are modular Baginsky’s “Thelxiapeia” for modified rototom [5], and can be administered to any hyperinstrument, new MacMurtie’s life size sculptures [6], Gordon Monohans interface or robotic system to inform composers about “Machine Matrix” [7], and Miles van Dorssen’s “Cell the strengths and limitation of different designs to guide Project” including an 8 octave Xylophone, Bamboo composition and performance.
    [Show full text]