The New Developments in Csound 6

Total Page:16

File Type:pdf, Size:1020Kb

The New Developments in Csound 6 ICMC 2015 – Sept. 25 - Oct. 1, 2015 – CEMI, University of North Texas The New Developments in Csound 6 John ffitch Victor Lazzarini Steven Yi University of Bath National University of Ireland, Maynooth Department of Computer Science Department of Music Michael Gogins Andres´ Cabrera Irreducible Productions University of California, Santa Barbara New York Media Arts and Technology ABSTRACT of Csound presented here, over the years [3] [4] [5] [6] [7] [8] [9]. However, it is well worth describing what Csound In this paper we introduce a major new version of Csound, is in a bit more detail, because 1) Csound has a long history the audio processing system and library. We begin with an of development, and much of the information describing it is overview of the current status of Csound (version 5), as well outdated; and 2) the motivation for the present directions will as its three layers of use (high, middle, and low). We then become clearer as we outline the present system. outline the design motivations and features of version 6. We The best way to describe Csound, in its version 5, is to continue by exploring external changes and discussing some present it as a series of layers, with various ‘modes of entry’ examples of use. We conclude by looking forward to the next for users and for related applications. steps in the development of Csound. At the lowest level, Csound is a self-contained audio pro- gramming language implemented in a cross-platform library, 1. INTRODUCTION with a well-defined API, which allows software developers to create programs for audio synthesis and processing, and com- In 2012, six years after the initial release of the first major puter music composition. Csound supports a variety of syn- re-engineering of the well-known and widely-used software thesis techniques in its orchestra language, and allows various sound synthesis system Csound [1], we (its developers) de- means/levels of internal and external control. Csound is ex- cided to embark on a further revision of many of its inter- tensible via plugin modules. Software that uses Csound can nal and external aspects. Developments since version 5.00 be written in C, C++, Objective-C, Java, Python, Lua, Tcl, [2] until the current release, 5.19, have been mostly incre- Lisp, and others. Csound runs on Windows, Linux, OSX, mental. They have also been limited by our commitment Solaris, Haiku, Android and iOS. to maintaining both binary and API (Application Program- ming Interface) compatibility with earlier versions (although The middle layer is characterized by writing programs in the the system has actually come through a binary upgrade, after Csound language for performance, composition, and other version 5.09). To allow for a number of requested changes, audio processing tasks such as sonification. At this level, the we decided a new major version was necessary, which would system allows composers to design computer music instru- mean a break in backwards compatibility (both API and bi- ments, and to control them in real time or deferred time. In- nary). This does not, however, mean a break in backwards teraction with the system comes via various frontends, many compatibility of Csound code and pieces. Older pieces and of which are third party (i.e. not maintained as part of the code will always continue to work with Csound 6. This pa- Csound releases). The ‘classic’ command-line interface (CLI) per discusses the motivation for Csound 6, its development is the basic frontend, where the system is controlled by a sin- process, and major features of the new system. gle terminal command. As this was the only original means of using the software, traditionally a number of frontends have been designed to provide a simpler wrapper around CLI 2. WHAT IS CSOUND? Csound. More commonly, today, frontends access the Csound For the ICMC audience, it might not seem necessary to de- library directly (via its API). These frontends provide diverse scribe such a well-known and established software package. modes of interaction. For example, Csound can be embedded After all, there have been a number of papers on the subject in graphical environments such as Pure Data via the csoundapi∼ frontend, and in Max/MSP via csound∼. Com- position environments such as blue use it as a sound engine. Copyright: c 2015 John ffitch et al. This is an open-access article dis- For more general-purpose uses, there are integrated devel- tributed under the terms of the Creative Commons Attribution License 3.0 opment environments (IDEs) for programming with Csound Unported, which permits unrestricted use, distribution, and reproduction in (such as CsoundQt and WinXsound), and plugin/application any medium, provided the original author and source are credited. generators, such as Cs-LADSPA [6] and Cabbage[8]. – 424 – ICMC 2015 – Sept. 25 - Oct. 1, 2015 – CEMI, University of North Texas At this level, Csound co-exists with a number of tools and cult now, especially with the availability of models that exist languages which add support for activities such as algorith- as open-source code. In fact, there is a multiplication of in- mic composition and graphical user interaction. cipient systems that claim to be the intelligent solution to per- ceived problems in existing software. The majority of these app users, do not cross the 80/20 divide of development. This occurs performers, possibly for a variety of reasons, but especially for the lack composers of an enthusiastic user (and developer) community. There is great value in the accumulated knowledge of the community Csound-based apps, high and the large body of existing code. We understand that mov- applications desktop programs composers, performers, ing Csound away from its origins as a system does not mean Csound IDEs, frontends, researchers ditching users and music along the way. The requirements of middle language plugin generators the community are paramount to where we want the software levels to go. By supporting the various levels of entry into the sys- programming languages low Csound library and tools tem, we aim to foster interest in the software and in computer music in general. This translates as well into the different lev- els of difficulty that the Csound language contains. It allows developers, researchers educators to provide a smooth learning curve for students, go- ing from the early (and simpler) set of language elements into Figure 1. The various levels of Csound, related software and end users. the expanded one that the system supports today. At the highest level, use of Csound occurs by developing 3. WHY CSOUND 6? applications based on the middle and lower levels. Here, the user might not even know that Csound is involved, as user By 2012, we began to feel that Csound 5’s incremental model programming is generally not involved. This is seen, for in- of development was becoming a limitation. At the 2011 In- stance, in some frontends, such as Cecilia and blue, where ternational Csound Conference in Hannover, users and devel- the user might only need to deal with parameter setting in the opers met to agree on a number of desired features that the graphical interface, in plugins or applications generated by software should have in future versions. Some of these (like Cabbage, in bundled packages such as Csound4Live (which support for mobile platforms and some additional language uses the csound∼ frontend), or in mobile applications for iOS features) were achievable in Csound 5 and indeed were soon and Android [9]. made available. Others have required a major re-engineering It is clear that Csound has attracted a diverse set of users, of the system. Among them, we can cite: from the expert programmer to the mobile app customer. In • the capacity of new orchestra code, ie. instruments and addition, thanks to Csound’s long history, especially by com- user-defined opcodes (UDOs), to be added to a running posers working at the middle level, there is a legacy of music instance of the engine (enhancing, for instance, live- written with Csound that is worth preserving (and that in fact coding support and interactive sound design); stretches back to 70s compositions written for Csound’s pre- decessors MUSIC 11 and MUSIC 360 [10]). This has focused • major additions to the orchestra language, for instance, our minds to provide a completely backwards-compatible sys- generic arrays, debugging/introspection, and a type sys- tem (as far as the language is concerned) as sine qua non con- tem; dition for future versions. Some may criticise this as recipe for an ever increasing lan- • rationalisation of the API to simplify its usage and to guage, with its associated complexity penalty, and, as often allow further features in frontends; vocalised in the detractors’ corner, ‘bloat.’ But although some • fast loadable (FASL-like) binary formats, API construc- ‘bloat’ is inevitable in a system nearly three decades old, the tion of instruments; Csound language is still syntactically very simple (it consists of just a few simple syntactical constructs), and the process- • further development of concurrency (enhancement of ing engine is generally efficient in terms of DSP and algo- existing support [7]). rithm implementations. What criticisms fail to consider is that Csound has fostered a vibrant community of users and This list was our starting point for the development of developers. Csound 6. We understand that community is the biggest asset a sys- tem like Csound can have. Without users, expert and non- 4. INTERNAL CHANGES IN VERSION 6.0 expert, a system withers and dies. It would not be a huge task (in comparative terms) to ditch the old system and re- A number of important changes have been made to the code create one whose language adapts completely to the flavour- base, which not only introduce significant improvements and of-the-moment software design.
Recommended publications
  • Synchronous Programming in Audio Processing Karim Barkati, Pierre Jouvelot
    Synchronous programming in audio processing Karim Barkati, Pierre Jouvelot To cite this version: Karim Barkati, Pierre Jouvelot. Synchronous programming in audio processing. ACM Computing Surveys, Association for Computing Machinery, 2013, 46 (2), pp.24. 10.1145/2543581.2543591. hal- 01540047 HAL Id: hal-01540047 https://hal-mines-paristech.archives-ouvertes.fr/hal-01540047 Submitted on 15 Jun 2017 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. A Synchronous Programming in Audio Processing: A Lookup Table Oscillator Case Study KARIM BARKATI and PIERRE JOUVELOT, CRI, Mathématiques et systèmes, MINES ParisTech, France The adequacy of a programming language to a given software project or application domain is often con- sidered a key factor of success in software development and engineering, even though little theoretical or practical information is readily available to help make an informed decision. In this paper, we address a particular version of this issue by comparing the adequacy of general-purpose synchronous programming languages to more domain-specific
    [Show full text]
  • Tosca: an OSC Communication Plugin for Object-Oriented Spatialization Authoring
    ICMC 2015 – Sept. 25 - Oct. 1, 2015 – CEMI, University of North Texas ToscA: An OSC Communication Plugin for Object-Oriented Spatialization Authoring Thibaut Carpentier UMR 9912 STMS IRCAM-CNRS-UPMC 1, place Igor Stravinsky, 75004 Paris [email protected] ABSTRACT or consensus on the representation of spatialization metadata (in spite of several initiatives [11, 12]) complicates the inter- The paper presents ToscA, a plugin that uses the OSC proto- exchange between applications. col to transmit automation parameters between a digital audio This paper introduces ToscA, a plugin that allows to commu- workstation and a remote application. A typical use case is the nicate automation parameters between a digital audio work- production of an object-oriented spatialized mix independently station and a remote application. The main use case is the of the limitations of the host applications. production of massively multichannel object-oriented spatial- ized mixes. 1. INTRODUCTION There is today a growing interest in sound spatialization. The 2. WORKFLOW movie industry seems to be shifting towards “3D” sound sys- We propose a workflow where the spatialization rendering tems; mobile devices have radically changed our listening engine lies outside of the workstation (Figure 1). This archi- habits and multimedia broadcast companies are evolving ac- tecture requires (bi-directional) communication between the cordingly, showing substantial interest in binaural techniques DAW and the remote renderer, and both the audio streams and and interactive content; massively multichannel equipments the automation data shall be conveyed. After being processed and transmission protocols are available and they facilitate the by the remote renderer, the spatialized signals could either installation of ambitious speaker setups in concert halls [1] or feed the reproduction setup (loudspeakers or headphones), be exhibition venues.
    [Show full text]
  • Proceedings of the Fourth International Csound Conference
    Proceedings of the Fourth International Csound Conference Edited by: Luis Jure [email protected] Published by: Escuela Universitaria de Música, Universidad de la República Av. 18 de Julio 1772, CP 11200 Montevideo, Uruguay ISSN 2393-7580 © 2017 International Csound Conference Conference Chairs Paper Review Committee Luis Jure (Chair) +yvind Brandtsegg Martín Rocamora (Co-Chair) Pablo Di Liscia John -tch Organization Team Michael Gogins Jimena Arruti Joachim )eintz Pablo Cancela Alex )ofmann Guillermo Carter /armo Johannes Guzmán Calzada 0ictor Lazzarini Ignacio Irigaray Iain McCurdy Lucía Chamorro Rory 1alsh "eli#e Lamolle Juan Martín L$#ez Music Review Committee Gusta%o Sansone Pablo Cetta Sofía Scheps Joel Chadabe Ricardo Dal "arra Sessions Chairs Pablo Di Liscia Pablo Cancela "olkmar )ein Pablo Di Liscia Joachim )eintz Michael Gogins Clara Ma3da Joachim )eintz Iain McCurdy Luis Jure "lo Menezes Iain McCurdy Daniel 4##enheim Martín Rocamora Juan Pampin Steven *i Carmelo Saitta Music Curator Rodrigo Sigal Luis Jure Clemens %on Reusner Index Preface Keynote talks The 60 years leading to Csound 6.09 Victor Lazzarini Don Quijote, the Island and the Golden Age Joachim Heintz The ATS technique in Csound: theoretical background, present state and prospective Oscar Pablo Di Liscia Csound – The Swiss Army Synthesiser Iain McCurdy How and Why I Use Csound Today Steven Yi Conference papers Working with pch2csd – Clavia NM G2 to Csound Converter Gleb Rogozinsky, Eugene Cherny and Michael Chesnokov Daria: A New Framework for Composing, Rehearsing and Performing Mixed Media Music Guillermo Senna and Juan Nava Aroza Interactive Csound Coding with Emacs Hlöðver Sigurðsson Chunking: A new Approach to Algorithmic Composition of Rhythm and Metre for Csound Georg Boenn Interactive Visual Music with Csound and HTML5 Michael Gogins Spectral and 3D spatial granular synthesis in Csound Oscar Pablo Di Liscia Preface The International Csound Conference (ICSC) is the principal biennial meeting for members of the Csound community and typically attracts worldwide attendance.
    [Show full text]
  • Computer Music
    THE OXFORD HANDBOOK OF COMPUTER MUSIC Edited by ROGER T. DEAN OXFORD UNIVERSITY PRESS OXFORD UNIVERSITY PRESS Oxford University Press, Inc., publishes works that further Oxford University's objective of excellence in research, scholarship, and education. Oxford New York Auckland Cape Town Dar es Salaam Hong Kong Karachi Kuala Lumpur Madrid Melbourne Mexico City Nairobi New Delhi Shanghai Taipei Toronto With offices in Argentina Austria Brazil Chile Czech Republic France Greece Guatemala Hungary Italy Japan Poland Portugal Singapore South Korea Switzerland Thailand Turkey Ukraine Vietnam Copyright © 2009 by Oxford University Press, Inc. First published as an Oxford University Press paperback ion Published by Oxford University Press, Inc. 198 Madison Avenue, New York, New York 10016 www.oup.com Oxford is a registered trademark of Oxford University Press All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted, in any form or by any means, electronic, mechanical, photocopying, recording, or otherwise, without the prior permission of Oxford University Press. Library of Congress Cataloging-in-Publication Data The Oxford handbook of computer music / edited by Roger T. Dean. p. cm. Includes bibliographical references and index. ISBN 978-0-19-979103-0 (alk. paper) i. Computer music—History and criticism. I. Dean, R. T. MI T 1.80.09 1009 i 1008046594 789.99 OXF tin Printed in the United Stares of America on acid-free paper CHAPTER 12 SENSOR-BASED MUSICAL INSTRUMENTS AND INTERACTIVE MUSIC ATAU TANAKA MUSICIANS, composers, and instrument builders have been fascinated by the expres- sive potential of electrical and electronic technologies since the advent of electricity itself.
    [Show full text]
  • Implementing Stochastic Synthesis for Supercollider and Iphone
    Implementing stochastic synthesis for SuperCollider and iPhone Nick Collins Department of Informatics, University of Sussex, UK N [dot] Collins ]at[ sussex [dot] ac [dot] uk - http://www.cogs.susx.ac.uk/users/nc81/index.html Proceedings of the Xenakis International Symposium Southbank Centre, London, 1-3 April 2011 - www.gold.ac.uk/ccmc/xenakis-international-symposium This article reflects on Xenakis' contribution to sound synthesis, and explores practical tools for music making touched by his ideas on stochastic waveform generation. Implementations of the GENDYN algorithm for the SuperCollider audio programming language and in an iPhone app will be discussed. Some technical specifics will be reported without overburdening the exposition, including original directions in computer music research inspired by his ideas. The mass exposure of the iGendyn iPhone app in particular has provided a chance to reach a wider audience. Stochastic construction in music can apply at many timescales, and Xenakis was intrigued by the possibility of compositional unification through simultaneous engagement at multiple levels. In General Dynamic Stochastic Synthesis Xenakis found a potent way to extend stochastic music to the sample level in digital sound synthesis (Xenakis 1992, Serra 1993, Roads 1996, Hoffmann 2000, Harley 2004, Brown 2005, Luque 2006, Collins 2008, Luque 2009). In the central algorithm, samples are specified as a result of breakpoint interpolation synthesis (Roads 1996), where breakpoint positions in time and amplitude are subject to probabilistic perturbation. Random walks (up to second order) are followed with respect to various probability distributions for perturbation size. Figure 1 illustrates this for a single breakpoint; a full GENDYN implementation would allow a set of breakpoints, with each breakpoint in the set updated by individual perturbations each cycle.
    [Show full text]
  • The Viability of the Web Browser As a Computer Music Platform
    Lonce Wyse and Srikumar Subramanian The Viability of the Web Communications and New Media Department National University of Singapore Blk AS6, #03-41 Browser as a Computer 11 Computing Drive Singapore 117416 Music Platform [email protected] [email protected] Abstract: The computer music community has historically pushed the boundaries of technologies for music-making, using and developing cutting-edge computing, communication, and interfaces in a wide variety of creative practices to meet exacting standards of quality. Several separate systems and protocols have been developed to serve this community, such as Max/MSP and Pd for synthesis and teaching, JackTrip for networked audio, MIDI/OSC for communication, as well as Max/MSP and TouchOSC for interface design, to name a few. With the still-nascent Web Audio API standard and related technologies, we are now, more than ever, seeing an increase in these capabilities and their integration in a single ubiquitous platform: the Web browser. In this article, we examine the suitability of the Web browser as a computer music platform in critical aspects of audio synthesis, timing, I/O, and communication. We focus on the new Web Audio API and situate it in the context of associated technologies to understand how well they together can be expected to meet the musical, computational, and development needs of the computer music community. We identify timing and extensibility as two key areas that still need work in order to meet those needs. To date, despite the work of a few intrepid musical Why would musicians care about working in explorers, the Web browser platform has not been the browser, a platform not specifically designed widely considered as a viable platform for the de- for computer music? Max/MSP is an example of a velopment of computer music.
    [Show full text]
  • Interactive Csound Coding with Emacs
    Interactive Csound coding with Emacs Hlöðver Sigurðsson Abstract. This paper will cover the features of the Emacs package csound-mode, a new major-mode for Csound coding. The package is in most part a typical emacs major mode where indentation rules, comple- tions, docstrings and syntax highlighting are provided. With an extra feature of a REPL1, that is based on running csound instance through the csound-api. Similar to csound-repl.vim[1] csound- mode strives to enable the Csound user a faster feedback loop by offering a REPL instance inside of the text editor. Making the gap between de- velopment and the final output reachable within a real-time interaction. 1 Introduction After reading the changelog of Emacs 25.1[2] I discovered a new Emacs feature of dynamic modules, enabling the possibility of Foreign Function Interface(FFI). Being insired by Gogins’s recent Common Lisp FFI for the CsoundAPI[3], I decided to use this new feature and develop an FFI for Csound. I made the dynamic module which ports the greater part of the C-lang’s csound-api and wrote some of Steven Yi’s csound-api examples for Elisp, which can be found on the Gihub page for CsoundAPI-emacsLisp[4]. This sparked my idea of creating a new REPL based Csound major mode for Emacs. As a composer using Csound, I feel the need to be close to that which I’m composing at any given moment. From previous Csound front-end tools I’ve used, the time between writing a Csound statement and hearing its output has been for me a too long process of mouseclicking and/or changing windows.
    [Show full text]
  • Expanding the Power of Csound with Integrated Html and Javascript
    Michael Gogins. Expanding the Power of Csound with Intergrated HTML and JavaScript EXPANDING THE POWER OF CSOUND WITH INTEGRATED HTML AND JAVA SCRIPT Michael Gogins [email protected] https://michaelgogins.tumblr.com http://michaelgogins.tumblr.com/ This paper presents recent developments integrating Csound [1] with HTML [2] and JavaScript [3, 4]. For those new to Csound, it is a “MUSIC N” style, user- programmable software sound synthesizer, one of the first yet still being extended, written mostly in the C language. No synthesizer is more powerful. Csound can now run in an interactive Web page, using all the capabilities of current Web browsers: custom widgets, 2- and 3-dimensional animated and interactive graphics canvases, video, data storage, WebSockets, Web Audio, mathematics typesetting, etc. See the whole list at HTML5 TEST [5]. Above all, the JavaScript programming language can be used to control Csound, extend its capabilities, generate scores, and more. JavaScript is the “glue” that binds together the components and capabilities of HTML5. JavaScript is a full-featured, dynamically typed language that supports functional programming and prototype-based object- oriented programming. In most browsers, the JavaScript virtual machine includes a just- in-time compiler that runs about 4 times slower than compiled C, very fast for a dynamic language. JavaScript has limitations. It is single-threaded, and in standard browsers, is not permitted to access the local file system outside the browser's sandbox. But most musical applications can use an embedded browser, which bypasses the sandbox and accesses the local file system. HTML Environments for Csound There are two approaches to integrating Csound with HTML and JavaScript.
    [Show full text]
  • DVD Program Notes
    DVD Program Notes Part One: Thor Magnusson, Alex Click Nilson is a Swedish avant McLean, Nick Collins, Curators garde codisician and code-jockey. He has explored the live coding of human performers since such Curators’ Note early self-modifiying algorithmic text pieces as An Instructional Game [Editor’s note: The curators attempted for One to Many Musicians (1975). to write their Note in a collaborative, He is now actively involved with improvisatory fashion reminiscent Testing the Oxymoronic Potency of of live coding, and have left the Language Articulation Programmes document open for further interaction (TOPLAP), after being in the right from readers. See the following URL: bar (in Hamburg) at the right time (2 https://docs.google.com/document/d/ AM, 15 February 2004). He previously 1ESzQyd9vdBuKgzdukFNhfAAnGEg curated for Leonardo Music Journal LPgLlCe Mw8zf1Uw/edit?hl=en GB and the Swedish Journal of Berlin Hot &authkey=CM7zg90L&pli=1.] Drink Outlets. Alex McLean is a researcher in the area of programming languages for Figure 1. Sam Aaron. the arts, writing his PhD within the 1. Overtone—Sam Aaron Intelligent Sound and Music Systems more effectively and efficiently. He group at Goldsmiths College, and also In this video Sam gives a fast-paced has successfully applied these ideas working within the OAK group, Uni- introduction to a number of key and techniques in both industry versity of Sheffield. He is one-third of live-programming techniques such and academia. Currently, Sam the live-coding ambient-gabba-skiffle as triggering instruments, scheduling leads Improcess, a collaborative band Slub, who have been making future events, and synthesizer design.
    [Show full text]
  • 62 Years and Counting: MUSIC N and the Modular Revolution
    62 Years and Counting: MUSIC N and the Modular Revolution By Brian Lindgren MUSC 7660X - History of Electronic and Computer Music Fall 2019 24 December 2019 © Copyright 2020 Brian Lindgren Abstract. MUSIC N by Max Mathews had two profound impacts in the world of music ​ synthesis. The first was the implementation of modularity to ensure a flexibility as a tool for the user; with the introduction of the unit generator, the instrument and the compiler, composers had the building blocks to create an unlimited range of sounds. The second was the impact of this implementation in the modular analog synthesizers developed a few years later. While Jean-Claude Risset, a well known Mathews associate, asserts this, Mathews actually denies it. They both are correct in their perspectives. Introduction Over 76 years have passed since the invention of the first electronic general purpose computer,1 the ENIAC. Today, we carry computers in our pockets that can perform millions of times more calculations per second.2 With the amazing rate of change in computer technology, it's hard to imagine that any development of yesteryear could maintain a semblance of relevance today. However, in the world of music synthesis, the foundations that were laid six decades ago not only spawned a breadth of multifaceted innovation but continue to function as the bedrock of important digital applications used around the world today. Not only did a new modular approach implemented by its creator, Max Mathews, ensure that the MUSIC N lineage would continue to be useful in today’s world (in one of its descendents, Csound) but this approach also likely inspired the analog synthesizer engineers of the day, impacting their designs.
    [Show full text]
  • Bringing Csound to a Modern Production Environment
    Mark Jordan-Kamholz and Dr.Boulanger. Bringing Csound to a Modern Production Environment BRINGING CSOUND TO A MODERN PRODUCTION ENVIRONMENT WITH CSOUND FOR LIVE Mark Jordan-Kamholz mjordankamholz AT berklee.edu Dr. Richard Boulanger rboulanger AT berklee.edu Csound is a powerful and versatile synthesis and signal processing system and yet, it has been far from convenient to use the program in tandem with a modern Digital Audio Workstation (DAW) setup. While it is possible to route MIDI to Csound, and audio from Csound, there has never been a solution that fully integrated Csound into a DAW. Csound for Live attempts to solve this problem by using csound~, Max and Ableton Live. Over the course of this paper, we will discuss how users can use Csound for Live to create Max for Live Devices for their Csound instruments that allow for quick editing via a GUI; tempo-synced and transport-synced operations and automation; the ability to control and receive information from Live via Ableton’s API; and how to save and recall presets. In this paper, the reader will learn best practices for designing devices that leverage both Max and Live, and in the presentation, they will see and hear demonstrations of devices used in a full song, as well as how to integrate the powerful features of Max and Live into a production workflow. 1 Using csound~ in Max and setting up Csound for Live Rendering a unified Csound file with csound~ is the same as playing it with Csound. Sending a start message to csound~ is equivalent to running Csound from the terminal, or pressing play in CsoundQT.
    [Show full text]
  • (CCM)2 - College-Conservatory of Music Center for Computer Music, University of Cincinnati
    (CCM)2 - College-Conservatory of Music Center for Computer Music, University of Cincinnati Mara Helmuth, Director (CCM)2 (1) Michael Barnhart, Carlos Fernandes, Cheekong Ho, Bonnie Miksch (1) Division of Composition, History and Theory - College-Conservatory of Music, University of Cincinnati [email protected] [email protected] http://meowing.ccm.uc.edu Abstract The College-Conservatory of Music Center for Computer Music at the University of Cincinnati is a new environment for computer music composition and research. Composers create sound on powerful systems with signal processing and MIDI, with an excellent listening environment. The technical and aesthetical aspects of computer music can be studied in courses for composers, theorists and performers. Innovative research activity in granular synthesis, live performance interfaces and sound/animation connections is evolving, incorporating collaborative work with faculty in the School of Art. CCM has traditionally had a lively performance environment, and the studios are extending the capabilities of performers in new directions. Computer music concerts and other listening events by students, faculty and visiting composers are becoming regular occurrences. 1 Introduction video cables connect the studios and allow removal of noisy computer CPUs from the listening spaces. Fiber network connections will be installed for internet An amorphous quality based on many aspects of the access and the studio website. The studios will also be environment, including sounds of pieces being created, connected by fiber to a new studio theater, for remote personalities of those who work there, software under processing of sound created and heard in the theater. construction, equipment, layout and room design, and The theater will be equipped with a sixteen-channel architecture creates a unique studio environment.
    [Show full text]