Instruments Numériques Et Performances Musicales: Enjeux

Total Page:16

File Type:pdf, Size:1020Kb

Instruments Numériques Et Performances Musicales: Enjeux Instruments numériques et performances musicales : enjeux ontologiques et esthétiques Madeleine Le Bouteiller To cite this version: Madeleine Le Bouteiller. Instruments numériques et performances musicales : enjeux ontologiques et esthétiques. Musique, musicologie et arts de la scène. Université de Strasbourg, 2020. Français. NNT : 2020STRAC002. tel-02860790 HAL Id: tel-02860790 https://tel.archives-ouvertes.fr/tel-02860790 Submitted on 8 Jun 2020 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. UNIVERSITÉ DE STRASBOURG ÉCOLE DOCTORALE DES HUMANITÉS – ED 520 ACCRA (Approches Contemporaines de la Création et de la Réflexion Artistique) GREAM (Groupe de Recherches Expérimentales sur l’Acte Musical) THÈSE présentée par : Madeleine LE BOUTEILLER soutenue le : 13 janvier 2020 pour obtenir le grade de : Docteur de l’université de Strasbourg Discipline/ Spécialité : Musicologie INSTRUMENTS NUMÉRIQUES ET PERFORMANCES MUSICALES Enjeux ontologiques et esthétiques THÈSE dirigée par : M. Alessandro ARBO Professeur, université de Strasbourg RAPPORTEURS : M. Philippe LE GUERN Professeur, université de Rennes 2 M. Bernard SÈVE Professeur émérite, université de Lille AUTRES MEMBRES DU JURY : Mme Anne SÈDES Professeur, université de Paris 8 M. Thomas PATTESON Professeur, Curtis Institute of Music 2 Remerciements Mes remerciements vont d’abord à Alessandro Arbo, mon directeur de thèse, sans qui ce travail n’aurait pas été possible. Ses conseils avisés et sa rigueur m’ont permis de mener ce projet à son terme. Je remercie aussi Eric Maestri et Anne Sèdes d’avoir accompagné mes recherches en constituant mon comité de suivi de thèse. Merci au groupe des jeunes chercheurs du GREAM, qui de cafés (sans café) en séminaires, rythme le temps et le travail et enrichit de perspectives nouvelles. Ce travail doit beaucoup aux artistes qui m’ont accordé du temps pour me parler de leurs instruments et de leurs expériences musicales. Je souhaite remercier en particulier Anne Hege, qui m’a relaté l’histoire de son rope-instrument, Lainie Fefferman qui m’a renseignée sur les procédés compositionnels de WALO, Chris Douthitt pour les précieuses informations et les documents concernant sa pièce Human Modular, Alyssa Weinberg et sa collègue de scène pour leur témoignage sur la création de Tethered, Tom Mays pour Le Patch Bien Tempéré III. Je remercie aussi Atau Tanaka, Serge de Laubier et Laetitia Sonami pour leur partage d’expérience, respectivement avec la BioMuse, le Méta- Instrument, le Lady’s glove et le Spring Spyre. Je remercie Jeff Snyder ainsi que les musiciens de PLOrk pour leur accueil lors de leurs répétitions, et enfin, Thomas Patteson et Adam Vidiksis pour leur intérêt pour mon travail et des échanges aussi stimulants qu’enrichissants. Merci aussi à tous ceux qui m’ont entourée pendant ces trois années : mes coloc’, mes amis, merci à Alessia, Camille et Elise, acolytes de doctorat, merci aux musiciens strasbourgeois et à ceux de Philadelphie – une gratitude particulièrement chaleureuse pour Espersan et le Bach Collegium. Merci à mes aînés du doctorat, chers physiciens dont l’amitié, même à distance, m’est précieuse. Merci à tous mes proches et à mes lointains, qui à la question « et alors la thèse, ça avance ? » se sont contentés, peut-être malgré eux, d’une absence de (vraie) réponse. Merci à Alexandre et Olivia, Véronique, Clément et Joachim, Pauline et Alex, Amélie, Céline, leur présence joyeuse et leur accueil ont su accompagner les dernières semaines de rédaction. Merci à mes parents qui m’ont toujours encouragée, ainsi qu’à mes frères et sœurs. Et enfin merci à Vangelis pour son soutien inconditionnel, pour sa patience et sa confiance tout au long de ces années. 3 4 Avertissement Toutes les traductions de citations en anglais sont de l’auteur. L’original n’est précisé qu’en cas de vocabulaire idiomatique difficilement traduisible. Les termes techniques sont référencés dans un glossaire situé à la fin du texte. Les mots y figurant sont signalés dans le texte par un astérisque*. Pour ne pas trop encombrer le texte, seules les premières occurrences dans chaque partie seront marquées d’une astérisque. Pour faciliter l’écoute et le visionnage des exemples musicaux, les liens vers les vidéos sont signalés par une numérotation (.23.…) et regroupés sur cette page : https://www.dropbox.com/s/c8p1s2m6fm9wa9m/liens.docx?dl=0 Le mot performance, en français, admet deux significations distinctes : dans la première (plus courante), une performance est une prouesse, un succès ou un exploit accompli par une personne, une machine, une organisation, ou bien le résultat d’une action évaluée (la performance d’un sportif par exemple). Le second sens du mot est un anglicisme : performance en anglais désigne une représentation ou action artistique, elle concerne principalement le théâtre et la musique. On peut aussi utiliser les mots « exécution » ou « interprétation », mais ils sous-entendent qu’un contenu musical déjà existant est transmis lors du jeu de la musique – or, ce n’est pas toujours vrai. Quant au terme « création », il est ambigu : s’il peut évoquer une exécution musicale sans référence à aucune œuvre, il se réfère aussi à l’acte de composition ; il désigne par ailleurs la première représentation d’une œuvre musicale. Dans le texte, nous préférerons donc l’anglicisme « performance », nous l’utiliserons dans le sens de performance musicale. Le performeur est celui qui réalise une performance. 5 6 Introduction « Je joue de l’ordinateur1 » : il est aujourd’hui courant d’entendre ces propos venant d’artistes des scènes musicales numériques*. Ces mots résonnent avec ceux qui titraient un article de Max Mathews pour le magazine Science, quelques quarante années plus tôt : « l’ordinateur numérique comme instrument de musique2 ». En effet, l’ordinateur est aujourd’hui utilisé lors de performances musicales au moyen d’interfaces* gestuelles : ce sont des objets équipés de divers capteurs* de mouvement qui permettent à celui qui les manipule de contrôler des algorithmes* de synthèse sonore* fonctionnant sur un ordinateur. On peut alors « jouer » de la musique numérique de manière gestuelle. Interface, ordinateur et haut-parleurs forment un genre d’instrument bien particulier. Il faut vraisemblablement donner tort à Diderot, qui écrivait dans l’Encyclopédie : « On n’invente plus d'instruments, et il y en a assez d’inventés ; mais je crois qu'il y a beaucoup de découvertes à faire sur leur facture3 ». On invente toujours de nouveaux instruments, et d’autant plus lorsque de nouveaux moyens techniques deviennent accessibles. Mais jusqu’où peut-on aller ? Le philosophe Bernard Sève pointe que « les musiciens se sont emparés d’objets techniques qui sont, quant à eux, le produit de théories scientifiques indépendantes et qui sont primitivement destinés à d’autres usages qu’à “faire de la musique” (magnétophones, ordinateurs), tous ces faits dessinent une pratique de la musique qui se situe au-delà de la frontière organologique traditionnelle4 ». Quant à Claude Cadoz, il ose être plus radical : « l’ordinateur n’est pas un instrument, mais une représentation d’instrument5 », affirmant qu’avec l’ordinateur, il revient à l’homme de « se placer dans une relation “de type instrumentale”, une représentation de relation instrumentale6 ». Emerge ici une contradiction : les ordinateurs et les instruments numériques sont-ils de véritables instruments de musique ? 1. Sergi JORDA, Digital Lutherie: Crafting musical computers for new musics’ performance and improvisation, thèse de doctorat (Universitat Pompeu Fabra, 2005), p. 1. 2. Max V. MATHEWS, « The digital computer as a musical instrument », Science 142, n° 3592 (1963), p. 553-557. 3. Denis DIDEROT, « INSTRUMENS (Musique et Lutherie) », dans Encyclopédie de Diderot [en ligne], Encyclopédie ou Dictionnaire raisonné des sciences, des arts et des métiers, 1772-1751 [http://www.encyclopédie.eu/index.php/logique/929124137-grammaire/1065186196-INSTRUMENT, consulté le 25/07/2019]. 4. Bernard SEVE, L’instrument de musique : une étude philosophique, L’Ordre philosophique (Paris: Seuil, 2013), p. 196. 5. Claude CADOZ, « Musique, geste, technologie », dans Les nouveaux gestes de la musique, éd. par Hugues Genevois et Raphaël de Vivo (Marseille: Parenthèses, 1999), p. 91. 6. CADOZ, « Musique, geste, technologie », p. 91. 7 La question se pose lorsque l’on constate que la relation du musicien à la production sonore est toute différente. Contrairement à un instrument acoustique, où les gestes du musicien servent à mettre en vibration cordes, membranes ou colonnes d’air, exigeant du musicien qu’il engage son énergie physique via ses mains, ses bras, son souffle ou même ses pieds7, il n’y a pas de lien direct entre un geste et le son résultant. Avec un instrument numérique, tout est codé dans un logiciel*. Le signal sonore*, que l’on peut contrôler grâce des interfaces, est généré de manière informatique : entre le geste et le son, il y a le logiciel. Les techniques musicales et les instruments ont évolué au cours des siècles, au rythme des avancées technologiques. Bien avant l’arrivée du numérique, dès la fin du XIXème siècle, on est capable de transformer des ondes sonores en ondes de courant électrique, on invente le téléphone et on sait enregistrer la musique sur des cylindres de cire8. Aux débuts de l’électronique arrivent les premiers instruments électroniques. Ainsi le célèbre thérémine, construit par Lev Sergueïevitch Termen en 1919. Les mains du musicien flottent dans les airs, leurs positions contrôlent respectivement la hauteur des notes et le volume. Le son, produit par des oscillateurs électroniques*, est modulé par les gestes de l’instrumentiste9.
Recommended publications
  • Implementing Stochastic Synthesis for Supercollider and Iphone
    Implementing stochastic synthesis for SuperCollider and iPhone Nick Collins Department of Informatics, University of Sussex, UK N [dot] Collins ]at[ sussex [dot] ac [dot] uk - http://www.cogs.susx.ac.uk/users/nc81/index.html Proceedings of the Xenakis International Symposium Southbank Centre, London, 1-3 April 2011 - www.gold.ac.uk/ccmc/xenakis-international-symposium This article reflects on Xenakis' contribution to sound synthesis, and explores practical tools for music making touched by his ideas on stochastic waveform generation. Implementations of the GENDYN algorithm for the SuperCollider audio programming language and in an iPhone app will be discussed. Some technical specifics will be reported without overburdening the exposition, including original directions in computer music research inspired by his ideas. The mass exposure of the iGendyn iPhone app in particular has provided a chance to reach a wider audience. Stochastic construction in music can apply at many timescales, and Xenakis was intrigued by the possibility of compositional unification through simultaneous engagement at multiple levels. In General Dynamic Stochastic Synthesis Xenakis found a potent way to extend stochastic music to the sample level in digital sound synthesis (Xenakis 1992, Serra 1993, Roads 1996, Hoffmann 2000, Harley 2004, Brown 2005, Luque 2006, Collins 2008, Luque 2009). In the central algorithm, samples are specified as a result of breakpoint interpolation synthesis (Roads 1996), where breakpoint positions in time and amplitude are subject to probabilistic perturbation. Random walks (up to second order) are followed with respect to various probability distributions for perturbation size. Figure 1 illustrates this for a single breakpoint; a full GENDYN implementation would allow a set of breakpoints, with each breakpoint in the set updated by individual perturbations each cycle.
    [Show full text]
  • Eastman Computer Music Center (ECMC)
    Upcoming ECMC25 Concerts Thursday, March 22 Music of Mario Davidovsky, JoAnn Kuchera-Morin, Allan Schindler, and ECMC composers 8:00 pm, Memorial Art Gallery, 500 University Avenue Saturday, April 14 Contemporary Organ Music Festival with the Eastman Organ Department & College Music Department Steve Everett, Ron Nagorcka, and René Uijlenhoet, guest composers 5:00 p.m. + 7:15 p.m., Interfaith Chapel, University of Rochester Eastman Computer Wednesday, May 2 Music Cente r (ECMC) New carillon works by David Wessel and Stephen Rush th with the College Music Department 25 Anniversa ry Series 12:00 pm, Eastman Quadrangle (outdoor venue), University of Rochester admission to all concerts is free Curtis Roads & Craig Harris, ecmc.rochester.edu guest composers B rian O’Reilly, video artist Thursday, March 8, 2007 Kilbourn Hall fire exits are located along the right A fully accessible restroom is located on the main and left sides, and at the back of the hall. Eastman floor of the Eastman School of Music. Our ushers 8:00 p.m. Theatre fire exits are located throughout the will be happy to direct you to this facility. Theatre along the right and left sides, and at the Kilbourn Hall back of the orchestra, mezzanine, and balcony Supporting the Eastman School of Music: levels. In the event of an emergency, you will be We at the Eastman School of Music are grateful for notified by the stage manager. the generous contributions made by friends, If notified, please move in a calm and orderly parents, and alumni, as well as local and national fashion to the nearest exit.
    [Show full text]
  • CD Program Notes
    CD Program Notes Contents tique du CEMAMu), the graphic mu- the final tape version was done at sic computer system developed by the Stichting Klankscap and the Xenakis and his team of engineers at Sweelinck Conservatory in Am- Part One: James Harley, Curator the Centre d’Etudes de Mathe´ma- sterdam. I would like to thank tique et Automatique Musicales (CE- Paul Berg and Floris van Manen for MAMu). The piece by Cort Lippe, their support in creating this work. completed in 1986, is mixed, for bass Sound material for the tape was Curator’s Note clarinet and sounds created on the limited to approximate the con- It probably would not be too far computer. My own piece, from 1987, fines one normally associates with wrong to state that most composers is also mixed, for flute and tape; in individual acoustic instruments in of the past 40 years have been influ- the excerpt presented here, the pre- order to create a somewhat equal enced in one way or another by the produced material, a mixture of re- relationship between the tape and music or ideas of Iannis Xenakis. corded flute and computer-generated the bass clarinet. Although con- The composers presented on this sounds, is heard without the live trasts and similarities between the disc have been touched more directly flute part. Richard Barrett’s piece tape and the instrument are evi- than most. It was my intention to was entirely produced on the UPIC. dent, musically a kind of intimacy gather a disparate collection of music The sound-world Xenakis created was sought, not unlike our to show the extent to which that in- in his music owes much to natural present-day ‘‘sense’’ of intimacy fluence has crossed styles.
    [Show full text]
  • A Tabletop Waveform Editor for Live Performance
    Proceedings of the 2008 Conference on New Interfaces for Musical Expression (NIME08), Genova, Italy A tabletop waveform editor for live performance Gerard Roma Anna Xambo´ Music Technology Group Music Technology Group Universitat Pompeu Fabra Universitat Pompeu Fabra [email protected] [email protected] ABSTRACT We present an audio waveform editor that can be oper- ated in real time through a tabletop interface. The system combines multi-touch and tangible interaction techniques in order to implement the metaphor of a toolkit that allows di- rect manipulation of a sound sample. The resulting instru- ment is well suited for live performance based on evolving loops. Keywords tangible interface, tabletop interface, musical performance, interaction techniques 1. INTRODUCTION The user interface of audio editors has changed relatively little over time. The standard interaction model is centered on the waveform display, allowing the user to select portions Figure 1: Close view of the waveTable prototype. of the waveform along the horizontal axis and execute com- mands that operate on those selections. This model is not very different to that of the word processor, and its basics Barcelona 2005) which inspired this project. The use of a are usually understood by computer users even with little standard audio editor on a laptop computer as a sophisti- or no experience in specialized audio tools. As a graphical cated looper served the performers’ minimalist glitch aes- representation of sound, the waveform is already familiar thetics. Perhaps more importantly, the projected waveform for many people approaching computers for audio and mu- provided a visual cue that helped the audience follow the sic composition.
    [Show full text]
  • Experiments in Sound and Electronic Music in Koenig Books Isbn 978-3-86560-706-5 Early 20Th Century Russia · Andrey Smirnov
    SOUND IN Z Russia, 1917 — a time of complex political upheaval that resulted in the demise of the Russian monarchy and seemingly offered great prospects for a new dawn of art and science. Inspired by revolutionary ideas, artists and enthusiasts developed innumerable musical and audio inventions, instruments and ideas often long ahead of their time – a culture that was to be SOUND IN Z cut off in its prime as it collided with the totalitarian state of the 1930s. Smirnov’s account of the period offers an engaging introduction to some of the key figures and their work, including Arseny Avraamov’s open-air performance of 1922 featuring the Caspian flotilla, artillery guns, hydroplanes and all the town’s factory sirens; Solomon Nikritin’s Projection Theatre; Alexei Gastev, the polymath who coined the term ‘bio-mechanics’; pioneering film maker Dziga Vertov, director of the Laboratory of Hearing and the Symphony of Noises; and Vladimir Popov, ANDREY SMIRNO the pioneer of Noise and inventor of Sound Machines. Shedding new light on better-known figures such as Leon Theremin (inventor of the world’s first electronic musical instrument, the Theremin), the publication also investigates the work of a number of pioneers of electronic sound tracks using ‘graphical sound’ techniques, such as Mikhail Tsekhanovsky, Nikolai Voinov, Evgeny Sholpo and Boris Yankovsky. From V eavesdropping on pianists to the 23-string electric guitar, microtonal music to the story of the man imprisoned for pentatonic research, Noise Orchestras to Machine Worshippers, Sound in Z documents an extraordinary and largely forgotten chapter in the history of music and audio technology.
    [Show full text]
  • Processes and Systems in Computer Music
    Processes Processes and systems and systems in computer in computer music; from music; from meta to matter meta to matter DR TONY MYATT DR TONY MYATT 05 2 The use of the terms ‘system’ and ‘process’ to generate works of music are often applied to the output of composers such as Steve Reich or Philip Glass, and also to Serial composers [1]2 from Schoenberg to Stockhausen, or to the works of experimentalists like John Cage. This may be where these terms are most clearly expressed, in either the material of the work or in the discourse that surrounds it, but few composers would not claim that systematic processes lie at the root of their work, in the methods they use to generate, manipulate, or control sound. The grand narrative of twentieth century classical music to move away from the tonal system of harmony has prompted experimental composers to push at the boundaries of music to discover approaches that create original musics. With the advent of the computer age and the use of computers to control and generate sound, it appeared to many that it was inevitable this new technology might take forward the concept of music. Computers, as we will see, were initially employed as tools that sustained traditional concepts of music, but as Cage hinted in 1959, and maybe for the same reason, perhaps this is now changing. In this essay, I will discuss how these early approaches to systematic process developed a canon of computer music, and in particular algorithmic systems and processes that reflected an underlying vision of what computer technologies were thought to hold in store, from a utopian and rationalist perspective.
    [Show full text]
  • A Festival of Unexpected New Music February 28March 1St, 2014 Sfjazz Center
    SFJAZZ CENTER SFJAZZ MINDS OTHER OTHER 19 MARCH 1ST, 2014 1ST, MARCH A FESTIVAL FEBRUARY 28 FEBRUARY OF UNEXPECTED NEW MUSIC Find Left of the Dial in print or online at sfbg.com WELCOME A FESTIVAL OF UNEXPECTED TO OTHER MINDS 19 NEW MUSIC The 19th Other Minds Festival is 2 Message from the Executive & Artistic Director presented by Other Minds in association 4 Exhibition & Silent Auction with the Djerassi Resident Artists Program and SFJazz Center 11 Opening Night Gala 13 Concert 1 All festival concerts take place in Robert N. Miner Auditorium in the new SFJAZZ Center. 14 Concert 1 Program Notes Congratulations to Randall Kline and SFJAZZ 17 Concert 2 on the successful launch of their new home 19 Concert 2 Program Notes venue. This year, for the fi rst time, the Other Minds Festival focuses exclusively on compos- 20 Other Minds 18 Performers ers from Northern California. 26 Other Minds 18 Composers 35 About Other Minds 36 Festival Supporters 40 About The Festival This booklet © 2014 Other Minds. All rights reserved. Thanks to Adah Bakalinsky for underwriting the printing of our OM 19 program booklet. MESSAGE FROM THE ARTISTIC DIRECTOR WELCOME TO OTHER MINDS 19 Ever since the dawn of “modern music” in the U.S., the San Francisco Bay Area has been a leading force in exploring new territory. In 1914 it was Henry Cowell leading the way with his tone clusters and strumming directly on the strings of the concert grand, then his students Lou Harrison and John Cage in the 30s with their percussion revolution, and the protégés of Robert Erickson in the Fifties with their focus on graphic scores and improvisation, and the SF Tape Music Center’s live electronic pioneers Subotnick, Oliveros, Sender, and others in the Sixties, alongside Terry Riley, Steve Reich and La Monte Young and their new minimalism.
    [Show full text]
  • Post-Xenakians” Interactive, Collaborative, Explorative, Visual, Immediate: Alberto De Campo, Russell Haswell, Florian Hecker
    Forum “Post-Xenakians” Interactive, Collaborative, Explorative, Visual, Immediate: Alberto de Campo, Russell Haswell, Florian Hecker Peter Hoffmann, Germany, [email protected] In Makis Solomos, (ed.), Proceedings of the international Symposium Xenakis. La Musique Électroacoustique / Xenakis. The Electroacoustic Music (Université Paris 8, May 2012) “A noise-ician should never rehearse.” [Toshiji Mikawa, quoted by Russell Haswell] Abstract: As a follow-up to a previous panel called “Post-Xenakian Applications” which was held at the International Xenakis Symposium 2011 at Goldsmiths in London, I presented three artists whose work was simply missing in my London talk. I was happy to be able to have a panel discussion with all the three of them, two of them visiting the Paris conference, one (Florian Hecker) present via Skype from Boston. Instead of a transcript of the panel discussion (which was not recorded), I give here my overview of their work as related to the conference’s subject. This is an elaboration of the slide show which was presented to introduce the subject and the people and to support our discussion with music and visual examples. You can find the examples on the slides which are published alongside with this paper. Introduction At the International Xenakis Conference last year 2011 in London, I gave a lecture about “Post-Xenakian Applications” [Hoffmann2011b]. I presented a selection of a dozen artists as “post-Xenakians”. What does it mean to be a “post-Xenakian”? Well, it means that one must be younger than Xenakis and be inspired by his work. In our case, inspired by his work in electroacoustics, since I will present to you three electroacoustic composers.
    [Show full text]
  • To Graphic Notation Today from Xenakis’S Upic to Graphic Notation Today
    FROM XENAKIS’S UPIC TO GRAPHIC NOTATION TODAY FROM XENAKIS’S UPIC TO GRAPHIC NOTATION TODAY FROM XENAKIS’S UPIC TO GRAPHIC NOTATION TODAY PREFACES 18 PETER WEIBEL 24 LUDGER BRÜMMER 36 SHARON KANACH THE UPIC: 94 ANDREY SMIRNOV HISTORY, UPIC’S PRECURSORS INSTITUTIONS, AND 118 GUY MÉDIGUE IMPLICATIONS THE EARLY DAYS OF THE UPIC 142 ALAIN DESPRÉS THE UPIC: TOWARDS A PEDAGOGY OF CREATIVITY 160 RUDOLF FRISIUS THE UPIC―EXPERIMENTAL MUSIC PEDAGOGY― IANNIS XENAKIS 184 GERARD PAPE COMPOSING WITH SOUND AT LES ATELIERS UPIC/CCMIX 200 HUGUES GENEVOIS ONE MACHINE— TWO NON-PROFIT STRUCTURES 216 CYRILLE DELHAYE CENTRE IANNIS XENAKIS: MILESTONES AND CHALLENGES 232 KATERINA TSIOUKRA ESTABLISHING A XENAKIS CENTER IN GREECE: THE UPIC AT KSYME-CMRC 246 DIMITRIS KAMAROTOS THE UPIC IN GREECE: TEN YEARS OF LIVING AND CREATING WITH THE UPIC AT KSYME 290 RODOLPHE BOUROTTE PROBABILITIES, DRAWING, AND SOUND TABLE SYNTHESIS: THE MISSING LINK OF CONTENTS COMPOSERS 312 JULIO ESTRADA THE UPIC 528 KIYOSHI FURUKAWA EXPERIENCING THE LISTENING HAND AND THE UPIC AND UTOPIA THE UPIC UTOPIA 336 RICHARD BARRETT 540 CHIKASHI MIYAMA MEMORIES OF THE UPIC: 1989–2019 THE UPIC 2019 354 FRANÇOIS-BERNARD MÂCHE 562 VICTORIA SIMON THE UPIC UPSIDE DOWN UNFLATTERING SOUNDS: PARADIGMS OF INTERACTIVITY IN TACTILE INTERFACES FOR 380 TAKEHITO SHIMAZU SOUND PRODUCTION THE UPIC FOR A JAPANESE COMPOSER 574 JULIAN SCORDATO 396 BRIGITTE CONDORCET (ROBINDORÉ) NOVEL PERSPECTIVES FOR GRAPHIC BEYOND THE CONTINUUM: NOTATION IN IANNIX THE UNDISCOVERED TERRAINS OF THE UPIC 590 KOSMAS GIANNOUTAKIS EXPLORING
    [Show full text]
  • Idiosyncratic Software for Sketching Spectral Music A
    UNIVERSITY OF CALIFORNIA, SAN DIEGO AudioChalk: Idiosyncratic Software For Sketching Spectral Music A Thesis submitted in partial satisfaction of the requirements for the degree Master of Arts in Music by Elliot M. Patros Committee in charge: Professor Tamara Smyth, Chair Professor Tom Erbe Professor Miller Puckette 2015 Copyright Elliot M. Patros, 2015 All rights reserved. The Thesis of Elliot M. Patros is approved, and it is accept- able in quality and form for publication on microfilm and electronically: Chair University of California, San Diego 2015 iii TABLE OF CONTENTS Signature Page . iii Table of Contents . iv List of Figures . vi List of Tables . vii Acknowledgements . viii Abstract of the Thesis . ix Chapter 1 Introduction . 1 1.1 Objectives For Software Developers . 3 1.2 Paradigms Of Experimental Electronic Composition . 4 Chapter 2 AudioChalk . 6 2.1 The Main Window . 7 2.2 Generating Tracks With Drawing Tools . 9 2.2.1 Anatomy Of A Track . 9 2.2.2 The Line Tool . 10 2.2.3 The Pencil Tool . 11 2.2.4 Drawing And Seeing Loudness . 11 2.3 Generating Tracks By Importing Audio . 14 2.3.1 Getting Tracks From Audio Files . 15 2.3.2 Rapidly Auditioning Imported Audio . 19 2.4 Track Editing Tools . 21 2.4.1 The Selector Tool . 21 2.4.2 Track Groups . 22 2.4.3 Linear Transform . 24 2.4.4 Time Stretch . 25 2.4.5 Pitch Shift . 26 2.4.6 Duplicate . 26 2.4.7 Swappable Preferences Window . 28 2.5 Rendering Audio . 29 2.6 AudioChalk: Conclusion .
    [Show full text]
  • Software Documentation V
    software documentation v. 0.9.20 This document is based on: Scordato, J. (2016). An introduction to IanniX graphical sequencer. Master's Thesis in Sound Art, Universitat de Barcelona. © 2016-2017 Julian Scordato Index 1. What is IanniX? p. 1 1.1 Main applications p. 2 1.2 Classification of IanniX scores p. 3 2. User interface p. 5 2.1 Main window p. 5 2.1.1 Visualization p. 6 2.1.2 Creation p. 7 2.1.3 Position p. 8 2.1.4 Inspection p. 9 2.2 Secondary windows p. 10 2.2.1 Script editor p. 11 3. IanniX objects p. 15 3.1 Curves p. 16 3.2 Cursors p. 17 3.3 Triggers p. 19 4. Control and communication with other devices p. 21 4.1 Sent messages p. 21 4.1.1 Cursor-related variables p. 22 4.1.2 Trigger-related variables p. 24 4.1.3 Transport variables p. 25 4.2 Received commands p. 25 4.2.1 Score instances and viewport p. 25 4.2.2 Object instances and identification p. 26 4.2.3 Object positioning p. 27 4.2.4 Object appearance p. 28 4.2.5 Object behavior p. 29 4.2.6 Sequencing p. 30 4.3 Network protocols p. 31 4.3.1 OSC p. 31 4.3.2 Raw UDP p. 33 4.3.3 TCP and WebSocket p. 34 4.3.4 HTTP p. 35 4.4 MIDI interface p. 36 4.5 Serial interface p. 37 4.6 Software interfaces p.
    [Show full text]
  • A Tutorial on Spectral Sound Processing Using Max/MSP and Jitter
    Jean-Franc¸ois Charles A Tutorial on Spectral 1 Route de Plampery´ 74470 Vailly, France [email protected] Sound Processing Using Max/MSP and Jitter For computer musicians, sound processing in the This article is intended as both a presentation of frequency domain is an important and widely used the potential of manipulating spectral sound data technique. Two particular frequency-domain tools as matrices and a tutorial for musicians who want of great importance for the composer are the phase to implement such effects in the Max/MSP/Jitter vocoder and the sonogram. The phase vocoder, an environment. Throughout the article, I consider analysis-resynthesis tool based on a sequence of spectral analysis and synthesis as realized by the overlapping short-time Fourier transforms, helps Fast Fourier Transform (FFT) and Inverse-FFT perform a variety of sound modifications, from time algorithms. I assume a familiarity with the FFT stretching to arbitrary control of energy distribution (Roads 1995) and the phase vocoder (Dolson 1986). through frequency space. The sonogram, a graphical To make the most of the examples, a familiarity representation of a sound’s spectrum, offers com- with the Max/MSP environment is necessary, and posers more readable frequency information than a a basic knowledge of the Jitter extension may be time-domain waveform. helpful. Such tools make graphical sound synthesis I begin with a survey of the software currently convenient. A history of graphical sound synthesis is available for working in this domain. I then show beyond the scope of this article, but a few important some improvements to the traditional phase vocoder figures include Evgeny Murzin, Percy Grainger, and used in both real time and performance time.
    [Show full text]