Phd Thesis, University of Helsinki, Department of Computer Science, November 2000

Total Page:16

File Type:pdf, Size:1020Kb

Phd Thesis, University of Helsinki, Department of Computer Science, November 2000 5 Quantisation hythm together with melody is one of the basic elements in music. According to Longuet-Higgins ([LH76]) human listeners are much more sensitive to the perception of rhythm than to the R perception of melody or pitch related features. Usually it is easier for trained and untrained listeners as well as for musicians to transcribe a rhythm, than details about heard intervals, harmonic changes or absolute pitch. “The term rhythm refers to the general sense of movement in time that characterizes our experience of music (Apel, 1972). Rhythm often refers to the organization of events in time, such that they combine perceptually into groups or induce a sense of meter (Cooper & Meyer, 1960; Lerdahl & Jackendoff, 1983). In this sense, rhythm is not an objective property of music, it is an experience that has both objective and subjective components. The experience of rhythm arises from the interaction of the various materials of music – pitch, intensity, timbre, and so forth – within the individual listener.” [LK94] In general the task of transcribing the rhythm of a performance is different from the previously described beat tracking or tempo detection issue. For estimating a tempo profile it is sufficient to infer score time positions for a set of dedicated anchor notes (i.e., beats respectively clicks), for the rhythm transcription task score time positions and durations for all performance notes must be inferred. Because the possible time positions in a score are specified as fractions using a rather small set of possible denominators, a score represents a grid of discrete time positions. The resolution of the grid is context and style dependent, common scores often uses only a resolution of 1/16th notes but higher resolutions (in most cases binary) or non-standard resolutions for arbitrary tuplets can always occur and might also be correct. In the context of computer aided transcription the process of rhythm transcription is called (musical) quantisation. It is equivalent to transferring timing information from a continuous domain (absolute performance timing in seconds) into a discrete domain (metrical score timing). Depending on the way an input file was created (e.g., free performance recording, recording synchronous to a MIDI click, mechanical performance) a tempo detection must be performed before the quantisation; for quantisation a given tempo profile is required. Because the tempo detection module might create a tempo profile that indicates only the score time positions for anchor notes (e.g., start of a measure or down beats) the score time information for notes between these anchor points can still be imprecise because of slight tempo fluctuations between two beats. These errors need now to be removed by the quantisation module. For example, an unquantised onset time at a score time position of 191/192 needs to be quantised to a reasonable grid position (e.g., quarter or eighth notes) to be displayable in conventional music notation. By choosing a too high resolution for the conversion from performance timing into score timing (e.g., 1/128 of a whole note), the resulting scores will be correct displayable but very hard to read because of complex note durations. If the resolution is chosen too low (e.g., a quarter note) the resulting score will contain overlap errors (i.e., notes that did not overlap in the performance have overlapping durations in the score) and it will be very inaccurate compared to the performance data. An human transcriber would here – if possible – prefer ‘simple’ transcriptions over complex solutions. But he would intuitively detect the point where a correct, ‘simple’ transcription is not possible and a more complex solution needs to be chosen. For example, five equal spaced notes cannot be transcribed as eighth notes if they were played all during a single half note, here a more complex quintuplet needs to be chosen. Musical quantisation can be divided into onset time quantisation and duration quantisation. It depends on the specific approach, if the durations of notes are treated and quantised as a duration or if they become quantised indirectly by quantising the respective offset time positions. In commercial sequencer 106 CHAPTER 5. QUANTISATION applications often also a quantisation approach – usually called swing- or groove-quantisation – is im- plemented. This type of quantisation is used for creating ‘performance-like’ MIDI files from quantised (mechanical) scores or from very inaccurate performance data. This performance simulation approach is not in the scope of this thesis and will therefore not be discussed in the following. Similar to the already discussed issue of tempo detection and the area of voice separation the basic problem for quantisation is indicated by the fact that the translation from score to performance and also the inverse task of transcription are highly ambiguous relations and no injective functions. A single score can result in different but ‘correct’ performances (including different tempo profiles), and a single (non-mechanical) performance can be transcribed correctly by different scores of different complexity. If we, for example, assume that a rhythmically complex performance is a mechanical performance (e.g., played by a notation software system), then its rhythmical complexity should be preserve in a transcription. If we would assume that this performance was created by an untrained, human performer, a much simpler transcription should be preferred, because the complexity of the performance has likely be caused only by inexact playing. An untrained performer would probably not be able to play a complex piece very precisely. An human transcriber would here intuitively evaluate the performance quality of the input data. If the whole piece was played with poor accuracy, he would allow more deviations between the written and the performed data as for a very exact played performance. This means that if a piece is played with low accuracy the grid size for the quantisation will be increased intuitively by the transcriber. As shown in Chapter 4 the quality or accurateness of a performance can be inferred algorithmically from the performance data itself before starting tempo detection or quantisation. Different from other known approaches our system uses this accuracy information and will adapt some resolution and search window parameters. As already shown in previous sections (e.g., Chapter 4), the position of the onset time of a note is usually much closer to the timing as specified in the score than the timing of the performed duration of a note. In [DH89] citation of [Pov77]: “. the note durations in performance can deviate by up to 50 percent from their metrical value.” Therefore the expected and allowed amount of inaccuracy (i.e., deviation between observed performance and inferred score) is higher for the duration of a note than for its onset time. As shown in [Cam00b] and in own experiments the output quality of a quantisation algorithm also depends on a correct voice separation (see Chapter 2). If because of an incorrect voice separation the correct local context of a note is not available (previous, or successive notes have been attached to other voice streams) the onset time position and even more the duration of this note might be quantised incorrect. Figure 5.1 shows a correct quantisation for the last note in the first measure (note c”) and a wrong quantisation if this note is attached incorrectly to the voice stream of the upper voice. correct wrong wrongcorrect E –EÚ– Ú –EÚ–EÚ E 4 4–EÚ– Ú –– –EÚ– Ú –XÚ– Ú a a¥¥ –– & 44 –– ‹‹ ‹‹‹‹ –– ‹‹ J–– \ —— —— —— 4 4—— —— —— XÛ—XÛ— w —— —— —— ww & 44 EÛ—Û—E XÛ—XÛ— XÛXÛ. ‹ ‹ ‹‹‹ ‹EÛ—Û— XÛ—XÛ— XÛXÛ ‹ ‹ \ \ ««« Figure 5.1: Possible effect of wrong voice separation to quantisation. In the remainder of this chapter we first give an overview on existing quantisation approaches and then describe an approach developed in the context of this thesis that consists of a pattern-based part and a quantisation model which evaluates the local context of notes. All described approaches work on a single voice stream (including chords, but no overlapping notes) as input data. 5.1. EXISTING APPROACHES 107 5.1 Existing Approaches In the following we give an introduction to musical quantisation. First we show and formally define the principles of the simplest type of quantisation called grid quantisation. Then we discuss the details of some more advanced approaches known from literature. 5.1.1 Grid Quantisation This simplest form of musical quantisation maps (shifts) all performance time information to the closest time position of a fixed, equidistant grid of quantised score positions. For a performance time position t always the closest grid time position is chosen as quantised time position. This method – also called hard quantisation – is easy to calculate and implemented in most commercial sequencer and notation software products. Given an arbitrary, unquantised score time position t and an ordered set of equidistant score time grid positions G with G = { g1, . , g|G| | gi = (i − 1) · r, i ∈ N }, (5.1) where r is the grid resolution given as a beat duration in score time units. For an arbitrary unquantised time position t and a given grid G (with g|G| ≥ t), a quantisation function tqpos returning the quantised time position for t can be defined as tqpos (t, G) = arg min{|g − t|}. (5.2) g∈G If |gi − t| = |gi+1 − t| (because t = (n + 1/2) · r) the earlier grid position gi should be returned as result of tqpos (t, G). In general, depending on the actual implementation, also other selection strategies are possible (e.g., latest grid position, avoid collisions between notes). This approach requires that the grid size r needs to be selected in advance (e.g., quavers) and kept fixed for the range of notes to which tqpos should be applied.
Recommended publications
  • An App to Find Useful Glissandos on a Pedal Harp by Bill Ooms
    HarpPedals An app to find useful glissandos on a pedal harp by Bill Ooms Introduction: For many years, harpists have relied on the excellent book “A Harpist’s Survival Guide to Glisses” by Kathy Bundock Moore. But if you are like me, you don’t always carry all of your books with you. Many of us now keep our music on an iPad®, so it would be convenient to have this information readily available on our tablet. For those of us who don’t use a tablet for our music, we may at least have an iPhone®1 with us. The goal of this app is to provide a quick and easy way to find the various pedal settings for commonly used glissandos in any key. Additionally, it would be nice to find pedal positions that produce a gliss for common chords (when possible). Device requirements: The application requires an iPhone or iPad running iOS 11.0 or higher. Devices with smaller screens will not provide enough space. The following devices are recommended: iPhone 7, 7Plus, 8, 8Plus, X iPad 9.7-inch, 10.5-inch, 12.9-inch Set the pedals with your finger: In the upper window, you can set the pedal positions by tapping the upper, middle, or lower position for flat, natural, and sharp pedal position. The scale or chord represented by the pedal position is shown in the list below the pedals (to the right side). Many pedal positions do not form a chord or scale, so this window may be blank. Often, there are several alternate combinations that can give the same notes.
    [Show full text]
  • Recognizing Key Signatures Music Fundamentals 14-119-T
    Recognizing Key Signatures Music Fundamentals 14-119-T Constructing the Key Signature: In the last lecture, we connected scales and tonality. We also found that in order to alleviate the labor of writing accidentals, key signatures Figure 1 were created. To review, key signatures are placed immediately after a clef, and they indicate what accidentals are needed throughout a composition, thus those accidentals would not have to be written next to a note in the music [see Figure 1]. Any note that does not have an accidental is considered natural. Key signatures will have either all sharps or all flats. Key signatures with both sharps and flats do not exist. Figure 2 Key Signatures 2 - page 1 The Circle of 5ths: One of the easiest ways of recognizing key signatures is by using the circle of 5ths [see Figure 2]. To begin, we must simply memorize the key signature without any flats or sharps. For a major key, this is C-major and for a minor key, A-minor. After that, we can figure out the key signature by following the diagram in Figure 2. By adding one sharp, the key signature moves up a perfect 5th from what preceded it. Therefore, since C-major has not sharps or flats, by adding one sharp to the key signa- ture, we find G-major (G is a perfect 5th above C). You can see this by moving clockwise around the circle of 5ths. For key signatures with flats, we move counter-clockwise around the circle. Since we are moving “backwards,” it makes sense that by adding one flat, the key signature is a perfect 5th below from what preceded it.
    [Show full text]
  • Musical Notation Codes Index
    Music Notation - www.music-notation.info - Copyright 1997-2019, Gerd Castan Musical notation codes Index xml ascii binary 1. MidiXML 1. PDF used as music notation 1. General information format 2. Apple GarageBand Format 2. MIDI (.band) 2. DARMS 3. QuickScore Elite file format 3. SMDL 3. GUIDO Music Notation (.qsd) Language 4. MPEG4-SMR 4. WAV audio file format (.wav) 4. abc 5. MNML - The Musical Notation 5. MP3 audio file format (.mp3) Markup Language 5. MusiXTeX, MusicTeX, MuTeX... 6. WMA audio file format (.wma) 6. MusicML 6. **kern (.krn) 7. MusicWrite file format (.mwk) 7. MHTML 7. **Hildegard 8. Overture file format (.ove) 8. MML: Music Markup Language 8. **koto 9. ScoreWriter file format (.scw) 9. Theta: Tonal Harmony 9. **bol Exploration and Tutorial Assistent 10. Copyist file format (.CP6 and 10. Musedata format (.md) .CP4) 10. ScoreML 11. LilyPond 11. Rich MIDI Tablature format - 11. JScoreML RMTF 12. Philip's Music Writer (PMW) 12. eXtensible Score Language 12. Creative Music File Format (XScore) 13. TexTab 13. Sibelius Plugin Interface 13. MusiXML: My own format 14. Mup music publication program 14. Finale Plugin Interface 14. MusicXML (.mxl, .xml) 15. NoteEdit 15. Internal format of Finale (.mus) 15. MusiqueXML 16. Liszt: The SharpEye OMR 16. XMF - eXtensible Music 16. GUIDO XML engine output file format Format 17. WEDELMUSIC 17. Drum Tab 17. NIFF 18. ChordML 18. Enigma Transportable Format 18. Internal format of Capella (ETF) (.cap) 19. ChordQL 19. CMN: Common Music 19. SASL: Simple Audio Score 20. NeumesXML Notation Language 21. MEI 20. OMNL: Open Music Notation 20.
    [Show full text]
  • Music Braille Code, 2015
    MUSIC BRAILLE CODE, 2015 Developed Under the Sponsorship of the BRAILLE AUTHORITY OF NORTH AMERICA Published by The Braille Authority of North America ©2016 by the Braille Authority of North America All rights reserved. This material may be duplicated but not altered or sold. ISBN: 978-0-9859473-6-1 (Print) ISBN: 978-0-9859473-7-8 (Braille) Printed by the American Printing House for the Blind. Copies may be purchased from: American Printing House for the Blind 1839 Frankfort Avenue Louisville, Kentucky 40206-3148 502-895-2405 • 800-223-1839 www.aph.org [email protected] Catalog Number: 7-09651-01 The mission and purpose of The Braille Authority of North America are to assure literacy for tactile readers through the standardization of braille and/or tactile graphics. BANA promotes and facilitates the use, teaching, and production of braille. It publishes rules, interprets, and renders opinions pertaining to braille in all existing codes. It deals with codes now in existence or to be developed in the future, in collaboration with other countries using English braille. In exercising its function and authority, BANA considers the effects of its decisions on other existing braille codes and formats, the ease of production by various methods, and acceptability to readers. For more information and resources, visit www.brailleauthority.org. ii BANA Music Technical Committee, 2015 Lawrence R. Smith, Chairman Karin Auckenthaler Gilbert Busch Karen Gearreald Dan Geminder Beverly McKenney Harvey Miller Tom Ridgeway Other Contributors Christina Davidson, BANA Music Technical Committee Consultant Richard Taesch, BANA Music Technical Committee Consultant Roger Firman, International Consultant Ruth Rozen, BANA Board Liaison iii TABLE OF CONTENTS ACKNOWLEDGMENTS ..............................................................
    [Show full text]
  • The Latex Graphics Companion / Michel Goossens
    i i “tlgc2” — 2007/6/15 — 15:36 — page iii — #3 i i The LATEXGraphics Companion Second Edition Michel Goossens Frank Mittelbach Sebastian Rahtz Denis Roegel Herbert Voß Upper Saddle River, NJ • Boston • Indianapolis • San Francisco New York • Toronto • Montreal • London • Munich • Paris • Madrid Capetown • Sydney • Tokyo • Singapore • Mexico City i i i i i i “tlgc2” — 2007/6/15 — 15:36 — page iv — #4 i i Many of the designations used by manufacturers and sellers to distinguish their products are claimed as trademarks. Where those designations appear in this book, and Addison-Wesley was aware of a trademark claim, the designations have been printed with initial capital letters or in all capitals. The authors and publisher have taken care in the preparation of this book, but make no expressed or implied warranty of any kind and assume no responsibility for errors or omissions. No liability is assumed for incidental or consequential damages in connection with or arising out of the use of the information or programs contained herein. The publisher offers discounts on this book when ordered in quantity for bulk purchases and special sales. For more information, please contact: U.S. Corporate and Government Sales (800) 382-3419 [email protected] For sales outside of the United States, please contact: International Sales [email protected] Visit Addison-Wesley on the Web: www.awprofessional.com Library of Congress Cataloging-in-Publication Data The LaTeX Graphics companion / Michel Goossens ... [et al.]. -- 2nd ed. p. cm. Includes bibliographical references and index. ISBN 978-0-321-50892-8 (pbk. : alk. paper) 1.
    [Show full text]
  • Universi^ Micfdmlms International
    INFORMATION TO USERS This reproduction was made from a copy of a document sent to us for microfilming. While the most advanced technology has been used to photograph and reproduce this document, the quality of the reproduction is heavily dependent upon the quality o f the material submitted. The following explanation o f techniques is provided to help clarify markings or notations which may appear on this reproduction. 1. The sign or “target” for pages apparently lacking from the document photographed is “Missing Page(s)”. If it was possible to obtain the missing page(s) or section, they are spliced into the film along with adjacent pages. This may have necessitated cutting through an image and duplicating adjacent pages to assure complete continuity. 2. When an image on the film is obliterated with a round black mark, it is an indication of either blurred copy because of movement during exposure, duplicate copy, or copyrighted materials that should not have been filmed. For blurred pages, a good image o f the page can be found in the adjacent frame. If copyrighted materials were deleted, a target note will appear listing the pages in the adjacent frame. 3. When a map, drawing or chart, etc., is part of the material being photographed, a definite method of “sectioning” the material has been followed. It is customary to begin filming at the upper left hand comer of a large sheet and to continue from left to right in equal sections with small overlaps. If necessary, sectioning is continued again—beginning below the first row and continuing on until complete.
    [Show full text]
  • Mutated! - Music Tagging Type Definition
    University of Glasgow Funded by the Department of Music Joint Information Systems Committee Call: Communications and Information Technology Standards MuTaTeD! - Music Tagging Type Definition A project to provide a meta-standard for music mark-up by integrating two existing music standards (Report Jan 2000) Dr. Stephen Arnold Carola Boehm Dr. Cordelia Hall University of Glasgow Department of Music Department of Music MuTaTeD! Contents CONTENTS: I. INTRODUCTION (BY DR. STEPHEN ARNOLD).................................................................. 3 II. ACHIEVEMENTS AND DELIVERABLES (BY THE MUTATED! TEAM) .............................. 5 1. Conference Papers and Workshops ..................................................................................... 5 2. Contact with relevant research groups .................................................................................. 5 3. Future Dissemination Possibilities ......................................................................................... 6 III. SYSTEMS FOR MUSIC REPRESENTATION AND RETRIEVAL (BY CAROLA BOEHM) 8 1. Abstract .................................................................................................................................. 8 2. The Broad Context ................................................................................................................. 8 3. The Music-Specific Context ................................................................................................... 8 4. Music Representation Standards – The Historical
    [Show full text]
  • Musixtex Using TEX to Write Polyphonic Or Instrumental Music Version T.111 – April 1, 2003
    c MusiXTEX Using TEX to write polyphonic or instrumental music Version T.111 – April 1, 2003 Daniel Taupin Laboratoire de Physique des Solides (associ´eau CNRS) bˆatiment 510, Centre Universitaire, F-91405 ORSAY Cedex E-mail : [email protected] Ross Mitchell CSIRO Division of Atmospheric Research, Private Bag No.1, Mordialloc, Victoria 3195, Australia Andreas Egler‡ (Ruhr–Uni–Bochum) Ursulastr. 32 D-44793 Bochum ‡ For personal reasons, Andreas Egler decided to retire from authorship of this work. Never- therless, he has done an important work about that, and I decided to keep his name on this first page. D. Taupin Although one of the authors contested that point once the common work had begun, MusiXTEX may be freely copied, duplicated and used in conformance to the GNU General Public License (Version 2, 1991, see included file copying)1 . You may take it or parts of it to include in other packages, but no packages called MusiXTEX without specific suffix may be distributed under the name MusiXTEX if different from the original distribution (except obvious bug corrections). Adaptations for specific implementations (e.g. fonts) should be provided as separate additional TeX or LaTeX files which override original definition. 1Thanks to Free Software Foundation for advising us. See http://www.gnu.org Contents 1 What is MusiXTEX ? 6 1.1 MusiXTEX principal features . 7 1.1.1 Music typesetting is two-dimensional . 7 1.1.2 The spacing of the notes . 8 1.1.3 Music tokens, rather than a ready-made generator . 9 1.1.4 Beams . 9 1.1.5 Setting anything on the score .
    [Show full text]
  • Key and Mode in Seventeenth- Century Music Theory Books
    204 KEY AND MODE IN SEVENTEENTH- One of the more remarkable cultural phenomena of modern times is the emergence of a generally accepted musico-syn- tactical system (a musical "language") known, among other ways, as the "common practice of the eighteenth and nineteenth centuries," or as the "major/minor system of harmony," or as "tonal [as opposed to modal] harmony. " While twentieth-cen- tury composers have sought new ways to structure music, in- terest in the major/minor system continues undiminished owing to a variety of reasons, among them the facts that the bulk of present day concert repertoire stems from the common prac- tice period, and that theories relevant originally to music of the eighteenth and nineteenth centuries have been profitably extended to more recent music. 205 CENTURY MUSIC THEORY BOOKS WALTER ATCHERSON Much has been written concerning the major/minor system. By far the most words have been expended in attempts to codify the syntax of major/minor and make it into a workable method of teaching composition. Numerous music theory text books of the eighteenth and nineteenth centuries fall into this cate- gory. Another category of writing concerning the major/minor system consists of those efforts to set forth philosophical, mathematical, or acoustical explanations of major/minor har- mony; this category often overlaps the first, as it does already in Jean-Philippe Rameau's TRAITE DE L'HARMONIE (1722). In the twentieth century the major/minor system has ceased to serve as a basis for composition instruction, and the search 206 for philosophical, mathematical, and acoustical explanations has proved unrewarding.
    [Show full text]
  • Key Signatures Meters Tempo Clefs and Transpositions Position Work
    The composition criteria for MSHSAA sight reading selections were revised in 2013-14. As a result, the committee determined that it would be beneficial to music directors throughout the state to have this criterion available to assist in preparations for sight reading. With this information in hand, directors will have a better understanding of the difficulty of the selection that will be used and will know what to expect in regard to rhythms, ranges, time signatures, etc. Key Signatures Key Signatures will be limited to the following: C, G, D, F, Bb major and relative natural minors. There should be at least one key signature change per piece. Meters Meters will be limited to the following: 2/4, 3/4, 4/4, Cut Time, 6/8 Tempo Tempi will be no slower than 72 beats per minute and no faster than 120 beats per minute. For continuous pieces, there will be at least one tempo change, but no more than two. Clefs and Transpositions Viola can have limited use of treble clef. Cello can have limited use of tenor clef. Position Work for Strings 3rd position and minimal use of 5th position may be used in the violin 1 part only. Minimal 3rd position work may be used in the violin 2 and viola parts. Half, 2nd, 3rd and 4th position work may be used in the cello part. Half position to the fourth position work may be used in the bass part. Divisi Limited use of divisi parts may be written. Repeats 1st and 2nd endings may be used.
    [Show full text]
  • B.A.Western Music School of Music and Fine Arts
    B.A.Western Music Curriculum and Syllabus (Based on Choice based Credit System) Effective from the Academic Year 2018-2019 School of Music and Fine Arts PROGRAM EDUCATIONAL OBJECTIVES (PEO) PEO1: Learn the fundamentals of the performance aspect of Western Classical Karnatic Music from the basics to an advanced level in a gradual manner. PEO2: Learn the theoretical concepts of Western Classical music simultaneously along with honing practical skill PEO3: Understand the historical evolution of Western Classical music through the various eras. PEO4: Develop an inquisitive mind to pursue further higher study and research in the field of Classical Art and publish research findings and innovations in seminars and journals. PEO5: Develop analytical, critical and innovative thinking skills, leadership qualities, and good attitude well prepared for lifelong learning and service to World Culture and Heritage. PROGRAM OUTCOME (PO) PO1: Understanding essentials of a performing art: Learning the rudiments of a Classical art and the various elements that go into the presentation of such an art. PO2: Developing theoretical knowledge: Learning the theory that goes behind the practice of a performing art supplements the learner to become a holistic practioner. PO3: Learning History and Culture: The contribution and patronage of various establishments, the background and evolution of Art. PO4: Allied Art forms: An overview of allied fields of art and exposure to World Music. PO5: Modern trends: Understanding the modern trends in Classical Arts and the contribution of revolutionaries of this century. PO6: Contribution to society: Applying knowledge learnt to teach students of future generations . PO7: Research and Further study: Encouraging further study and research into the field of Classical Art with focus on interdisciplinary study impacting society at large.
    [Show full text]
  • Lecture Notes on Transposition and Reading an Orchestral Score
    Lecture Notes on Transposition and Reading an Orchestral Score John Paul Ito Orchestral Scores The amount of information presented in an orchestral score can seem daunting, but a few simple things about the way the score is arranged can make it much more manageable. The instruments are grouped by families, in the following order: Woodwinds Horns Brass Percussion [Soloists, Singers] Strings Within each family, instruments are grouped by type, ordered with the higher instrument types higher within the family: Woodwinds: Flutes Oboes Clarinets Bassoons Brass: Trumpets Trombones Tubas Strings: Violin 1 Violin 2 Viola Cello Bass Notes on Transposition – © 2010 John Paul Ito 2 Similarly, individual instruments within an instrument type are ordered from high to low: Flutes Piccolo Flute Alto Flute Oboes Oboe Cor anglais Clarinets Eb Clarinet Clarinet Bass Clarinet Note that grouping by instrument type first means that the score will not reflect a consistent progression from high to low within instrument families. For example, among the woodwinds, the cor anglais will appear above the Eb clarinet, even though the Eb clarinet has a higher register. For the purposes of harmonic analysis, it is very convenient that scores are laid out with the strings on the bottom. This reflects an 18th-century concept of orchestration, in which the strings served as a foundation, with winds and brass added for extra color. Even into the 19th century, it remains true that all of the information needed for harmonic analysis is found in the strings, at the bottom of the score. This makes analysis easier, because string instruments are not transposing instruments.
    [Show full text]