<<

AN INTELLIGENT MULTI-TRACK AUDIO EDITOR

Roger B. Dannenberg Carnegie Mellon University School of Computer Science

ABSTRACT breakthroughs. To illustrate the potential for intelligent, automated editing, and to explore problems that arise in allows multi-track recordings to practice, the author has constructed a working prototype be manipulated by moving notes, correcting pitch, and that: (1) identifies and labels all notes in a multi-track making other fine adjustments, but this is a tedious recording, using a standard MIDI file as a score; (2) es- process. An “intelligent audio editor” uses a machine- timates the performed pitch and pitch error of each note; readable score as a specification for the desired (3) applies pitch correction to all notes; (4) locates all performance and automatically makes adjustments to simultaneous onsets in the score; (5) aligns recorded note pitch, timing, and dynamic level. notes that are not synchronized within a degree of toler- ance; (6) estimates the average recording level for each 1. INTRODUCTION instrument; and (7) balances the instruments according to user-selected ratios (initially 1:1). The result of this Multi-track recording in the studio is a standard produc- fully automated process is a new set of “corrected” tion method for a variety of musical styles, including tracks that can be mixed as is, or edited further by hand. classical, experimental, and popular forms. Multi-track While these steps can be performed manually using techniques are used in many recording situations, in- modern audio editing software, the present work is the cluding ensembles that record all tracks at once, ensem- first to automate these steps, making use of a score bles that record tracks in several passes, and even indi- (MIDI file) to specify the desired result. viduals that play different parts in succession. One of The following sections describe each step of the sys- the advantages of multi-track recording is the control it tem, named IAED for Intelligent Audio Editor. Section 5 gives after the music is recorded. Balance, , describes some experience with the IAED, and this is reverberation, and many other parameters can be altered followed by a discussion, summary, and conclusions. without the time and expense of re-recording. Originally, allowed individual musi- cians to correct mistakes quickly by “punching in,” i.e., 2. RELATED WORK replacing as little as one note of one track with new ma- terial. has ushered in many more pos- The IAED combines and applies various forms of in- sibilities. Now, individual notes or whole phrases can be formation derived from audio and scores. Earlier work shifted in time, stretched, transposed to improve intona- by the author and others offers many ways to perform tion, or even replaced, perhaps using material from the the necessary analysis. One of the purposes of the IAED same performance. is to gain experience combining these techniques in or- As with many technologies, new capabilities work der to guide further development. quickly to establish a new set of expectations. Now, vir- To align tracks to scores, both monophonic and poly- tually every music recording is carefully edited and re- phonic score-to-audio alignment techniques are useful. fined using digital techniques. Ironically, the efficiency Note-based techniques used in monophonic score fol- and effectiveness of digital editors often results in much lowing [3] can be applied. There is also work on mono- more time spent in production than ever before. phonic and polyphonic alignment [11], [15] that can be Meanwhile, there have been many advances in audio used directly. To refine the alignment, onset detection is and music analysis, machine listening, audio-to-score used. Previous work by Rodet and Jaillet [13], Plumbley, alignment, and music understanding. Tasks such as Brossier and Bello [12], Dannenberg and Hu [5], and finding notes, aligning them in time, correcting intona- many others is relevant. tion problems, and balancing dynamic levels can be per- Time alignment and pitch shifting often rely on time formed automatically, at least in principle, and this is the stretching techniques that include SMS [16], phase vo- objective of the present research. In the production ver- coder [8], and PSOLA [14] techniques. sion of this work, one starts with audio recordings and a Finally, it should be noted that the practice of fine- machine-readable score. An automatic process identifies grain editing by hand is well established using commer- possible problems. Using an interface similar to that of a cial editors such as ProTools (www.digidesign.com), and spelling checker, the user is prompted to audition, ac- plug-ins for time stretching and pitch correction such as cept, or reject each suggested edit. The user can also Antares Auto-Tune (www.antarestech.com). Non- simply accept all suggestions and listen to the final mix. commercial editors such as Sonic Visualizer [2] and While this vision of a highly automated and easy-to- CLAM Annotator [1] offer some assistance labeling use editing system will require considerable effort to pitch and onsets, but are not intended for audio editing. realize, the basic techniques require no fundamental

89

3. ANALYSIS linear interpolation between beats. Then, the estimated onset times are refined as described in the previous para- The first stage of IAED processing obtains information graph. about each note as well as summary information about To verify the labeling, clicks can be synthesized at the the tracks. For simplicity, assume that every track is a note onset times. Then by listening to the rhythm and recording of a pitched monophonic instrument. Other timing of the clicks, the user can quickly detect any ma- cases will be discussed later. jor errors. Figure 1 shows three input tracks and labels (below each audio track) generated automatically by 3.1. Labeling IAED. The number shown by each onset time is the in- dex of the note in the track. Seconds are shown at the The process of identifying notes in the recordings and very top. associating them with notes in the score is called label- ing. Labeling can be performed in two ways by the 3.2. Dynamics and F0 Estimation IAED. The first method is based on audio-to-score alignment techniques [6]. An alignment program con- Once note onsets are identified, the IAED labels each verts both score and audio time frames into sequences of note with an indication of dynamic level and funda- chroma vectors and aligns them using dynamic time mental frequency (F0). Since dynamic level is estimated warping. [4] This method works for polyphonic record- for the purposes of balance between tracks, it is assumed ings as well as monophonic ones. Each track is sepa- that the subjective impression is created by the loudest rately aligned with its corresponding standard MIDI file part of the overall note. For simplicity, RMS is used track, resulting in a mapping between MIDI time and rather than a perceptually based measure. The dynamics audio time. This mapping is smoothed and used to map estimation algorithm, which operates on each note, is as each MIDI key-down message to an onset time in the follows: audio recording. • Pre-compute an RMS value for each 5ms frame (a The score alignment is not precise enough for editing, by-product of the labeling step described above). so additional refinement is performed. (Another motiva- • Find the location of the highest RMS value between tion for the refinement is that we are most interested in the current note onset and the next note onset. notes that occur early or late. Their onset times may not • Scan backward from the peak RMS point until the be predicted very well if the score alignment is presented RMS falls to half the peak value (but stop at the note as a smooth tempo curve.) An onset detection function, onset if it is reached). Call this time point A. based mainly on change in the local RMS energy is used • Scan forward from the peak RMS point until the to estimate the onset probability of each 5ms frame in RMS falls to half, stopping if necessary at the next the recorded audio. For each note in the score, this prob- note onset time. Call this time point B. ability is weighted by a Gaussian centered on the ex- • Compute the mean RMS over the interval [A, B]. pected note. This is a roughly Bayesian model where the Fundamental frequency is estimated using the YIN al- RMS feature represents an observation probability and gorithm [7]. YIN generates an F0 estimate and a confi- the Gaussian represents the prior probability based on dence level for each 5ms frame. On the assumption that the score. The location where their product is maximized intonation is most important when notes are loudest, we determines a refined estimate of the note onset. This is a average the YIN estimates over the same interval [A, B] partial implementation of a bootstrapping method [5] used to average RMS values, but ignoring frames where that proved to work extremely well for onset labeling. the YIN confidence level is low. The second method for note onset detection dispenses with the score alignment step and uses labeled beats in- 3.3. Coincident Notes stead. Beats can be labeled simply by tapping along with the music, or the music might already have a click track The IAED makes many timing decisions based on notes or a synchronized MIDI track, in which case the beat that should sound at the same time in different tracks. locations are already known. The score is warped using These coincident notes are detected by scanning the score to find note onsets that match within a small toler- ance.

4. EDITING AUTOMATION

After the analysis stage, the IAED computes plans for editing the tracks and then carries out these plans to produce new tracks. In principle, these plans could be reviewed and edited by a user, but the IAED prototype operates in a fully automated batch mode. The editing operations are pitch correction, time alignment, and bal- ancing. Figure 1. Three tracks (Bb Tpts) with onset labels.

90

4.1. Pitch Correction represented as a function from score position (measured in beats) to tempo, then the tempo curves of all tracks Pitch adjustments are based on comparing each F0 label can be averaged to form a composite tempo curve. Con- to the MIDI key number of the corresponding note in sider, however, the case where many instruments are the score. A variable-rate, high-quality resampling algo- either holding notes or not playing while the tempo rithm, based on sync interpolation [17], is used to warp changes. If tempo curves are constant between note on- the original track. Each note is stretched or compressed sets, these instruments will contribute constant values to according to how much sharper or flatter it is relative to the average tempo even though the tempo should be the reference note in the score. The idea behind time- changing. A possible solution is to use a weighted aver- varying resampling is to map each sample time in the age to suppress the influence of instruments that are not output to a continuous time in the source signal (a re- establishing the tempo [9]. On the other hand, it is com- corded track). The source signal is then interpolated to mon for musicians to exhibit tempo variation in moving obtain a value at this time point. This process is repeated lines due to rhythmic variation or technical difficulties. to generate each output sample. Overall, the system should find a balance between The mapping from output sample time to source sig- maintaining a smooth average tempo and allowing for nal time is computed in several steps because it is sim- expressive timing variations that can appear as local pler to think in terms of mapping from source to output. tempo fluctuations. First, a frequency ratio is computed for each note, repre- A third possibility is to link instruments to a “master” senting the amount by which the signal should be track. For example, three lower voices could be tied to stretched or compressed to produce a “correct” pitch. Let the timing of the top voice in a quartet. The IAED pro- ratio Ri = Mi / ωi, where Mi is the fundamental frequency totype uses this technique. The timing of the lower indicated by the MIDI score for note i, and ωi is the es- voices is altered so that coincident notes in the score are timated actual frequency of note i (see Section 3.2). A coincident in the adjusted audio. function is then constructed to represent the time-varying Timing adjustments are computed in several steps. stretch factor to be applied to the track: First, the coincident notes between the master track and a E(t) = Ri s.t. i satisfies Oi ≤ t < Oi+1 (1) “slave” track are identified. For each pair of coincident where Oi is the onset time for note i. E(t) is just a piece- notes, the onset difference tells how the slave timing wise constant function with the stretch factors as values. should be adjusted. Let (Mi, Si) be pairs of onset times If we integrate E(t), we get a mapping from the original for coincident notes in the master and slave tracks. Then signal times to corresponding time points in the output the sequence [ Mi − Si | 1 ≤ i ≤ N ] describes the amounts signal. Taking the inverse, we get a mapping from by which to shift coincident notes in the slave track. stretched time to original time. Letting X be the original Timing will be altered by time stretching and shrink- recording and Y the pitch-corrected signal: ing sections of music. The ith section spans the time in- Y = X ° W, or Y(t) = X(W(t)), where (2) terval [Si-1, Si]. When one section of music is time- –1 W = V , and (3) stretched, all following notes are affected. Therefore, the amount of time stretching must be decreased by the run- V(t) = ∫E(t)dt (4) ning total of all previous time stretching: In the implementation, E is represented numerically as a function sampled at the frame rate of about 200Hz. This (1) is easily integrated, and since the result is monotonically increasing, the inverse is also easy to compute as an- We now have a list of audio segments [Si-1, Si] and the other sampled function. The output is just Y(t) computed amounts Δi by which to lengthen them (negative values at discrete sample times. W(t), the “warp” function, is mean to shorten them). evaluated by linear interpolation to determine a time Ideally, the stretching or shrinking of segments should point in the source signal X. Then X(W(t)) is calculated be based on a high-quality time-stretching algorithm. For from X using sync interpolation. monophonic tracks, one method, based on PSOLA [14], As a result of resampling, the note onset times are is to identify pitch periods in areas where the signal is changed. The note onset labels are changed accordingly highly periodic. Then, duplicate a selected set of these from Oi to W(Oi). periods to stretch the signal, or delete them to shrink the signal. Because only isolated whole periods are deleted 4.2. Time Alignment or inserted, this method does not suffer from phase dis- continuities or other frame-rate artifacts. Other tech- One of the interesting questions for editing is how to niques such as SMS [16] and the phase- [8] can determine what note onset times should be. A simple also be applied. answer is to quantize note onsets to match the times In the prototype IAED, an even simpler system for given in the symbolic score, but in most cases, this pro- time stretching is used. If there are rests in the segment, duces a very mechanical and lifeless result. silence is simply inserted or deleted from the rests. If Another approach computes the instantaneous tempo there are n rests in segment i, each rest is modified by of each track as a piece-wise constant function between Δi/n. (Locating these rests in the actual signal is easy each pair of adjacent note onsets. If this “tempo curve” is since each rest is terminated by a labeled note onset). If

91

the segment has no rests and is too long, the last note of The labeling process generally works well, but editing the segment is truncated and cross-faded with the note- has very little tolerance for errors. Initial attempts to use onset of the following segment. If the segment is too score alignment left a few mistakes, primarily where short, a bit of silence is inserted at the end of the seg- there were repeated notes. To solve this problem, prior ment. In general, time adjustments are small, so this research on accurate onset detection [5] will be inte- simple method works reasonably well. grated and tested in this environment. This earlier work achieved 100% accuracy labeling trumpet onsets. It was not integrated into the IAED simply for ease of imple- mentation and the (over-optimistic) assumption that a simpler approach would be adequate. The alternate labeling method uses taps rather than score alignment for initial note placement. Using this Figure 2. Editing is performed by splitting source system, there were also a few problems. The original tracks on selected note onsets, time stretching, and re- recordings used for testing were made without a click assembling to form an edited track. track and had some rather large timing errors. Figure 3 The signal processing required for time alignment is illustrates a section where the top voice, recorded first, essentially a matter of cutting out segments of audio has no moving line for 6 beats. When the second track from the original track and reassembling them, using was recorded, the moving line took too long, leading to a short cross-fades to avoid clicks at the splice points. (See proverbial “train wreck” where the parts come together. Figure 2.) The IAED computes a list of all cutting and Figure 1 begins around the last 2 beats of the music pasting operations. Each operation is described by a tri- shown in Figure 3. This is an extreme case where there ple, (Ti, Fi, Di), where Ti is the time to play clip i, which really is no common beat shared by all the parts, and is taken from the original track at offset Fi and duration tapping to suddenly shifting time is not an easy job. To Di. resolve this problem, the click track was edited manually To incorporate other forms of time stretching, the to shift a few clicks into synchrony with the music on the same approach can be taken. The essence is to break up track. In a production system, the obvious solution is to each source track into contiguous segments (perhaps provide for interactive editing, but only where more with a little overlap to allow cross-fades), time-stretch automated techniques break down. the segments using any available technique, and then reassemble the segments to form result tracks.

4.3. Balancing Levels

Track levels are often critical, and setting them requires musical knowledge and at least a good model of human auditory perception. Fine adjustments are left to the hu- man user, but the IAED can at least make an initial Figure 3. Fragment from score of test piece (m. 9-11). rough setting. For each track, the IAED computes Lavg, the average of all RMS labels (see Section 3.2). The Once note onsets and pitches are properly labeled, whole track is then scaled by Lnom/Lavg , where Lnom is the pitch correction works very well. Even when note onsets desired, nominal track level. It might be the case that are not perfectly detected, the pitch-corrected version there are desired differences in levels when an instru- sounds very good. Probably the instability of pitch dur- ment is playing solo vs. in ensemble. As an enhance- ing note transients and the fact that pitch changes are on ment, IAED computes Lavg using only notes that are in the order of 1% masks any artifacts that are introduced. the coincident set, i.e. notes where the player should be Since pitch correction is applied continuously, there are trying to balance levels with other players. When the no splices or other forms of editing, and the resulting adjusted RMS levels of individual coincident notes are pitch-corrected signal is continuous, which also mini- not well matched, the interactive version of the IAED mizes any artifacts. can suggest further refinement. Time correction introduces the most dramatic changes. Figure 4 shows the output after editing the in- 5. RESULTS put shown in Figure 1 (the “train wreck”). Even a quick visual inspection reveals how much better the parts are aligned. While some extreme changes are made in the The IAED is fully implemented in prototype form. The test tracks, and considering that very simple time- purpose of the prototype is to test the various compo- stretching techniques are used, the result is quite good. nents and set the stage for further research and devel- The fact that edits are made only milliseconds before opment. So far, it has only been used for small test ex- coincident attacks undoubtedly helps to mitigate any amples. The evaluation is necessarily subjective, but the signal processing artifacts. There do seem to be some results are very encouraging. The results from labeling, musical artifacts: Even though time correction only pitch correction, time alignment, and balancing levels changes note durations by tens of milliseconds, this can are described below.

92

like a graphical equalizer to control tempo variation on different time scales. Determining how best to adjust note timing is an in- teresting area for future work. A good illustration of the subtlety of this problem is measure 2 of staff 2 in Figure 3. These notes should be aligned to the implied beat from measure 1 of staff 1, even though there are no shared onset times. This example illustrates the tension between

the opposing goals of steady tempo and expressive per- Figure 4. Output generated by the IAED. formance timing. be rhythmically important. Better time stretching that It is tempting to synchronize every pair of coincident distributes the stretch evenly over segments (which may notes, but systematic deviations from synchrony can be range from one to many notes) should help in this area. musically important. Sometimes melody leads the ac- In the test data, the average levels, Lavg, of all parts companiment, perhaps to make the melody more salient, were approximately equal, so track level adjustments but musicians also “lay back,” placing melodic onsets were insignificant. However, the resulting balance is later in time than the accompaniment. The best way to acceptable, indicating that this is at least a plausible way handle this may be to give the user the ability to apply to automatically set initial levels. time correction selectively. Moreover, the very notion of As important as it is to “correct” the original perform- aligning note onsets is only an approximation to a more ance material, it is crucial to retain its desirable musical sophisticated view that considers perceptual onset times and expressive attributes. Since pitch is not corrected as well as timbral and musical implications of onset within notes, expressive and characteristic pitch changes timing. Perhaps desired onset timings can be learned are retained and sound natural. Since levels are not ad- though a statistical analysis of the performance. Assum- justed between notes but only between tracks, expressive ing that onset timing is good for the most part, it is pos- dynamics are also retained. The one area where expres- sible to detect and “correct” outliers to match the mean siveness seems to be compromised is the rhythmic rather than simply line up all onsets. “feel.” This is an inherent problem even with manual Pitch correction is another interesting area for future editing, and overall, the IAED processing improves the research. The IAED simply adjusts pitches to equal tem- recording. As automatic editing moves from research to perament, but future systems might use an automatic practice, there will be opportunities to identify the harmonic analysis of the score [20] to achieve better sources of rhythmic “feel” and to better retain this “feel” intonation. For example, major thirds could be tuned to a even while correcting timing. 5/4 ratio, based on “just” intonation, which is substan- tially smaller than an equal-tempered third. The IAED uses a particular model to predict perceived pitch, but 6. DISCUSSION now that the problem is posed, surely future research will develop better models. Finally, the IAED intention- The results are quite encouraging. IAED can start with a ally does not modify pitch within a note, but this is easily rather sloppy multi-tracked performance, a MIDI ver- done by changing E(t) in Equation 1. This could be used sion of the score, and some tapped beat tracks. Without to manipulate vibrato, pitch change during the attack, further intervention, IAED generates an edited version and other pitch variation with better intonation, greatly improved rhythmic accu- IAED does not adjust note durations or the dynamics racy, and a good initial balance. The result retains much within a note. In the test data, there are some crescendos of the musicality and expressivity of the original tracks. that are not perfectly matched. One could simply adjust The prototype has revealed many interesting issues, the RMS envelope of one note and the cutoff time to setting the stage for future research and development. match another’s. This is just one example of how One of the most interesting issues is tempo and beat articulation might be altered to improve the ensemble. placement. There are a variety of sometimes-conflicting These techniques may not apply very well to poly- goals, including: phonic instruments such as , or to non-pitched in- • Avoid long-term tempo changes. struments, e.g. drum sets. However, onset detection and • Allow slight tempo changes for expression. alignment might still be possible if the editing decisions • Avoid phrases that rush or drag. are made manually. The user interface and support for Make rhythms accurate. • manipulating these sorts of tracks are left to future work. Allow expressive timing deviations, e.g. for swing. • The next step will be to process music recorded by a Align note onsets to the expected onset times. • jazz octet. These octet recordings include a large amount Align note onsets to those in other tracks. • of notated music for three saxophones, trumpet, and If tempo is viewed as a signal, we can think of filtering trombone, while the piano, bass, and drums are largely the tempo to achieve different effects. A high-pass filter unnotated. The arrangements were created using a music allows local timing deviations and eliminates long-term notation editor that can save standard MIDI files. There- changes. A low-pass filter eliminates local time defor- fore, IAED can read the scores and operate on the horn mations, but admits longer-term tempo changes. It tracks. might be advantageous to have an interface something

93

7. CONCLUSIONS [8] Dolson, M. “The Phase Vocoder: A Tutorial.” Computer Music Journal 10(4), 1987, pp. 14-27. Detailed editing of multi-track recordings is a standard [9] Grubb, L., and R. B. Dannenberg. “Automated practice in modern music production. Editing is used to Accompaniment of Musical Ensembles." perfect professional recordings and, with perhaps more Proceedings of the Twelfth National Conference on work, editing can make amateur recordings sound pro- Artificial Intelligence. AAAI, 1994, pp. 94-99. fessional. For the most part, editing operations are quite [10] Hu, N., R. B. Dannenberg, and G. Tzanetakis. mechanical and systematic. Using state-of-the-art tech- “Polyphonic Audio Matching and Alignment for niques for audio analysis and processing, the Intelligent Music Retrieval.” In 2003 IEEE Workshop on Audio Editor, or IAED, automates these tedious opera- Applications of Signal Processing to Audio and tions. IAED uses a score to help analyze the audio and Acoustics (WASPAA). New York: IEEE, 2003, pp. also to serve as a specification for the desired music. 185-188. Edits correct pitch errors, timing errors, and overall sig- [11] Orio, N. and Schwarz, D. “Alignment of nal imbalance while retaining most of the expressive Monophonic and Polyphonic Music to a Score,” in nuance, including expressive timing, articulation, and Proceedings of the 2001 ICMC. San Francisco: pitch variation such as vibrato. Many interesting issues International Computer Music Association, 2001, are raised in this work, suggesting future research on pp. 155-158. score-to-audio alignment, onset detection, performance timing, and perceived pitch. Additional work may also [12] Plumbley, M., Brossier, P., and Bello, J. “Fast labelling of notes in music signals.” In ISMIR 2004 be needed on interfaces that integrate audio with music Fifth International Conference on Music notation and allow users to request, constrain, and adjust Information Retrieval Proceedings. Barcelona, “intelligent” editing operations. Spain: Universitat Pompeu Fabra, 2004, pp. 331-6. 8. REFERENCES [13] Rodet, X., and F. Jaillet. "Detection and Modeling of Fast Attack Transients." Proceedings of the [1] Amtriain, X., J. Massaguer, D. Garcia, and I. 2001 International Computer Music Conference. Mosquera. “The CLAM Annotator: A Cross- International Computer Music Association, 2001, platform Audio Descriptors Editing Tool.” In pp. 30-33. Proceedings of 6th International Conference on [14] Schnell, N., Peeters, G., Lemouton, S., Manoury Music Information Retrieval. London: Queen P., Rodet, X. “Synthesizing a in Real-Time Mary, University of London, 2005, pp. 426-9. Using Pitch Synchronous Overlap Add (PSOLA)”, [2] Cannam, C., C. Landone, M. Sandler, J. P. Bello. In Proceedings of the International Computer “The Sonic Visualizer: A Visualization Platform Music Conference 2000, San Francisco: for Semantic Descriptors from Musical Signals.” International Computer Music Association, 2000. th In ISMIR 2006 7 International Conference on [15] Schwarz, D., Cont, A., and Schnell, N. “From Music Information Retrieval Proceedings. Boulez to Ballads: Training IRCAM’s Score Victoria, BC, Canada: University of Victoria, Follower,” in Proceedings of International 2006, pp. 324-7. Computer Music Conference. San Francisco: [3] Dannenberg, R. “Real-Time Scheduling and International Computer Music Association, 2005. Computer Accompaniment.” In Mathews, M. and [16] Serra, X. and Smith, J. Spectral Modeling J. Pierce, eds. Current Research in Computer Synthesis: A Sound Analysis/Synthesis System Music. Cambridge, MIT Press, 1989. Based on a Deterministic Plus Stochastic [4] Dannenberg, R. and Hu, N. “Polyphonic Audio Decomposition. Computer Music Journal, 14(4), Matching for Score Following and Intelligent 1990, pp. 12-24. Audio Editors,” in Proceedings of the 2003 [17] Smith, J. and Gossett, P. “A Flexible Sampling- International Computer Music Conference, San Rate Conversion Method,” In ICASSP'84, 19.4.1- Francisco: International Computer Music 19.4.4. San Diego, CA, March 1984, http://www- Association, 2003, pp. 27-34. crma.stanford.edu/ jos/resample. [5] Dannenberg, R. and Hu, N. “Bootstrap Learning [18] Soulez, F., X. Rodet, and D. Schwarz. “Improving for Accurate Onset Detection,” Machine Learning Polyphonic and Poly-Instrumental Music To Score 65:2-3, December 2006, pp. 457-471. Alignment.” In ISMIR 2003 Proceedings of the [6] Dannenberg, R. and Raphael, C. “Music Score Fourth International Conference on Music Alignment and Computer Accompaniment,” Information Retrieval. Baltimore: Johns Hopkins Communications of the ACM 49:8, August 2006, University, 2003, pp. 143-150. pp 38-43. [19] Spiegl, S. “Three for the Money” in Jazz [7] de Cheveigné, A. and Kawahara, H. “Yin, A Inventions: Duets, Trios for Trumpet. N. Fundamental Frequency Estimator for Speech and Hollywood Maggio Music Press, pp. 28-34. Music.” Journal of the Acoustical Society of [20] Temperley, D. The Cognition of Basic Musical America, 111(4), 2002. Structures. MIT Press, 2001.

94