<<

ADAPTIVE TIME–FREQUENCY SCATTERING FOR PERIODIC MODULATION RECOGNITION IN MUSIC SIGNALS

Changhong Wang1, Emmanouil Benetos1, Vincent Lostanlen2, Elaine Chew3 1Centre for Digital Music, Queen Mary University of London, UK 2Music and Audio Research Laboratory, New York University, NY, USA 3CNRS-UMR9912/STMS IRCAM, Paris, France {changhong.wang,emmanouil.benetos}@qmul.ac.uk; [email protected]; [email protected]

ABSTRACT tation that remains stable to time shifts, time warps, and frequency transpositions. Time–frequency scattering [2] , tremolos, trills, and flutter-tongue are techniques provides such mathematical guarantees. Besides the local frequently found in vocal and instrumental music. A com- invariance to translation and stability to deformation pro- mon feature of these techniques is the periodic modulation vided by the scattering transform [1, 10], time–frequency in the time–frequency domain. We propose a representa- scattering goes further by applying frequential scattering tion based on time–frequency scattering to model the inter- along the log-frequency axis. This operation provides in- class variability for fine discrimination of these periodic variance to frequency transposition and captures regularity modulations. Time–frequency scattering is an instance of along log-frequency dimension. the scattering transform, an approach for building invari- Prior work in the representation of and tremolo ant, stable, and informative signal representations. The can be divided into three broad categories: F0-based rep- proposed representation is calculated around the wavelet resentations [4,14,20], template-based techniques [5], and subband of maximal acoustic energy, rather than over all modulation spectra [13, 15, 17]. The error-prone stage of the wavelet bands. To demonstrate the feasibility of this fundamental frequency estimation hinders the performance approach, we build a system that computes the represen- of F0-based methods. Template-based methods may work tation as input to a machine learning classifier. Whereas for vibratos with a large modulation extent (frequency vari- previously published datasets for playing technique analy- ation), while for subtly-modulated vibratos, both the def- sis focus primarily on techniques recorded in isolation, for inition of templates and the matching between templates ecological validity, we create a new dataset to evaluate the and test segments are problematic. The modulation spec- system. The dataset, named CBF-periDB, contains full- tra is another well-known representation for modulated length expert performances on the Chinese bamboo flute sounds, which is averaged on the audio clip level [17]. It that have been thoroughly annotated by the players them- may work well for long-term music information retrieval selves. We report F-measures of 99% for flutter-tongue, tasks such as genre classification or instrument recog- 82% for trill, 69% for vibrato, and 51% for tremolo detec- nition, but struggling with providing temporal positions tion, and provide explanatory visualisations of scattering for short-duration playing technique recognition. To our coefficients for each of these techniques. knowledge, there is not yet any computational research that compares and discriminates between these periodic modu- 1. INTRODUCTION lations in real-world music pieces. Besides the question of coming up with an adequate Expressive performances of instrumental music or singing signal representation, there is a critical need for human- voice often abound with vibratos, tremolos, trills, and annotated playing techniques in audio recordings. Up to flutter-tongue. A common feature of these four playing now, most of the available research literature has focused techniques is that they all result in some periodic mod- on playing techniques that have been recorded in highly ulation in the time–frequency domain. However, from a controlled environments [8, 16, 19]. Yet, recent findings musical standpoint, these techniques convey distinct stylis- demonstrate that, in the context of a music piece, play- tic effects. Discriminating between these spectrotempo- ing techniques exhibit considerable variations as compared ral patterns requires a compact and informative represen- to when they are played in isolation [18]. For periodic modulations, these variations are more evident in folk mu- c Changhong Wang, Emmanouil Benetos, Vincent sic, which highly depends on the interpretation of the per- Lostanlen, Elaine Chew. Licensed under a Creative Commons Attribu- former. Such inter-performer variability in folk music per- tion 4.0 International License (CC BY 4.0). Attribution: Changhong formance necessitates data collection with full pieces. Wang, Emmanouil Benetos, Vincent Lostanlen, Elaine Chew. “Adaptive Time–frequency Scattering for Periodic Modulation Recognition in This paper includes three contributions: representa- Music Signals”, 20th International Society for Music Information tion, application, and dataset. We propose a representa- Retrieval Conference, Delft, The Netherlands, 2019. tion based on time–frequency scattering to model the inter-

809 Proceedings of the 20th ISMIR Conference, Delft, Netherlands, November 4-8, 2019 class variability for fine discrimination of vibrato, tremolo, Type Rate (Hz) Extent Shape trill, and flutter-tongue. Rather than decomposing all the Flutter-tongue 25-50 < 1 semitone Sawtooth-like wavelet bands as the scattering transform, we calculate a Vibrato 3-10 < 1 semitone Sinusoidal (FM) time–frequency scattering around the wavelet subband of maximal acoustic energy, i.e. the transform is calculated Tremolo 3-8 ≈ 0 semitone Sinusoidal (AM) adaptively on the predominant frequency. On the applica- Trill 3-10 Note level Square-like tion side, to our knowledge this is the first attempt at cre- ating a system for detecting and classifying periodic mod- Table 1. Characteristic statistics of four periodic modula- ulations in music signals. To evaluate our methodology, tions in music signals. we create a dedicated dataset of the Chinese bamboo flute, Spectrogram of flutter-tongue, vibrato, tremolo, and trill also known as the or zhudi, and thereafter abbreviated 8 7 as CBF. This dataset, named CBF-periDB, contains full- 6 length solo performances recorded by professional CBF 5 4 players and has been thoroughly annotated by the players 3 Frequency (kHz) themselves. 2 1

The rest of this paper is organised as follows. The char- 0 1 2 3 4 5 6 7 8 9 10 Time (s) acteristics of each periodic modulation and how this infor- (a) Flutter-tongue (b) Vibrato (c) Tremolo (d) Trill mation can be represented by an adaptive time–frequency 6.5 6.5 6.5 6.5 scattering are described in Section 2. Section 3 shows 6 6 6 6 details of the feature extraction process and the proposed 5.5 5.5 5.5 5.5 Frequency (kHz) Frequency (kHz) Frequency (kHz) Frequency (kHz) recognition system. The dataset, evaluation methodology, 5 5 5 5 2 2.2 2.4 4 4.2 4.4 6 6.2 6.4 8.6 8.8 9 and results are discussed in Section 4. Section 5 presents Time (s) Time (s) Time (s) Time (s) our conclusions and directions for future research. Figure 1. Visual comparison of four periodic modulations. Top: Spectrogram of flutter-tongue, vibrato, tremolo, and 2. SCATTERING REPRESENTATION OF trill; bottom: partially enlarged spectrogram of each mod- PERIODIC MODULATIONS ulation for detailed comparison. Prior to discriminating between the four periodic modu- lations, we analyse characteristics of each modulation in 2.2 Scattering Transform Section 2.1. A short introduction of the scattering trans- form is provided in Section 2.2. Section 2.3 describes Proposed by [10], the scattering transform is a cascade of the proposed representation for modelling periodic mod- wavelet transforms and nonlinearities. Its structure is sim- ulations. ilar to a deep convolutional network. The difference is that its weights are not learnt but can be hand-crafted to encode 2.1 Characteristic Statistics of Periodic Modulations prior knowledge about the task at hand. Since little energy is captured by the scattering transform with orders higher The characteristic statistics of each modulation in discus- than two [2], we focus in this paper to the second order. sion are shown in Table 1. As can be seen, flutter tongu- Let ψλ denote the wavelet filter bank obtained from a ing has a much higher modulation rate as compared to the mother wavelet ψ, where λ is the centre frequency of each other three modulations; thus, the modulation rate can be wavelet in the filter bank. Likewise, ψλm refers to the used as a main feature to distinguish it from others. For wavelet filter bank of the mth-order scattering transform, the other three techniques with similar modulation rate, for m ≥ 1. The second-order temporal scattering trans- the discriminative information lies in the modulation ex- form of a time-domain signal x is defined as: tent and shape of the modulation unit. The modulation unit refers to the unit pattern that repeats periodically within the t t t x ∗ ψλ1 ∗ ψλ2 ∗ φT , (1) modulation. It can be one-dimensional, either amplitude modulation (AM) or frequency modulation (FM), or two- t t dimensional as a spectro-temporal modulation. This can where ∗ is the wavelet convolution along time. |x ∗ ψλm | be intuitively observed from the partially enlarged spec- is the modulus of the mth-order wavelet transform, which trograms given in Fig. 1. Trills are note-level modula- hereafter we refer to as the mth-order wavelet modulus tions, for which the frequency variations are larger than transform. The temporal scattering coefficients are ob- one semitone. This extent of modulation is much larger tained by an averaging at a time scale T by means of a than vibratos and tremolos. The shape of the modulation low-pass filter φT . unit for trill is more square-like rather than sinusoidal ones In addition to the invariance to time-shifts and which vibratos and tremolos exhibit. The difference be- time-warps provided by the scattering transform, time– tween vibrato and tremolo is that vibratos are FMs, while frequency scattering provides frequency transposition in- tremolos are AMs. We show later how this discriminative variance by adding a wavelet transform along log- information is encoded into the proposed representation in frequency axis [7]. The specific time–frequency scattering Section 2.3. we apply is a separable scattering proposed in [3]. Here,

810 Proceedings of the 20th ISMIR Conference, Delft, Netherlands, November 4-8, 2019 separable refers to separate steps of temporal and frequen- (a) Spectrogram of different periodic modulations tial operations of wavelet scattering, arranged in a cas- cade. The separable scattering representation comprises a second-order temporal scattering and a first-order frequen- Frequency tial scattering. The latter is calculated by another wavelet 0 5 10 15 20 25 Time(s) transform along the log-frequency dimension on top of the (b) Second-order temporal scattering representation second-order temporal scattering:

t t t fr fr x ∗ ψλ1 ∗ ψλ2 ∗ φT ∗ ψγ1 ∗ φF . (2) Modulation scale

5 10 15 20 25 Time(s) fr where ∗ is the wavelet convolution along log-frequency. (c) First-order frequential scattering representation

ψγ1 is the wavelet filter bank applied in the first-order fre- quential scattering. The frequential scattering coefficients

are obtained by an averaging of the frequential wavelet Modulation scale modulus transform with transposition invariance of F (in 5 10 15 20 25 Time(s) octave unit) using a low-pass filter φF . All scattering co- efficients in this paper are normalised and have their log- Figure 2. Separable scattering representation of different arithm calculated, to capture only the temporal structure periodic modulations. From left to right, the first four are and to motivate auditory perception [2]. Hereafter, we regular cases: vibrato, tremolo, trill, flutter-tongue based use Morlet wavelets throughout the whole scattering net- on stable pitches; the last three are time-varying cases: work for wavelet convolutions. This is because Morlet rate-changing trill, rate- and extent-changing trill, flutter- wavelets have an exactly null average while reaching a tongue with time-varying pitch. quasi-optimal tradeoff in time–frequency localisation [9]. Our source code is based on the ScatNet toolbox 1 . one partial is sufficient to capture modulation informa- 2.3 Scattering Representation of Periodic tion. Fig. 2 (b) shows the second-order temporal scattering Modulations representation decomposed only from the frequency band with the highest energy. Flutter-tongue is the most discrim- Periodic modulation recognition, as suggested by the anal- inable one with the highest modulation rate. For the other ysis above, is a pitch invariant task. The core discrimina- three patterns with close modulation rate value, other char- tive information is based on the modulation itself, which is acteristic information is considered. By dominant band de- indicated by its modulation rate, extent, and shape. Fig. 2 composition, the trill can also be discriminated because of shows respectively (a) the spectrogram, (b) the second- its large modulation extent. This can be interpreted by fil- order temporal scattering representation, and (c) the first- ters with bandwidth larger than one semitone, which blurs order frequential scattering representation of a series of pe- other subtle modulations. riodic modulation examples in CBF-periDB. The spectro- To specifically detect vibrato or tremolo, frequency gram used here is only for illustration purposes. The first bands less than one semitone should be obtained. We then four examples are regular cases (modulations based on sta- make use of their modulation shape information by intro- ble pitch or with constant parameters): vibrato, tremolo, ducing a band-expanding technique. Assume we have fre- trill, and flutter-tongue. The last three are cases with time- quency bands of 1/16 octave bandwidth in the first-order varying parameters (modulations based on time-varying wavelet modulus transform. Ideally for tremolo, the mod- pitch or with time-varying rate, extent, or shape): rate- ulation information is contained only in the dominant fre- changing trill, rate- and extent-changing trill, and flutter- quency band since it is an AM. This is verified by the sec- tongue with time-varying pitch. We use these examples to ond example in Fig. 2 (b), which has almost only the fun- show how the characteristic information of each pattern is damental modulation rate with no upper harmonics in the captured and discriminated by a separable scattering trans- second-order temporal scattering representation. However, form, which consists of a second-order temporal scattering vibratos are FMs, which means the modulation informa- and a first-order frequential scattering transform. tion spreads over neighbouring frequency bands. Decom- 2.3.1 Second-order temporal scattering posing neighbouring frequency bands above or below the dominant band provides additional information to distin- Different from a standard two-order temporal scatter- guish vibrato from tremolo. All this discriminative infor- ing transform, we do not decompose all the frequency mation can be visualised from the fundamental modulation bands (exchangeable with wavelet bands) in the first-order rate and the richness of the harmonics in the second-order wavelet modulus transform. As can be observed from temporal scattering representation in Fig. 2 (b). Fig. 2 (a), the patterns of these modulations, either in the regular case or time-varying case are similar for each har- 2.3.2 First-order frequential scattering monic partial. This indicates that the decomposition of The temporal scattering transform is sensitive to attacks 1 https://www.di.ens.fr/data/software/scatnet/ and amplitude modulations, which results in the high en-

811 Proceedings of the 20th ISMIR Conference, Delft, Netherlands, November 4-8, 2019 ergy part of the boundaries as clearly observed in Fig. 2 averaging scale for the temporal scattering coefficients. (b). To suppress this noisy information while retaining the This parameter is useful for discriminating modulations frequential structure offered by the second-order - with large differences on modulation rate, for example, ral scattering, we use frequential scattering along the log- on distinguishing flutter-tongue from other low-rate peri- frequency axis on top of Fig. 2 (b). Frequential scattering odic modulations. Averaging scales covering at least four has a similar framework as temporal scattering while the unit patterns are recommended for reliable estimation of former captures regularity along log-frequency. As shown the modulation rate. Q1 are the filters per octave in the in Fig. 2 (c), we obtain a clearer representation without first-order temporal scattering transform. Since the modu- reducing the discriminative information necessary for the lations discussed here are all oscillatory patterns, setting task. Although the last example is flutter-tongue bounded Q1 should ensure that each of the modulations are not to time-varying pitch, its modulation rate is relatively sta- blurred in the first-order wavelet modulus transform. Here, ble. This verifies our method of using just the dominant we use Q1 > 12 for the first-order temporal scattering to frequency band or expanded frequency bands from the support subtly-modulated vibratos and tremolos, of which first-order wavelet modulus transform. The rate-changing the modulation extent is less that one semitone. N is the and rate-extent changing cases show that the time-varying number of neighbouring frequency bands besides the dom- modulation rates are also captured. inant band decomposed from the first-order wavelet mod- ulus transform. N = 0 refers to dominant band decom- position only while N > 0 means expanded band decom- 3. PERIODIC MODULATION RECOGNITION position. This is a key parameter to encode the unit shape With the proposed representation prepared, we build a information of subtle modulations. However, if the task at recognition system consisting of four binary classification hand is only to detect modulations with high modulation schemes that each predicts one modulation type. Section rate or with large extent, this parameter is not necessary. 3.1 describes the feature extraction process. The calculated features are then fed to a machine learning classifier illus- Parameter Notation Main information encoded trated in Section 3.2. Averaging scale T Modulation rate Filters per octave Q Modulation extent 3.1 Feature Input 1 = 0, temporal shape As described in Section 2.3, we adapt the scattering trans- Expanded bands N > 0, spectro-temporal shape form by decomposing the dominant and its neighboring frequency bands in the first-order wavelet modulus trans- Table 2. Parameters encoding modulation information in form. The feature extraction process of dominant band de- the adaptive time–frequency scattering framework. composition is shown in Fig. 3. Using a waveform as input, we first obtain the first-order wavelet modulus transform, Other parameters involved in the feature calculation in- where the frequency band with the highest energy for each clude frame-size h, filters per octave Q in the second- time frame is localised. Decomposing these bands, we ob- 2 order scattering decomposition, frequency bands l ex- tain a second-order temporal scattering. A first-order fre- tracted from the second-order temporal scattering. For fre- quential scattering is then conducted on top of the second- quential scattering, we apply a single scale wavelet trans- order temporal scattering coefficients. Concatenating the form. The frame size is inversely log-proportional to two representations in a frame-wise manner, we obtain the the oversampling parameter α by h = T/2α (samples), feature input to the classifiers. For expanded band decom- which is designed to compensate for the low temporal res- position, additional features are calculated similarly by de- olution resulting from the large averaging scales. Since composing the neighbouring frequency bands around the these parameters carry little discriminative information, we dominant band. set them consistently for all classification schemes, with α = 4, Q2 = 8, and l = 2Q2 based on experimental results. A general example with T = 215 corresponds to frame size of h = T/2α = 2048 samples (46ms, as- suming the sampling rate is 44.1kHz). The dimensional- ity of the final representation at each time frame equals to (N + 1) × 2l = 4(N + 1)Q2.

3.2 Recognition System Due to the existence of combined playing techniques, such as the combination of flutter-tongue and vibrato, and the Figure 3. Feature extraction process. combination of tremolo with trill, one frame of the input may have multiple labels. Multi-label classification is con- Table 2 gives the parameters which encode the core sidered beyond the scope of this paper and is regarded as discriminative information for the recognition. T is the future work. Here, we conduct binary classifications for

812 Proceedings of the 20th ISMIR Conference, Delft, Netherlands, November 4-8, 2019 each modulation, which enables us to explicitly encode the analysis dataset, CBF-periDB, comprises monophonic characteristic information specifically for the correspond- performances recorded by ten professional CBF players ing pattern. Four binary classifiers are constructed using from the China Conservatory of Music. All data is support vector machines (SVMs) with Gaussian kernels. recorded in a professional recording studio using a Zoom The model parameters to be optimized in the training pro- H6 recorder at 44.1kHz/24-bits. Each of the ten players cess are the error penalty parameter and the width of the performs both isolated periodic modulations covering Gaussian kernel [6]. The best parameters selected in the all notes on the CBF and two full-length pieces selected validation stage are used for testing. The input feature from Busy Delivering Harvest «扬鞭催马运粮忙», Jolly to the classifiers is the proposed adaptive time–frequency Meeting «喜相逢», Morning «早晨», and Flying Partridge transform of the current time frame. «鹧鸪飞». Players are grouped by flute type (C and G, Taking flutter-tongue as an example, its relatively high the most representative types for Southern and Northern modulation rate (25-50Hz) can be emphasized by setting styles, respectively) and each player uses their own flute. T = 8192 (sampling rate is 44.1kHz). The modulation ex- This dataset is an extension of the CBF-glissDB dataset tent which is less than one semitone is interpreted by set- in [18], with ten pieces containing periodic modulations ting Q1 = 16. Fig. 4 shows the detection result of flutter- added. The playing techniques are thoroughly annotated tongue from a piece in CBF-periDB using dominant band by the players themselves. Details of both isolated tech- decomposition (N = 0), which can be clearly observed niques and full-piece (performed) recordings are shown from the harmonic structure in Fig. 4 (a). This is then en- in Table 3. The dataset and annotations can be downloaded forced by removing the noisy attacks using a frequential from c4dm.eecs.qmul.ac.uk/CBFdataset.html. scattering transform as shown in Fig. 4 (b). Concatenat- ing the two representations, we form frame-wise separa- Isolated Performed ble scattering feature vectors with (N + 1) × 2l = 32 di- Type Length Piece, number Length α mensions and frame size of h = T/2 = 512 samples Flutter-tongue 4.9 Mo, 3 16.0 (12ms). Fig. 4 (c) visualises the binary classification re- Vibrato 7.3 BH, 7 28.0 sult of flutter-tongue compared with the ground truth for an example excerpt. Overall it can be seen that the pro- Tremolo 5.0 JM, 4 12.4 posed approach is successful at detecting flutter-tongue, Trill 12.3 FP, 6 51.9 even in short segments, although occasionally the output is over-fragmented. Similarly, binary classifiers can be im- Table 3. Length of both isolated techniques and full-piece plemented for detecting vibratos, tremolos, and trills, with recordings in CBF-periDB (Mo=Morning; JM=Jolly Meet- parameters fine-tuned to the corresponding modulations. ing; BH=Busy Delivering Harvest; FP=Flying Partridge; all numbers for length are measured in minutes). (a) Second-order temporal scattering representation In the recognition implementation process, the dataset is split into a 6:2:2 ratio according to players (players are

Modulation scale 10 20 30 40 50 60 70 80 randomly initialised) and a 5-fold cross-validation is con- Time (s) ducted. This way of data splitting ensures a non-overlap

(b) First-order frequential scattering representation between players in the train, validation, and test sets. Since each player uses their own flute during recording, there is no overlap across flutes. Due to the limited piece types, non-overlap of pieces is not feasible at the current stage.

Modulation scale 10 20 30 40 50 60 70 80 We regard this as future work with a dataset expansion plan Time (s) to address piece diversity. Reference (c) Reference and detection result Detected 4.2 Metrics Due to the short duration and periodic nature of vibratos, tremolos, trills, and flutter-tongue, feature dependencies

Detected-reference 0 10 20 30 40 50 60 70 80 Time (s) over sequential frames are slightly required. The precision P = TP , recall R = TP , and F-measure F = Figure 4. Binary classification result of flutter-tongue in TP +FP TP +FN 2PR an example excerpt of the CBF-periDB. P+R are used here for frame-based evaluation, where TP,FP,FN are true positives, false positives, and false negatives respectively [12]. Assigned labels by SVMs are 4. EVALUATION then compared to the ground truth annotations in a frame- wise manner. 4.1 Dataset 4.3 Baseline To verify the proposed system, we focus on folk music recordings which have more inter-performer variations According to our knowledge, there is not yet any previous than Western music. The proposed periodic modulation work on discriminating between these four periodic mod-

813 Proceedings of the 20th ISMIR Conference, Delft, Netherlands, November 4-8, 2019

Dominant band Expanded band Type Temporal scattering Separable scattering Temporal scattering Separable scattering P(%) R(%) F(%) P(%) R(%) F(%) P(%) R(%) F(%) P(%) R(%) F(%) Flutter-tongue 96.1 99.8 97.9 96.3 99.8 98.0 96.7 99.6 98.1 97.8 99.5 98.7 Trills 87.1 66.7 75.1 87.4 68.2 76.2 89.5 73.3 80.4 89.8 76.3 82.3 Vibrato 75.2 17.5 26.4 72.2 33.1 45.3 75.9 59.4 66.5 75.1 64.7 69.3 Tremolo 92.5 1.21 2.2 80.9 6.7 10.6 70.8 38.5 49.1 67.6 41.4 50.7

Table 4. Performance comparison of binary classification for flutter-tongue, vibratos, tremolos, and trills in CBF-periDB using separable scattering and temporal scattering representations based on the dominant frequency band and expanded frequency band decomposition (P=precision; R=recall; F=F-measure). ulations; thus we compare the proposed systems against a tion, expanded band decomposition works much better. state-of-art detection method for vibrato. The filter diag- Generally, detection performance on vibrato and onalisation method (FDM), which efficiently extracts high tremolo is worse than that for flutter-tongue and trill detec- resolution spectral information for short time signals, was tion. Identifying the errors in the original audio, we find first applied to vibrato detection in erhu performance [20]. that in most cases these are combined techniques, i.e. sub- Based on the high similarity of the music style between tle frequency variations are accompanied with amplitude erhu and CBF, both being traditional Chinese instruments, modulations or vice versa. In the case of CBF, such com- we use FDM as a baseline method for vibrato detection. binations are common because of the instrumental gestures Using automatically estimated fundamental frequency by of vibrato and tremolo. Vibrato can be generated by finger- pYIN [11] as input for FDM, we try different parame- ing or , while tremolos are commonly produced ter ranges for vibrato rate and extent based on the vibrato by breathing variations. Performers are also expressively characteristics of the CBF. The best result we obtain based intended to add tremolo effects on top of other playing on a 256ms-frame-wise evaluation is P=36.5%, R=58.7%, techniques. F=45.0%. The rate range and extent range we use are 3- 10 Hz and 5-20 cent, respectively. 5. CONCLUSIONS AND FUTURE WORK 4.4 Results In order to explicitly show information captured in each Periodic modulation recognition is a pitch invariant task classification task, a small set of parameter settings for T , and should only capture modulation information. This is Q1, and N is used. Besides the different meta-parameter realised by calculating a time–frequency scattering around settings specifically designed for each modulation classifi- the wavelet subband of the maximum acoustic energy. We cation, we run further experiments by using the same pa- found that the proposed representation decomposed from rameters for all four binary classification processes: T = the dominant band is sufficient to detect modulations with 15 2 , Q1 = 16, and N = 6 frequency bands symmetrically high modulation rate (flutter-tongue) and large modula- expanded around the dominant band. This corresponds to tion extent (trill), while expanded band decomposition cap- frame size of 46ms and feature dimension of 224. Note tures subtle frequency modulations (vibrato and tremolo). that for specific modulation detection, meta parameters of We introduce the ecologically valid dataset, CBF-periDB, the scattering transform can be fine-tuned to have a much to evalate the recognition system. Results show that the lower feature dimension, as the flutter-tongue detection ex- proposed representation captures discriminative informa- ample demonstrated in Section 3.2. The binary classifica- tion between vibratos, tremolos, trills, and flutter-tongue tion results for each pattern are given in both the domi- in real-world pieces. nant band decomposition and expanded band decomposi- The current work only considers continuous periodic tion, as shown in Table 4. The detection results using the modulations; other periodic patterns, such as tonguing, second-order temporal scattering coefficients only as fea- will be considered as future work due to their noncontinu- ture are also provided. The comparison between temporal ous nature and more complicated parameter variations. In- scattering and separable scattering verifies our analysis in spired by research on polyphonic music transcription, cur- Section 2.3 that for periodic modulation recognition, fre- rent work may be expanded to polyphonic periodic mod- quential scattering in the separable scattering representa- ulation detection. Additional comparison of the current tions provides additional discriminative information by re- result with other equivalent representations such as the moving noisy information. Better performance for flutter- modulation spectra, will be conducted. We conclude that tongue and trill detection shows that for modulations with the scattering transform presents a versatile and compact high modulation rate and large extent, decomposition of representation for analysing periodic modulations in per- the dominant band is sufficient. For discriminating be- formed music and opens up new avenues for computational tween temporal modulation and spectro-temporal modula- research on playing techniques.

814 Proceedings of the 20th ISMIR Conference, Delft, Netherlands, November 4-8, 2019

6. ACKNOWLEDGEMENTS [12] M. Müller. Fundamentals of music processing: Audio, analysis, algorithms, applications. Springer, 2015. Changhong Wang is funded by the China Scholarship Council (CSC). Emmanouil Benetos is supported by a [13] Y. Panagakis, C. Kotropoulos, and G. R. Arce. Music RAEng Research Fellowship RF/128. Elaine Chew is sup- genre classification via sparse representations of au- ported by the European Union’s Horizon 2020 research ditory temporal modulations. In 17th IEEE European and innovation program (grant agreement No 788960) Signal Processing Conference (Eusipco), pages 1–5, under the European Research Council (ERC) Advanced 2009. Grant (ADG) project COSMOS. [14] R. Panda, R. Malheiro, and R. Paiva. Novel audio fea- tures for music emotion recognition. IEEE Transac- 7. REFERENCES tions on Affective Computing, (1):1–1, 2018. [1] J. Andén and S. Mallat. Scattering representation of [15] B. L. Sturm and P. Noorzad. On automatic music genre modulated sounds. 15th International Conference on recognition by sparse representation classification us- Digital Audio Effects (DAFx), 2012. ing auditory temporal modulations. Proc. Computer [2] J. Andén and S. Mallat. Deep scattering spec- Music Modeling and Retrieval (CMMR), 2012. trum. IEEE Transactions on Signal Processing, [16] L. Su, H. M. Lin, and Y. H. Yang. Sparse modeling of 62(16):4114–4128, 2014. magnitude and phase-derived spectra for playing tech- [3] C. Baugé, M. Lagrange, J. Andén, and S. Mallat. Rep- nique classification. IEEE/ACM Transactions on Au- resenting environmental sounds using the separable dio, Speech and Language Processing, 22(12):2122– scattering transform. In IEEE International Conference 2132, 2014. on Acoustics, Speech and Signal Processing (ICASSP), [17] S. Sukittanon, L. E. Atlas, and J. W. Pitton. pages 8667–8671, 2013. Modulation-scale analysis for content identifica- tion. IEEE Transactions on Signal Processing, [4] Y. P. Chen, L. Su, and Y. H. Yang. Electric play- 52(10):3023–3035, 2004. ing technique detection in real-world recording based on F0 sequence pattern recognition. In Proc. Conf. [18] C. Wang, E. Benetos, X. Meng, and E. Chew. Hmm- International Society for Music Information Retrieval based detection for recordings of chinese (ISMIR), pages 708–714, 2015. bamboo flute. In Proc. Conf. Sound and Music Com- puting (SMC), pages 545–550, May 2019. [5] J. Driedger, S. Balke, S. Ewert, and M. Müller. Template-based vibrato analysis in music signals. In [19] J. Wilkins, P. Seetharaman, A. Wahl, and B. Pardo. Vo- Proc. Conf. International Society for Music Informa- calset: A singing voice dataset. In Proc. Conf. Inter- tion Retrieval (ISMIR), pages 239–245, 2016. national Society for Music Information Retrieval (IS- MIR), pages 468–474, 2018. [6] T. Hastie, R. Tibshirani, and J. Friedman. The elements of statistical learning: data mining, inference, and pre- [20] L. Yang, K. Rajab, and E. Chew. The filter diagonalisa- diction, 2009. tion method for music signal analysis: frame-wise vi- brato detection and estimation. Journal of Mathematics [7] V. Lostanlen. Convolutional operators in the time- and Music, 11(1):42–60, 2017. frequency domain. PhD thesis, PSL Research Univer- sity, 2017.

[8] V. Lostanlen, J. Andén, and M. Lagrange. Extended playing techniques: the next milestone in musical in- strument recognition. In 5th International Conference on Digital Libraries for Musicology (DLfM), 2018.

[9] S. Mallat. A Wavelet Tour of Signal Processing, Third Edition: The Sparse Way. Academic Press, 2008.

[10] S. Mallat. Group invariant scattering. Communications on Pure and Applied Mathematics, 65(10):1331–1398, 2012.

[11] M. Mauch and S. Dixon. pYIN: A fundamental fre- quency estimator using probabilistic threshold distribu- tions. In IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 659– 663, 2014.

815