<<

GENERATION AND TRANSFORMATION WITH MOMENT MATCHING-SCATTERING INVERSE NETWORKS

Mathieu Andreux and Stephane´ Mallat Departement´ d’informatique de l’ENS, Ecole´ normale superieure,´ CNRS, PSL Research University, 75005 Paris, France

ABSTRACT do not take as input a Gaussian white but estimate a with a factorization, We introduce a Moment Matching-Scattering Inverse which computes conditional probabilities given from Network (MM-SIN) to generate and transform musical past values. The generated have an outstanding . The MM-SIN generator is similar to a variational quality but these neural architectures are complex and or an adversarial network. However, the en- lack interpretability. Several effective techniques have coder or the discriminator are not learned, but computed been introduced to modify audio generations [4–6, 9] with a scattering transform defined from prior information by modifying the latent code to change the probability on sparse time- audio properties. The genera- distribution or by using reference signals as targets, which tor is trained by jointly minimizing the reconstruction loss is more complex than arithmetic operations used for of an inverse problem, and a generation loss which com- images. putes a distance over scattering moments. It has a similar This paper introduces a simplified neural network ar- causal architecture as a WaveNet and provides a simpler chitecture which synthesizes and modifies music signals mathematical model related to time-frequency decomposi- from Gaussian white noise, with two key contributions. tions. Numerical experiments demonstrate that this MM- As opposed to image GANs or variational , SIN generates new realistic musical signals. It can trans- we do not learn a discriminator or an encoder: both are form low-level musical attributes such as pitch with a lin- provided by prior information on audio signals, which is ear transformation in the embedding space of scattering co- captured by their time-frequency regularity. This is done efficients. by adapting a result obtained in [2] for images, and intro- ducing a moment matching technique. The second con- 1. INTRODUCTION tribution is the introduction of a causal computational ar- chitecture allowing to progressively synthesize audio sig- This paper investigates musical generation and nals, and which can be parallelized in GPU’s. The resulting transformation with a simplified algorithmic architecture, Scattering Autoencoder architecture has some similarities which relates generative networks to time-frequency rep- with a Parallel WaveNet [16]. Similarly to warping proper- resentations. Image generation has led the way through ties over images, we show that arithmetic transformations the development of Generative Adversarial Networks in the latent space produce time-frequency deformations, (GANs) [7] and Variational Autoencoders [13] where im- which can modify the pitch of musical notes or interpolate ages are generated from a Gaussian white noise vector, music. Despite a lower synthesis quality than state-of-the- which defines a latent space. Arithmetic operations in this art generating methods, these preliminary results pave the latent space lead to controlled transformations over the im- way for a new approach to synthesize audio signals, with- ages such as aging of faces or transforming women in men. out learning encoders or discriminators. The problem is however different for audio signals which must take into account time causality properties. Some au- thors have applied image generation algorithms over spec- 2. SCATTERING AUTOENCODER trograms [3, 10] but it then requires to invert the spectro- grams with a vocoder or a Griffin-Lim algorithm which is This section introduces the principles of a scattering au- long and has a reduced quality. toencoder and its computational architecture. The encoder Deep autoregressive neural networks such as is not learned but computed based on prior information on WaveNet [15, 16] or SampleRNN [14] have achieved audio time-frequency properties. Audio and musical sig- exceptional synthesis of music and speech signals. They nals have sparse representations over time-frequency dic- tionaries such as audio [20]. Their perceptual properties are not much affected by small time-frequency c Mathieu Andreux and Stephane´ Mallat. Licensed under warpings. We define a embedding which takes ad- a Creative Commons Attribution 4.0 International License (CC BY 4.0). vantage of these characteristics. Attribution: Mathieu Andreux and Stephane´ Mallat. “Music Genera- tion and Transformation with Moment Matching-Scattering Inverse Net- The architecture of a scattering autoencoder is illus- works”, 19th International Society for Music Information Retrieval Con- trated in Figure 1. The random input audio signal X[t] ference, Paris, France, 2018. is first transformed into a nearly Gaussian random vec-

327 328 Proceedings of the 19th ISMIR Conference, Paris, France, September 23-27, 2018

Latent Input Noise Z[t ] Z[2J n] Signal Fixed Encoder Φ J code tJ Scattering SJ Whitening L −1 J L X[t] Z[2 n] Layer XJ [tJ ] ReLU −1 CNN VAR L tJ Temporal W Trained Generator G J Vector ReLU

tJ−1 W Figure 1. The audio is encoded with a time- Zero insertion J−1 ReLU frequency scattering SJ followed by a linear whitening op- erator L. The Scattering Inverse Network (SIN) G restores tJ−2 W1 a signal from a Gaussian white noise by inverting LSJ on Layer X0[t0] training data. t0 J J 2 X0[2 n] J tor SJ (X)[2 n] by applying the time-frequency scattering transform SJ , where J ≥ 0 is a hyperparameter. Gaus- Figure 2. An MM-SIN is a linear recurrent network fol- sianization is achieved through a time averaging over a lowed by a causal deep convolutional network with J lay- sufficiently long time interval thanks to the central limit ers. It takes as input a vector of Gaussian white noise theorem. In order to preserve enough information on the Z[2J n] (top right, red), and computes the corresponding original signal X after averaging, multiple sparse time- J −1 scattering vector XJ [2 n] by applying L . Intermedi- frequency channels are built by applying iterative ate layers Xj[tj] are then computed with causal convolu- transforms [1]. The Gaussian scattering SJ (X) is then tions denoted by blue arrows and zero insertions (white mapped into a Gaussian white noise with a whitening op- points). The single vector Z[2J n] outputs 2J values for erator L, which outputs the linear prediction errors of a fu- X0[t0], marked with red points. ture scattering vector from its neighboring past. Section 4 details the whitened scattering transformation LSJ . The generator G inverts the linear whitening operator network layer is composed of vectors Xj[tj] having kj and the joint scattering transform by synthesizing an ap- channels and sampled at intervals 2j, for 0 ≤ j ≤ J. All proximation of X[t] from Z[2J n]. It first applies a vec- layers in Figure 2 appear to be aligned but to understand tor autoregressive (VAR) filter L−1, which can be deduced the causality structure one must realize that each layer is j from L. This operation is followed by a causal convolu- indexed by a time index tj which is shifted by 2 relatively tional neural network (CNN), with the convolutions acting to the absolute time variable t of the original input signal j along time. The network has J layers to invert a scattering X[t]: tj = t − 2 . Using the absolute time t, we thus use at scale 2J . Section 3 describes the architecture which has the noise vector at a time t = 2J (n+1) to generate 2J new similarities with a Parallel WaveNet [16]. The optimiza- output signal values at times 2J n − 2J + 1 < t ≤ 2J n + 1. tion of the generator G amounts to inverting the scattering The first layer maps Z[tJ ] to XJ [tJ ] with a vector au- −1 transform SJ in an adapted metric. The of syn- toregressive filter L which inverts the whitening opera- thesized signals are constrained by ensuring that they have tor L, followed by a ReLU non-linearity ρ(u) = max(u, 0) the same moments in the scattering space as the input sig- −1  nal X[t]. This network is thus called a Moment Matching- XJ = ρ L Z . (1) Scattering Inverse Network (MM-SIN). A new noise vector Z[2J n] outputs a new vector J J−j XJ [2 n]. At depth j, this initial vector gives rise to 2 J J−j J 3. MOMENT MATCHING-SCATTERING temporal vectors Xj[tj] for 2 n − 2 < tj ≤ 2 n. INVERSE GENERATOR At layer j > 0, the layer Xj is mapped to Xj−1 with an A Moment Matching-Scattering Inverse Network G is a` trous convolution followed by a ReLU non-linearity plus computed by inverting a scattering embedding computed bias. We first double the size of Xj with a zero insertion: at a scale 2J . It is a causal network which takes as input X˜ [n2j] = X [n2j] and X˜ [n2j + 2j−1] = 0. (2) a nearly Gaussian white noise vector Z computed with a j j j scattering transform, to recover an estimation of the signal Each Xj−1 is then calculated with a causal convolution X. This network is trained with a loss which incorporates along time and a linear operator along channels with a bias, the inverse problem loss computed with a scattering met- Wj, followed by a ReLU: ric, regularized by a moment matching term that can be

interpreted as a discriminative metric. Xj−1 = ρ(WjX˜j). (3) The MM-SIN generator is a linear recurrent neural net- work L−1 followed by a causal convolutional network im- Except for the autoregressive layer, the ReLU is pre- plemented with a cascade of J convolutions and pointwise ceded by a batch normalization [11]. The last convolu- non-linearities, illustrated in Figure 2. Each intermediate tion is not followed by a ReLU so that we can output a Proceedings of the 19th ISMIR Conference, Paris, France, September 23-27, 2018 329

signal with negative values. The final output is X0[t0] J J J for 2 n − 2 < t0 ≤ 2 n which corresponds to X[t] for X Log-Spectrogr. 2J n − 2J + 1 < t ≤ 2J n + 1. As in standard generative networks [18], the number of the channels kj decreases 1 ψ` with j according to a geometric law fitted such that X0 has only one channel and XJ has kJ channels, which is the 2D filter 2 hξ ⊗ ψ 0 dimensionality of each vector Z[2J n]. Modulus ` The parameters of the network G are optimized by in- Averaging verting the scattering transform followed by the causal Subsampling whitening operator L. The loss L minimized by G is a Scattering J sum of two terms, weighted by a hyperparameter λ > 0. φJ 2 The first term Linv measures the accuracy of the inversion. The second discriminative term LMM measures the dis- tance between the scattering moments of the synthesized Figure 3. Time-Frequency Scattering transform. The log- signals and the scattering moments of the original signals. 1 is obtained with a first wavelet transform ψ` followed by a modulus. A joint time-frequency filtering of min L = Linv + λLMM. (4) G 2 this log-spectrogram with the filters hξ⊗ψ`0 regularizes the The inverse problem loss Linv computes the reconstruc- time-frequency deformations of the signal. The low-pass tion error on each training example xi from its embedding convolution with φJ Gaussianizes the resulting tensor. zi computed with the scattering transform SJ followed by the linear whitening operator L. The reconstruction er- ror is calculated over scattering coefficients computed at ing L, which results in the embedding Z[2J n]. Figure 3 a scale 2K < 2J : sketches the different computational steps. This transform N relies on priors on musical signals in order to build a time- 1 X J L = kS (x ) − S G(z )k with z = LS (x ). dependent vector representation SJ X[2 n] which is ap- inv N K i K i 1 i J i i=1 proximately Gaussian and linearizes small time-frequency (5) deformations. 1 The l norm promotes the sparsity of the responses, while Musical signals admit a sparse decomposition in time- the scattering SK allows to generate signals which may be frequency representations with a spectrogram. Here, we locally deformed, but are perceptually similar to the orig- first compute a spectrogram with sampled inal ones. In order to avoid useless computations, we do on a logarithmic scale thanks to a wavelet filterbank −1 not apply L zi but directly input the vectors SJ (xi) to 1 {ψ` }0≤`

LMM = SK xi[t] − SK G(zi)[t] (6) formations This is obtained with a joint 2D filtering of the spectrogram in the time and log-frequency axis. One The codes {zi} correspond to a batch of random vectors which is renewed at each iteration of the gradient descent can prove that the resulting representation is Lipschitz- algorithm. continuous to these deformations, while preserving invert- ibility thanks to the use of filters spanning all the energy The loss LMM is similar to the Maximum Dis- crepancy regularization introduced in [19]. The moment of the signal [1]. This will imply a form a linearization of matching term can be interpreted as a distance with a scat- these deformations, paving the way for meaningful arith- tering transform kernel [8]. However, in this case it can metic in the latent space. directly be implemented as a difference of moments. The time-frequency filters are built as a separable prod- 2 uct hξ ⊗ ψ`0 of frequential filters hξ and temporal fil- 2 ters ψ 0 . The frequential filters h are localized Fourier 4. WHITENED TIME-FREQUENCY SCATTERING ` ξ atoms with a Hann window whose size P matches one oc- This section details the time-frequency scattering trans- tave, P = Q = 12. The convolution is computed in half- 2 form SJ (X) originally introduced in [1] and its whiten- overlaps over the frequency axis. The temporal filters ψ`0 330 Proceedings of the 19th ISMIR Conference, Paris, France, September 23-27, 2018

are also Gammatone wavelets, with only Q2 = 1 wavelet available at http://github.com/AndreuxMath/ per octave. After performing the convolution, a modulus ismir2018, where the reader may also find the audio non-linearity is also applied in order to remove the local recordings corresponding to the figures. phase, thereby regularizing the representation. The time-frequency scattering transform SJ is com- The time-frequency filtering of the log-spectrogram re- puted with an averaging window of size 2J = 210 = 1024. sults in a large tensor with one temporal axis. Its co- It is implemented on GPU with a code inspired from [17]. efficients are sparse along the channel axis and typically In the case of the loss (4), we employ a first-order scat- decorrelate as they get farther apart in time. To obtain a tering SK , which that it is an averaged scalogram, variable which is more Gaussian, we use the central limit with an averaging window of size 2K = 25 = 32. The fil- theorem which says that the averaging of a large number of terbanks are normalized so as to have responses of average independant variables converges towards a Gaussian ran- equal magnitude in each band over the training dataset. dom variable. We perform this averaging with a window The architecture of the SIN G is defined as follows. The size 2J which should be larger than the typical decorrela- whitening operator L has a past size M = 4. All subse- tion length in order to average over enough independent quent convolutions have a kernel size equal to 7. These val- events. This averaging is carried out with a low-pass filter ues were not tuned: results could likely be improved with φJ along the temporal axis. It is followed by a subsam- a careful hyperparameter search. Each network is trained pling by 2J in order to remove redundant information. by Adam [12] with a learning rate of 5 × 10−4 for 1200 The time-frequency scattering transform SJ X is de- epochs and batches of size 128. fined as: We use two different musical datasets: NSynth [5] and Beethoven [14]. NSynth is a dataset consisting of anno- J h J 1 J SJ X[2 n] = x ? φJ [2 n], |x ? ψ` | ? hξ ? φJ [2 n], tated musical notes from multiple instruments, thereby al- 2 J i lowing to perform carefully controlled transformation ex- |x ? ψ | ? (h ⊗ ψ 0 ) ? φ [2 n] , (8) ` ξ ` J periments. All recordings begin with the onset of the note and last 4s. We restrict ourselves to two types of acoustic for all `, ξ, `0, where convolutions with h should be in- ξ instruments, keyboards and flutes, totalling 40 different in- terpreted along the frequential ` axis and other convolu- struments with MIDI pitches ranging 20 − 110, leading to tions along time. S X is subsampled in time by a fac- J a varied and well-balanced dataset. In the original dataset, tor 2J . Each vector S X[2J n] has dimension k = J J instruments of the training and testing sets do not overlap. 1 + Q(2J − 3) + Q(J − 2)2. For J = 10 and Q = 12, In this paper, we use an alternative split based on the veloc- this amounts to 973 channels. The scale 2J is chosen as ity’s attributes of the training samples: for each instrument, a trade-off between the Gaussianization condition which a random velocity is picked to define the test set. We only improves as J increases, and the stability of the scattering use the first 2s of the recordings, as they concentrate most invertibility which improves when J decreases. of the energy of the signals. Thanks to the local averaging, the vectors S X tend to J The Beethoven dataset is closer to an actual musical a Gaussian distribution. However, this distribution might composition than NSynth, insofar as it consists in 8s ex- not be white, i.e. it has temporal and channel-wise cor- tracts of Beethoven’s piano sonata. Therefore, it is a good relations. The whitening operator L is a causal vector testbed for music generation experiments. We use the autoregressive linear filter which removes this correlation train-test split provided by the authors. structure. It is trained by minimizing a prediction error For both datasets, the amplitudes of all recordings are of S X[2J n] given previous vectors S X[2J (n − m)] J J normalized in [−1, 1]. The sampling rate is reduced from for 1 ≤ m ≤ M. The whitening operator L outputs 16000Hz to 4096Hz so as to reduce the computational the innovations of the fitted vector autoregressive process complexity. It is very likely that the quality of the syn- {Z[2J n]} , which have an approximately Gaussian white n thesis could be improved by increasing this sampling rate. distribution. 5.2 Waveform generation 5. NUMERICAL EXPERIMENTS We first show that the MM-SIN generator is able to recon- We show that the MM-SIN is able to reconstruct wave- struct and to generate realistic musical samples. forms from their embeddings and generate new decent In Figure 4, we display two reconstructions of wave- waveforms from noise. Furthermore, we show that it is forms from their embeddings, along with the correspond- possible to manipulate low-level attributes of sounds such ing log-, for each of the studied datasets. This as pitch with a simple arithmetic in the embedding. In shows that the network is able to generalize to a test set, addition, a simple arithmetic in the embedding allows to and to adapt to the specifics of a given dataset. The recon- merge the contents of different inputs, while preserving the struction is not perfect. The network can introduce small musical structure of the resulting signal. time-frequency deformations because of the scattering en- coder and the use of a scattering loss. As witnessed in the 5.1 Methods log-spectrograms, the time-frequency content of the sig- We describe the numerical details which lay out experi- nals is correctly retrieved, and perceptually the two signals ments. The source code supporting experiments is freely sound similar, up to minor artifacts. Proceedings of the 19th ISMIR Conference, Paris, France, September 23-27, 2018 331

Figure 4. Reconstruction with the MM-SIN G. Left: Original test sample X. Right: Reconstruction G(Z) for Z = LSJ X. Top: NSynth dataset. Bottom: Beethoven dataset. A different network was trained for each dataset.

Table 1 provides quantitative reconstruction results on the Beethoven dataset. Training a network with the mean- 2 square error (MSE) metric kxi − G(zi)k2 instead of the proposed perceptual metric within the inverse problem loss (5) negatively impacts results, both on the training and testing sets. Further, the moment-matching term has a pos- itive effect on the reconstruction: even though it degrades Figure 5. Musical signal G(Z) generated from a white reconstruction on the training set, it improves the general- noise Z, where G is learned on the Beethoven dataset. ization on the test set both in absolute and relative terms. Each line corresponds to an independent sample obtained We now investigate the ability of the network to gen- from a different white noise realization. The resulting sig- erate new waveforms from Gaussian white noise. Fig- nals last about 4s. ure 5 displays several samples generated from white noise through a network trained on the Beethoven dataset. De- σ = 6.69, whereas σ = 3.51 for the distribution generated spite the input being a pure white noise, the network is able from white noise. This shows that the generated samples to generate samples which alternate silences and more ac- exhibit a non-negligible variability, even though it is lower tive phases. Further, the fundamental frequency which is than the one expressed in the training set. played varies through the samples. The effect of the moment matching loss (6) on the gen- In order to measure the variability of the generated erated samples is difficult to assess qualitatively, so we samples, we measure the spread σ of the distribution of resort to quantitative measures. In the case of a network the time-averaged scattering coefficients SK X of the sam- trained without this loss, the moment matching distance ples. This spread corresponds to the average Euclidean between generated samples and the training set was equal distance between the time-averaged scattering of the wave- to 38.7, whereas the same distance was equal to 0.176 forms and the average scattering coefficients of this distri- when also optimizing this loss. As a comparison, the test- bution. In the case of the training distribution, we obtained ing set has a distance of 0.334 with respect to the training set. Thus, using this loss term brings generated samples much closer to the natural signals’ statistics. Loss Linv Linv Linv + λLMM Metric MSE SK SK 5.3 Pitch modification Train error 0.56 0.16 0.23 Test error 0.77 0.37 0.31 We now study the ability of the algorithm to transform the Gap test/train pitch of musical signals with an arithmetic operations in 1.36 3.53 1.21 (dB) the latent space. We use the NSynth dataset, whose care- ful construction allows to perform modifications with fixed Table 1. Reconstruction errors on the Beethoven dataset, factors of variability. In the test set, we pick two sam- expressed in terms of the perceptual loss (5). MSE denotes ples belonging to the same instrument, but with a pitch 2 the mean-square error metric kxi − G(zi)k2. Using the separated by 5 MIDI scales. We compute their embed- perceptual metric for training instead of the MSE metric dings Z1 and Z2, their mean embedding (Z1 + Z2)/2, and reduces the error. Further, adding the moment-matching reconstruct the corresponding signals with the generator: term during training improves the reconstruction results G(Z1), G(Z2) and G((Z1 + Z2)/2). and the generalization. The results are displayed in Figure 6. The interpolation 332 Proceedings of the 19th ISMIR Conference, Paris, France, September 23-27, 2018

Figure 7. Interpolations in the latent and signal space. Top two signals: G(Z1) and G(Z2), where Z1,Z2 are Gaussian white noise realizations and G is trained on the Beethoven dataset. Bottom left: Latent interpolation G((Z1 +Z2)/2). Bottom right: (G(Z1) + G(Z2))/2. The latent interpo- lation is able to merge both signals while preserving the musical structure.

5.4 Waveform interpolation Let us show results when interpolating waveforms from the Beethoven dataset, which have a high density of musical events. We take two random white noise realizations Z1 and Z2, and compare the effect on the waveforms of an interpolation in the latent space and in the signal space, with a network G trained on the Beethoven dataset. Figure 6. Pitch interpolation. Left column: G(Z1). Mid- The results are represented in Figure 7. The top two dle column: G((Z1 + Z2)/2). Right column: G(Z2). Z1 signals are the original signals G(Z1) and G(Z2), while and Z2 are the embeddings of samples from the test set. the bottom left is the latent interpolation G((Z1 + Z2)/2) The generator interpolates the fundamental frequency with and the bottom right the linear interpolation (G(Z1) + a simple arithmetic. The frequential displacement from left G(Z2))/2. The latent interpolation incorporates patterns to right corresponds to 5 MIDI scales. from both signals but it respects the musical structure. It has successive musical notes with their harmonics, and it recovers a sound with silences. On the opposite, the linear in the latent space does not result in a linear interpolation interpolation merges both signals, which eliminates the si- which would double the number of harmonics. It yields lence regions while producing a cluttered log-spectrogram. one fundamental frequency in each case. Furthermore, this fundamental frequency is indeed interpolated by this sim- 6. CONCLUSION ple arithmetic. Observe that this is also the case of the This paper introduces a causal musical synthesis network partials, as can be seen in particular in the bottom exam- optimized through an inverse problem and which thus in- ple. However, this interpolation suffers from some arti- volves no learned encoder or discriminator. The encoder is facts. For instance, in the middle example, the partials at defined from time-frequency signal priors in order to Gaus- highest frequencies are cluttered and the resulting signal sianize the input signal. The generator network maps back misses harmonicity. Yet, these results showcase the ability the resulting codes to raw waveforms. This network inverts to transform signals via linear interpolations in the latent the encoder and generates new signals whose scattering space with a simple unsupervised learning procedure and a moments match those of the original signals. The resulting predefined embedding. system synthesizes new realistic musical signals and per- The algorithm owes the ability to perform such pitch forms the transformation of low-level attributes, such as interpolations to the time-frequency scattering transform pitch, by simple linear combinations in the latent space. SJ used as an encoder, which regularizes small time- Synthesized signals do not reach the quality of state-of- frequency deformations. As such, the pitch interval on the-art generating architectures but these first results show which the interpolations can be performed is bounded by that this approach is a new promising avenue to synthesize the size P of the Hann window used to filter the scalogram audio signals directly from Gaussian white noise, without along the frequency axis. learning encoders or discriminators. Proceedings of the 19th ISMIR Conference, Paris, France, September 23-27, 2018 333

7. ACKNOWLEDGEMENTS [13] D. P. Kingma and M. Welling. Auto-encoding varia- tional bayes. In International Conference on Learning This work is supported by the ERC InvariantClass grant Representations, 2014. 320959 and an AMX grant from the MNESR. [14] S. Mehri, K. Kumar, I. Gulrajani, R. Kumar, S. Jain, J. Sotelo, A. Courville, and Y. Bengio. SampleRNN: 8. REFERENCES An unconditional end-to-end neural audio generation [1] J. Anden, V. Lostanlen, and S. Mallat. Joint time- model. In International Conference on Learning Rep- frequency scattering for audio classification. In Proc. resentations, 2017. of IEEE MLSP, 2015. [15] A. Van Den Oord, S. Dieleman, H. Zen, K. Simonyan, [2] Tomas` Angles and Stephane´ Mallat. Generative net- O. Vinyals, A. Graves, N. Kalchbrenner, A. Senior, and works as inverse problems with scattering transforms. K. Kavukcuoglu. WaveNet: A generative model for In International Conference on Learning Representa- raw audio. arXiv preprint arXiv:1609.03499, 2016. tions, 2018. [16] A. Van Den Oord, Y. Li, I. Babuschkin, K. Simonyan, O. Vinyals, K. Kavukcuoglu, G. van den Driessche, [3] M. Blaauw and J. Bonada. Modeling and transforming E. Lockhart, L. C. Cobo, F. Stimberg, et al. Parallel speech using variational autoencoders. In Interspeech, WaveNet: Fast high-fidelity speech synthesis. arXiv pages 1770–1774, 2016. preprint arXiv:1711.10433, 2017.

[4] J. Chorowski, R.J. Weiss, R. A. Saurous, and S. Ben- [17] E. Oyallon, E. Belilovsky, and S. Zagoruyko. Scal- gio. On using backpropagation for speech texture gen- ing the scattering transform: Deep hybrid networks. In International Confer- eration and voice conversion. In Proc. of ICCV, 2017. ence on Audio and Speech Processing (ICASSP), 2018. [18] A. Radford, L. Metz, and R. Chintala. Unsupervised [5] J. Engel, C. Resnick, A. Roberts, S. Dieleman, representation learning with deep convolutional gener- M. Norouzi, D. Eck, and K. Simonyan. Neural au- ative adversarial networks. In International Conference dio synthesis of musical notes with WaveNet autoen- on Learning Representations, 2016. coders. In Proceedings of the 34th International Con- ference on , pages 1068–1077, 2017. [19] I. Tolstikhin, O. Bousquet, S. Gelly, and B. Scholkopf.¨ Wasserstein auto-encoders. In International Confer- [6] L. Gatys, A. S. Ecker, and M. Bethge. Texture synthe- ence on Learning Representations, 2018. sis using convolutional neural networks. In Advances in Neural Information Processing Systems, pages 262– [20] A. Venkitaraman, A. Adiga, and C. S. Seelaman- 270, 2015. tula. Auditory-motivated gammatone wavelet trans- form. , 94:608–619, 2014. [7] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y.Bengio. Generative adversarial nets. Advances in Neural Infor- mation Processing Systems, pages 2672–2680, 2014.

[8] A. Gretton, K. M. Borgwardt, M. Rasch, B. Scholkopf,¨ and A. J. Smola. A kernel method for the two-sample- problem. In Advances in neural information processing systems, pages 513–520, 2007.

[9] E. Grinstein, N. Duong, A. Ozerov, and P. Perez.´ Audio style transfer. HAL preprint hal-01626389, 2017.

[10] W.-N. Hsu, Y. Zhang, and J. Glass. Learning latent rep- resentations for speech generation and transformation. In Interspeech, pages 1273–1277, 2017.

[11] S. Ioffe and C. Szegedy. Batch normalization: Accel- erating deep network training by reducing internal co- variate shift. In International Conference on Machine Learning, pages 448–456, 2015.

[12] D. P. Kingma and J. Ba. Adam: A method for stochas- tic optimization. arXiv preprint arXiv:1412.6980, 2014.