Improving Sequence-To-Sequence Speech Recognition Training with On-The-Fly Data Augmentation

Improving Sequence-To-Sequence Speech Recognition Training with On-The-Fly Data Augmentation

IMPROVING SEQUENCE-TO-SEQUENCE SPEECH RECOGNITION TRAINING WITH ON-THE-FLY DATA AUGMENTATION Thai-Son Nguyen1, Sebastian Stuker¨ 1, Jan Niehues2, Alex Waibel1 1 Institute for Anthropomatics and Robotics, Karlsruhe Institute of Technology 2Department of Data Science and Knowledge Engineering (DKE), Maastricht University ABSTRACT tation method together with a long training schedule to reduce overfitting, they have achieved a large gain in performance su- Sequence-to-Sequence (S2S) models recently started to show perior to many modifications in network architecture. state-of-the-art performance for automatic speech recognition To date, there have been different sequence-to-sequence (ASR). With these large and deep models overfitting remains encoder-decoder models [12, 13] reporting superior perfor- the largest problem, outweighing performance improvements mance over the HMM hybrid models on standard ASR bench- that can be obtained from better architectures. One solution marks. While [12] uses Long Short-Term Memory (LSTM) to the overfitting problem is increasing the amount of avail- networks, for both encoder and decoder, [13] employs self- able training data and the variety exhibited by the training attention layers to construct the whole S2S network. data with the help of data augmentation. In this paper we ex- In this paper, we investigate three on-the-fly data aug- amine the influence of three data augmentation methods on mentation methods for S2S speech recognition, two of which the performance of two S2S model architectures. One of the are proposed in this work and the last was recently discov- data augmentation method comes from literature, while two ered [12]. We contrast and analyze both LSTM-based and other methods are our own development – a time perturbation self-attention S2S models that were trained with the proposed in the frequency domain and sub-sequence sampling. Our ex- augmentation methods by performing experiments on the periments on Switchboard and Fisher data show state-of-the- Switchboard (SWB) and Fisher telephone conversations task. art performance for S2S models that are trained solely on the We found that not only the two models behave differently speech training data and do not use additional text data. with the augmentation methods, but also the combination of Index Terms— Sequence-to-sequence, Self-attention, different augmentation methods and network architectures Data Augmentation, Speed Perturbation, Sub-sequence can significantly reduce word error rate (WER). Our final S2S model achieved a WER of 5.2% on the SWB test set and 1. INTRODUCTION 10.2% WER on the Callhome (CH) test set. This is already on par with human performance. We made our source code In automatic speech recognition (ASR), data augmentation available as open source1, as well as the model checkpoints has been used for producing additional training data in order of the experiments in this paper. to increase the quality of the training data, i.e. their amount and variety. This then improves the robustness of the models 2. DATA AUGMENTATION and avoids overfitting. As in [1, 2], both unsupervised and artificial training data has been augmented to improve model We investigated three data augmentation methods for sequence- training in low-resource conditions. The addition of training to-sequence encoder-decoder models. The first two modify data with perturbation of the vocal tract length [3] or audio the input sequences from different inspirations and aim to speed [4] helps ASR models to be robust to speaker vari- improve the generalization of the log-mel spectrogram en- ations. Simulated far-field speech [5] and noisy speech [6] coder. The third approach improves the decoder by adding have been used to supplement clean close-talk training data. sub-samples of target sequences. All of the proposed methods Sequence-to-sequence attention-based models [7, 8] were are computationally cheap and can be performed on-the-fly introduced as a promising approach for end-to-end speech and can be optimized together with the model. recognition. Several advances [9, 10, 11] have been proposed for improving the performance of S2S models. While many 2.1. Dynamic Time Stretching works focus on designing better network architectures, the au- Many successful S2S models adopt log-mel frequency fea- thors in [12] have recently pointed out that overfitting is the tures as input. In the frequency domain, one major difficulty most critical issue when training their sequence-to-sequence model on popular benchmarks. By proposing a data augmen- 1The source code is available at https://github.com/thaisonngn/pynn 978-1-5090-6631-5/20/$31.00 ©2020 IEEE 7689 ICASSP 2020 Authorized licensed use limited to: KIT Library. Downloaded on August 20,2020 at 13:46:50 UTC from IEEE Xplore. Restrictions apply. SS Variant 2 for the recognition models is to recognize temporal patterns SS Variant 3 which occur with varying duration. To make the models more robust to temporal variations, the addition of audio data with huh you learn something every day speed perturbation in the time domain such as in [4] has been T1 T2 T3 T4 T5 T6 shown to be effective. In contrast, in our work we manipulate directly the time series of the frequency vectors which are the SS Variant 1 features of our S2S models, in order to achieve the effect of Fig. 1. Sub-sequence Sampling. speed perturbation. Specifically, given a sequence of consecu- tive feature vectors seq, we stretch every window of w feature aligned. The alignment can be also estimated automatically vectors by a factor of s obtained from an uniform distribution via the traditional force-alignment process. Taking advantage of range [low; high], resulting in a new window of size w ∗ s. of this property, we experiment with the ability to sub-sample There are different approaches to perform window stretching, training utterances to have more variants of target sequences. in this work we adopt nearest-neighbor interpolation for its Since the approach of generating sub-sequences with arbi- speed, as it is fast enough to augment many speech utterances trary lengths does not work, we propose a constraint sampling on a CPU while model training for other utterances is being depicted in Figure 1. Basically, given an utterance, we allow performed on a GPU. The dynamic time stretching algorithm three different variants of sub-sequences with equal distribu- is implemented by the following python code: tions. The first and second variants constraint sub-sequences to having either the same start or end as the original sequence def t i m e stretch(seq ,w,low=0.8,high=1.25): while the third variant needs to have their start and end point ids = None; time l e n = l e n ( seq ) within the utterance. All sub-sequences need to have at least f o r i in range ( t i m e l e n / / w + 1 ) : half the size of the original sequence. During training, we s = random.uniform(low, high) randomly select a training sample with probability alpha e = min ( t i m e l e n , w∗( i + 1 ) ) r = numpy.arange(w∗ i , e −1, s ) and replace it with one of the sampled sub-sequence vari- r = numpy . round (r).astype( i n t ) ants. We also allow static mode in which only one fixed ids = numpy.concatenate((ids , r)) instance of sub-sequence per utterance per variant is gener- return seq [ i d s ] ated. This mode is equivalent to statically adding three sets of sub-sequences to the original training set. 2.2. SpecAugment 3. MODEL Recently [12] found that LSTM-based S2S models tend to We use two different S2S models to investigate the on-the-fly overfit easily to the training data, even when regularization data augmentation methods proposed in Section 2. In the first methods such as Dropout [14] are applied. Inspired by the model, we use LSTMs and a new approach for building the data augmentation from computer vision, [12] proposed to de- decoder network. For the second model, we follow the work form the spectrogram input with three cheap operations such in [13] to replace LSTMs with deep self-attention layers in as time warping, frequency and time masking before feeding both the encoder and decoder. it to their sequence-to-sequence models. Time warping shifts a random point in the spectrogram input with a random dis- 3.1. LSTM-based S2S tance, while frequency and time masking apply zero masks to some consecutive lines in both the frequency and the time Before the LSTM layers in the encoder, we place a two-layer dimensions. In this work, we study the two most effective Convolutional Neural Network (CNN) with 32 channels and a operations which are the frequency and time masking. Ex- time stride of two to down-sample the input spectrogram by a perimenting on the same dataset, we benefit from optimized factor of four. In the decoder, we adopt two layers of unidirec- configurations in [12]. Specifically, we consider T 2 [1; 2; 3] tional LSTMs as language modeling for the sequence of sub- – the number of times that both, frequency and time masking, word units and the approach of Scaled Dot-Product (SDP) At- are applied. For each time, f consecutive frequency channels tention [15] to generate context vectors from the hidden states and t consecutive time steps are masked where f and t are of the two LSTM networks. Specifically, our implementation randomly chosen from [0; 70] and [0; 7]. When T = 2, we for LSTM-based S2S works as follows: obtain a similar setting for 40 log-mel features as the SWB enc = LST M(CNN(spectrogram)) mild (SM) configuration in [12]. We experimentally find T emb = Embedding(subwords) for different model architectures in our experiments.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    5 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us