Leveraging Native Language Speech for Accent Identification Using Deep Siamese Networks

Leveraging Native Language Speech for Accent Identification Using Deep Siamese Networks

LEVERAGING NATIVE LANGUAGE SPEECH FOR ACCENT IDENTIFICATION USING DEEP SIAMESE NETWORKS Aditya Siddhanty, Preethi Jyothix, Sriram Ganapathyz∗ yCarnegie Mellon University, Pittsburgh, PA, USA xIndian Institute of Technology Bombay, Mumbai, India zIndian Institute of Science, Bengaluru, India ABSTRACT seen in the training data. The accented nature of speech The problem of automatic accent identification is important can be primarily attributed to the influence of the speaker’s for several applications like speaker profiling and recogni- native language. In this work we focus on the problem of tion as well as for improving speech recognition systems. accent identification, where the user’s native language is au- The accented nature of speech can be primarily attributed to tomatically determined from their non-native speech. This the influence of the speaker’s native language on the given can be viewed as a first step towards building accent-aware speech recording. In this paper, we propose a novel accent voice-driven systems. identification system whose training exploits speech in native Accent identification from non-native speech bears re- languages along with the accented speech. Specifically, we semblance to the task of language identification [1]. How- develop a deep Siamese network based model which learns ever, accent identification is a harder task as many cues about the association between accented speech recordings and the the speaker’s native language are lost or suppressed in the native language speech recordings. The Siamese networks non-native speech. Nevertheless, one may expect that the are trained with i-vector features extracted from the speech speaker’s native language is reflected in the acoustics of the recordings using either an unsupervised Gaussian mixture individual phones used in non-native language speech, along model (GMM) or a supervised deep neural network (DNN) with pronunciations of words and grammar. In this work, we model. We perform several accent identification experiments focus on the acoustic characteristics of an accent induced by using the CSLU Foreign Accented English (FAE) corpus. a speaker’s native language. In these experiments, our proposed approach using deep Our main contributions: Siamese networks yield significant relative performance im- provements of 15.4% on a 10-class accent identification task, • We develop a novel deep Siamese network based model over a baseline DNN-based classification system that uses which learns the association between accented speech GMM i-vectors. Furthermore, we present a detailed error and native language speech. analysis of the proposed accent identification system. • We explore i-vector features extracted using both an un- Index Terms— Accent identification, i-vectors, Deep supervised Gaussian mixture model (GMM) and a su- Siamese networks, Multi-lingual modeling. pervised deep neural network (DNN) model. arXiv:1712.08992v2 [cs.CL] 18 Jun 2018 • We present a detailed error analysis of the proposed 1. INTRODUCTION system which reveals that the confusions among accent predictions are contained within the language family of Over the recent years, many of voice-driven technologies the corresponding native language. have achieved significant robustness needed for mass deploy- ment. This is largely due to significant advances in automatic Section 3 outlines the i-vector feature extraction process. speech recognition (ASR) technologies and deep learning Section 4 describes our Siamese network-based model for ac- algorithms. However, the variability in speech accents pose cent identification. Our experimental results are detailed in a significant challenge to state-of-the-art speech systems. In Section 5 and Section 6 provides an error analysis of our pro- particular, large sections of the English-speaking population posed approach. in the world face difficulties interacting with voice-driven agents in English due to the mis-match in speech accents 2. RELATED WORK ∗This work was carried out with the help of a research grant awarded by Microsoft Research India (MSRI) for the Summer Workshop on Artificial The prior work on foreign accent identification has drawn in- Social Intelligence. spiration from techniques used in language identification [2]. The phonotactic model based approaches [3] and acoustic ponent c. The sufficient statistics are then computed as, model based approaches [4] have been explored for accent H(s) identification in the past. More recently, i-vector based repre- X N (s) = p (cjxs) sentations, which is part of the state-of-the-art speaker recog- c λ i i=1 nition [5] and language recognition [6] systems, have been (2) H(s) applied to the task of accent recognition. The i-vector sys- X S (s) = p (cjxs)(xs − µ ) tems that used GMM-based background models were found X;c λ i i c i=1 to outperform other competitive baseline systems [7, 8, 9]. In the recent years, the language recognition systems Let N(s) denote the D×D block diagonal matrix with diag- and speaker recognition systems have shown promising re- onal blocks N1(s)I, N2(s)I,..,NC (s)I where I is the F × F sults with the use of deep neural network (DNN) model identity matrix. Let SX (s) denote the D × 1 vector obtained based i-vector extraction [10, 11]. Also, none of the previ- by splicing SX;1(s),..,SX;C (s). ous approaches have exploited speech in native languages It can be easily shown [12] that the posterior distribution of the i-vector pλ(y(s)jX (s)) is Gaussian with covariance while training accent identification systems to the best of our −1 −1 ∗ −1 knowledge. This work attempts to develop accent recognition l (s) and mean l (s)V Σ SX (s), where systems using both these components. l(s) = I + V ∗Σ−1N(s)V (3) The optimal estimate for the i-vector y(s) obtained as 3. FACTOR ANALYSIS FRAMEWORK FOR argmaxy pλ(y(s)jX (s)) is given by the mean of the poste- I-VECTOR EXTRACTION rior distribution. For re-estimating the V matrix, the maximization of the The techniques outlined here are derived from previous work expected value of the log-likelihood function (EM algorithm), on joint factor analysis (JFA) and i-vectors [12, 13, 14]. We gives the following relation [12], follow the notations used in [12]. The training data from all S S the speakers is used to train a GMM with model parameters X ∗ X ∗ N(s) V E y(s)y (s) = SX (s)E y (s) (4) λ = fπc; µc; Σcg where πc, µc and Σc denote the mixture s=1 s=1 component weights, mean vectors and covariance matrices re- spectively for c = 1; ::; C mixture components. Here, µ is a where E[:] denotes the posterior expectation operator. The c solution for Eq. (4) can be computed for each row of V . Thus, vector of dimension F and Σc is of assumed to be a diagonal matrix of dimension F × F . the i-vector estimation is performed by iterating between the estimation of posterior distribution and the update of the total variability matrix (Eq. (4)). 3.1. GMM-based i-vector Representations 3.2. DNN i-vectors Let M0 denote the universal background model (UBM) su- pervector which is the concatenation of µc for c = 1; ::; C Instead of using a GMM-UBM based computation of i- and is of dimensionality D × 1 (where D = C · F ). Let Σ vectors, we can also use DNN-based context dependent state denote the block diagonal matrix of size D × D whose diago- (senone) posteriors to generate the sufficient statistics used in s nal blocks are Σc. Let X (s) = fxi ; i = 1 ; :::; H (s)g denote the i-vector computation [15, 10]. The GMM mixture com- the low-level feature sequence for input recording s where i ponents will be replaced with the senone classes present at s denotes the frame index. Here H(s) denotes the number of the output of the DNN. Specifically, pλ(cjxi ) used in Eq. (2) s frames in the recording. Each xi is of dimension F × 1. is replaced with the DNN posterior probability estimate of s Let M(s) denote the recording supervector which is the the senone c given the input acoustic feature vector xi and concatenation of speaker adapted GMM means µc(s) for c = the total number of senones is the parameter C. The other pa- 1; ::; C for the speaker s. Then, the i-vector model is, rameters of the UBM model λ = fπc; µc; Σcg are computed as M(s) = M0 + V y(s) (1) PS PH(s) s s=1 i=1 p(cjxi ) where V denotes the total variability matrix of dimension πc = PC PS PH(s) s D × M and y(s) denotes the i-vector of dimension M. The c=1 s=1 i=1 p(cjxi ) i-vector is assumed to be distributed as N (0; I). PS PH(s) s s s=1 i=1 p(cjxi )xi µc = (5) In order to estimate the i-vectors, the iterative EM algo- PC PS PH(s) p(cjxs) rithm is used. We begin with a random initialization of the c=1 s=1 i=1 i s PS PH(s) p(cjxs)(xs − µ )(xs − µ )∗ total variability matrix V . Let pλ(cjxi ) denote the alignment s=1 i=1 i i c i c Σc = probability of assigning the feature vector xs to mixture com- PC PS PH(s) s i c=1 s=1 i=1 p(cjxi ) LANGUAGE Accented speech Native language Training Dev Test speech Margin-based BP 92 30 31 198 Contrastive Loss HI 66 22 22 206 FA 50 17 16 182 F(ia) F(in) GE 55 18 18 161 HU 51 17 17 187 IT 34 11 12 168 MA 52 18 18 189 Deep Deep RU 44 14 14 172 Feedforward Feedforward SP 31 9 10 140 Neural Network Neural Network TA 37 13 12 128 shared Table 1.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    8 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us