Determining the Temporal Interval of Segments with the Help of F0 Contours
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Recognition of Nasalized and Non- Nasalized Vowels
RECOGNITION OF NASALIZED AND NON- NASALIZED VOWELS Submitted By: BILAL A. RAJA Advised by: Dr. Carol Espy Wilson and Tarun Pruthi Speech Communication Lab, Dept of Electrical & Computer Engineering University of Maryland, College Park Maryland Engineering Research Internship Teams (MERIT) Program 2006 1 TABLE OF CONTENTS 1- Abstract 3 2- What Is Nasalization? 3 2.1 – Production of Nasal Sound 3 2.2 – Common Spectral Characteristics of Nasalization 4 3 – Expected Results 6 3.1 – Related Previous Researches 6 3.2 – Hypothesis 7 4- Method 7 4.1 – The Task 7 4.2 – The Technique Used 8 4.3 – HMM Tool Kit (HTK) 8 4.3.1 – Data Preparation 9 4.3.2 – Training 9 4.3.3 – Testing 9 4.3.4 – Analysis 9 5 – Results 10 5.1 – Experiment 1 11 5.2 – Experiment 2 11 6 – Conclusion 12 7- References 13 2 1 - ABSTRACT When vowels are adjacent to nasal consonants (/m,n,ng/), they often become nasalized for at least some part of their duration. This nasalization is known to lead to changes in perceived vowel quality. The goal of this project is to verify if it is beneficial to first recognize nasalization in vowels and treat the three groups of vowels (those occurring before nasal consonants, those occurring after nasal consonants, and the oral vowels which are vowels that are not adjacent to nasal consonants) separately rather than collectively for recognizing the vowel identities. The standard Mel-Frequency Cepstral Coefficients (MFCCs) and the Hidden Markov Model (HMM) paradigm have been used for this purpose. The results show that when the system is trained on vowels in general, the recognition of nasalized vowels is 17% below that of oral vowels. -
Acoustic-Phonetics of Coronal Stops
Acoustic-phonetics of coronal stops: A cross-language study of Canadian English and Canadian French ͒ Megha Sundaraa School of Communication Sciences & Disorders, McGill University 1266 Pine Avenue West, Montreal, QC H3G 1A8 Canada ͑Received 1 November 2004; revised 24 May 2005; accepted 25 May 2005͒ The study was conducted to provide an acoustic description of coronal stops in Canadian English ͑CE͒ and Canadian French ͑CF͒. CE and CF stops differ in VOT and place of articulation. CE has a two-way voicing distinction ͑in syllable initial position͒ between simultaneous and aspirated release; coronal stops are articulated at alveolar place. CF, on the other hand, has a two-way voicing distinction between prevoiced and simultaneous release; coronal stops are articulated at dental place. Acoustic analyses of stop consonants produced by monolingual speakers of CE and of CF, for both VOT and alveolar/dental place of articulation, are reported. Results from the analysis of VOT replicate and confirm differences in phonetic implementation of VOT across the two languages. Analysis of coronal stops with respect to place differences indicates systematic differences across the two languages in relative burst intensity and measures of burst spectral shape, specifically mean frequency, standard deviation, and kurtosis. The majority of CE and CF talkers reliably and consistently produced tokens differing in the SD of burst frequency, a measure of the diffuseness of the burst. Results from the study are interpreted in the context of acoustic and articulatory data on coronal stops from several other languages. © 2005 Acoustical Society of America. ͓DOI: 10.1121/1.1953270͔ PACS number͑s͒: 43.70.Fq, 43.70.Kv, 43.70.Ϫh ͓AL͔ Pages: 1026–1037 I. -
On the Acoustic and Perceptual Characterization of Reference Vowels in a Cross-Language Perspective Jacqueline Vaissière
On the acoustic and perceptual characterization of reference vowels in a cross-language perspective Jacqueline Vaissière To cite this version: Jacqueline Vaissière. On the acoustic and perceptual characterization of reference vowels in a cross- language perspective. The 17th International Congress of Phonetic Sciences (ICPhS XVII), Aug 2011, China. pp.52-59. halshs-00676266 HAL Id: halshs-00676266 https://halshs.archives-ouvertes.fr/halshs-00676266 Submitted on 8 Mar 2012 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. ICPhS XVII Plenary Lecture Hong Kong, 17-21 August 2011 ON THE ACOUSTIC AND PERCEPTUAL CHARACTERIZATION OF REFERENCE VOWELS IN A CROSS-LANGUAGE PERSPECTIVE Jacqueline Vaissière Laboratoire de Phonétique et de Phonologie, UMR/CNRS 7018 Paris, France [email protected] ABSTRACT 2. IPA CHART AND THE CARDINAL Due to the difficulty of a clear specification in the VOWELS articulatory or the acoustic space, the same IPA symbol is often used to transcribe phonetically 2.1. The IPA vowel chart, and other proposals different vowels across different languages. On the The first International Phonetic Alphabet (IPA) basis of the acoustic theory of speech production, was proposed in 1886 by a group of European this paper aims to propose a set of focal vowels language teachers led by Paul Passy. -
UC Berkeley Dissertations, Department of Linguistics
UC Berkeley Dissertations, Department of Linguistics Title The Aeroacoustics of Nasalized Fricatives Permalink https://escholarship.org/uc/item/00h9g9gg Author Shosted, Ryan K Publication Date 2006 eScholarship.org Powered by the California Digital Library University of California The Aeroacoustics of Nasalized Fricatives by Ryan Keith Shosted B.A. (Brigham Young University) 2000 M.A. (University of California, Berkeley) 2003 A dissertation submitted in partial satisfaction of the requirements for the degree of Doctor of Philosophy in Linguistics in the GRADUATE DIVISION of the UNIVERSITY OF CALIFORNIA, BERKELEY Committee in charge: John J. Ohala, Chair Keith Johnson Milton M. Azevedo Fall 2006 The dissertation of Ryan Keith Shosted is approved: Chair Date Date Date University of California, Berkeley Fall 2006 The Aeroacoustics of Nasalized Fricatives Copyright 2006 by Ryan Keith Shosted 1 Abstract The Aeroacoustics of Nasalized Fricatives by Ryan Keith Shosted Doctor of Philosophy in Linguistics University of California, Berkeley John J. Ohala, Chair Understanding the relationship of aerodynamic laws to the unique geometry of the hu- man vocal tract allows us to make phonological and typological predictions about speech sounds typified by particular aerodynamic regimes. For example, some have argued that the realization of nasalized fricatives is improbable because fricatives and nasals have an- tagonistic aerodynamic specifications. Fricatives require high pressure behind the suprala- ryngeal constriction as a precondition for high particle velocity. Nasalization, on the other hand, vents back pressure by allowing air to escape through the velopharyngeal orifice. This implies that an open velopharyngeal port will reduce oral particle velocity, thereby potentially extinguishing frication. By using a mechanical model of the vocal tract and spoken fricatives that have undergone coarticulatory nasalization, it is shown that nasal- ization must alter the spectral characteristics of fricatives, e.g. -
Part 1: Introduction to The
PREVIEW OF THE IPA HANDBOOK Handbook of the International Phonetic Association: A guide to the use of the International Phonetic Alphabet PARTI Introduction to the IPA 1. What is the International Phonetic Alphabet? The aim of the International Phonetic Association is to promote the scientific study of phonetics and the various practical applications of that science. For both these it is necessary to have a consistent way of representing the sounds of language in written form. From its foundation in 1886 the Association has been concerned to develop a system of notation which would be convenient to use, but comprehensive enough to cope with the wide variety of sounds found in the languages of the world; and to encourage the use of thjs notation as widely as possible among those concerned with language. The system is generally known as the International Phonetic Alphabet. Both the Association and its Alphabet are widely referred to by the abbreviation IPA, but here 'IPA' will be used only for the Alphabet. The IPA is based on the Roman alphabet, which has the advantage of being widely familiar, but also includes letters and additional symbols from a variety of other sources. These additions are necessary because the variety of sounds in languages is much greater than the number of letters in the Roman alphabet. The use of sequences of phonetic symbols to represent speech is known as transcription. The IPA can be used for many different purposes. For instance, it can be used as a way to show pronunciation in a dictionary, to record a language in linguistic fieldwork, to form the basis of a writing system for a language, or to annotate acoustic and other displays in the analysis of speech. -
3. Acoustic Phonetic Basics
Prosody: speech rhythms and melodies 3. Acoustic Phonetic Basics Dafydd Gibbon Summer School Contemporary Phonology and Phonetics Tongji University 9-15 July 2016 Contents ● The Domains of Phonetics: the Phonetic Cycle ● Articulatory Phonetics (Speech Production) – The IPA (A = Alphabet / Association) – The Source-Filter Model of Speech Production ● Acoustic Phonetics (Speech Transmission) – The Speech Wave-Form – Basic Speech Signal Parameters – The Time Domain: the Speech Wave-Form – The Frequency Domain: simple & complex signals – Pitch extraction – Analog-to-Digital (A/D) Conversion ● Auditory Phonetics (Speech Perception) – The Auditory Domain: Anatomy of the Ear The Domains of Phonetics ● Phonetics is the scientific discipline which deals with – speech production (articulatory phonetics) – speech transmission (acoustic phonetics) – speech perception (auditory phonetics) ● The scientific methods used in phonetics are – direct observation (“impressionistic”), usually based on articulatory phonetic criteria – measurement ● of position and movement of articulatory organs ● of the structure of speech signals ● of the mechanisms of the ear and perception in hearing – statistical evaluation of direct observation and measurements – creation of formal models of production, transmission and perception Abidjan 9 May 2014 Dafydd Gibbon: Phonetics Course 3 Abidjan 2014 3 The Domains of Phonetics: the Phonetic Cycle A tiger and a mouse were walking in a field... Abidjan 9 May 2014 Dafydd Gibbon: Phonetics Course 3 Abidjan 2014 4 The Domains of Phonetics: the Phonetic Cycle Sender: Articulatory Phonetics A tiger and a mouse were walking in a field... Abidjan 9 May 2014 Dafydd Gibbon: Phonetics Course 3 Abidjan 2014 5 The Domains of Phonetics: the Phonetic Cycle Sender: Articulatory Phonetics A tiger and a mouse were walking in a field.. -
Issues in the Distribution of the Glottal Stop
Theoretical Issues in the Representation of the Glottal Stop in Blackfoot* Tyler Peterson University of British Columbia 1.0 Introduction This working paper examines the phonemic and phonetic status and distribution of the glottal stop in Blackfoot.1 Observe the phonemic distribution (1), four phonological operations (2), and three surface realizations (3) of the glottal stop: (1) a. otsi ‘to swim’ c. otsi/ ‘to splash with water’ b. o/tsi ‘to take’ (2) a. Metathesis: V/VC ® VV/C nitáó/mai/takiwa nit-á/-omai/takiwa 1-INCHOAT-believe ‘now I believe’ b. Metathesis ® Deletion: V/VùC ® VVù/C ® VVùC kátaookaawaatsi káta/-ookaa-waatsi INTERROG-sponsor.sundance-3s.NONAFFIRM ‘Did she sponsor a sundance?’ (Frantz, 1997: 154) c. Metathesis ® Degemination: V/V/C ® VV//C ® VV/C áó/tooyiniki á/-o/too-yiniki INCHOAT-arrive(AI)-1s/2s ‘when I/you arrive’ d. Metathesis ® Deletion ® V-lengthening: V/VC ® VV/C ® VVùC kátaoottakiwaatsi káta/-ottaki-waatsi INTEROG-bartender-3s.NONAFFIRM ‘Is he a bartender?’ (Frantz, 1997) (3) Surface realizations: [V/C] ~ [VV0C] ~ [VùC] aikai/ni ‘He dies’ a. [aIkaI/ni] b. [aIkaII0ni] c. [aIkajni] The glottal stop appears to have a unique status within the Blackfoot consonant inventory. The examples in (1) (and §2.1 below) suggest that it appears as a fully contrastive phoneme in the language. However, the glottal stop is put through variety of phonological processes (metathesis, syncope and degemination) that no other consonant in the language is subject to. Also, it has a variety of surfaces realizations in the form of glottalization on an adjacent vowel (cf. -
HCS 7367 Speech Perception
Course web page HCS 7367 • http://www.utdallas.edu/~assmann/hcs6367 Speech Perception – Course information – Lecture notes – Speech demos – Assigned readings Dr. Peter Assmann – Additional resources in speech & hearing Fall 2010 Course materials Course requirements http://www.utdallas.edu/~assmann/hcs6367/readings • Class presentations (15%) • No required text; all readings online • Written reports on class presentations (15%) • Midterm take-home exam (20%) • Background reading: • Term paper (50%) http://www.speechandhearing.net/library/speech_science.php • Recommended books: Kent, R.D. & Read, C. (2001). The Acoustic Analysis of Speech. (Singular). Stevens, K.N. (1999). Acoustic Phonetics (Current Studies in Linguistics). M.I.T. Press. Class presentations and reports Suggested Topics • Pick two broad topics from the field of speech • Speech acoustics • Vowel production and perception perception. For each topic, pick a suitable (peer- • Consonant production and perception reviewed) paper from the readings web page or from • Suprasegmentals and prosody • Speech perception in noise available journals. Your job is to present a brief (10-15 • Auditory grouping and segregation minute) summary of the paper to the class and • Speech perception and hearing loss • Cochlear implants and speech coding initiate/lead discussion of the paper, then prepare a • Development of speech perception written report. • Second language acquisition • Audiovisual speech perception • Neural coding of speech • Models of speech perception 1 Finding papers Finding papers PubMed search engine: Journal of the Acoustical Society of America: http://www.ncbi.nlm.nih.gov/entrez/ http://scitation.aip.org/jasa/ The evolution of speech: The evolution of speech: a comparative review a comparative review Primate vocal tract W. Tecumseh Fitch Primate vocal tract W. -
Chapter 9 Consonant Substitution in Child Language (Ikwere) Roseline I
Chapter 9 Consonant substitution in child language (Ikwere) Roseline I. C. Alerechi University of Port Harcourt The Ikwere language is spoken in four out of the twenty-three Local Government Areas (LGAs) of Rivers State of Nigeria, namely, Port Harcourt, Obio/Akpor, Emohua and Ikwerre LGAs. Like Kana, Kalabari and Ekpeye, it is one of the major languages of Rivers State of Nigeria used in broadcasting in the electronic media. The Ikwere language is classified asan Igboid language of the West Benue-Congo family of the Niger-Congo phylum of languages (Williamson 1988: 67, 71, Williamson & Blench 2000: 31). This paper treats consonant substi- tution in the speech of the Ikwere child. It demonstrates that children use of a language can contribute to the divergent nature of that language as they always strive for simplification of the target language. Using simple descriptive method of data analysis, the paper identifies the various substitutions of consonant sounds, which characterize the Ikwere children’s ut- terances. It stresses that the substitutions are regular and rule governed and hence implies the operation of some phonological processes. Some of the processes are strengthening and weakening of consonants, loss of suction of labial implosives causing them to become labial plosives, devoicing of voiced consonants, etc. While some of these processes are identical with the adult language, others are peculiar to children, demonstrating the relationships between the phonological processes in both forms of speech. It is worthy of note that high- lighting the relationships and differences will make for effective communication between children and adults. 1 Introduction The Ikwere language is spoken in four out of the twenty-three Local Government Areas (LGAs) of Rivers State of Nigeria, namely, Port Harcourt, Obio/Akpor, Emohua and Ik- werre LGAs. -
Towards Zero-Shot Learning for Automatic Phonemic Transcription
The Thirty-Fourth AAAI Conference on Artificial Intelligence (AAAI-20) Towards Zero-Shot Learning for Automatic Phonemic Transcription Xinjian Li, Siddharth Dalmia, David R. Mortensen, Juncheng Li, Alan W Black, Florian Metze Language Technologies Institute, School of Computer Science Carnegie Mellon University {xinjianl, sdalmia, dmortens, junchenl, awb, fmetze}@cs.cmu.edu Abstract Hermann and Goldwater 2018), which typically uses an un- supervised technique to learn representations which can be Automatic phonemic transcription tools are useful for low- resource language documentation. However, due to the lack used towards speech processing tasks. of training sets, only a tiny fraction of languages have phone- However, those unsupervised approaches could not gener- mic transcription tools. Fortunately, multilingual acoustic ate phonemes directly and there has been few works study- modeling provides a solution given limited audio training ing zero-shot learning for unseen phonemes transcription, data. A more challenging problem is to build phonemic tran- which consist of learning an acoustic model without any au- scribers for languages with zero training data. The difficulty dio data or text data for a given target language and unseen of this task is that phoneme inventories often differ between phonemes. In this work, we aim to solve this problem to the training languages and the target language, making it in- transcribe unseen phonemes for unseen languages without feasible to recognize unseen phonemes. In this work, we ad- dress this problem by adopting the idea of zero-shot learning. considering any target data, audio or text. Our model is able to recognize unseen phonemes in the tar- The prediction of unseen objects has been studied for a get language without any training data. -
Velar Segments in Old English and Old Irish
In: Jacek Fisiak (ed.) Proceedings of the Sixth International Conference on Historical Linguistics. Amsterdam: John Benjamins, 1985, 267-79. Velar segments in Old English and Old Irish Raymond Hickey University of Bonn The purpose of this paper is to look at a section of the phoneme inventories of the oldest attested stage of English and Irish, velar segments, to see how they are manifested phonetically and to consider how they relate to each other on the phonological level. The reason I have chosen to look at two languages is that it is precisely when one compares two language systems that one notices that structural differences between languages on one level will be correlated by differences on other levels demonstrating their interrelatedness. Furthermore it is necessary to view segments on a given level in relation to other segments. The group under consideration here is just one of several groups. Velar segments viewed within the phonological system of both Old English and Old Irish cor relate with three other major groups, defined by place of articulation: palatals, dentals, and labials. The relationship between these groups is not the same in each language for reasons which are morphological: in Old Irish changes in grammatical category are frequently indicated by palatalizing a final non-palatal segment (labial, dental, or velar). The same function in Old English is fulfilled by suffixes and /or prefixes. This has meant that for Old English the phonetically natural and lower-level alternation of velar elements with palatal elements in a palatal environ ment was to be found whereas in Old Irish this alternation had been denaturalized and had lost its automatic character. -
Package 'Phontools'
Package ‘phonTools’ July 31, 2015 Title Tools for Phonetic and Acoustic Analyses Version 0.2-2.1 Date 2015-07-30 Author Santiago Barreda Maintainer Santiago Barreda <[email protected]> Depends R (>= 2.15) Description Contains tools for the organization, display, and analysis of the sorts of data fre- quently encountered in phonetics research and experimentation, including the easy cre- ation of IPA vowel plots, and the creation and manipulation of WAVE audio files. License BSD_2_clause + file LICENSE URL http://www.santiagobarreda.com/rscripts.html NeedsCompilation no Repository CRAN Date/Publication 2015-07-31 01:00:47 R topics documented: a96..............................................3 b95 .............................................4 combocalc . .4 createtemplate . .5 errorbars . .6 f73..............................................7 f99..............................................8 fastacf . .9 Ffilter . 10 findformants . 11 FIRfilter . 13 formanttrack . 14 freqresponse . 16 h95 ............................................. 17 hotelling.test . 18 1 2 R topics documented: imputeformants . 20 interpolate . 21 ldboundary . 22 ldclassify . 23 loadsound . 25 loadtable . 26 lpc.............................................. 26 makeFIR . 27 makesound . 29 multiplot . 30 normalize . 31 normalize.compare . 33 ntypes . 34 p73 ............................................. 35 pb52............................................. 36 peakfind . 37 phasor . 37 pickIPA . 38 pitchtrack . 40 playsound . 41 polezero . 42 powertrack . 43 preemphasis . 44