Analysis of Nasal Consonants Using Perceptual Linear Prediction

Total Page:16

File Type:pdf, Size:1020Kb

Analysis of Nasal Consonants Using Perceptual Linear Prediction Analysis of nasal consonants using perceptual linear prediction YingyongQi Departmentof Speech and Hearing Sciences and Department of E!ectrical and Computer Engineering, Universityof •4rizona, Tucson, •4 rizona 85721 RobertA. Fox Diuisionof Speech and Hearing Science& Ohio State Unioersit•, Columbus, Ohio 43210 (Received31 May 1991;accepted for publication19 November 1991 ) Until recently,speech analysis techniques have been built aroundthe all-polelinear predictive model.This studyexamines the effectivenessof usingthe perceptuallinear predictive method for analyzingnasal consonants. Six speakers(three men and three women) produced300 CV syllableswith initial nasalconsonants/m/and/n/. A threshold-basedboundary detection algorithmwas developed to extractnasal segments from the CV contexts.Poles of a fifth-order perceptuallinear predictivemodel were calculated and the frequencyof the secondpole was usedto characterizethe placeof articulationof nasalconsonants. Results indicated that the frequencyfor the secondtransformed pole was significantly lower for/m/than for/n/and wasindependent of factorssuch as vowelcontext and genderof the speaker.A nasal identificationrate of 86% wasobtained based on the frequencyof the secondpole. The useof the perceptuallinear predictive method may thusovercome some difficulties associated with analyzingnasal consonants. PACS numbers:43.71.Es, 43.71.Cq, 43.72.Ar INTRODUCTION salconsonants and measured the frequency of theantireson- The progressivephonetic refinement of broadcategories anceusing the methodof analysisby synthesis.This study of speechsounds (stop, fricative, and sonorant) is animpor- provided,for the first time, valuableexperimental evidence tant step in implementingspeech understanding and/or on the predictedacoustic characteristics of nasal conson- speechrecognition systems (Reddy, 1976); however, ants.As pointedout by the author,however, "the procedure knowledgeabout the acousticcues that allow for reliable ... is by no meanssimple or mechanical... and a machine extractionof phoneticcharacteristics of speechfrom the couldhardly replacethe humanoperator at the stageof this acousticwaveform is limited (Harrington, 1988). For exam- study." ple, robustfeatures that enableautomatic within-class dif- Kurowski and Blumstein(1987) reportedthat the place ferentiation of nasal consonants have not been well estab- of articulationof nasalconsonants in CV syllablescould be lished. differentiatedby thepattern of spectralenergy change in the The acousticstructure of nasal consonantshas long vicinityof theoral releaseonce the spectra were transformed beenpredicted by the acoustictheory of speechproduction usingan algorithmbased on the psychoacousticcharacteris- (Fant, 1960;Fujimura, 1962; Flanagan,1972). The pres- ticsof humanauditory system. Utilizing criticalband analy- enceof a side-branchingresonator (the blockedoral cavity) sis based on the Bark scale (Zwicker, 1970), Kurowski and will introducean antiresonance(zero) in the spectrumof Blumsteinexamined the spectraof two glottalpulses of the nasal consonants.Theoretically, the antiresonancecan be nasalmurmur immediatelypreceding oral releaseof the na- usedto identifythe placeof articulationof nasalconsonants salstop closure and the firsttwo glottalpulses after the nasal becausethe frequencyof the antiresonanceis determinedby release.The spectralenergy increase was reported to belarg- the dimensionof the side-branchingresonator. For example, er between 5-7 Bark than between 11-14 Bark for the nasal thefrequency of theantiresonance should be lower for/m/ /m/; whereasthe spectral energy increase was reported to be than for/n/because the blockedoral cavityattached to the larger between11-14 Bark than between5-7 Bark for the pharyngeal-nasalpassage is longerfor/m/than for/ru t. nasal/n/. Although the auditory-basedspectral transfor- Such differences,however, are difficult to detect using con- mation provides a reasonablealternative for analyzing the ventionaltechniques of spectralanalysis. The presenceof the nasalconsonants, these features were inevitablycontext de- spectralzero introducesnonlinear equations to parametric pendentbecause spectral information of the followingvowel methodsof spectralanalysis (Kay, 1987). The harmonic was an important part of the measurement.In addition, structure,the absence of a prominentspectral valley, and the there wasa certaindegree of arbitrarinessin the selectionof significantlydamped high-frequency energy in nasalspectra the frequencyrange for comparingenergy increase around are the sources of some of the difficulties associated with nasalrelease since the selectionwas neithersystematically analyzingnasal consonants using nonparametric (periodo- optimizedgiven the transformedspectrum nor directlypre- gram) methods. scribedby theory. Fujimura (1962) examinedthe acoustic structure of ha- Hermansky(1990) proposeda new approach,called 1718 J. Acoust.Soc. Am. 91 (3), March1992 0001-4966/92/031718-09500.80 (• 1992Acoustical Society of America 1718 Downloaded 21 Sep 2011 to 128.146.89.68. Redistribution subject to ASA license or copyright; see http://asadl.org/journals/doc/ASALIB-home/info/terms.jsp perceptuallinear predictive (PLP) method,for thespectral cyand bandwidth of oneequivalent spectral peak. The real transformationbased on characteristicsof the auditorysys- pole specifiesthe overall spectraltilt of the transformed tem.While the underlying principles of thespectral transfor- spectrum.Because nasal spectra are dominatedby only one mationwas similar to thosepreviously reported (Kewley- low-frequencypole (250-300 Hz) (Fujimura, 1962;Fuji- Port and Luce, 1984; Kurowski and Blumstein, 1987), the mura arid Lindqvist, 1971), the 5th-order PLP model is finalall-pole modeling of thetransformed spectrum parame- adoptedhere for the analysis.However, the realpole is not tricallyspecified the spectrumand, thus,simplified the ex- consideredin the finalanalysis because spectral tilt caneasi- tractionof acousticcharacteristics of speechsignals. Her- ly be influencedby factorssuch as recording conditions and mansky (1990) demonstrated that a fifth-order PLP glottal characteristicsof a particular speakerand is often spectraltransformation improved the performanceof a tem- consideredas a phoneticallynonessential factor (Klatt, plate-basedspeech recognition system. 1982). In light of the newlyproposed PLP methodof spectral analysis,the presentstudy was undertakento determine A. Subjects and recordings whether acoustic characteristics of nasal consonants could Sixnative speakers of AmericanEnglish (three men and be found that were independentof the following vowel con- three women) providedthe speechsamples. Each speaker text in a CV syllableand to evaluatewhether place of articu- producedten C¾ syllableswith initial nasalconsonants/m/ lation couldbe automaticallyidentified on the basisof these and/n/followed by five differentvowels/i/,/a•/,/o/,/a/, features.The paper is organizedas follows. SectionI in- and/u/. Eachsyllable was repeated five times. The syllables cludesa brief rationalefor usingthe PLP method and the wererandomly ordered. The recordingswere made in a quiet implementationdetails of the PLP analysis.Section II con- room with backgroundnoise at about 30 dB (SPL). The tainsa descriptionof thecontext-independent acoustic char- microphonewas placed approximately10 cm from the acteristicsof nasalconsonants revealed by the PLP analysis mouth of the speaker.Recorded tokens were low-passfil- and the results of nasal identification based on these features. tered (f• = 4.5 kHz) and digitizedat a samplingrate of 10 Section III is the discussion and conclusion. kHz with a 16-bitquantization level. I. METHOD B. Nasal segment selection The integrationof spectralenergy as part of the audi- A 25.6-mssegment of the nasalconsonant before the tory-basedspectral transformation provides a uniqueadvan- oral releaseinto the followingvowel (i.e., within the nasal tage in analyzingthe spectralcharacteristics of nasalcon- murmur) was selectedfor the PLP analysis.The boundary sonants. As stated earlier, the absenceof a prominent betweenthe nasalconsonant and the following vowel was spectralvalley and the reducedspectral energy in the mid- to automaticallylocated from the acousticwaveform. The ex- high-frequencyranges make it difficultto accuratelylocate act algorithmis asfollows. The signalwas bandpass filtered the center frequencyof the antiresonanceand to use this ( 1-3.5 k Hz, fourth-order Butterworth filter) to enhancethe frequencyto distinguishnasal consonants. There are, how- energydiscontinuity of the boundarysince energy for the ever, less-intensivespectral perturbations at differentfre- nasalmurmur is mostlyin the frequencyrange below about quencyranges for differentnasal consonants due to the side- 800 Hz (Mermelstein,1977). The upperfrequency limit for branch (oral) coupling. For example, there is a relatively the bandpassfilter wasused to eliminateunpredictable high- broad energyreduction in the mid-frequency(1000-2000 frequencyinfluences. The boundarywas determined using a Hz) rangewithin the nasalmurmur of nasal/n/that is not short-term amplitude signal calculated from the filtered asapparent for the nasal/m/. When a nonuniformintegra- waveform.A hammingwindow, with a windowsize of 10ms tion of poweris undertakenas part of the spectraltransfor- anda windowstep of 2 ms,was used in the calculations.The mation, someotherwise unreliable energy concentrations short-termamplitude was calculated as the absolutesum of are expectedto
Recommended publications
  • Laryngeal Features in German* Michael Jessen Bundeskriminalamt, Wiesbaden Catherine Ringen University of Iowa
    Phonology 19 (2002) 189–218. f 2002 Cambridge University Press DOI: 10.1017/S0952675702004311 Printed in the United Kingdom Laryngeal features in German* Michael Jessen Bundeskriminalamt, Wiesbaden Catherine Ringen University of Iowa It is well known that initially and when preceded by a word that ends with a voiceless sound, German so-called ‘voiced’ stops are usually voiceless, that intervocalically both voiced and voiceless stops occur and that syllable-final (obstruent) stops are voiceless. Such a distribution is consistent with an analysis in which the contrast is one of [voice] and syllable-final stops are devoiced. It is also consistent with the view that in German the contrast is between stops that are [spread glottis] and those that are not. On such a view, the intervocalic voiced stops arise because of passive voicing of the non-[spread glottis] stops. The purpose of this paper is to present experimental results that support the view that German has underlying [spread glottis] stops, not [voice] stops. 1 Introduction In spite of the fact that voiced (obstruent) stops in German (and many other Germanic languages) are markedly different from voiced stops in languages like Spanish, Russian and Hungarian, all of these languages are usually claimed to have stops that contrast in voicing. For example, Wurzel (1970), Rubach (1990), Hall (1993) and Wiese (1996) assume that German has underlying voiced stops in their different accounts of Ger- man syllable-final devoicing in various rule-based frameworks. Similarly, Lombardi (1999) assumes that German has underlying voiced obstruents in her optimality-theoretic (OT) account of syllable-final laryngeal neutralisation and assimilation in obstruent clusters.
    [Show full text]
  • Nasal Release, Nasal Finals and Tonal Contrasts in Hanoi Vietnamese: an Aerodynamic Experiment Alexis Michaud, Tuân Vu-Ngoc, Angelique Amelot, Bernard Roubeau
    Nasal release, nasal finals and tonal contrasts in Hanoi Vietnamese: an aerodynamic experiment Alexis Michaud, Tuân Vu-Ngoc, Angelique Amelot, Bernard Roubeau To cite this version: Alexis Michaud, Tuân Vu-Ngoc, Angelique Amelot, Bernard Roubeau. Nasal release, nasal finals and tonal contrasts in Hanoi Vietnamese: an aerodynamic experiment. Mon-Khmer Studies, 2006, 36, pp. 121-137. hal-00138758v2 HAL Id: hal-00138758 https://hal.archives-ouvertes.fr/hal-00138758v2 Submitted on 26 Jan 2008 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Nasal release, nasal finals and tonal contrasts in Hanoi * Vietnamese: An aerodynamic experiment Alexis MICHAUD Tuaán VUÕ NGOïC Laboratoire de Phonétique et Phonologie Laboratoire d’Informatique pour CNRS-Sorbonne Nouvelle, France la Me¤canique et les Sciences de l’Inge¤nieur CNRS-Universite¤s Paris-Sud et Pierre et Marie Curie, France Ange¤lique AMELOT Bernard ROUBEAU Laboratoire de Phonétique et Phonologie Service ORL, Ho^pital Tenon, France CNRS-Sorbonne Nouvelle, France Abstract The present research addresses three
    [Show full text]
  • Part 1: Introduction to The
    PREVIEW OF THE IPA HANDBOOK Handbook of the International Phonetic Association: A guide to the use of the International Phonetic Alphabet PARTI Introduction to the IPA 1. What is the International Phonetic Alphabet? The aim of the International Phonetic Association is to promote the scientific study of phonetics and the various practical applications of that science. For both these it is necessary to have a consistent way of representing the sounds of language in written form. From its foundation in 1886 the Association has been concerned to develop a system of notation which would be convenient to use, but comprehensive enough to cope with the wide variety of sounds found in the languages of the world; and to encourage the use of thjs notation as widely as possible among those concerned with language. The system is generally known as the International Phonetic Alphabet. Both the Association and its Alphabet are widely referred to by the abbreviation IPA, but here 'IPA' will be used only for the Alphabet. The IPA is based on the Roman alphabet, which has the advantage of being widely familiar, but also includes letters and additional symbols from a variety of other sources. These additions are necessary because the variety of sounds in languages is much greater than the number of letters in the Roman alphabet. The use of sequences of phonetic symbols to represent speech is known as transcription. The IPA can be used for many different purposes. For instance, it can be used as a way to show pronunciation in a dictionary, to record a language in linguistic fieldwork, to form the basis of a writing system for a language, or to annotate acoustic and other displays in the analysis of speech.
    [Show full text]
  • Aerodynamic and Acoustic Evidence for the Articulations of Complex Nasal Consonants Didier Demolin, Katharina Haude, Luciana Storto
    Aerodynamic and acoustic evidence for the articulations of complex nasal consonants Didier Demolin, Katharina Haude, Luciana Storto To cite this version: Didier Demolin, Katharina Haude, Luciana Storto. Aerodynamic and acoustic evidence for the articu- lations of complex nasal consonants. Revue PAROLE, Université de Mons-Hainaut, 2006, pp.177-205. halshs-00692079 HAL Id: halshs-00692079 https://halshs.archives-ouvertes.fr/halshs-00692079 Submitted on 3 May 2012 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Aerodynamic and acoustic evidence for the articulation of complex nasal consonants Didier Demolin*#, Katharina Haude° and Luciana Storto* Universidade de São Paulo*, Universitat Köln° & Université libre de Bruxelles# 1. Introduction Indigenous languages contribute in an essential way to understand variation in speech production and perception and to identify the constraints that act on natural languages. These constraints play a crucial role to make, to test or to revise hypotheses made in phonology. The paper examines data from Karitana, a Tupi language of Brasil, from Movima, an isolated language from the Bolivian Amazon, and from Rwanda, a Bantu language from Africa. These languages show rather uncommon data. Karitiana has pre and post oralized consonants, e.g.
    [Show full text]
  • THE INTERNATIONAL PHONETIC ALPHABET (Revised to 2015)
    THE INTERNATIONAL PHONETIC ALPHABET (revised to 2015) CONSONANTS (PULMONIC) © 2015 IPA Bilabial Labiodental Dental Alveolar Postalveolar Retroflex Palatal Velar Uvular Pharyngeal Glottal Plosive Nasal Trill Tap or Flap Fricative Lateral fricative Approximant Lateral approximant Symbols to the right in a cell are voiced, to the left are voiceless. Shaded areas denote articulations judged impossible. CONSONANTS (NON-PULMONIC) VOWELS Clicks Voiced implosives Ejectives Front Central Back Close Bilabial Bilabial Examples: Dental Dental/alveolar Bilabial Close-mid (Post)alveolar Palatal Dental/alveolar Palatoalveolar Velar Velar Open-mid Alveolar lateral Uvular Alveolar fricative OTHER SYMBOLS Open Voiceless labial-velar fricative Alveolo-palatal fricatives Where symbols appear in pairs, the one to the right represents a rounded vowel. Voiced labial-velar approximant Voiced alveolar lateral flap Voiced labial-palatal approximant Simultaneous and SUPRASEGMENTALS Voiceless epiglottal fricative Primary stress Affricates and double articulations Voiced epiglottal fricative can be represented by two symbols Secondary stress joined by a tie bar if necessary. Epiglottal plosive Long Half-long DIACRITICS Some diacritics may be placed above a symbol with a descender, e.g. Extra-short Voiceless Breathy voiced Dental Minor (foot) group Voiced Creaky voiced Apical Major (intonation) group Aspirated Linguolabial Laminal Syllable break More rounded Labialized Nasalized Linking (absence of a break) Less rounded Palatalized Nasal release TONES AND WORD ACCENTS Advanced Velarized Lateral release LEVEL CONTOUR Extra Retracted Pharyngealized No audible release or high or Rising High Falling Centralized Velarized or pharyngealized High Mid rising Mid-centralized Raised ( = voiced alveolar fricative) Low Low rising Syllabic Lowered ( = voiced bilabial approximant) Extra Rising- low falling Non-syllabic Advanced Tongue Root Downstep Global rise Rhoticity Retracted Tongue Root Upstep Global fall Typeface: Doulos SIL .
    [Show full text]
  • Phonemes and Allophones of English Consonants
    Rough definition of phoneme • Phoneme (Concise Dictionary of English consonants: Linguistics, Oxford U. Press 1997) Phonemes and Allophones • “The smallest distinct sound unit in a given language: e.g. /»tIp/ in English realizes the Effects related to aspiration and three successive phonemes, represented in ‘devoiced’ voiced sounds and a few spelling by the letters t, i, and p. other issues Phonemic differences vs. Phonemes allophonic differences • Strict, detailed definitions of the term phoneme are • Differences in speech sound that can signal complex differences between two different words are – Not part of this course phonemic differences – Take phonology courses to fight over the details • Other differences in speech sound that are • Rough and ready idea is indispensable for practical phonetics clearly audible are only allophonic – Must make a distinction between phonemic and differences allophonic differences – ‘pronunciation variants’ that cannot signal different words. Representing allophonic Answer: ‘pie, spy, buy’ differences • ‘Broad’ (= coarse-grained) transcription enough Phonemes in ‘/’ (slash or solidus, pl solidi) for phonemic representation marks – Choose simple symbol for a ‘representative’ (allo)phone /p/ /b/ • ‘Narrow’ (= fine-grained) transcription often requires diacritics • Diacritics for stops [p] [pH] [b] [b8] pH - aspirated p p| - ‘p with inaudible release’ (‘unreleased p’) b8 - ‘(partially) devoiced b’ Phones in square brackets Examples ‘Stop.’, ‘Stop!’, Examples: ‘pie, spy, buy’ ‘Stop!!’, ‘Stob!’ • ‘pie’ [»pHaj]
    [Show full text]
  • LINGUISTICS 401 Advanced Phonetics
    LINGUISTICS 401 Advanced Phonetics GUIDE TO TRANSCRIPTION PRACTICES #1 ASPIRATION: Voiceless stops are aspirated at the beginning of a stressed syllable. Optional aspiration: a. at the beginning of a syllable with secondary stress b. word-finally Symbol: [ ˙] #2 ADVANCED ARTICULATION: Velars have more forward articulation before front vowels. Symbol: [+] #3 UNRELEASED STOP ARTICULATION: a. A stop is not released before another stop articulated at the same place b. A stop has inaudible release before a stop articulated at a different place (= overlapping articulation or double articulation) c. Word-final stops may have inaudible release Symbols: [ ¬] for (a), (c) [ í ] for (b) #4 DENTAL ARTICULATION: Alveolars become dental before dental consonants. Symbol: [ ∞∞ ] #5 VELARIZATION: The alveolar lateral approximant becomes velarized word-finally or before another consonant in the same syllable. Symbol: [ " ] #6 OBSTRUENT DEVOICING: Obstruents become devoiced word-finally. Symbol: [ ≤] 1 #7 FLAPPING: Alveolar stops are articulated as flaps between vowels if the first vowel is stressed. Symbol: [‰] #8 LIQUID AND GLIDE DEVOICING: Liquids and glides become devoiced when preceded by a syllable-initial voiceless stop. Symbol: [ ≤] #9 NASAL ASSIMILATION: Nasals may become homorganic to the following consonant. #10 RETROFLEX ARTICULATION: Alveolar stops become retroflex when followed by a rhotic approximant which, in turn, will also have retroflex articulation. #11 SYLLABIC NASALS AND LIQUIDS: Nasals and liquids become syllabic word-finally when preceded by a consonant. Symbol: [»] #12 LATERAL RELEASE: Alveolar stops are laterally released before an alveolar lateral approximant. Symbol: [ l] #13 LATERAL ONSET The alveolar stops following the alveolar lateral approximant will have a lateral onset. Symbol: [ ¬] (after the lateral approximant) #14 NASAL RELEASE: Stops are nasally released before a nasal of the same place of articulation.
    [Show full text]
  • Articulatory Settings of English Consonants
    ARTICULATORY SETTINGS OF ENGLISH CONSONANTS Edward T. Erazmus University of Kansas Tiiis paper proposes to describe the articulation of English consonants following the articulatory setting principles of Beatrice Honikman (1964). As such it is a sequel to an ear- lier paper done on the articulatory settings of English vowels (1983) and completes an articulatory settings analysis of English phonemes. Tiie articulation of English consonants is conveniently sum- marized in the literature by the use of a consonant chart which displays in two dimensions the convergence of points of articu- lation in the vocal tract (physiological dimension) versus the manner of articulation {phonotactic dimension) with the phonemes represented by phonetic symbols at the points of con- vergence. From these data the individual phonemes are described briefly in terms of the convergence, as, for example, /p/ is a voiceless bilabial stop and /d/ a voiced alveolar stop, as we note in Table 1. While the consonant chart has proved an invaluable tool to theorists and students for the conceptualization and descrip- tion of phonological parameters, it has one major weakness and that is its failure to specify the activity of .the tongue, which is, after all, the principal articulator in language. Tiie activity of the tongue must be inferred from the data which can be readily done with one's own native language but offers little help to the grasp of the phonology of other languages in a practical sense. This paper will utilize the consonant chart as a point of departure for the discussion of articula- tory parameter·s and make an attempt to modify it in the direc- tion of providing more comprehensive physiological data, par- ticularly tongue activity.
    [Show full text]
  • Computer-Coding the IPA: a Proposed Extension of SAMPA J.C.Wells, University College London
    Computer-coding the IPA: a proposed extension of SAMPA J.C.Wells, University College London 1. Computer coding. When an ASCII file (a DOS text file) is sent as an e-mail message, the only characters that are sure to be properly transmitted are those with ASCII/ANSI numbers between 32 and 126. These comprise upper-case A..Z, lower-case a..z, numerals 0..9, punctuation marks ! " ' ( ) , - . / : ; ? [ ] { }, other marks # $ % & * + < = > @ \ ^ _ ` | ~, and space. If we want to transmit phonetic symbols, we must therefore recode them using only these characters. It is not even the case that all the 'other marks' mentioned will necessarily reappear correctly at the receiving end: on my own British screen, for example, an incoming character originally transmitted as a hash mark (#) appears as pound sterling (£). But at least there is a consistent one- for-one substitution, so that information is not lost. On the other hand an outgoing character falling outside the range 32..126 is very likely to be converted into something else: a pound sterling sign (£, ASCII 156, ANSI 0163) transmitted from a British keyboard may be received (even in the UK) as an exclamation mark (!), hash (#), or other substitute. 2. The SAM Phonetic Alphabet (SAMPA) conventions were drawn up in 1988-1991 (with subsequent minor revisions and extensions) by the SAM (Speech Assessment Methods) consortium, comprising speech scientists from nine countries of the European Community (Wells et al, 1992). The purpose of SAMPA was to form the basis of an international standard machine-readable phonetic alphabet for purposes of international collaboration in speech research.
    [Show full text]
  • The Unicode Cookbook for Linguists
    The Unicode Cookbook for Linguists Managing writing systems using orthography profiles Steven Moran Michael Cysouw language Translation and Multilingual Natural science press Language Processing 10 Translation and Multilingual Natural Language Processing Editors: Oliver Czulo (Universität Leipzig), Silvia Hansen-Schirra (Johannes Gutenberg-Universität Mainz), Reinhard Rapp (Johannes Gutenberg-Universität Mainz) In this series: 1. Fantinuoli, Claudio & Federico Zanettin (eds.). New directions in corpus-based translation studies. 2. Hansen-Schirra, Silvia & Sambor Grucza (eds.). Eyetracking and Applied Linguistics. 3. Neumann, Stella, Oliver Čulo & Silvia Hansen-Schirra (eds.). Annotation, exploitation and evaluation of parallel corpora: TC3 I. 4. Czulo, Oliver & Silvia Hansen-Schirra (eds.). Crossroads between Contrastive Linguistics, Translation Studies and Machine Translation: TC3 II. 5. Rehm, Georg, Felix Sasaki, Daniel Stein & Andreas Witt (eds.). Language technologies for a multilingual Europe: TC3 III. 6. Menzel, Katrin, Ekaterina Lapshinova-Koltunski & Kerstin Anna Kunz (eds.). New perspectives on cohesion and coherence: Implications for translation. 7. Hansen-Schirra, Silvia, Oliver Czulo & Sascha Hofmann (eds). Empirical modelling of translation and interpreting. 8. Svoboda, Tomáš, Łucja Biel & Krzysztof Łoboda (eds.). Quality aspects in institutional translation. 9. Fox, Wendy. Can integrated titles improve the viewing experience? Investigating the impact of subtitling on the reception and enjoyment of film using eye tracking and questionnaire data. 10. Moran, Steven & Michael Cysouw. The Unicode cookbook for linguists: Managing writing systems using orthography profiles ISSN: 2364-8899 The Unicode Cookbook for Linguists Managing writing systems using orthography profiles Steven Moran Michael Cysouw language science press Steven Moran & Michael Cysouw. 2018. The Unicode Cookbook for Linguists: Managing writing systems using orthography profiles (Translation and Multilingual Natural Language Processing 10).
    [Show full text]
  • QWERTY DB (Version 0.8) a Multilingual Unicode Keyboard
    QWERTY DB (version 0.8) A multilingual Unicode keyboard This keyboard layout is loosely based on the US International keyboard. It preserves all the key combinations of a standard US or UK keyboard. (The only difference between the US and the UK keyboard is the placement of " and @; however, " should hardly be needed anymore with this keyboard because you can directly enter “ and ” instead.) However, it offers a great variety of additional possibilities: 1. Better typography Key combination Comment Shift + AltGr + N → – dash, (U+2013, so-called “n-dash”) Shift + AltGr + M → — long dash (U+2014, so-called “m-dash”) Shift + AltGr + B → — typographically correct long dash with hair spaces before and after (U+ 200A, U+2014, U+200A) Shift + AltGr + ] → − minus sign (U+2212; you will never need to use the hyphen for this again: “7 + 8 − 3 = 12”, not “7 + 8 - 3 = 12”) AltGr + - → ‑ no-break space (U+2011), prevents word division and is not only independent of the software used but also behaves much better than the special character in Word Shift + AltGr + - → (SHY) soft hyphen (U+00AD), enables you to enter optional word breaks even outside text processors AltGr + J → (ZWJ) zero-width joiner (U+200D), can form ligatures in certain contexts Shift + AltGr + J → (ZWNJ) zero-width non-joiner (U+200C), can prevent the formation of ligatures in certain contexts AltGr + space → no-break space (U+00A0), prevents a line break (and with justified alignment usually also (NBSP) preserves the width of the space from being changed) Shift + space → narrow no-break space (U+202F) for arranging numbers, e.g.
    [Show full text]
  • THE INTERNATIONAL PHONETIC ALPHABET (Revised to 2015)
    THE INTERNATIONAL PHONETIC ALPHABET (revised to 2015) CONSONANTS (PULMONIC) © 2015 IPA Bilabial Labiodental Dental Alveolar Postalveolar Retroflex Palatal Velar Uvular Pharyngeal Glottal Plosive Nasal Trill Tap or Flap Fricative Lateral fricative Approximant Lateral approximant Symbols to the right in a cell are voiced, to the left are voiceless. Shaded areas denote articulations judged impossible. CONSONANTS (NON-PULMONIC) VOWELS Clicks Voiced implosives Ejectives Front Central Back Close Bilabial Bilabial Examples: Dental Dental/alveolar Bilabial Close-mid (Post)alveolar Palatal Dental/alveolar Palatoalveolar Velar Velar Open-mid Alveolar lateral Uvular Alveolar fricative OTHER SYMBOLS Open Voiceless labial-velar fricative Alveolo-palatal fricatives Where symbols appear in pairs, the one to the right represents a rounded vowel. Voiced labial-velar approximant Voiced alveolar lateral flap Voiced labial-palatal approximant Simultaneous and SUPRASEGMENTALS Voiceless epiglottal fricative Primary stress Affricates and double articulations Voiced epiglottal fricative can be represented by two symbols Secondary stress joined by a tie bar if necessary. Epiglottal plosive Long Half-long DIACRITICS Some diacritics may be placed above a symbol with a descender, e.g. Extra-short Voiceless Breathy voiced Dental Minor (foot) group Voiced Creaky voiced Apical Major (intonation) group Aspirated Linguolabial Laminal Syllable break More rounded Labialized Nasalized Linking (absence of a break) Less rounded Palatalized Nasal release TONES AND WORD ACCENTS Advanced Velarized Lateral release LEVEL CONTOUR Extra Retracted Pharyngealized No audible release or high or Rising High Falling Centralized Velarized or pharyngealized High Mid rising Mid-centralized Raised ( = voiced alveolar fricative) Low Low rising Syllabic Lowered ( = voiced bilabial approximant) Extra Rising- low falling Non-syllabic Advanced Tongue Root Downstep Global rise Rhoticity Retracted Tongue Root Upstep Global fall Typefaces: Doulos SIL (metatext); Doulos SIL, IPA Kiel, IPA LS Uni (symbols) .
    [Show full text]