Silent Communication: Whispered Speech-To-Clear Speech Conversion Viet-Anh Tran

Total Page:16

File Type:pdf, Size:1020Kb

Silent Communication: Whispered Speech-To-Clear Speech Conversion Viet-Anh Tran Silent Communication: whispered speech-to-clear speech conversion Viet-Anh Tran To cite this version: Viet-Anh Tran. Silent Communication: whispered speech-to-clear speech conversion. Computer Sci- ence [cs]. Institut National Polytechnique de Grenoble - INPG, 2010. English. tel-00614289 HAL Id: tel-00614289 https://tel.archives-ouvertes.fr/tel-00614289 Submitted on 10 Aug 2011 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. INSTITUT POLYTECHNIQUE DE GRENOBLE N° attribué par la bibliothèque |__|__|__|__|__|__|__|__|__|__| T H E S E pour obtenir le grade de DOCTEUR DE L’INP Grenoble Spécialité : Signal, Image, Parole, Télécoms préparée au : Département de Parole et Cognition du laboratoire GIPSA-Lab dans le cadre de l’Ecole Doctorale : Electronique, Electrotechnique, Automatique et Traitement du Signal présentée et soutenue publiquement par Viet Anh Tran le 28 Janvier 2010 TITRE Silent Communication : whispered speech-to-clear speech conversion Communication silencieuse : conversion de la parole chuchotée en parole claire DIRECTEUR DE THESE Hélène Lœvenbruck CO-DIRECTEUR DE THESE Christian Jutten & Gérard Bailly JURY M. Eric Moulines , Président Mme. Régine André-Obrecht , Rapporteure M. Gérard Chollet , Rapporteur Mme. Hélène Lœvenbruck , Directrice de thèse M. Christian Jutten , Co-encadrant M. Gérard Bailly , Co-encadrant M. Yves Laprie , Examinateur Acknowledgements I would like to express my deepest appreciation to my advisors Dr. Hélène Lœvenbruck, Dr. Gérard Bailly and Prof. Christian Jutten, whose constant guidance, encouragement and support from the initial to the final level enabled me to develop an understanding of this research in GIPSA-Lab. During my thesis, I had a great opportunity to work with Dr. Tomoki Toda, assistant professor at Nara Institute of Science and Technology (NAIST), Japan. I would like to thank Dr. Tomoki Toda for letting me use the NAM-to-speech system, developed at NAIST, and especially for his continuous support, valuable advice, discussions and collaboration throughout this research. I would like to express my gratitude to the various members of the jury. I would like to acknowledge Prof Régine André-Obrecht and Dr. Gérard Chollet for having accepted to be examiners of my thesis and for their helpful comments and suggestions in the writing of the thesis. Special thanks go to Prof. Eric Moulines for accepting to be the president of the defense committee as well as Dr. Yves Laprie for his participation in the jury. I am grateful to Coriandre Vilain, Christophe Savariaux, Lionel Granjon and Alain Arnal for their help in fixing the hardware problems and in data acquisition. Thanks also go to Pierre Badin for valuable advice about the presentation of my thesis. I also wish to thank Dr. Panikos Heracleous with whom I was fortunate to discuss and collaborate on the Non-Audible Murmur recognition problem. Finally, I would like to acknowledge my colleagues: Nino, Nadine, Mathieu, Oliver, Xavier, Anahita, Noureddine, Stephan, Atef, Amélie, Keigo, Yamato and others for many good times with them at Gipsa-Lab and at NAIST, and of course, my family and my friends for their endless encouragement and support. Contents Contents ................................................................................................................... 1 List of figures........................................................................................................... 4 List of tables............................................................................................................. 7 Abstract.................................................................................................................... 8 Résumé ................................................................................................................... 10 Introduction........................................................................................................... 12 Motivation of research ........................................................................................ 12 Scope of thesis .................................................................................................... 12 Signal-to-signal mapping ................................................................................ 13 HMM-based recognition-synthesis................................................................. 13 Organization of thesis ......................................................................................... 14 Related project .................................................................................................... 15 Chapter 1 Silent speech interfaces and silent-to-audible speech conversion. 17 1.1 Introduction............................................................................................. 17 1.2 Silent speech........................................................................................... 18 1.3 Silent speech interfaces (SSI) ................................................................. 20 1.3.1 Electro-encephalographic (EEG) sensors ....................................... 20 1.3.2 Surface Electromyography (sEMG)................................................ 23 1.3.3 Tongue displays.............................................................................. 28 1.3.4 Non-Audible Murmur (NAM) microphone .................................... 32 1.4 Conversion of silent speech to audible speech........................................ 34 1.4.1 Direct signal-to-signal mapping...................................................... 35 1.4.2 Combination of speech recognition and speech synthesis.............. 45 1.5 Summary ................................................................................................. 51 Chapter 2 Whispered speech ............................................................................. 53 2.1 Introduction............................................................................................. 53 2.2 The speech production system ................................................................ 53 2.3 Whispered speech production................................................................. 55 2.4 Acoustic characteristics of whispered speech......................................... 57 2.5 Perceptual characteristics of whisper...................................................... 59 2.5.1 Pitch-like perception ............................................................................. 59 2.5.2 Intelligibility ................................................................................... 61 2.5.3 Speaker identification..................................................................... 62 2.5.4 Multimodal speech perception........................................................ 62 2.6 Summary ................................................................................................. 66 Chapter 3 Audio and audiovisual corpora ........................................................ 67 3.1 Introduction............................................................................................. 67 1 2 Contents 3.2 French corpus..........................................................................................69 3.2.1 Text material....................................................................................69 3.2.2 Recording protocol ................................................................................70 3.3 Audiovisual Japanese corpus...................................................................72 3.3.1 Text material....................................................................................72 3.3.2 Materials and recording...................................................................72 3.4 Summary..................................................................................................74 Chapter 4 Improvement to the GMM-based whisper-to-speech system........75 4.1 Introduction .............................................................................................75 4.2 Original NAIST NAM-to-Speech system ...............................................76 4.2.1 Spectral estimation ..........................................................................77 4.2.2 Source excitation estimation............................................................82 4.3 Improvement to the GMM-based whisper-to-speech conversion ...........83 4.3.1 Aligning data...................................................................................84 4.3.2 Feature extraction............................................................................85 4.3.3 Voiced/unvoiced detection..............................................................86 4.3.4 F0 estimation ...................................................................................89 4.3.5 Influence of spectral variation on the predicted accuracy ...............91 4.3.6 Influence of visual information for the conversion .........................96 4.4 Perceptual evaluations...........................................................................102
Recommended publications
  • 6. Laryngeal Features
    6. Laryngeal features Before you study this chapter, check whether you are familiar with the following terms: allomorph, allophone, aspiration, devoicing, frequency, glottalization, glottal stop, glottis, hiatus, larynx, organs of speech, phoneme, pulmonic egressive airstream, root, sibilant, suffix, syllabic consonant, utterance, vocal cords/folds, voice assimilation, voiced, voiceless, voicing In this chapter we take a look at the articulatory role of the glottis, the vocal cords/folds and all the different phenomena that are related to the operation of the larynx. This includes voicing and voicelessness, (a comparison of English and Hungarian) voice assimilation, devoicing, aspiration and glottalization, and the effect of voicelessness on preceding vowels, Pre-fortis Clipping. Recall from Chapter 1 that the basic mechanism that is used to produce speech sounds in English and Hungarian is a pulmonic egressive airstream mechanism. Having left the lungs, the air continues upward in the windpipe up to the larynx – the front, shield-like part of which is called Adam's apple – then into the pharynx and the supraglottal cavities, the oral and nasal cavities. In the larynx it has to pass between the vocal cords/folds, two bundles of muscle, which may produce numerous different effects. Before discussing these, let us take a look at the structure and parts of the larynx and their different configurations. As it can be seen in the diagram, the larynx consists of some cartilages – one fixed in the front and two movable ones in the back, plus one more on the top not shown in the diagram, the epiglottis, which can close the windpipe, and a circular one serving as the frame for the larynx –, the Laryngeal features vocal cords connecting the cartilages, and the opening between them, the glottis.
    [Show full text]
  • Learning [Voice]
    University of Pennsylvania ScholarlyCommons Publicly Accessible Penn Dissertations Fall 2010 Learning [Voice] Joshua Ian Tauberer University of Pennsylvania, [email protected] Follow this and additional works at: https://repository.upenn.edu/edissertations Part of the First and Second Language Acquisition Commons Recommended Citation Tauberer, Joshua Ian, "Learning [Voice]" (2010). Publicly Accessible Penn Dissertations. 288. https://repository.upenn.edu/edissertations/288 Please see my home page, http://razor.occams.info, for the data files and scripts that make this reproducible research. This paper is posted at ScholarlyCommons. https://repository.upenn.edu/edissertations/288 For more information, please contact [email protected]. Learning [Voice] Abstract The [voice] distinction between homorganic stops and fricatives is made by a number of acoustic correlates including voicing, segment duration, and preceding vowel duration. The present work looks at [voice] from a number of multidimensional perspectives. This dissertation's focus is a corpus study of the phonetic realization of [voice] in two English-learning infants aged 1;1--3;5. While preceding vowel duration has been studied before in infants, the other correlates of post-vocalic voicing investigated here --- preceding F1, consonant duration, and closure voicing intensity --- had not been measured before in infant speech. The study makes empirical contributions regarding the development of the production of [voice] in infants, not just from a surface- level perspective but also with implications for the phonetics-phonology interface in the adult and developing linguistic systems. Additionally, several methodological contributions will be made in the use of large sized corpora and data modeling techniques. The study revealed that even in infants, F1 at the midpoint of a vowel preceding a voiced consonant was lower by roughly 50 Hz compared to a vowel before a voiceless consonant, which is in line with the effect found in adults.
    [Show full text]
  • Aerodynamic and Durational Cues of Phonological Voicing in Whisper Yohann Meynadier, Yulia Gaydina
    Aerodynamic and durational cues of phonological voicing in whisper Yohann Meynadier, Yulia Gaydina To cite this version: Yohann Meynadier, Yulia Gaydina. Aerodynamic and durational cues of phonological voicing in whisper. Interspeech, Aug 2013, Lyon, France. pp.335-339. hal-01211117 HAL Id: hal-01211117 https://hal.archives-ouvertes.fr/hal-01211117 Submitted on 5 Oct 2015 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Aerodynamic and durational cues of phonological voicing in whisper Yohann Meynadier, Yulia Gaydina Aix-Marseille Université, CNRS UMR 7309, LPL, Aix-en-Provence, France [email protected], [email protected] consonants and pre-consonant vowels has long been observed: Abstract (i) vowels are longer before voiced than before voiceless This paper presents analyses on the phonological voicing consonants and, (ii) voiceless obstruents are longer than contrast in whispered speech, which is characterized by the voiced ones [2], see [35, 36] for a review. absence of vocal fold vibrations. In modal speech, besides While the voiced-voiceless difference in duration of pre- glottal vibration, the contrast between voiced and unvoiced consonantal vowels is not still well understood [2, 37, 38], the consonants is realized by other phonetic correlates: e.g.
    [Show full text]
  • Laryngeal Physiology and Terminology in CCM Singing
    Faculty of Education Ingvild Vestfall Master’s thesis Laryngeal physiology and terminology in CCM singing A thesis investigating research on the underlying laryngeal physiology of CCM singing techniques, and experiences of teaching CCM genres to adolescents Stemmefysiologi og terminologi i CCM/rytmisk sang En studie av forskning på stemmefysiologi knyttet til sangteknikker i CCM/rytmiske sjangere, og erfaringer med å undervise ungdommer i CCM/rytmisk sang Master in Culture and Language 2018 Consent to lending by University College Library YES ☒ NO ☐ Consent to accessibility in digital archive Brage YES ☒ NO ☐ ii TABLE OF CONTENTS TABLE OF CONTENTS.................................................................................................................. III LIST OF TABLES ........................................................................................................................... V LIST OF FIGURES ........................................................................................................................ VI ABSTRACT ................................................................................................................................. VII SAMMENDRAG (IN NORWEGIAN) .............................................................................................. VIII PREFACE ..................................................................................................................................... IX 1 INTRODUCTION ........................................................................................................................
    [Show full text]
  • Ingressive Phonation in Contemporary Vocal Music, Works by Helmut Lachenmann, Georges Aperghis, Michael Baldwin, and Nicholas
    © 2012 Amanda DeBoer Bartlett All Rights Reserved iii ABSTRACT Jane Schoonmaker Rodgers, Advisor The use of ingressive phonation (inward singing) in contemporary vocal music is becoming more frequent, yet there is limited research on the physiological demands, risks, and pedagogical requirements of the various ingressive phonation techniques. This paper will discuss ingressive phonation as it is used in contemporary vocal music. The research investigates the ways in which ingressive phonation differs acoustically, physiologically, and aesthetically from typical (egressive) phonation, and explores why and how composers and performers use the various ingressive vocal techniques. Using non-invasive methods, such as electroglottograph waveforms, aerodynamic (pressure, flow, flow resistance) measures, and acoustic analyses of recorded singing, specific data about ingressive phonation were obtained, and various categories of vocal techniques were distinguished. Results are presented for basic vocal exercises and tasks, as well as for specific excerpts from the repertoire, including temA by Helmut Lachenmann and Ursularia by Nicholas DeMaison. The findings of this study were applied to a discussion surrounding pedagogical and aesthetic applications of ingressive phonation in contemporary art music intended for concert performance. Topics of this discussion include physical differences in the production and performance of ingressive phonation, descriptive information regarding the various techniques, as well as notational and practical recommendations for composers. iv This document is dedicated to: my husband, Tom Bartlett my parents, John and Gail DeBoer and my siblings, Mike, Matt, and Leslie DeBoer Thank you for helping me laugh through the process – at times ingressively – and for supporting me endlessly. v ACKNOWLEDGEMENTS I have endless gratitude for my advisor and committee chair, Dr.
    [Show full text]
  • The Effect of Whisper and Creak Vocal Mechanisms on Vocal Tract Resonances
    The effect of whisper and creak vocal mechanisms on vocal tract resonances ͒ Yoni Swerdlin, John Smith,a and Joe Wolfe School of Physics, University of New South Wales, Sydney, New South Wales 2052, Australia ͑Received 12 May 2009; revised 20 January 2010; accepted 21 January 2010͒ The frequencies of vocal tract resonances estimated using whisper and creak phonations are compared with those in normal phonation for subjects who produced pairs of these phonations in the same vocal gesture. Peaks in the spectral envelope were used to measure the frequencies of the first four resonances ͑R1–R4͒ for the non-periodic phonations, and broadband excitation at the mouth was used to measure them with similar precision in normal phonation. For resonances R1–R4, whispering raises the average resonant frequencies by 255 Hz with standard deviation 90 Hz, 115Ϯ105, 125Ϯ125, and 75Ϯ120 Hz, respectively. A simple one dimensional model of the vocal tract is presented and used to show how an effective glottal area can be estimated from shifts in resonance frequency measured during the same vocal gesture. Calculations suggest that the effective glottal area thus defined increases to 40Ϯ30 mm2 during whispering. Creak phonation raised significantly only the first and third resonant frequencies, by 45Ϯ50 and 65Ϯ120 Hz respectively. It thus appears possible to use creak phonation to determine resonances with more precision than is available from the spectral envelope of voiced speech, and this supports its use in teaching resonance tuning to singers. © 2010 Acoustical Society of America. ͓DOI: 10.1121/1.3316288͔ PACS number͑s͒: 43.70.Gr, 43.70.Aj, 43.70.Bk ͓CHS͔ Pages: 2590–2598 I.
    [Show full text]
  • Phonetics and Phonology
    DiSUm, a.a. 2019–20, Scienze e lingue per la comunicazione (L20) Lingua e comunicazione inglese 1, prof. Iain Halliday Lezione martedì 7 gennaio 2020 (26 di 27), aula A1, ore 08:00–10:00 Phonetics and phonology Chapters 2 and 3 of English Language, respectively titled “Phonetics” (by Kevin Watson) and “Phonology” (by Sam Kirkham and Claire Nance) are very detailed in their descriptions of the structures and mechanisms that work together in the production of speech – phonetics regarding the production of speech sounds, phonology regarding the organization of those speech sounds within the language. In your reading of these chapters, I don’t want you to be daunted by the amount of technical information made available and discussed by the authors. Those of you who feel inspired by the topics may wish to absorb the information profoundly, perhaps with a view to future study in this field of linguistics. What I am more interested in is that you should have a general awareness of how speech is created and perceived and how speech is organized. Watson early on introduces the components of the vocal tract, asking his readers to imagine instructions from the dentist to “say aahh” (ə) and “say eeee” (e) vowel sounds – the position of the tongue is very different in the two cases and can facilitate the dentist’s access to certain areas of the mouth. Apart from the obvious “tongue” – synonym both in English and in Italian for “language” – and which together with the larynx receives particular attention in the chapter, the component parts are the lips, the teeth, the alveolar ridge, the roof of the mouth (hard palate), the velum (the soft palate), the uvula, the pharynx (the throat), the larynx containing the two vocal cords (vocal folds), with the glottis separating them.
    [Show full text]
  • Silentvoice: Unnoticeable Voice Input by Ingressive Speech
    SilentVoice: Unnoticeable Voice Input by Ingressive Speech Masaaki FUKUMOTO Microsoft Research Beijing, China [email protected] ABSTRACT accuracy [5,31,34]; smartphones [3,24] and smart speakers SilentVoice is a new voice input interface device that [1] also incorporate voice input interfaces into daily life. penetrates the speech-based natural user interface (NUI) in However, we rarely see people using voice input in public daily life. The proposed "ingressive speech" method 1 enables placement of a microphone very close to the front spaces and offices . One major reason is that voice leakage of the mouth without suffering from pop-noise, capturing annoys surrounding people, and may risk disseminating very soft speech sounds with a good S/N ratio. It realizes private information to unintended audiences. These are not ultra-small (less than 39dB(A)) voice leakage, allowing us technical issues but social issues for which there is no easy to use voice input without annoying surrounding people in fix even if performance of the voice recognition system is public and mobile situations as well as offices and homes. greatly improved. By measuring airflow direction, SilentVoice can easily be RELATED WORKS separated from normal utterances with 98.8% accuracy; no There exist some "silent" voice input systems that can be activation words are needed. It can be used for voice- used without being noticed by surroundings [8]. Soft activated systems with a specially trained voice recognizer; whisper voices can be detected by using a stethoscopic evaluation results yield word error rates (WERs) of 1.8% microphone that contacts the skin behind the ear (called (speaker-dependent condition), and 7.0% (speaker- NAM: Non-Audible Mummer) [26], or a throat microphone independent condition) with a limited dictionary of 85 [17].
    [Show full text]
  • Non-Invasive Mechanosensory Perturbation to Test Voice-Related Motor and Somatosensory Cortical Responses in Spasmodic Dysphonia
    Non-Invasive Mechanosensory Perturbation to Test Voice-Related Motor and Somatosensory Cortical Responses in Spasmodic Dysphonia Michael J. Hammer, M.A., Ph.D., CCC-SLP Director, Airway Sensory Physiology Laboratory Director, Center for Clinical Speech, Voice, and Swallow Physiology University of Wisconsin – Whitewater Supported by NIH Grants DC014519, DC010900, RR025012 and the Dystonia Coalition Background •Mechanosensation (sense of touch) in the larynx has often been associated with ballistic or defensive gestures including those for airway protection. •The laryngeal mechanosensorium may also be directly involved in the regulation of fine motor control of the larynx for voice. Laryngeal Mechanosensory Mechanisms Internal branch of superior laryngeal nerve (CN X) Dependent upon rapidly adapting mechanoreceptors in laryngeal mucosa Primary location for tissue strain associated with medialization of the arytenoids/vocal folds (Andreatta, Mann, Poletto & Ludlow; Davis & Nail, 1988; Tanaka, Yoshida & Hirano, 1993; Yoshida, Tanaka, Hirano, & Nakashima, 2000) The trouble is – of course – the sensory receptors in the larynx are a challenge to access. Primary location for tissue strain associated with medialization of the arytenoids/vocal folds Supported by NIH Grants DC014519, DC010900, RR025012 and the Dystonia Coalition Voice-Related Modulation of Mechanosensory Detection Healthy Mechanisms Hammer MJ, Krueger MA. (2014). Experimental Brain Research. 232, 13-20. NIH Extramural Nexus Thursday - October 2, 2014 7:04 PM Subject Line: “NIH Update: Look at science funding across agencies, Considering sex as a biological research variable, Peer review webinars, Subaccounting for NIH awards” Adductor Spasmodic Dysphonia (ADSD) Adductor Spasmodic Dysphonia (ADSD) is a chronic, speech-specific focal laryngeal dystonia characterized by intermittent hyperadduction of the vocal folds, voice breaks, strained-strangled voice, and increased effort to speak.
    [Show full text]
  • Diction for Singers
    DICTION FOR SINGERS: A COMPREHENSIVE ASSESSMENT OF BOOKS AND SOURCES DOCUMENT Presented in Partial Fulfillment of the Requirements for the Degree Doctor of Musical Arts in the Graduate School of The Ohio State University By Cynthia Lynn Mahaney, B.M., M.A. * * * * * The Ohio State University 2006 Document Committee: Approved by Professor Emeritus, Eileen Davis, Adviser Dr. Karen Peeler ___________________________ Dr. R. J. David Frego Adviser Graduate Program in Music Copyright by Cynthia Lynn Mahaney 2006 ABSTRACT A common dilemma for today’s college voice professor is how to teach vocal diction effectively to the undergraduate student in the limited time allotted to these courses in a college music program. The college voice professor may rely on familiar and previously used texts, though other excellent resources have become available in the last decade. It is the purpose of this study to identify the diction books and supplemental sources currently used in the United States, and provide assessment of their suitability for teaching undergraduate voice students. A survey was conducted to determine which books and sources college professors currently use in vocal diction classes. The survey concentrated primarily on diction instruction resources for the Italian, German, and French languages, since these are the first languages that need to be mastered by the undergraduate voice student. The survey instrument was sent to all 1,733 institution members of the College Music Society in the United States. The 118 completed surveys which were returned formed the basis of this study. From the 118, twenty-two interviews were conducted with instructors who used different diction texts.
    [Show full text]
  • Paralinguistic Qualifiers: Our Many Voices
    Languaae & Communrcurion. Vol. 1 I. No. 3. pp. 181-195. 1991 0271-5309191 $3.00 + .OO Printed in Great Enrain. Pcrgamon Press plc PARALINGUISTIC QUALIFIERS: OUR MANY VOICES FERNANDO POYATOS Prince Tsugu Akihito [. .] on Aug. 15, 1945 [. .j listened silently [. .I asthe ready voice of his father, Emperor Hirohito. crackled on the radio, telling a shocked Japanese people that he had decided to surrender (;Yew York Tinres, 8 January 1989) William Faulkner was a small man [. .I His voice was soft and whispery but had carrying power and he spoke fast. His laugh was a chuckle, almost a snort (Morris, W., ‘Faulkner’s Mississippi’, N4tionol Geogruphic, March 1989) 1. Introduction Apart from the paralinguistic primary qualities (timbre, resonance, loudness, tempo, pitch (level, range, registers, intervals), intonation range, syllabic duration, and rhythm) that are always present in the human voice, what truly colors our speech is a complex series of voice qualities that should be identified as paralinguistic voice qualifiers. They modify from syllables to longer speech segments, or a whole deliverance, qualifying words, paralinguistic diffcrentiatiors (e.g. strident laughter, muffled crying) and paralinguistic alternants (e.g. a laryngealized moan), and can also appear as permanent traits (e.g. a person’s husky voice), their own determining factors being biological (e.g. normal or abnormal anatomical configuration of larynx, lips, etc.), physiologicof (e.g. type of vocal band vibration), psychological and emotional (e.g. shyness, anger, scorn) and sociocultural, that is, voice effects established universally (e.g. whispering of intimacy) and culturally (e.g. the creakiness and falsetto of British surprised and incredulous ‘What!?‘).
    [Show full text]
  • Yapese Reference Grammar PALI LANGUAGE TEXTS: MICRONESIA
    Yapese Reference Grammar PALI LANGUAGE TEXTS: MICRONESIA Social Sciences and Linguistics Institute University of Hawaii Donald M. Topping Editor Yapese Reference Grammar JOHN THAYER JENSEN with the assistance of LEO DAVID PUGRAM JOHN BAPTIST IOU RAPHAEL DEFEG The University Press of Hawaii HONOLULU Open Access edition funded by the National Endowment for the Humanities / Andrew W. Mellon Foundation Humanities Open Book Program. Licensed under the terms of Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 In- ternational (CC BY-NC-ND 4.0), which permits readers to freely download and share the work in print or electronic format for non-commercial purposes, so long as credit is given to the author. Derivative works and commercial uses require per- mission from the publisher. For details, see https://creativecommons.org/licenses/by-nc-nd/4.0/. The Cre- ative Commons license described above does not apply to any material that is separately copyrighted. Open Access ISBNs: 9780824881283 (PDF) 9780824881290 (EPUB) This version created: 20 May, 2019 Please visit www.hawaiiopen.org for more Open Access works from University of Hawai‘i Press. The research reported herein and the publication of this book were supported by the Government of the Trust Territory of the Pacific. The assistance of Leo David Pugram, John Baptist Iou, and Raphael Defeg was made possible through support received from the Trust Territory Government and the East-West Culture Learning Institute. Copyright © 1977 by The University Press of Hawaii All rights reserved. For the Yapese people, with the hope that it might be a means of practical help, and further appreciation of the “riches of the glory of God” in His creation.
    [Show full text]