During Monaural and Dichotic Listening Neural Coding of Continuous

Total Page:16

File Type:pdf, Size:1020Kb

During Monaural and Dichotic Listening Neural Coding of Continuous Neural coding of continuous speech in auditory cortex during monaural and dichotic listening Nai Ding and Jonathan Z. Simon J Neurophysiol 107:78-89, 2012. First published 5 October 2011; doi: 10.1152/jn.00297.2011 You might find this additional info useful... This article cites 63 articles, 26 of which you can access for free at: http://jn.physiology.org/content/107/1/78.full#ref-list-1 This article has been cited by 2 other HighWire-hosted articles: http://jn.physiology.org/content/107/1/78#cited-by Updated information and services including high resolution figures, can be found at: http://jn.physiology.org/content/107/1/78.full Downloaded from Additional material and information about Journal of Neurophysiology can be found at: http://www.the-aps.org/publications/jn This information is current as of October 23, 2012. http://jn.physiology.org/ at Univ Texas At Dallas on October 23, 2012 Journal of Neurophysiology publishes original articles on the function of the nervous system. It is published 24 times a year (twice monthly) by the American Physiological Society, 9650 Rockville Pike, Bethesda MD 20814-3991. Copyright © 2012 the American Physiological Society. ESSN: 1522-1598. Visit our website at http://www.the-aps.org/. J Neurophysiol 107: 78–89, 2012. First published October 5, 2011; doi:10.1152/jn.00297.2011. Neural coding of continuous speech in auditory cortex during monaural and dichotic listening Nai Ding1 and Jonathan Z. Simon1,2 1Department of Electrical and Computer Engineering and 2Department of Biology, University of Maryland, College Park, Maryland Submitted 1 April 2011; accepted in final form 28 September 2011 Ding N, Simon JZ. Neural coding of continuous speech in auditory phase-locked to the speech envelope, i.e., the slow modulations cortex during monaural and dichotic listening. J Neurophysiol 107: 78–89, summed over a broad spectral region (Abrams et al. 2008; Downloaded from 2012. First published October 5, 2011; doi:10.1152/jn.00297.2011.—The Ahissar et al. 2001; Aiken and Picton 2008; Lalor and Foxe cortical representation of the acoustic features of continuous speech is 2010; Luo and Poeppel 2007). Temporal locking to features of the foundation of speech perception. In this study, noninvasive mag- speech has also been supported by intracranial recordings from netoencephalography (MEG) recordings are obtained from human subjects actively listening to spoken narratives, in both simple and human core auditory cortex (Nourski et al. 2009). The temporal cocktail party-like auditory scenes. By modeling how acoustic fea- features of speech contribute significantly to speech intelligi- tures of speech are encoded in ongoing MEG activity as a spectro- bility, as do key spectrotemporal features in speech such as http://jn.physiology.org/ temporal response function, we demonstrate that the slow temporal upward and downward formant transitions. The neural coding modulations of speech in a broad spectral region are represented of spectrotemporal modulations in natural soundtracks has bilaterally in auditory cortex by a phase-locked temporal code. For been studied invasively in human auditory cortex using intra- speech presented monaurally to either ear, this phase-locked response cranial extracellular recordings (Bitterman et al. 2008), where is always more faithful in the right hemisphere, but with a shorter latency in the hemisphere contralateral to the stimulated ear. When the spectrotemporal tuning of individual neurons was found to different spoken narratives are presented to each ear simultaneously be generally complex and sometimes very fine in frequency. At (dichotic listening), the resulting cortical neural activity precisely a neural network level, the blood oxygen level-dependent encodes the acoustic features of both of the spoken narratives, but (BOLD) activity measured by functional magnetic resonance at Univ Texas At Dallas on October 23, 2012 slightly weakened and delayed compared with the monaural response. imaging (fMRI) also shows complex spectrotemporal tuning Critically, the early sensory response to the attended speech is con- and possesses no obvious spatial map (Schönwiesner and siderably stronger than that to the unattended speech, demonstrating Zatorre 2009). Which spectrotemporal features of speech are top-down attentional gain control. This attentional gain is substantial encoded in the large-scale synchronized neural activity mea- even during the subjects’ very first exposure to the speech mixture and therefore largely independent of knowledge of the speech content. surable by MEG and EEG, however, remain unknown and are Together, these findings characterize how the spectrotemporal fea- the focus of the current study. tures of speech are encoded in human auditory cortex and establish a When investigating the neural coding of speech, there are single-trial-based paradigm to study the neural basis underlying the several key issues that deserve special consideration. One cocktail party phenomenon. arises from the diversity of speech: language is a productive system permitting the generation of novel sentences. In every- speech segregation; attention; spectrotemporal response function; magnetoencephalography day life, human listeners constantly decode spoken messages they have never heard. In most neurophysiological studies of speech processing, however, small sets of sentences are re- SPOKEN LANGUAGE IS THE DOMINANT form of human communica- peated tens or hundreds of times (although see Lalor and Foxe tion, and human listeners are superb at tracking and under- 2010). This is primarily due to methodological constraints: standing speech even in the presence of interfering speakers neurophysiological recordings, especially noninvasive record- (Bronkhorst 2000; Cherry 1953). The critical acoustic features ings, are quite variable, and so integrating over trials is neces- of speech are distributed across several distinct spectral and sary to obtain a valid estimate of the neural response. An often temporal scales. The slow temporal modulations and coarse neglected cost of repeated stimuli, however, is that the listener spectral modulations reflect the rhythm of speech and contain has obtained complete knowledge of the entire stimulus speech syllabic and phrasal level segmentation information (Green- after only a few repetitions. Without the demands of speech berg 1999) and are particularly important for speech intelligi- comprehension, the encoding of this repeated speech might be bility (Shannon et al. 1995). The neural tracking of slow quite different from the neural coding of novel speech under temporal modulations of speech (e.g., 1–10 Hz) in human natural listening conditions. It is pressing, therefore, to develop auditory cortex can be studied noninvasively using magneto- experimental paradigms that do not require repeating stimuli encephalography (MEG) and electroencephalography (EEG). many times, to study how speech is encoded in a more The low-frequency, large-scale synchronized neural activity ecologically realistic manner. recorded by MEG/EEG has been demonstrated to be synchro- Second, speech communication is remarkably robust against nized by speech stimulus (Luo and Poeppel 2007) and is interference. When competing speech signals are present, hu- man listeners can actively maintain attention on a particular Address for reprint requests and other correspondence: J. Z. Simon, Univ. of speech target and comprehend it. The superior temporal gyrus Maryland, College Park, MD 20742 (e-mail: [email protected]). has been identified as a region heavily involved in processing 78 0022-3077/12 Copyright © 2012 the American Physiological Society www.jn.org CORTICAL SPEECH ENCODING IN HUMAN LISTENERS 79 concurrent speech signals (Scott et al. 2009). Recent EEG left and right auditory pathways. Moreover, previous studies results have shown that human auditory cortex can selectively have only demonstrated that speech is encoded in MEG/EEG amplify the low-frequency neural correlates of the speech activity with sufficient fidelity to discriminate among two or signal being attended to (Kerlin et al. 2010). This attentional three sentences (Kerlin et al. 2010; Luo and Poeppel 2007). modulation of low-frequency neural activity has been sug- With a long-duration, discourse-level stimulus, we can test the gested as a general mechanism for sensory information selec- limit of this fidelity by quantifying the maximum number of tion (Schroeder and Lakatos 2009). Because speech compre- speech stimuli that can be discriminated based on MEG re- hension is a complex hierarchical process involving multiple sponses. brain regions, it is unclear whether the attentional effect seen in Inspired by research on single-unit neurophysiology (de- the auditory cortex directly modulates feedforward auditory Charms et al. 1998; Depireux et al. 2001), the analysis of MEG processing or reflects only feedback from language areas, or activity was performed using the spectrotemporal response even motor areas (Hickok and Poeppel 2007). One approach to function (STRF), which can reveal neural coding mechanisms test whether feedforward processing is involved in speech by analyzing the relationship between ongoing neural activity segregation is to investigate the latency of the attentional and the corresponding continuous stimuli (Fig. 1). The prop- Downloaded from effect. If the attentional modulation of MEG/EEG response has erties of network-level cortical activity, which plays an impor- a relatively short latency, e.g., 100 ms, then
Recommended publications
  • Linguistic Processing of Task-Irrelevant Speech at a Cocktail Party Paz Har-Shai Yahav*, Elana Zion Golumbic*
    RESEARCH ARTICLE Linguistic processing of task-irrelevant speech at a cocktail party Paz Har-shai Yahav*, Elana Zion Golumbic* The Gonda Center for Multidisciplinary Brain Research, Bar Ilan University, Ramat Gan, Israel Abstract Paying attention to one speaker in a noisy place can be extremely difficult, because to- be-attended and task-irrelevant speech compete for processing resources. We tested whether this competition is restricted to acoustic-phonetic interference or if it extends to competition for linguistic processing as well. Neural activity was recorded using Magnetoencephalography as human participants were instructed to attend to natural speech presented to one ear, and task- irrelevant stimuli were presented to the other. Task-irrelevant stimuli consisted either of random sequences of syllables, or syllables structured to form coherent sentences, using hierarchical frequency-tagging. We find that the phrasal structure of structured task-irrelevant stimuli was represented in the neural response in left inferior frontal and posterior parietal regions, indicating that selective attention does not fully eliminate linguistic processing of task-irrelevant speech. Additionally, neural tracking of to-be-attended speech in left inferior frontal regions was enhanced when competing with structured task-irrelevant stimuli, suggesting inherent competition between them for linguistic processing. Introduction *For correspondence: The seminal speech-shadowing experiments conducted in the 50 s and 60 s set the stage for study- [email protected] (PH-Y); ing one of the primary cognitive challenges encountered in daily life: how do our perceptual and lin- [email protected] (EZG) guistic systems deal effectively with competing speech inputs? (Cherry, 1953; Broadbent, 1958; Treisman, 1960).
    [Show full text]
  • Dichotic Training in Children with Auditory Processing Disorder
    International Journal of Pediatric Otorhinolaryngology 110 (2018) 114–117 Contents lists available at ScienceDirect International Journal of Pediatric Otorhinolaryngology journal homepage: www.elsevier.com/locate/ijporl Dichotic training in children with auditory processing disorder T ∗ Maryam Delphia, Farzaneh Zamiri Abdollahib, a Musculoskeletal Rehabilitation Research Center, Ahvaz Jundishapur University of Medical Sciences, Ahvaz, Iran b Audiology Department, Tehran University of Medical Sciences, Tehran, Iran ARTICLE INFO ABSTRACT Keywords: Objectives: Several test batteries have been suggested for auditory processing disorder (APD) diagnosis. One of Dichotic listening the important tests is dichotic listening tests. Significant ear asymmetry (usually right ear advantage) can be Auditory training indicative of (APD). Two main trainings have been suggested for dichotic listening disorders: Differential Auditory processing Interaural Intensity Difference (DIID) and Dichotic Offset Training (DOT). The aim of the present study was comparing the efficacy of these two trainings in resolving dichotic listening disorders. Methods: 12 children in the age range of 8 to 9 years old with APD were included (mean age 8.41 years old ± 0.51). They all had abnormal right ear advantage based on established age-appropriate norms for Farsi dichotic digit test. Then subjects were randomly divided into two groups (each contained 6 subjects): group 1 received DIID training (8.33 years old ± 0.51) and group 2 received DOT training (8.50 years old ± 0.54). Results: Both trainings were effective in improvement of dichotic listening. There was a significant difference between two trainings with respect to the length of treatment (P-value≤0.001). DOT needed more training sessions (12.83 ± 0.98 sessions) than DIID (21.16 ± 0.75 sessions) to achieve the same amount of performance improvement.
    [Show full text]
  • Neural Encoding of Attended Continuous Speech Under Different Types of Interference
    Neural Encoding of Attended Continuous Speech under Different Types of Interference Andrea Olguin, Tristan A. Bekinschtein, and Mirjana Bozic Abstract ■ We examined how attention modulates the neural encoding Critically, however, the type of the interfering stream significantly of continuous speech under different types of interference. In modulated this process, with the fully intelligible distractor an EEG experiment, participants attended to a narrative in English (English) causing the strongest encoding of both attended and while ignoring a competing stream in the other ear. Four different unattended streams and latest dissociation between them and types of interference were presented to the unattended ear: a nonintelligible distractors causing weaker encoding and early dis- different English narrative, a narrative in a language unknown to sociation between attended and unattended streams. The results the listener (Spanish), a well-matched nonlinguistic acoustic inter- were consistent over the time course of the spoken narrative. ference (Musical Rain), and no interference. Neural encoding of These findings suggest that attended and unattended information attended and unattended signals was assessed by calculating can be differentiated at different depths of processing analysis, cross-correlations between their respective envelopes and the with the locus of selective attention determined by the nature EEG recordings. Findings revealed more robust neural encoding of the competing stream. They provide strong support to flexible for the attended envelopes compared with the ignored ones. accounts of auditory selective attention. ■ INTRODUCTION “late selection” approaches. The early selection theory Directingattentiontoasinglespeakerinamultitalker (Broadbent, 1958) argued that, because of our limited environment is an everyday occurrence that we manage processing capacity, attended and unattended informa- with relative ease.
    [Show full text]
  • Individual Differences in Working Memory Capacity and Divided Attention in Dichotic Listening
    Psychonomic Bulletin & Review 2007, 14 (4), 699-703 Individual differences in working memory capacity and divided attention in dichotic listening GREGORY J. H. COLFLESH University of Illinois, Chicago, Illinois AND ANDREW R. A. CONWAY Princeton University, Princeton, New Jersey The controlled attention theory of working memory suggests that individuals with greater working memory capacity (WMC) are better able to control or focus their attention than individuals with lesser WMC. This re- lationship has been observed in a number of selective attention paradigms including a dichotic listening task (Conway, Cowan, & Bunting, 2001) in which participants were required to shadow words presented to one ear and ignore words presented to the other ear. Conway et al. found that when the participant’s name was presented to the ignored ear, 65% of participants with low WMC reported hearing their name, compared to only 20% of participants with high WMC, suggesting greater selective attention on the part of high WMC participants. In the present study, individual differences in divided attention were examined in a dichotic listening task, in which participants shadowed one message and listened for their own name in the other message. Here we find that 66.7% of high WMC and 34.5% of low WMC participants detected their name. These results suggest that as WMC capacity increases, so does the ability to control the focus of attention, with high WMC participants being able to flexibly “zoom in” or “zoom out” depending on task demands. Since Baddeley and Hitch (1974) first published their tasks). They found no difference between high and low seminal chapter on working memory (WM), many theo- WMC participants in the prosaccade task but high WMC ries regarding the construct have been proposed.
    [Show full text]
  • Broadbent's Filter Theory Cherry: the Cocktail Party Problem
    194 Part II ● Cognitive psychology KEY STUDY EVALUATION — Cherry Cherry: The cocktail party problem The research by Colin Cherry is a very good example of how a psycholo- gist, noticing a real-life situation, is able to devise a hypothesis and carry Cherry (1953) found that we use physical differences between out research in order to explain a phenomenon, in this case the “cocktail the various auditory messages to select the one of interest. party” effect. Cherry tested his ideas in a laboratory using a shadowing These physical differences include differences in the sex of the technique and found that participants were really only able to give infor- speaker, in voice intensity, and in the location of the speaker. mation about the physical qualities of the non-attended message When Cherry presented two messages in the same voice to (whether the message was read by a male or a female, or if a tone was both ears at once (thereby removing these physical differences), used instead of speech). Cherry’s research could be criticised for having moved the real-life phenomenon into an artificial laboratory setting. the participants found it very hard to separate out the two However, this work opened avenues for other researchers, beginning with messages purely on the basis of meaning. Broadbent, to elaborate theories about focused auditory attention. Cherry (1953) also carried out studies using a shadowing task, in which one auditory message had to be shadowed (repeated back out aloud) while a second auditory message was presented to the other ear. Very little information seemed to be obtained from the second or non-attended message.
    [Show full text]
  • Evaluating CAPDOTS Training for Children and Adolescents with APD by Jay R
    ONLINE EXCLUSIVE Evaluating CAPDOTS Training for Children and Adolescents with APD By Jay R. Lucker, EdD, CCC-A/SLP, FAAA; Cydney Fox, AuD; and Bea Braun, AuD uditory processing disorders (APD) in school-age children can lead to learning problems.1 Audiolo- gists may determine the presence or absence of APD in this population and make specific thera- Apeutic recommendations. However, many therapies for APD have not been peer-reviewed or examined with statistical analysis to determine efficacy. The present study evaluates a therapeutic option called CAPDOTS-Integrated (also referred to as CAPDOTS), an online training program that aims to im- prove dichotic listening problems. According to Jerger, “Dichotic listening (DL) tests are at the core of the diagnostic evaluation of auditory processing disorder... [Such tests] have been used for decades both as 2 screening tools and as diagnostic tests in APD evaluation.” Thus, audiologists evaluating school-aged children for APD AdobeStock may find these children to have dichotic listening problems. Those diagnosed with dichotic listening problems are rec- subjects.2 The 27-year-old subject identified as diagnosed ommended to try a dichotic listening training program, such with “binaural integration deficit” only completed 80 percent as CAPDOTS-Integrated.1 Presently, CAPDOTS-Integrated of the CAPDOTS training. The 16-year-old subject is re- claims to treat binaural auditory integration problems known ported diagnosed with a binaural integration (dichotic listen- as dichotic listening difficulties. CAPDOTS is a treatment ing) auditory processing disorder. Results of mid-latency program requiring access to the internet and good quality electrophysiological measures revealed increased response headphones.
    [Show full text]
  • Attentional Strategies in Dichotic Listening
    Memory & Cognition 1979, Vol. 7 (6),511.520 Attentional strategies in dichotic listening JACK BOOKBINDER and ELI OSMAN Brooklyn College ofthe City University ofNew York, Brooklyn, New York 11210 A person can attend to a message in one ear while seemingly ignoring a simultaneously presented verbal message in the other ear. There is considerable controversy over the extent to which the unattended message is actually processed. This issue was investigated by presenting dichotic messages to which the listeners responded by buttonpressing (not shadow­ ing) to color words occurring in the primary ear message while attempting to detect a target word in either the primary ear or secondary ear message. Less than 40% of the target words were detected in the secondary ear message, whereas for the primary ear message (and also for either ear in a control experiment), target detection was approximately 80%. Furthermore. there was a significant negative correlation between buttonpressing performance and secondary ear target-detection performance. The results were interpreted as being inconsistent with automatic processing theories of attention. The dichotic listening paradigm, beginning with the processing system. Early selection models (Broadbent, experiments of Broadbent (1958) and Cherry (1953), 1958,1971; Treisman, 1960, 1964; Treisman & Geffen, has played a central role in the study of attention during 1967) allow for analyses of simple physical features the past two decades. In a dichotic presentation, differ­ simultaneously (e.g., spatial location, voice pitch), while ent auditory messages are simultaneously presented to it is postulated that the simultaneous higher order (Le., the subject's right and left ears by means of stereo head­ semantic context) analysis of all inputs would overload phones.
    [Show full text]
  • An Optimal Dichotic-Listening Paradigm for the Assessment of Hemispheric
    1 Research article 2 An optimal dichotic-listening paradigm for the assessment of hemispheric 3 dominance for speech processing 4 5 René Westerhausen (1,2)* & Fredrik Samuelsen (2) 6 7 (1) Department of Psychology, University of Oslo, Norway 8 (2) Department of Biological and Medical Psychology, University of Bergen, Norway 9 10 Running title: Optimal dichotic-listening paradigm 11 12 13 *Corresponding author 14 René Westerhausen 15 Department of Psychology, University of Oslo, 16 POB 1094 Blindern, 0317 Oslo, Norway 17 email: [email protected], Phone: (+47) 228 45230 18 1 19 Abstract 20 Dichotic-listening paradigms are widely accepted as non-invasive tests of hemispheric 21 dominance for language processing and represent a standard diagnostic tool for the 22 assessment of developmental auditory and language disorders. Despite its popularity in 23 research and clinical settings, dichotic paradigms show comparatively low reliability, 24 significantly threatening the validity of conclusions drawn from the results. Thus, the aim of 25 the present work was to design and evaluate a novel, highly reliable dichotic-listening 26 paradigm for the assessment of hemispheric differences. Based on an extensive literature 27 review, the paradigm was optimized to account for the main experimental variables which are 28 known to systematically bias task performance or affect random error variance. The main 29 design principle was to minimize the relevance of higher cognitive functions on task 30 performance in order to obtain stimulus-driven laterality estimates. To this end, the key 31 design features of the paradigm were the use of stop-consonant vowel (CV) syllables as 32 stimulus material, a single stimulus pair per trial presentation mode, and a free recall (single) 33 response instruction.
    [Show full text]
  • A Listening Program, in the Treatment Of
    The use of LACETM, a listening program, in the treatment of problems associated with dichotic listening disorders. A Thesis Presented in Partial Fulfillment of the Requirements for The Degree Bachelors of Arts in Speech and Hearing Science with Honors in Research Distinction in the College of Arts and Sciences of The Ohio State University By Mariah N. Cheyney **** The Ohio State University 2009 Dr. Gail M. Whitelaw, Adviser Abstract Dichotic listening, or how two ears work together as a team, is critical for localizing sound sources and when listening in the presence of complex background noise. Disorders of dichotic listening can be caused by a number of issues, including neurological diseases such as multiple sclerosis and traumatic brain injury, and can result in communication difficulties in daily listening situations. A number of diagnostic protocols and management programs have recently been developed to address this population, based in part from interest in veterans returning from the Middle East with these types of disorders. Currently, there is no standard for assessment or rehabilitation of dichotic listening disorders. This study utilized a single-subject research design to address the potential effectiveness of a treatment program for dichotic listening difficulties. The subject presented with clinical deficits following a stroke, despite having normal hearing acuity. Auditory processing evaluation revealed severe deficits in the area of dichotic listening, most remarkably for the left ear. The patient was enrolled in the LACETM (Listening and Communications Enhancement) program, a computer-based aural rehabilitation program developed to assist patients with hearing loss acclimatize to hearing aids. The program is designed to improve listening skills through use of an adaptive program that addresses a number of auditory processing skills.
    [Show full text]
  • What Can Be Learned About Brain Function from Dichotic Listening ?
    Revista Española de Neuropsicología Copyright ©2000 de REN Vol. 2, No. 3, 62-84 ISSN: 1139-9872 What can be learned about brain function from dichotic listening ?. Kenneth Hugdahl University of Bergen A series of studies on the use of dichotic listening as a method to study brain-behavior interactions, including hemispheric asymmetry are reviewed. The dichotic listening method is presented from a historical perspective, followed by an outline of the empirical procedure used in our laboratory. Different methods of validating the dichotic listening procedure against both invasive and non-invasive techniques, including PET blood flow recordings are then presented. The paper is ended with some examples of clinical applications of the dichotic listening technique on brain damaged patients. A major argument in the present paper is that the dichotic listening technique is a method to study the interaction between bottom-up, or stimulus-driven, versus top-down, or instruction-driven laterality. This opens up for a more dynamic and interactive view of brain laterality than the traditional static view that the brain is lateralized only for specific stimuli and stimulus properties. ¿Qué se puede Aprender sobre el Funcionamiento del Cerebro a partir de la Escucha Dicótica? Se revisan una serie de estudios sobre el uso del estudio de la escucha dicótica como método para estudiar las interacciones entre cerebro y comportamiento, incluyendo la asimetría interhemisférica. Se presenta el método de la escucha dicótica desde una perspectiva histórica seguido por una revisión de los métodos empíricos utilizados en nuestro laboratorio. Posteriormente, se presentan diferentes métodos para validar este procedimiento en contra de las técnicas invasivas y no invasivas.
    [Show full text]
  • Dichotic Listening Deficits in Amblyaudia Are Characterized by Aberrant Neural Oscillations in Auditory Cortex
    Clinical Neurophysiology 132 (2021) 2152–2162 Contents lists available at ScienceDirect Clinical Neurophysiology journal homepage: www.elsevier.com/locate/clinph Dichotic listening deficits in amblyaudia are characterized by aberrant neural oscillations in auditory cortex ⇑ Sara Momtaz a, , Deborah Moncrieff a, Gavin M. Bidelman a,b,c a School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA b Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA c University of Tennessee Health Sciences Center, Department of Anatomy and Neurobiology, Memphis, TN, USA article info highlights Article history: Amblyaudia is characterized by a stark interaural asymmetry in dichotic listening performance. Accepted 29 April 2021 Stronger phase-locked b/c oscillations within the bilateral auditory cortex for stimuli presented to the Available online 4 June 2021 right ear in amblyaudia. Weaker functional connectivity in amblyaudia from right to left auditory cortex, despite overall stron- Keywords: Auditory processing disorders (APD) ger neural responses. Event-related brain potentials (ERPs) Gamma/beta band response Hemispheric asymmetries abstract Phase-locking Time-frequency analysis Objective: Children diagnosed with auditory processing disorder (APD) show deficits in processing com- plex sounds that are associated with difficulties in higher-order language, learning, cognitive, and com- municative functions. Amblyaudia (AMB) is a subcategory of APD characterized by abnormally large ear asymmetries in dichotic listening tasks. Methods: Here, we examined frequency-specific neural oscillations and functional connectivity via high- density electroencephalography (EEG) in children with and without AMB during passive listening of non- speech stimuli. Results: Time-frequency maps of these ‘‘brain rhythms” revealed stronger phase-locked beta-gamma (~35 Hz) oscillations in AMB participants within bilateral auditory cortex for sounds presented to the right ear, suggesting a hypersynchronization and imbalance of auditory neural activity.
    [Show full text]
  • Language Processing and the Human Brain
    01:615:201 Introduction to Linguistic Theory Adam Szczegielniak Language Processing and the Human Brain Copyright in part: Cengage learning The Human Mind at Work • Psycholinguistics: the study of linguistic performance in speech production and comprehension – Linking psychological mechanisms to grammar to understand language production and comprehension Comprehension: The Speech Signal • When we push air through the glos, vibrang vocal cords produce variaons in air pressure – The speed of these variaons in air pressure determines the fundamental frequency of the sounds • Fundamental frequency is perceived as pitch by the hearer – The magnitude of the variaons (or intensity) determines the loudness of the sound Comprehension: The Speech Signal • An image of the speech signal is displayed in a spectrogram • Vowels are indicated by dark bands called formants Parsing of sound • Categorical percep5on – We do not perceive linguis5c sounds as a connuum • Duplex percepon – Are able to integrate ‘spliced’ parts of a sound played into each ear VOT • Voice onset 5me differen5ates voiced sounds from voiceless p b t d k g The six English plosives [p, b, t, d, k, g], each followed by the vowel [a]. Top picture is of spectrograms. The y-axis represents frequency range 0-5kHz, with each 1kHz marked by a horizontal gray line. The x-axis is time - about 4s overall. The bottom picture is the same data in time aligned waveforms. ©Kate Morton 1995 http://www.essex.ac.uk/speech/material/kate/ref/vot-13.html VOT percepon %identifying identical cat ba pa VOT Formants - how to splice a sound Frequency response curves (indicating the preferred resonating frequencies of the vocal tract).
    [Show full text]