Intersensory Conflict Between Vision and Touch: the Response Modality Dominates When Precise, Attention-Riveting Judgments Are Required

Total Page:16

File Type:pdf, Size:1020Kb

Intersensory Conflict Between Vision and Touch: the Response Modality Dominates When Precise, Attention-Riveting Judgments Are Required Perception & Psychophysics 1999,61 (7), 1384-1398 Intersensory conflict between vision and touch: The response modality dominates when precise, attention-riveting judgments are required MORTON A. HELLER Winston-Salem State University, Winston-Salem, North Carolina JEFFREY A. CALCATERRA Wright State University, Dayton, Ohio and SHAVONDA L. GREEN and LATONYA BROWN Winston-Salem State University, Winston-Salem, North Carolina In four experiments, reducing lenses were used to minifyvision and generate intersensory size con­ flicts between vision and touch. Subjects made size judgments, using either visual matching or haptic matching. In visual matching, the subjects chose from a set of visible squares that progressively in­ creased in size. In haptic matching, the subjects selected matches from an array of tangible wooden squares. In Experiment 1,it was found that neither sense dominated when subjects exposed to an in­ tersensory discrepancy made their size estimates by using either visual matching or haptic matching. Size judgments were nearly identical for conflict subjects making visual or haptic matches. Thus, matching modality did not matter in Experiment 1. In Experiment 2, it was found that subjects were influenced by the sight of their hands, which led to increases in the magnitude of their size judgments. Sight of the hands produced more accurate judgments, with subjects being better able to compensate for the illusory effects of the reducing lens. In two additional experiments, it was found thatwhen more precise judgments were required and subjects had to generate their own size estimates, the response modality dominated. Thus, vision dominated in Experiment 3, where size judgments derived from view­ inga metric ruler, whereas touch dominated in Experiment 4, where subjects made size estimates with a pincers posture of their hands. It is suggested that matching procedures are inadequate for assessing intersensory dominance relations. These results qualifythe position (Hershberger & Misceo, 1996)that the modality of size estimates influences the resolution of intersensory conflicts. Only when required to self-generate more precise judgments did subjects rely on one sense, either vision or touch. Thus, task and attentional requirements influence dominance relations, and vision does not invariably pre­ vail over touch. Researchers have long found that vision is dominant sense of vision and the sense of touch (Rock & Harris, over other modalities when conflicting information is 1967; Rock & Victor, 1964). The subjects in Rock's ex­ presented (e.g., Ryan, 1940, p. 687). Rock and his col­ periments felt a square through a cloth while looking at leagues reported an influential series ofexperiments dem­ it through a reducing lens, which minifies vision, and made onstrating visual dominance, given a conflict between the a single judgment about the size (or shape) ofthe stimu­ lus object. Rock reported that subjects generally relied on vision for their judgments, despite the fact that this yielded Preparation of this report and some of the research were supported nonveridical perception. These experiments have been in­ by NIH Grant 2 S06 RR-08040. Faith Heller provided helpful com­ fluential; they have prompted a great deal ofresearch and ments on an earlier version of this manuscript. Cindy Flynt assisted have convinced many people that vision is invariably dom­ with data collection and analyses. We are grateful to Richard Williams inant over touch. for advice about optics and optical measurements on the lens used in Experiments I and 2. Lester Krueger, Mark Hollins, and two anonymous The visual dominance literature has relied heavily on the reviewers made helpful comments on this manuscript. Correspondence discrepancy paradigm introduced by Rock and Victor concerning this article should be addressed to M. A. Heller, Depart­ (1964), which presupposes that subjects will assume that ment ofPsychology, Eastern Illinois University, 119 Physical Science they are feeling and seeing the same object. There is no real Bldg., 600 Lincoln Avenue, Charleston, IL 61920-3099 (e-mail: cfmah [email protected]). conflict between vision and touch ifwe look at one object and feel another. Rock and Victor used a variety ofresponse -r-Accepted by previous editor, Myron L. Braunstein measures in their study. Their subjects showed visual dom- Copyright 1999 Psychonomic Society, Inc. 1384 INTERSENSORY CONFLICT 1385 inance when making haptic matches, when drawing the in matching procedures is overloaded. Under many nor­ stimuli using sight, and when making visual matches. mal circumstances, people see their hands as they feel The Rock and Victor (1964) discrepancy paradigm objects (Heller, 1993) and this visual information may makes the assumption that subjects will have no difficulty be critical for the resolution of intersensory discrepan­ making accurate judgments about length with both vi­ cies. Thus, ifsubjects see their hands move in a direction sion and touch. However, things are not that simple, and that is counter to proprioception, most do not believe vi­ there are a number of reasons to doubt that subjects can sion (Heller, 1992). make comparably accurate simultaneous or sequential comparisons through both modalities (see Freides, 1974, Hershberger and Misceo's (1996) Study 1975). There are reports indicating that conditions ofde­ A recent, interesting report by Hershberger and Mis­ lay can influence haptic or cross-modal conditions more ceo (1996) is clearly relevant to the present issue. Hersh­ than they influence visual judgments (Abravanel, 1973; berger and Misceo used a variant of Rock and Victor's Cashdan, 1968; Cashdan & Zung, 1970; Zung, Butter, & (1964) paradigm and had subjects feel a square while Cashdan, 1974). Thus, memory demands could affect viewing it through a negative meniscus lens that halved resolution ofconflicts between vision and touch, because the size ofthe visual image. They reported that their sub­ the Rock and Harris (1967) discrepancy paradigm forces jects resolved the conflict between vision and touch in subjects to remember the stimulus experienced during favor oftouch, in haptic matching conditions. When vi­ the conflict trial. It is very possible that the results are sual matches were required, the subjects showed visual limited to differential memory skills for the senses ofvi­ dominance. When the data were examined on the first trial sion and touch. Other studies indicate that subjects have block, mean matching size varied with modality of the difficulty making size estimates with touch. Bartley match in discrepancy conditions (see Hershberger & Mis­ (1953) assumed that tactual judgments were generally ceo, 1996, Figure 2). Subjects making visual matches in modified by visual imagery and, also, that visual judg­ conflict conditions judged the stimuli to be slightly less ments ofsize were influenced by true distance. Schlater, than 3.5 cm (see Figure 2 in Hershberger & Misceo, 1996). Baker, and Wapner ( 1981) reported that perceived arm Visual controls (intramodal, vision alone, with vision of length was different when subjects were active or passive the standard stimuli through a reducing lens) judged the and actually shrank when subjects were passively touched. stimuli to be less than 2.5 ern. Note that, in the intersen­ These reports conflict with recent empirical evidence that sory conflict literature, the term control refers to condi­ suggests that people are able to make veridical, equiva­ tions in which there is no conflict. Conflict judgments lentjudgments oflength with vision and touch (Abravanel, with haptic matching (M greater than 4 em) were closer 1971; Teghtsoonian & Teghtsoonian, 1970). to the mean haptic control (haptic standards and haptic Recently, Power (1981) verified many ofRock's (Rock matches) judgments (M slightly greater than 4.5 em; see & Harris, 1967; Rock & Victor, 1964) findings and found Figure 2 in Hershberger & Misceo, 1996). that visual dominance occurs with a familiar object­ Also, in conflict conditions, subjects felt the squares namely, a coin. Furthermore, the phenomenon ofvisual (4,4.5,5,5.5, and 6 cm-) through an intervening cloth. dominance seemed robust, since it even occurred with the Hershberger and Misceo (1996) assumed that the pres­ bare hand. However, a strong lens tended to reduce the ence of the cloth made it impossible for subjects to see reliance on vision when the bare hand was visible (Power, their hands. However, one can readily see the movement 1980). Power used a lens that distorted the shape of vis­ of one's fingers through an intervening opaque cloth, ible objects, causing a square to appear rectangular. It is since motions ofthe cloth are visible (Heller, 1983). Con­ possible that this distortion was responsible for a shift in ceivably, individuals are then able to perceive a reduction apparent reliance on the senses of vision and touch. in finger span, using sight. If so, it is possible that changes There are limits to visual dominance. It is not evident in perceived size over trial block could reflect the influ­ with texture perception (Heller, 1982, 1989; Lederman ence ofthis knowledge ofminification, acquired through & Abbott, 1981) or when vision is blurry (Heller, 1983). vision, rather than the possible influence of haptics on Also, haptic dominance can occur in pattern perception resolution ofthe conflict. This possibility was examined when vision is manipulated by a mirror, rather than by a in Experiment 2 ofthe present study. lens (Heller, 1992).
Recommended publications
  • Crossmodal Attention Jon Driver and Charles Spence†
    245 Crossmodal attention Jon Driver∗ and Charles Spence² Most selective attention research has considered only a multimodal representations of space, within which atten- single sensory modality at a time, but in the real world, tion can be directed. our attention must be coordinated crossmodally. Recent studies reveal extensive crossmodal links in attention across the various modalities (i.e. audition, vision, touch Attending to one sensory modality versus and proprioception). Attention typically shifts to a common another location across the modalities, despite the vast differences The most basic crossmodal issue concerning attention is in their initial coding of space. These spatial synergies whether people can attend selectively to one modality in attention can be maintained even when receptors are at the expense of others, or whether the modalities realigned across the modalities by changes in posture. are so independent that concentrating on one has no Some crossmodal integration can arise preattentively. The implications for the others. As long ago as 1893, Wundt mechanisms underlying these crossmodal links can be [1] claimed that attention can speed up the processing of examined in a convergent manner by integrating behavioural a relevant modality, while delaying irrelevant modalities; studies of normal subjects and brain-damaged patients with many subsequent authors have concurred. However, as •• neuroimaging and neurophysiological studies. we have pointed out elsewhere [2 ], most studies on this issue have suffered from methodological ¯aws. For instance, stimuli in different modalities were typically Addresses presented from different spatial positions, so attention to ∗ Institute of Cognitive Neuroscience, Department of Psychology, a modality was confounded with attention to a location.
    [Show full text]
  • Variance Adaptation in Navigational Decision Making Ruben Gepner1, Jason Wolk1, Digvijay Shivaji Wadekar1, Sophie Dvali1, Marc Gershow1,2,3*
    RESEARCH ARTICLE Variance adaptation in navigational decision making Ruben Gepner1, Jason Wolk1, Digvijay Shivaji Wadekar1, Sophie Dvali1, Marc Gershow1,2,3* 1Department of Physics, New York University, New York, United States; 2Center for Neural Science, New York University, New York, United States; 3Neuroscience Institute, New York University, New York, United States Abstract Sensory systems relay information about the world to the brain, which enacts behaviors through motor outputs. To maximize information transmission, sensory systems discard redundant information through adaptation to the mean and variance of the environment. The behavioral consequences of sensory adaptation to environmental variance have been largely unexplored. Here, we study how larval fruit flies adapt sensory-motor computations underlying navigation to changes in the variance of visual and olfactory inputs. We show that variance adaptation can be characterized by rescaling of the sensory input and that for both visual and olfactory inputs, the temporal dynamics of adaptation are consistent with optimal variance estimation. In multisensory contexts, larvae adapt independently to variance in each sense, and portions of the navigational pathway encoding mixed odor and light signals are also capable of variance adaptation. Our results suggest multiplication as a mechanism for odor-light integration. DOI: https://doi.org/10.7554/eLife.37945.001 Introduction The world is not fixed but instead varies dramatically on multiple time scales, for example from cloudy to sunny, day to night, or summer to winter; and contexts: from sun to shade, or burrowing *For correspondence: [email protected] to walking. Nervous systems must respond appropriately in varied environments while obeying bio- physical constraints and minimizing energy expenditure (Niven and Laughlin, 2008).
    [Show full text]
  • Abstract the EFFECTS of PRESENTATION MODALITY ON
    Abstract THE EFFECTS OF PRESENTATION MODALITY ON LEARNING AND MEMORY PERFORMANCE IN CHILDREN WITH SPECIFIC LEARNING DISABILITIES IN THE AREA OF LANGUAGE By Marijo F. O’Connell This study investigated the learning and memory patterns in low-income, African- American children ages 9-13 with specific learning disabilities in the area of language (SLD/SLI) as compared to typically developing peers. All subjects (n=34) were administered the California Verbal Learning Test for Children (CVLT-C) and the experimental tasks. The experimental tasks measured supralist free recall under three presentation modalities; auditory, visual and simultaneous presentation of auditory and visual. Results indicate SLD/SLI children show significantly slower rates of learning than controls. Additionally, the visual presentation modality, with or without auditory information, demonstrates significantly better learning rates than the auditory modality alone. The results support the hypotheses stating that SLD/SLI children learn verbal information at a slower rate than typical children and the superiority of the visual presentation modality for verbal learning and memory. Effect of Presentation Modality on Learning and Memory Performance in Children with Specific Learning Disabilities in the Area of Language A Thesis Submitted to Miami University in partial fulfillment of the requirements for the degree of Master of Arts Department of Speech Pathology and Audiology by Marijo F. O’Connell Miami University Oxford, Ohio 2005 Advisor: ______________________________
    [Show full text]
  • Peri-Personal Space As a Prior in Coupling Visual and Proprioceptive
    www.nature.com/scientificreports OPEN Peri-personal space as a prior in coupling visual and proprioceptive signals Received: 24 April 2018 Jean-Paul Noel1,2, Majed Samad1,3, Andrew Doxon1, Justin Clark1, Sean Keller1 & Accepted: 7 October 2018 Massimiliano Di Luca1,4 Published: xx xx xxxx It has been suggested that the integration of multiple body-related sources of information within the peri-personal space (PPS) scafolds body ownership. However, a normative computational framework detailing the functional role of PPS is still missing. Here we cast PPS as a visuo-proprioceptive Bayesian inference problem whereby objects we see in our environment are more likely to engender sensations as they come near to the body. We propose that PPS is the refection of such an increased a priori probability of visuo-proprioceptive coupling that surrounds the body. To test this prediction, we immersed participants in a highly realistic virtual reality (VR) simulation of their right arm and surrounding environment. We asked participants to perform target-directed reaches toward visual, proprioceptive, and visuo-proprioceptive targets while visually displaying their reaching arm (body visible condition) or not (body invisible condition). Reach end-points are analyzed in light of the coupling prior framework, where the extension of PPS is taken to be represented by the spatial dispersion of the coupling prior between visual and proprioceptive estimates of arm location. Results demonstrate that if the body is not visible, the spatial dispersion of the visuo-proprioceptive coupling relaxes, whereas the strength of coupling remains stable. By demonstrating a distance-dependent alteration in visual and proprioceptive localization attractive pull toward one another (stronger pull at small spatial discrepancies) when the body is rendered invisible – an efect that is well accounted for by the visuo- proprioceptive coupling prior – the results suggest that the visible body grounds visuo-proprioceptive coupling preferentially in the near vs.
    [Show full text]
  • Control Systems and Homeostasis
    About This Chapter • General properties of sensory systems • Somatic senses • Chemoreception: smell and taste • The ear: hearing • The ear: equilibrium • The eye and vision © 2016 Pearson Education, Inc. Sensory Pathways • Stimulus as physical energy sensory receptor – Receptor acts as a transducer • Intracellular signal usually change in membrane potential • Stimulus threshold action potential to CNS • Integration in CNS cerebral cortex or acted on subconsciously © 2016 Pearson Education, Inc. © 2016 Pearson Education, Inc. Receptors to Particular Forms of Energy • Naked (“free”) nerves • Complex neural receptors encased in connective tissue capsules • Smell receptors are neurons • Non-neural receptors for four special senses © 2016 Pearson Education, Inc. Receptors to Particular Forms of Energy • Chemoreceptors respond to chemical ligands: taste, smell • Mechanoreceptors respond to mechanical energy pressure and sound: hearing • Thermoreceptors respond to temperature • Photoreceptors for vision respond to light © 2016 Pearson Education, Inc. Figure 10.1 Simple, complex, and nonneural sensory receptors Simple receptors are neurons Complex neural receptors have nerve Most special senses receptors are cells with free nerve endings. They endings enclosed in connective tissue capsules. that release neurotransmitter onto sensory may have myelinated or This illustration shows a Pacinian corpuscle, neurons, initiating an action potential. The unmyelinated axons. which senses touch. cell illustrated is a hair cell, found in the ear. Stimulus
    [Show full text]
  • Sensory Receptors A17 (1)
    SENSORY RECEPTORS A17 (1) Sensory Receptors Last updated: April 20, 2019 Sensory receptors - transducers that convert various forms of energy in environment into action potentials in neurons. sensory receptors may be: a) neurons (distal tip of peripheral axon of sensory neuron) – e.g. in skin receptors. b) specialized cells (that release neurotransmitter and generate action potentials in neurons) – e.g. in complex sense organs (vision, hearing, equilibrium, taste). sensory receptor is often associated with nonneural cells that surround it, forming SENSE ORGAN. to stimulate receptor, stimulus must first pass through intervening tissues (stimulus accession). each receptor is adapted to respond to one particular form of energy at much lower threshold than other receptors respond to this form of energy. adequate (s. appropriate) stimulus - form of energy to which receptor is most sensitive; receptors also can respond to other energy forms, but at much higher thresholds (e.g. adequate stimulus for eye is light; eyeball rubbing will stimulate rods and cones to produce light sensation, but threshold is much higher than in skin pressure receptors). when information about stimulus reaches CNS, it produces: a) reflex response b) conscious sensation c) behavior alteration SENSORY MODALITIES Sensory Modality Receptor Sense Organ CONSCIOUS SENSATIONS Vision Rods & cones Eye Hearing Hair cells Ear (organ of Corti) Smell Olfactory neurons Olfactory mucous membrane Taste Taste receptor cells Taste bud Rotational acceleration Hair cells Ear (semicircular
    [Show full text]
  • The Effects of Stimulus Modality and Task Integrality: Predicting Dual-Task Performance
    THE EFFECTS OF STIMULUS MODALITY AND TASK INTEGRALITY: PREDICTING DUAL-TASK PERFORMANCE. AND WORKLOAD FROM SINGLETASK LEVELS Sandra G. Hart, Robert J. Shively hlichael A. Vidulich NASA-Ames Research Center Moffett Field. CA Ronald C. Miller Informatics General Corporation Palo Alto. CA ABSTRACT The influence of stimulus modality and task difficulty on workload and perfor- mance was investigated in the current study. The goal ums to quantify the "cost" (in terms of response time and ezperienced workload) incurred when essentially serial task components shared common elements (e.g., the response to one initiated the other) which could be accomplished in parallel. The ezperimental tasks were based on the "Fittsberg" paradigm; the solution to a SternBERG-type memory task deter- mines which of two identical FITTS targets are acquired. Previous research suggested that such functionally integrated "dual" tasks are performed with substan- tially less workload and faster response times than would be predicted by sum- ming single-task components when both are presented in the same stimulus modality (visual). In the current study, the physical integration of task elements ums varied (although their functional relationship remained the same) to determine whether dual-task facilitation would persist if task components were presented in different sensory modalities. Again, ii was found that the cost of performing the two-stage task was considerably less than the sum of component single-task levels when both were presented visually. Less facilitation was found when task elements were presented in different sensory modalities. These results suggest the impor- tance of distinguishing between concurrent tasks that compete for limited resources from those that beneficially share common resources when selecting the stimulus modalities jor information displays.
    [Show full text]
  • Self-Prioritization in Perception 1 RUNNING HEAD: Self-Prioritization
    Self-Prioritization in Perception RUNNING HEAD: Self-Prioritization in Perception Self-prioritization beyond perception Sarah Schäfer1, Dirk Wentura2, & Christian Frings1 1 University of Trier, 2 Saarland University Word count main text (5321 words) Correspondence: Christian Frings University of Trier Cognitive Psychology D-54286 Trier, Germany [email protected] 1 Self-Prioritization in Perception Abstract Recently, Sui and colleagues (2012) introduced a new paradigm to measure perceptual self-prioritization processes. It seems that arbitrarily tagging shapes to self-relevant words (I, my, me, and so on) leads to speeded verification times when matching self-relevant word shape pairings (e.g., me – triangle) as compared to non-self-relevant word shape pairings (e.g., stranger – circle). In order to analyze the level at which self-prioritization takes place we analyzed whether the self-prioritization effect is due to a tagging of the self-relevant label and the particular associated shape or due to a tagging of the self with an abstract concept. In two experiments participants showed standard self-prioritization effects with varying stimulus features or different exemplars of a particular stimulus category suggesting that self- prioritization also works at a conceptual level. Keywords: self-prioritization, feature variance, conceptual processing 2 Self-Prioritization in Perception A function of self-relevance is that it guides our attention and behavior as self-relevant stimuli indicate important sensory inputs for our cognitive system. Accordingly, modulations of attentional processing due to self-relevance has been shown in different paradigms of selec- tive attention (e.g., Alexopoulos, Muller, Ric, & Marendaz, 2012; Bargh, 1982; Giesbrecht, Sy, & Lewis, 2009; Shapiro, Caldwell, & Sorensen, 1997).
    [Show full text]
  • Stimulus Modality and Short-Term Memory in Rats
    Animal Learning & Behavior 1980,8 (1),10-]6 Stimulus modality and short-term memory in rats JULIA WALLACE Princeton University, Princeton, New Jersey 08540 and PAMELA A. STEINERT, STANLEY R. SCOBIE, and NORMAN E. SPEAR State University ofNew York, Binghamton, New York 13901 Two experimental paradigms are presented aimed at determining the retention of auditory and visual information over brief delay intervals. First, a conditional delayed matching-to­ sample procedure was used in which rats were required to symbolically match the modality of the sample stimulus with one of two comparison stimuli. In the second experiment, subjects were trained and tested using a Konorski-type procedure. Despite the conceptual and procedural differences between the two procedures, subjects in both experiments showed steeper forgetting functions for visual events than for auditory events, while performance levels at O-sec delay intervals were equivalent for both stimuli. These results, when taken together with related research conducted with pigeons, suggest that content of memory may have important influenceson the short-term retention abilities ofanimal subjects. Surprisingly little is known about forgetting of oped for rats is especially important because so much recent events in the laboratory rat. Although Hunter complementary knowledge is available about other (1913) reported that rats could retain the spatial behavioral and physiological aspects of learning and location of visual stimulus representing food for retention in this animal. a maximumof 10 sec, successful performance appeared In the present research, two experimental procedures to be dependent upon the use of response-directing whichhave proven useful for the investigation of short­ behaviors which mediated the delay.
    [Show full text]
  • Encoding and Representation of Intranasal CO2 in the Mouse Olfactory Cortex
    The Journal of Neuroscience, August 21, 2013 • 33(34):13873–13881 • 13873 Behavioral/Cognitive Encoding and Representation of Intranasal CO2 in the Mouse Olfactory Cortex Kaitlin S. Carlson,* Christina Z. Xia,* and Daniel W. Wesson Department of Neurosciences, Case Western Reserve University School of Medicine, Cleveland, Ohio 44106 Intranasal trigeminal sensory input, often perceived as a burning, tingling, or stinging sensation, is well known to affect odor perception. While both anatomical and functional imaging data suggest that the influence of trigeminal stimuli on odor information processing may occur within the olfactory cortex, direct electrophysiological evidence for the encoding of trigeminal information at this level of process- ingisunavailable.Here,inagreementwithhumanfunctionalimagingstudies,wefoundthat26%ofneuronsinthemousepiriformcortex (PCX)displaymodulationinfiringtocarbondioxide(CO2 ),anodorlessstimulantwithknowntrigeminalcapacity.Interestingly,CO2 was represented within the PCX by distinct temporal dynamics, differing from those evoked by odor. Experiments with ascending concen- trations of isopentyl acetate, an odorant known to elicit both olfactory and trigeminal sensations, resulted in morphing of the temporal dynamics of stimulus-evoked responses. Whereas low concentrations of odorant evoked responses upon stimulus onset, high concen- trations of odorant and/or CO2 often evoked responses structured to stimulus offset. These physiological experiments in mice suggest that PCX neurons possess the capacity to encode for stimulus modality (olfactory vs trigeminal) by differential patterns of firing. These data provide mechanistic insights into the influences of trigeminal information on odor processing and place constraints on models of olfactory-trigeminal sensory integration. Introduction considerable perceptual interplay between the olfactory and tri- Intranasal perception of airborne chemicals is mediated by both geminal systems that together result in our perception of odors the olfactory and trigeminal systems.
    [Show full text]
  • Multiple Timescales Account for Adaptive Responses Across Sensory Cortices
    The Journal of Neuroscience, December 11, 2019 • 39(50):10019–10033 • 10019 Systems/Circuits Multiple Timescales Account for Adaptive Responses across Sensory Cortices Kenneth W. Latimer,1* XDylan Barbera,2* Michael Sokoletsky,3 Bshara Awwad,5 XYonatan Katz,3 XIsrael Nelken,4,5† X Ilan Lampl,3† Adriene L. Fairhall,1† and Nicholas J. Priebe2† 1Departments of Physiology and Biophysics, University of Washington, Seattle, Washington 98195, 2Department of Neuroscience and Center for Learning and Memory, University of Texas at Austin, Austin, Texas 78712, 3Department of Neurobiology, Weizmann Institute of Science, Rehovot 7610001, Israel, 4Department of Neurobiology, Hebrew University, Edmond J. Safra Campus, 91904 Jerusalem, Israel, and 5The Edmond and Lily Safra Center for Brain Sciences, Hebrew University of Jerusalem, Jerusalem, Israel Sensory systems encounter remarkably diverse stimuli in the external environment. Natural stimuli exhibit timescales and amplitudes of variation that span a wide range. Mechanisms of adaptation, a ubiquitous feature of sensory systems, allow for the accommodation of this range of scales. Are there common rules of adaptation across different sensory modalities? We measured the membrane potential responses of individual neurons in the visual, somatosensory, and auditory cortices of male and female mice to discrete, punctate stimuli delivered at a wide range of fixed and nonfixed frequencies. We find that the adaptive profile of the response is largely preserved across these three areas, exhibiting attenuation and responses to the cessation of stimulation, which are signatures of response to changes in stimulus statistics. We demonstrate that these adaptive responses can emerge from a simple model based on the integration of fixed filters operating over multiple time scales.
    [Show full text]
  • Beta Oscillations Reflect Supramodal Information During Perceptual Judgment
    Beta oscillations reflect supramodal information during perceptual judgment Saskia Haegensa,b,1, José Vergarac, Román Rossi-Poolc, Luis Lemusc, and Ranulfo Romoc,d,1 aDepartment of Neurosurgery, Columbia University Medical Center, New York, NY 10032; bDonders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen, 6500HB Nijmegen, The Netherlands; cInstituto de Fisiología Celular-Neurociencias, Universidad Nacional Autónoma de México, 04510 Mexico D.F., Mexico; and dEl Colegio Nacional, 06020 Mexico D.F., Mexico Contributed by Ranulfo Romo, November 15, 2017 (sent for review August 18, 2017; reviewed by Felix Blankenburg and Tobias H. Donner) Previous work on perceptual decision making in the sensorimotor subjects, using visual, tactile, and auditory unimodal flutter dis- system has shown population dynamics in the beta band, corre- crimination, reported parametric modulation of prefrontal beta sponding to the encoding of stimulus properties and the final power by stimulus frequency, regardless of modality, during WM decision outcome. Here, we asked how oscillatory dynamics in the retention (20). Thus, these studies suggest that task-relevant medial premotor cortex (MPC) contribute to supramodal perceptual stimulus properties are encoded in a modality-nonspecific way decision making. We recorded local field potentials (LFPs) and spikes in prefrontal regions, allowing for subsequent supramodal perceptual in two monkeys trained to perform a tactile–acoustic frequency judgment. discrimination task, including both unimodal and crossmodal condi- Here, we asked how oscillatory dynamics contribute to the tions. We studied the role of oscillatory activity as a function of supramodal perceptual decision process. We recorded LFPs and stimulus properties (frequency and sensory modality), as well as spikes in medial premotor cortex (MPC) in two animals per- decision outcome.
    [Show full text]