Examplecode Documentation Release 0.1

Total Page:16

File Type:pdf, Size:1020Kb

Examplecode Documentation Release 0.1 examplecode Documentation Release 0.1 Tim Apr 23, 2021 CONTENTS: 1 About 1 2 Arterial Spin Labeling (ASL)3 3 Directions 7 4 DIY fMRI 9 5 DTI 17 6 Fieldmaps 19 7 JPEG Format Variations 25 8 Legacy Web Pages 29 9 Optimizing FSL and SPM 31 10 Sensation and Perception (PSYC450) 37 11 Image to Inference (PSYC589/888) 49 12 Perfusion-weighted imaging (PWI) 55 13 spmScripts 61 14 Slice Time Correction (STC) 63 15 Publications 67 16 Indices and tables 69 i ii CHAPTER ONE ABOUT 1.1 Topics • Attention and perception. Our senses flood the brain with an overwhelming amount of information – how do we select the relevant information? Clinical syndromes such as spatial neglect (where individuals ignore information on their left side) provide insight into how the brain achieves this. • Speech and language. Communication is invaluable for sharing information, planning and coordinating actions in a group. Human language is quantitatively and qualitatively a quantum leap from that seen in other species. Cognitive neuroscience is able to employ new techniques to understand language. This work will help reveal who we are, and may help people who have suffered profound communication difficulties following brain injury. 1.2 Tools • Behavioral Tasks. Each of our studies requires us to develop sensitive behavioral tasks: for example in an fMRI study of time perception we will want to compare tasks where the person makes temporal judgments (e.g. which item appeared first) to perceptually identical tasks where the participant judges a different domain (for example the shape of the items). We have extensive skill in designing and implementing these tasks. • MRI scans use radio signals to take pictures of the brain. fMRI is a type of MRI scan that is is sensitive to oxygenation concentration, allowing us to infer brain function. Typically, we have people perform simple tasks in the scanner while we collect fMRI scans. We have used this technique to identify the brain areas involved with speech and perception. In addition, we have used fMRI to examine recovery from brain injury. • Lesion behavior mapping associates the location of brain injury with the resulting symptoms. For example, we use this technique to identify the brain injuries that result in speech impairment. We can also use this technique to identify the best targets for neurosurgery. 1 examplecode Documentation, Release 0.1 • Transcranial Direct Current Stimulation. tDCS applies weak electrical currents to the scalp. It appears that tDCS can induce subtle changes in brain activity, with regions near the positive electrode showing slightly increased firing rates, whereas regions under the negative electrode show small decreases in firing rate. Curiously, these changes seem to persist for many minutes after the stimulation ends. Because this technique is very safe and inexpensive, this technique offers potential for the helping people recover from brain injury as well as revealing the function of the healthy brain. We have devised methods for double-blind testing of tDCS (where neither the participant nor the experimenter knows the type of stimulation used) to investigate this mysterious but promising technique. • TMS uses a brief magnetic pulse to stimulate parts of the brain near the TMS coil. The region of stimulation is relatively focused. By introducing TMS pulses while participants are conducting a task we can 2 Chapter 1. About CHAPTER TWO ARTERIAL SPIN LABELING (ASL) Arterial Spin Labeling (ASL) is a Magnetic Resonance Imaging (MRI) technique for measuring blood flow. Whereas conventional Perfusion Weighted Imaging (PWI) uses an external agent like Gadolinium (Gd) to tag blood, ASL directly tags the blood entering the brain. Conventional PWI has high signal to noise, but we tend to only track the perfusion of a single bolus (e.g. Gd is injected once into the arm, and we measure the latency and amount of this agent reaching different parts of the brain). In contrast, with ASL we have low signal to noise, but can easily acquire and compare hundreds of images. The rest of this page focuses on ASL, for more details on conventional PWI, please visit my PWI page.. ASL can be used to measure make quantitative measures of perfusion, such as the relative cerebral blood flow (rCBF). In addition, ASL scans can be used to infer brain function, similar to T2*-weighted fMRI. In general, ASL fMRI has slower acquisition, reduced field of view, and worse spatial distortions relative to T2* fMRI. However, it does provide a more direct measure of blood flow that may be helpful in cases where the canonical hemodynamic response has been disrupted. In any case, FSL makes it pretty easy to analyses ASL fMRI in a manner that is very similar to T2* fMRI. The FSL web page provides more details. Therefore, the rest of this web page describes the analysis of quantitative ASL data. There are many different types of ASL sequences. Your sequences will be limited by the type of scanner you have, as well as the sequence licenses you have available. Several major variations are CASL (continuous ASL), PASL (pulsed ASL) and pCASL (pseudo-Continuous ASL). At the MCBI we have the official Siemens PASL sequence (PICORE Q2T) and the pCASL from JJ Wang and his team. The Siemens sequence is elegant, as it automatically creates a rCBF map. However, we tend to prefer the pCASL sequence for our studies. A crucial step when acquiring ASL data is to set the correct post-label delay time. This is the time between when the blood is tagged in the neck and when the image of the brain is acquired. If the delay is too short, the blood will not have time to transit into the image, and if it is too long it will already have washed out of the image. This is especially important, as we will acquire pairs of images: one labelled and one unlabelled. One could imaging that with a very brief post-label delay and a short time between volumes (TR), the tagged blood might not get to the head in time for 3 examplecode Documentation, Release 0.1 the ‘labelled’ image, but be clearly present during the ‘unlabelled’ image acquisition. I strongly suggest consulting the people who developed your sequence to get their suggestions for post-label delay times. For the pCASL sequence we have, Ze Wang has suggested a delay time in the range of 700-1000ms for healthy children and young adults, while for older individuals (65 or older) he suggests 1200-1500ms, finally for stroke patients or patients with vascular diseases he notes that 1800ms might be required. In any case, this selection should be standardized for a single study. For example, a shorter delay may be required for a study of stroke that hopes to examine both the intact and injured hemisphere. However, if you plan to acquire images from special populations (e.g. people with strokes) you may want to consult your physicist. As you adjust the delay time, the minimum TR is also influenced. Ze Wang suggests that your actual TR should always be at least 100ms longer than the minimum TR since the labeling pulses induce Magnetization Transfer (MT) effects to the brain regions to be imaged, so before the spins go back to the steady state, they are suppressed to some extent by the labeling pulses. Longer TRs provide more signal (more time for spins to relax), though at the cost of fewer acquisitions (and more difficulty temporally interpolating data for fMRI-like task based paradigms). In general, a TR of 3500ms seems appropriate (unless your population requires a very long delay time). Another thing you should bear in mind with the CfN pCASL sequence is the bandwidth (indeed, bandwidth is an important decision for echo-planar imaging [EPI] protocols). With regards to the CfN pCASL sequence, Ze Wang notes that high bandwidths can lead to severe eddy currents leading to phase accumulation and a N/2 ghost artifact. Performance varies between scanners, but he suggests that 2232 to 2694 Hz/pixel should be appropriate for most Siemens Trios (you should also check that images from your scanner do not show aliasing artifacts, if you see artifacts then you should collect images without iPAT [as this can also cause artifacts] and iteratively take images while decreasing the bandwidth until the artifacts go away). You will also want to specify your labeling time, for example if your protocol PDF reports 80 blocks, the Labeltime = 80*0.0185, since the CFN pCASL RF block duration is ALWAYS 0.0185s (20 RF pulses with gaps). For our protocol, we use 80 RF blocks, a bandwidth of 2442 Hz/px, and acquire 17 slices. With these settings the minimum TR is 2090ms plus the delay time (so since slicetime=[minTR- labelingtime-delaytime]/#slices, we can compute that our Slicetime is 36.35294118ms). For example, with a 1200ms post label delay the minimum TR is 3290ms, and we typically acquire with a TR of 3500ms. 2.1 pCASL Analysis Simplified This page is old. While the scripts below work, new users may want to consider using FSL’s BASIL. We have set up a simple script for processing our pCASL data. This script requires that you have the following installed: • Matlab (no toolboxes required) • SPM12 • ASLtbx – Since this script uses 4D NIfTI format files, you need a recent version of ASLtbx (the asl_perf_subtract.m text file should report being version May 2 2012 or later). • One NIfTI format T1-weighted anatomical scan per participant • One NIfTI format 4D ASL file per session (each participant may have multiple sessions). • My asl_process_subj.m matlab script (this needs to be in your Matlab path, you might as well put it into your ASLtbx folder,download includes script and sample images).
Recommended publications
  • Lecture 4 Objects
    4 Perceiving and Recognizing Objects ClickChapter to edit 4 Perceiving Master title and style Recognizing Objects • What and Where Pathways • The Problems of Perceiving and Recognizing Objects • Middle Vision • Object Recognition ClickIntroduction to edit Master title style How do we recognize objects? • Retinal ganglion cells and LGN = Spots • Primary visual cortex = Bars How do spots and bars get turned into objects and surfaces? • Clearly our brains are doing something pretty sophisticated beyond V1. ClickWhat toand edit Where Master Pathways title style Objects in the brain • Extrastriate cortex: Brain regions bordering primary visual cortex that contains other areas involved in visual processing. V2, V3, V4, etc. ClickWhat toand edit Where Master Pathways title style Objects in the brain (continued) • After extrastriate cortex, processing of object information is split into a “what” pathway and a “where” pathway. “Where” pathway is concerned with the locations and shapes of objects but not their names or functions. “What” pathway is concerned with the names and functions of objects regardless of location. Figure 4.2 The main visual areas of the macaque monkey cortex Figure 4.3 A partial “wiring” diagram for the main visual areas of the brain Figure 4.4 The main visual areas of the human cortex ClickWhat toand edit Where Master Pathways title style The receptive fields of extrastriate cells are more sophisticated than those in striate cortex. They respond to visual properties important for perceiving objects. • For instance, “boundary ownership.” For a given boundary, which side is part of the object and which side is part of the background? Figure 4.5 Notice that the receptive field sees the same dark edge in both images ClickWhat toand edit Where Master Pathways title style The same visual input occurs in both (a) and (b) and a V1 neuron would respond equally to both.
    [Show full text]
  • Illusion Free Download
    ILLUSION FREE DOWNLOAD Sherrilyn Kenyon | 464 pages | 05 Feb 2015 | Little, Brown Book Group | 9781907411564 | English | London, United Kingdom Illusion (Ability) Visually perceived images that differ from objective reality. This urge to see depth is probably so strong because our ability to use two-dimensional information to infer a three dimensional world is essential Illusion allowing us to operate in the world. The exact height of the bars is Illusion important here, but the relative heights should look something like this:. A hallucination is a perception Illusion a thing or quality Illusion has no physical counterpart: Under the influence of LSD, Terry had hallucinations that the living-room floor was rippling. MD GtI. Name that government! Water-color illusions consist of object-hole effects and coloration. Then, using some impressive mental geometry, our brain adjusts the experienced length of the top line to be consistent with the size it would have if it were that far away: Illusion two lines are the same length on my retina, but different distances from me, the more distant line must be in Illusion longer. Choose the Right Synonym for illusion delusionillusionhallucinationmirage mean something that is believed to be true or real Illusion that is actually false or unreal. ISSN Illusion Ponzo illusion is an example of an illusion which uses monocular cues of depth perception to fool the eye. The phi phenomenon is yet another example of how the brain perceives motion, which is most often created by blinking lights in close succession. The cubic texture induces a Necker-cube -like optical illusion.
    [Show full text]
  • Innateness and (Bayesian) Visual Perception Reconciling Nativism and Development
    3 BRIAN J. SCHOLL Innateness and (Bayesian) Visual Perception Reconciling Nativism and Development 1 A Research Strategy Because innateness is such a complex and controversial issue when applied to higher level cognition, it can be useful to explore how nature and nurture interact in simpler, less controversial contexts. One such context is the study of certain aspects of visual perception—where especially rigorous models are possible, and where it is less controversial to claim that certain aspects of the visual system are in part innately specified. The hope is that scrutiny of these simpler contexts might yield lessons that can then be applied to debates about the possible innateness of other aspects of the mind. This chapter will explore a particular way in which visual processing may involve innate constraints and will attempt to show how such processing overcomes one enduring challenge to nativism. In particular, many challenges to nativist theories in other areas of cognitive psychology (e.g., ‘‘theory of mind,’’ infant cognition) have focused on the later development of such abili- ties, and have argued that such development is in conflict with innate origins (since those origins would have to be somehow changed or overwritten). Innate- ness, in these contexts, is seen as antidevelopmental, associated instead with static processes and principles. In contrast, certain perceptual models demonstrate how the very same mental processes can both be innately specified and yet develop richly in response to experience with the environment. In fact, this process is entirely unmysterious, as is made clear in certain formal theories of visual per- ception, including those that appeal to spontaneous endogenous stimulation, and those based on Bayesian inference.
    [Show full text]
  • Contour Junctions Underlie Neural Representations of Scene Categories
    bioRxiv preprint doi: https://doi.org/10.1101/044156; this version posted March 16, 2016. The copyright holder for this preprint (which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under aCC-BY-NC-ND 4.0 International license. Contour junctions underlie neural representations of scene categories in high-level human visual cortex Contour junctions underlie neural code of scenes Heeyoung Choo1, & Dirk. B. Walther1 1Department of Psychology, University of Toronto 100 St. George Street, Toronto, ON, M5S 3G3 Canada Correspondence should be addressed to Heeyoung Choo, 100 St. George Street, Toronto, ON, M5S 3G3 Canada. Phone: +1-416-946-3200 Email: [email protected] Conflict of Interest: The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. 1 bioRxiv preprint doi: https://doi.org/10.1101/044156; this version posted March 16, 2016. The copyright holder for this preprint (which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under aCC-BY-NC-ND 4.0 International license. 1 Abstract 2 3 Humans efficiently grasp complex visual environments, making highly consistent judgments of entry- 4 level category despite their high variability in visual appearance. How does the human brain arrive at the 5 invariant neural representations underlying categorization of real-world environments? We here show that 6 the neural representation of visual environments in scenes-selective human visual cortex relies on statistics 7 of contour junctions, which provide cues for the three-dimensional arrangement of surfaces in a scene.
    [Show full text]
  • Object Perception As Bayesian Inference 1
    Object Perception as Bayesian Inference 1 Object Perception as Bayesian Inference Daniel Kersten Department of Psychology, University of Minnesota Pascal Mamassian Department of Psychology, University of Glasgow Alan Yuille Departments of Statistics and Psychology, University of California, Los Angeles KEYWORDS: shape, material, depth, perception, vision, neural, psychophysics, fMRI, com- puter vision ABSTRACT: We perceive the shapes and material properties of objects quickly and reliably despite the complexity and objective ambiguities of natural images. Typical images are highly complex because they consist of many objects embedded in background clutter. Moreover, the image features of an object are extremely variable and ambiguous due to the effects of projection, occlusion, background clutter, and illumination. The very success of everyday vision implies neural mechanisms, yet to be understood, that discount irrelevant information and organize ambiguous or \noisy" local image features into objects and surfaces. Recent work in Bayesian theories of visual perception has shown how complexity may be managed and ambiguity resolved through the task-dependent, probabilistic integration of prior object knowledge with image features. CONTENTS OBJECT PERCEPTION: GEOMETRY AND MATERIAL : : : : : : : : : : : : : 2 INTRODUCTION TO BAYES : : : : : : : : : : : : : : : : : : : : : : : : : : : : 3 How to resolve ambiguity in object perception? : : : : : : : : : : : : : : : : : : : : : : 3 How does vision deal with the complexity of images? : : : : :
    [Show full text]
  • A Dictionary of Neurological Signs.Pdf
    A DICTIONARY OF NEUROLOGICAL SIGNS THIRD EDITION A DICTIONARY OF NEUROLOGICAL SIGNS THIRD EDITION A.J. LARNER MA, MD, MRCP (UK), DHMSA Consultant Neurologist Walton Centre for Neurology and Neurosurgery, Liverpool Honorary Lecturer in Neuroscience, University of Liverpool Society of Apothecaries’ Honorary Lecturer in the History of Medicine, University of Liverpool Liverpool, U.K. 123 Andrew J. Larner MA MD MRCP (UK) DHMSA Walton Centre for Neurology & Neurosurgery Lower Lane L9 7LJ Liverpool, UK ISBN 978-1-4419-7094-7 e-ISBN 978-1-4419-7095-4 DOI 10.1007/978-1-4419-7095-4 Springer New York Dordrecht Heidelberg London Library of Congress Control Number: 2010937226 © Springer Science+Business Media, LLC 2001, 2006, 2011 All rights reserved. This work may not be translated or copied in whole or in part without the written permission of the publisher (Springer Science+Business Media, LLC, 233 Spring Street, New York, NY 10013, USA), except for brief excerpts in connection with reviews or scholarly analysis. Use in connection with any form of information storage and retrieval, electronic adaptation, computer software, or by similar or dissimilar methodology now known or hereafter developed is forbidden. The use in this publication of trade names, trademarks, service marks, and similar terms, even if they are not identified as such, is not to be taken as an expression of opinion as to whether or not they are subject to proprietary rights. While the advice and information in this book are believed to be true and accurate at the date of going to press, neither the authors nor the editors nor the publisher can accept any legal responsibility for any errors or omissions that may be made.
    [Show full text]
  • Object Recognition, Which Is Solved in Visual Areas Beyond V1
    Object vision (Chapter 4) Lecture 9 Jonathan Pillow Sensation & Perception (PSY 345 / NEU 325) Princeton University, Fall 2017 1 Introduction What do you see? 2 Introduction What do you see? 3 Introduction What do you see? 4 Introduction How did you recognize that all 3 images were of houses? How did you know that the 1st and 3rd images showed the same house? This is the problem of object recognition, which is solved in visual areas beyond V1. 5 Unfortunately, we still have no idea how to solve this problem. Not easy to see how to make Receptive Fields for houses the way we combined LGN receptive fields to make V1 receptive fields! house-detector receptive field? 6 Viewpoint Dependence View-dependent model - a model that will only recognize particular views of an object • template-based model e.g. “house” template Problem: need a neuron (or “template”) for every possible view of the object - quickly run out of neurons! 7 Middle Vision Middle vision: – after basic features have been extracted and before object recognition and scene understanding • Involves perception of edges and surfaces • Determines which regions of an image should be grouped together into objects 8 Finding edges • How do you find the edges of objects? • Cells in primary visual cortex have small receptive fields • How do you know which edges go together and which ones don’t? 9 Middle Vision Computer-based edge detectors are not as good as humans • Sometimes computers find too many edges • “Edge detection” is another failed theory (along with Fourier analysis!) of what V1 does.
    [Show full text]
  • Amodal Completion in the Absence of Image Tangent Discontinuities
    Perception, 1998, volume 27, pages 455-464 Amodal completion in the absence of image tangent discontinuities Peter Ulric Tse% Marc K Albert# Vision Sciences Laboratory, Department of Psychology, Harvard University, 33 Kirkland Street, Cambridge, MA 02138, USA; e-mail: [email protected]; [email protected] Received 26 August 1997, in revised form 31 December 1997 Abstract. It has been widely believed since Helmholtz that tangent discontinuities in image contours, such as T-junctions or L-junctions, will occur when one object occludes another. Here we describe a class of occlusion relationships where changes in 'border ownership' and amodal completion take place in the absence of tangent discontinuities in the image. We propose that even subtle curvature discontinuities can be a signal to potential changes in border ownership, and are valid ecological cues for occlusion in certain scenes. 1 Introduction When image fragments are perceived to be the visible portions of a single partially occluded object, the object is said to 'amodally complete' behind the occluder (eg Michotte 1964; Kanizsa 1979). It is generally believed that amodal completion requires tangent or 'first-order' discontinuities (ie discontinuities in the orientation of the tangent line, or first derivative, along the contour) at all visible junctions between the contours 'owned' by the occluded surface and those owned by the occluder (eg Clowes 1971; Huffman 1971; Waltz 1975; Lowe 1985; Malik 1987; Nakayama et al 1989; Kellman and Shipley 1991). Assuming no accidental alignment of surface borders, each contour in the image can only belong to one of the two surfaces adjacent to it.
    [Show full text]
  • Ldmk-Nak 859..877
    Perception, 2009, volume 38, pages 859 ^ 877 doi:10.1068/ldmk-nak Nakayama, Shimojo, and Ramachandran's 1990 paper [Nakayama K, Shimojo S, Ramachandran V S, 1990 ``Transparency: relation to depth, subjective contours,luminance,andneoncolorspreading''Perception19497^513.Originalpaperreprinted in the appendix.] Authors' update Surfaces revisited The target paper reviewed in this article was titled ``Transparency: relation to depth, subjective contours, luminance, and neon colour spreading'' coauthored by K Nakayama, S Shimojo, and V S Ramachandran, published in 1990. This paper, one of the first in a series on surface perception, examined how in untextured stereograms, local dispar- ity and luminance contrast can drastically change surface quality, subjective contours, and the effect of neon colour spreading. When we began to conceive this and related work, the ascendant view on visual perception was derived from the pioneering studies of the response properties of visual neurons with microelectrodes, including those of Barlow (1953), Lettvin et al (1959), and Hubel and Wiesel (1959, 1962). All suggested that there are remarkable operations on the image by earliest stages of the visual pathway, which bestowed selectivity to colour, orientation, motion direction, spatial frequency, binocular disparity, etc. As such, it would seem that an understanding of vision would come through more systematic description of the properties of single neuron selectivities. This viewpoint was well summarised by Horace Barlow in his famous neuron doctrine paper (Barlow 1972), which emphasised the importance of analysing the image in successive stages of processing by neurons with specific classes of receptive fields. Later work altered this conception somewhat by seeing receptive fields as linear filters.
    [Show full text]
  • Neural Correlates of Dynamic Object Recognition
    Neural correlates of dynamic object recognition Katja Martina Mayer Institute of Neuroscience, Newcastle University Thesis submitted for the degree Doctor of Philosophy September 2010 Table of contents Acknowledgements .................................................................................................. vii List of Figures ......................................................................................................... viii List of Tables ............................................................................................................ ix Abbreviations ............................................................................................................ xi 1. Abstract ...................................................................................................................... 1 2. Summary of previous research on recognition of multi-featured objects .................. 2 2.1. Motivation for research in object recognition .................................................... 2 2.2. The role of shape, colour, and motion in object recognition .............................. 3 2.3. Theories of object recognition ............................................................................ 7 2.3.1. Structural description models ...................................................................... 7 2.3.2. Image-based models .................................................................................... 9 2.3.3. Extensions to image-based models ............................................................
    [Show full text]
  • An Investigation Into the Neural Basis of Convergence Eye Movements
    An Investigation into the Neural Basis of Convergence Eye Movements Dissertation Presented in Partial Fulfillment of the Requirements for the Degree Doctor of Philosophy in the Graduate School of The Ohio State University By Emmanuel Owusu Graduate Program in Vision Science The Ohio State University 2018 Dissertation Committee Marjean T. Kulp, Advisor Nicklaus F. Fogt, Co-Advisor Andrew J. Toole Bradley Dougherty Michael J. Earley Xiangrui Li 1 Copyrighted by Emmanuel Owusu 2018 2 Abstract Introduction: Different components of convergence are tonic convergence, disparity convergence, accommodative convergence and proximal. However, it is not clear whether these different components ultimately draw on similar innervational control. Better understanding of the neurology of convergence eye movements could lead to improvement in interventions for deficits in convergence eye movements. Therefore, the purpose of this dissertation is to investigate the neural basis for convergence eye movements in binocularly normal young adults. Methods: Two approaches were used. First, clinical measurements were used to determine the correlations among accommodative, disparity and proximal convergence eye movements, as well as between proximal convergence and vergence facility. These correlations were used as an index of the extent of overlaps in their neurological control. Second, functional magnetic resonance imaging (fMRI) was performed on a group of adults with normal binocular function as they converged their eyes in response to stimuli for accommodative, disparity, proximal and voluntary convergence eye movements. Results: In the clinical study, stimulus gradient accommodative convergence was negatively correlated with far-near proximal convergence (Spearman’s correlation = -0.6111, p < 0.0001). However, the correlation may be at least in part attributable to the inclusion of gradient AC/A as a component of the calculation of far-near proximal convergence.
    [Show full text]
  • The Psychology of Perception
    The Psychology of Perception Perception Study Booklet 1........................................................................................................................Introduction 2........................................................................................................................Audition 3.....................................................................................The Body and Chemical Senses 4......................................................................Lights, Eyes the Brain and Spatial Vision 5...........................................................................................................Depth Perception 6.........................................................................................................Motion Perception 7...................................................................................................................Color Vision 8..............................................................................................................Face Perception 9...........................................................................................................Object Perception 10.............................................................................................Multi-sensory Integration 11.................................................................................................................Assessments 12.............................................................................................................Online Quizzes 13...........................................................................................................................Exams
    [Show full text]