Audioviewer: Learning to Visualize Sound

Total Page:16

File Type:pdf, Size:1020Kb

Audioviewer: Learning to Visualize Sound AudioViewer: Learning to Visualize Sound by Yuchi Zhang A THESIS SUBMITTED IN PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR THE DEGREE OF Master of Science in THE FACULTY OF GRADUATE AND POSTDOCTORAL STUDIES (Computer Science) The University of British Columbia (Vancouver) December 2020 c Yuchi Zhang, 2020 The following individuals certify that they have read, and recommend to the Faculty of Graduate and Postdoctoral Studies for acceptance, the thesis entitled: AudioViewer: Learning to Visualize Sound submitted by Yuchi Zhang in partial fulfillment of the requirements for the degree of Master of Science in Computer Science. Examining Committee: Helge Rhodin, Computer Science Supervisor Kwang Moo Yi, Computer Science Examining Committee Member ii Abstract Sensory substitution can help persons with perceptual deficits. In this work, we attempt to visualize audio with video. Our long-term goal is to create sound perception for hearing impaired people, for instance, to facilitate feedback for training deaf speech. Different from existing models that translate between speech and text or text and images, we target an immediate and low-level translation that applies to generic environment sounds and human speech without delay. No canonical mapping is known for this artificial translation task. Our design is to translate from audio to video by compressing both into a common latent space with a shared structure. Our core contribution is the development and evaluation of learned mappings that respect human perception limits and maximize user comfort by enforcing priors and combining strategies from unpaired image translation and disentanglement. We demonstrate qualitatively and quantitatively that our AudioViewer model maintains important audio features in the generated video and that generated videos of faces and numbers are well suited for visualizing high- dimensional audio features since they can easily be parsed by humans to match and distinguish between sounds, words, and speakers. iii Lay Summary We develop AudioViewer, an automated that translates speech into videos of faces or hand-written numbers. Without audio feedback, it’s challenging for hearing impaired people to learn speech. To help with the situation, we propose AudioViewer to generate videos from utterances as a visual feedback in substitution. Our current work is a proof of concept. AudioViewer is designed to generate consistent and recognizable visualization of pronunciation. For example, the word ‘water’ will be visualized as a short animation of a portrait of a young brunette lady with a short hair style facing rightward with a neutral expression, which is gradually transforming into a smiling young blonde lady facing forward. From our preliminary user study, we demonstrate that people can tell the difference between the visualization of different pronunciations and recognize the consistency of the identical ones. iv Preface The work presented here is original work done by the author Yuchi Zhang in collaboration with Willis Peng and Dr. Bastian Wandt under the supervision of Dr. Helge Rhodin. A shorter version of this work is under review as a conference submission. Design, implementation and experiments were done by Yuchi Zhang. Dr. Helge Rhodin provided suggestions and feedback during the development and Willis Peng helped with the user study evaluation. The initial draft of the paper is done by all of us and revised by Dr. Bastian Wandt. The user studies ran for this work have been approved by the UBC Behavioural Research Ethics Board (Certificate Number: H19-03006) v Table of Contents Abstract . iii Lay Summary . iv Preface . v Table of Contents . vi List of Tables . viii List of Figures . x Acknowledgments . xiii 1 Introduction . 1 2 Related Work . 4 3 Method . 9 3.1 Audio Encoding . 10 3.2 Structuring the Audio Encoding . 11 3.3 Image Encoding for Video Generation . 13 3.4 Linking Audio and Visual Spaces . 14 4 Experiments . 16 4.1 Temporal Smoothness Effectiveness . 18 4.2 Information Throughput . 20 vi 4.3 Visual Quality and Phonetics . 22 4.4 Visual Domain Influence . 22 4.5 Results on Environment Sounds . 23 4.6 User study . 25 4.6.1 Design of the questionnaire . 25 4.6.2 Results . 28 5 Limitations and Future Work . 31 6 Conclusion . 35 Bibliography . 36 vii List of Tables Table 4.1 Audio encoding analysis. The average SNR of reconstructed amplitude Mel spectrogram on the TIMIT test set and the average speed and acceleration between the latent vector (dim = 128) of neighbouring frames (∆t = 0:04s). The SNR of the autoencoder drops with the additional constraints we enforce (smoothness and disentanglement). This is a trade-off for improving velocity and acceleration properties as well as for inducing interpretability. 18 Table 4.2 Throughput lower bound. The throughput from audio to video with AudioViewer can be bounded by mapping back to the audio. It is evident that the cycle consistency improves greatly while the enforcement of additional constraints with Lp;log MSE, and Lrr, takes only a small dip. 21 Table 4.3 Information throughput on environment sounds, showing that the reconstruction error increases when evaluating on a test set that contains sounds vastly different from the training set (speech vs. environment sounds). 24 Table 4.4 Audio VAE Mel spectrum reconstruction. The average SNR of autoencoding and decoding Mel spectrograms on ESC-50 shows a significant reconstruction loss.The average speed and acceleration between the latent vector (dim = 128) of neigh- bouring frames (∆t = 0:04s) confirms the experiments in the main document, that smoothness comes at the cost of lower reconstruction accuracy. 24 viii Table 4.5 A breakdown of the questionnaire format by sound type and tested factors, listing their frequency of occurrence. 27 Table 4.6 User study results. Values indicate mean accuracy and standard deviation for distinguishing between visualizations of the tested factor across participants as a percentage. 28 ix List of Figures Figure 1.1 AudioViewer: A tool developed towards the long-term goal of helping hearing impaired users to see what they can not hear. We map an audio stream to video and thereby use faces or numbers for visualizing the high-dimensional audio features intuitively. Different from the principle of lip reading, this can encode general sound and pass on information on the style of spoken language, for instance, the timbre of the speaker. 2 Figure 2.1 An illustration of harmonics and formants of speech voice in a filter bank image. The above mel-scaled power spectro- gram shows the word ‘party’ spoken by a male speaker. The vertical axis indicates mel-scaled filter banks and the horizontal axis is time. The narrow stripes in the spectrogram are the harmonics which are the integral multiples of the fundamental frequency (F0), reflecting the vibration of the vocal fold. Note that the change in F0 reflects the intonation during speech. The wide band peaks are the formants, reflecting the resonant fre- quencies of the vocal tract. Formant patterns are different when differnt vowels (e.g., aa, r and iy) are spoken. 5 Figure 3.1 Overview. A joint latent encoding is learned with audio and video VAEs that are linked with a cycle constraint and aug- mented with a smoothness constraint (∆v annotation). 10 x Figure 3.2 Disentanglement, by mixing content encodings from two speak- 0 ers saying the same phones (zci with zci ) and style encodings 0 from the same speaker saying different words (zsa with zsa ). 12 Figure 3.3 Cycle constraint. We link audio and video latent spaces with a cycle constraint that ensures that the audio signal content is preserved through video decoding and encoding. 14 Figure 4.1 Smoothness evaluation by plotting the latent embedding of an audio snipped in 3D via dimensionality reduction. Left: PCA baseline model, consecutive audio sample embeddings are scattered. Middle: Similar performance of our model without temporal smoothness constraint. Right: Our full model leads to smoother trajectories. 19 Figure 4.2 Throughput. The information throughput can be measured by mapping from audio to video and back. Playing the depicted Mel Spectrogram shows that the speech is still recognizable. 20 Figure 4.3 Word analysis. Instances (different rows) of the same words are encoded with a similar video sequence (single row), yet the show pronunciation variations if spoken by different speakers. 22 Figure 4.4 Phoneme similarity analysis. The same phonemes occurring at the beginning of different words are mapped to visually similar images (colums) while different speakers have minimal influence (rows). 23 Figure 4.5 Environment sounds visualization. Although not trained for it, our approach also visualizes natural sounds. Here showing a clear difference between chirp and alarm clock. The visualiza- tion looks brighter and intenser when a target sound occurs. 25 Figure 4.6 Matching Example. Examples from the CelebA-combined and MNIST-content questionnaire of a matching question asked to participants in the user study. 26 Figure 4.7 Grouping Example. Examples from the CelebA-combined and MNIST-content questionnaire of a grouping question asked to participants in the user study. 26 xi Figure 4.8 Model Comparison on content questions. The mean user accuracy and standard deviation for distinguishing phoneme sequences is compared between the models, annotated with the significance level (y for 0:05 < p < 0:1, * for p < 0:05, ** for p < 0:01). The mean accuracy for random guessing is 0.384. 29 Figure 4.9 Model Comparison on style questions. The mean user accu- racy and standard deviation for distinguishing speaker sex (left) and dialect(right) is compared between the models, annotated with the significance level (y for 0:05 < p < 0:1, *** for p < 0:001). The mean accuracies for random guessing are 0.43 and 0.5 respectively. 30 Figure 5.1 Recombined reconstruction sample. Top left: The Mel spec- trograms of a sentence spoken by a female speaker and its re- construction by our audio model with disentanglement training.
Recommended publications
  • SENSORY SUBSTITUTION: LIMITS and PERSPECTIVES Charles Lenay, Olivier Gapenne, Sylvain Hanneton, Catherine Marque, Christelle Genouëlle
    SENSORY SUBSTITUTION: LIMITS AND PERSPECTIVES Charles Lenay, Olivier Gapenne, Sylvain Hanneton, Catherine Marque, Christelle Genouëlle To cite this version: Charles Lenay, Olivier Gapenne, Sylvain Hanneton, Catherine Marque, Christelle Genouëlle. SEN- SORY SUBSTITUTION: LIMITS AND PERSPECTIVES. Yvette Hatwell; Arlette Streri; Edouard Gentaz. Touching for Knowing, 53, John Benjamins Publishers, pp.275-292, 2004, Advances in Con- sciousness Research, 9789027251855. 10.1075/aicr.53.22len. hal-02434266 HAL Id: hal-02434266 https://hal.archives-ouvertes.fr/hal-02434266 Submitted on 9 Jan 2020 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. SENSORY SUBSTITUTION: LIMITS AND PERSPECTIVES Charles Lenay, Olivier Gapenne, Sylvain Hanneton1, Catherine Marque et Christelle Genouëlle 29/08/2003 Université de Technologie de Compiègne COSTECH - BIM Groupe Suppléance Perceptive2 A quarter of a century ago, in the preface to « Brain Mechanisms in Sensory Substitution », Paul Bach y Rita wrote: "This monograph thus risks becoming outdated in a very short time since the development of refined sensory substitution systems should allow many of the question raised here to be answered, and some of the conclusions may appear naive to future readers." (BACH Y RITA, 1972) As it turns out, this prediction is far from having been fulfilled: in spite of their scientific and social interest, their real effectiveness and a certain technological development, prosthetic devices employing the principle of "sensory substitution" are not widely used by the blind persons for whom they were originally destined.
    [Show full text]
  • Perceiving Invisible Light Through a Somatosensory Cortical Prosthesis
    ARTICLE Received 24 Aug 2012 | Accepted 15 Jan 2013 | Published 12 Feb 2013 DOI: 10.1038/ncomms2497 Perceiving invisible light through a somatosensory cortical prosthesis Eric E. Thomson1,2, Rafael Carra1,w & Miguel A.L. Nicolelis1,2,3,4,5 Sensory neuroprostheses show great potential for alleviating major sensory deficits. It is not known, however, whether such devices can augment the subject’s normal perceptual range. Here we show that adult rats can learn to perceive otherwise invisible infrared light through a neuroprosthesis that couples the output of a head-mounted infrared sensor to their soma- tosensory cortex (S1) via intracortical microstimulation. Rats readily learn to use this new information source, and generate active exploratory strategies to discriminate among infrared signals in their environment. S1 neurons in these infrared-perceiving rats respond to both whisker deflection and intracortical microstimulation, suggesting that the infrared repre- sentation does not displace the original tactile representation. Hence, sensory cortical prostheses, in addition to restoring normal neurological functions, may serve to expand natural perceptual capabilities in mammals. 1 Department of Neurobiology, Duke University, Box 3209, 311 Research Drive, Bryan Research, Durham, North Carolina 27710, USA. 2 Edmond and Lily Safra International Institute for Neuroscience of Natal (ELS-IINN), Natal 01257050, Brazil. 3 Department of Biomedical Engineering, Duke University, Durham, North Carolina 27710, USA. 4 Department of Psychology and Neuroscience, Duke University, Durham, North Carolina 27710, USA. 5 Center for Neuroengineering, Duke University, Durham, North Carolina 27710, USA. w Present address: University of Sao Paulo School of Medicine, Sao Paulo 01246-000, Brazil. Correspondence and requests for materials should be addressed to M.A.L.N.
    [Show full text]
  • Electrocorticography Evidence of Tactile Responses in Visual Cortices
    UvA-DARE (Digital Academic Repository) Electrocorticography Evidence of Tactile Responses in Visual Cortices Gaglianese, A.; Branco, M.P.; Groen, I.I.A.; Benson, N.C.; Vansteensel, M.J.; Murray, M.M.; Petridou, N.; Ramsey, N.F. DOI 10.1007/s10548-020-00783-4 Publication date 2020 Document Version Final published version Published in Brain Topography License CC BY Link to publication Citation for published version (APA): Gaglianese, A., Branco, M. P., Groen, I. I. A., Benson, N. C., Vansteensel, M. J., Murray, M. M., Petridou, N., & Ramsey, N. F. (2020). Electrocorticography Evidence of Tactile Responses in Visual Cortices. Brain Topography, 33(5), 559–570. https://doi.org/10.1007/s10548-020-00783-4 General rights It is not permitted to download or to forward/distribute the text or part of it without the consent of the author(s) and/or copyright holder(s), other than for strictly personal, individual use, unless the work is under an open content license (like Creative Commons). Disclaimer/Complaints regulations If you believe that digital publication of certain material infringes any of your rights or (privacy) interests, please let the Library know, stating your reasons. In case of a legitimate complaint, the Library will make the material inaccessible and/or remove it from the website. Please Ask the Library: https://uba.uva.nl/en/contact, or a letter to: Library of the University of Amsterdam, Secretariat, Singel 425, 1012 WP Amsterdam, The Netherlands. You will be contacted as soon as possible. UvA-DARE is a service provided by the library of the University of Amsterdam (https://dare.uva.nl) Download date:03 Oct 2021 Brain Topography (2020) 33:559–570 https://doi.org/10.1007/s10548-020-00783-4 ORIGINAL PAPER Electrocorticography Evidence of Tactile Responses in Visual Cortices Anna Gaglianese1,2,3 · Mariana P.
    [Show full text]
  • Sensory Substitution and the Human
    tavvJ TRENDSin CognitiveScierc€s Vot.7No.l2 D6csmbe.2OO3 Sensorysubstitution and the human- machineinterface Paul Baeh-y-Ritaland Stephen W. Kercelz l?e?.anmglE-9fonhopedics and RehabiritationMedicine, and BiomedicarEngineering, univorsity of wisconsin, MadisonWl 53706.USA 2Endogenous SystemsResearch Group, New EnglandInstitute, University ot New England,Biddeford ME 04005,USA Recent advances in the instrumentation technology ol advances have led to the possibility of new prosthetic sensory substitution presemed have new opportunities devices being potentially accessible at much lower cost to to dsv€lop systems for compsnsation of sansory loss, millions ofpatients. ThiB is Btimulating the interest ofboth ,n sensory substitution {€.g. ot sight or vestibular func. research groups and industry leading to the establish- tion), intormation trom an artificial roceptor is coupled ment of new research and development efrorts in many tho to brain via a human-machine interfaco. The brain countries. Thia recent exploeion of interest in sensorv is able to use this intormation in place of that usually substitution suggest€that now is a good time to review transmitted from an intaqt ssnss organ. Both auditory progress in the area. and tactilo systems show promiss for practical sensory substitution intartace sitos. This research provides Brain plasticity experimental tools tor examining brain plasticity and Brain plasticity can be defined as the adaptiee capacities has implications fot porceptual and cognition studios ofthe central nervous system - its ability to modify its own more gonerally. structural organization and fiuctioning, IlT. It permits atr adaptive (or a maladaptive) responseto functional demand, Persons who become blind do not lose tÄe capacity to Mechanisms of brain plasticity include neurochemical, see.
    [Show full text]
  • Robustness in Neurological Systems 13 - 15 November, 2015 Center for Philosophy of Science
    Robustness in Neurological Systems 13 - 15 November, 2015 Center for Philosophy of Science ABSTRACTS Alison BArth, Professor, DepArtment of Biology, AnD HeAD, BArth LAb, CArnegie Mellon University Functional Connectivity is Regulated by SOM Interneurons Spontaneous Activity UnDerstAnDing the DynAmic rAnge for synAptic trAnsmission is A criticAl component of builDing A functionAl circuit DiAgrAm for the mammaliAn brAin. We find that excitAtory synaptic strength between neocortical neurons is markeDly suppresseD During network Activity in mouse somatosensory cortex, with smaller EPSP AmplituDes AnD high fAilure rAtes thAn previously reporteD. This phenomenon is regulAteD by tonic ActivAtion of presynAptic GABAb receptors viA the Activity of somatostAtin-expressing interneurons. Optogenetic suppression of somatostAtin neurAl firing wAs sufficient to enhAnce EPSP AmplituDe AnD reDuce fAilure rAtes, effects thAt were fully reversible AnD Also occluDeD by GABAb AntAgonists. These DAtA inDicAte thAt somatostAtin-expressing interneurons can rapidly and reversibly rewire neocortical networks through synAptic silencing, And suggest A critical role for these neurons in gating perception And plAsticity. Emilio Bizzi, DepArtment of BrAin AnD Cognitive Sciences And McGovern Institute for Brain Research, MIT Muscle Synergies, Concept, Principles, and Potential use in Neurorehabilitation When the centrAl nervous system (CNS) generAtes voluntAry movement, many muscles, eAch comprising thousAnDs of motor units, Are simultAneously ActivAteD AnD coorDinAteD. ComputAtionAlly, this is A DAunting tAsk, AnD investigAtors hAve striveD to unDerstAnD whether AnD how the CNS’s burDen is reDuceD to A much smaller set of vAriAbles. In the lAst few yeArs, we and our collaborators have seArcheD for physiologicAl eviDence of simplifying strAtegies by exploring whether the motor system makes use of motor moDules, to construct A lArge set of movement.
    [Show full text]
  • Brain Uploading-Related Group Mind Scenarios
    MIRI MACHINE INTELLIGENCE RESEARCH INSTITUTE Coalescing Minds: Brain Uploading-Related Group Mind Scenarios Kaj Sotala University of Helsinki, MIRI Research Associate Harri Valpola Aalto University Abstract We present a hypothetical process of mind coalescence, where artificial connections are created between two or more brains. This might simply allow for an improved formof communication. At the other extreme, it might merge the minds into one in a process that can be thought of as a reverse split-brain operation. We propose that one way mind coalescence might happen is via an exocortex, a prosthetic extension of the biological brain which integrates with the brain as seamlessly as parts of the biological brain in- tegrate with each other. An exocortex may also prove to be the easiest route for mind uploading, as a person’s personality gradually moves away from the aging biological brain and onto the exocortex. Memories might also be copied and shared even without minds being permanently merged. Over time, the borders of personal identity may become loose or even unnecessary. Sotala, Kaj, and Harri Valpola. 2012. “Coalescing Minds: Brain Uploading-Related Group Mind Scenarios.” International Journal of Machine Consciousness 4 (1): 293–312. doi:10.1142/S1793843012400173. This version contains minor changes. Kaj Sotala, Harri Valpola 1. Introduction Mind uploads, or “uploads” for short (also known as brain uploads, whole brain emu- lations, emulations or ems) are hypothetical human minds that have been moved into a digital format and run as software programs on computers. One recent roadmap chart- ing the technological requirements for creating uploads suggests that they may be fea- sible by mid-century (Sandberg and Bostrom 2008).
    [Show full text]
  • The Sound of Vision Project: on the Feasibility of an Audio-Haptic Representation of the Environment, for the Visually Impaired
    brain sciences Review The Sound of Vision Project: On the Feasibility of an Audio-Haptic Representation of the Environment, for the Visually Impaired Ómar I. Jóhannesson 1,*, Oana Balan 2, Runar Unnthorsson 3, Alin Moldoveanu 2 and Árni Kristjánsson 1 1 Laboratory of Visual Perception and Visuo-motor control, Faculty of Psychology, School of Health Sciences, University of Iceland, Reykjavik 101, Iceland; [email protected] 2 Faculty of Automatic Control and Computers, Computer Science and Engineering Department, University Politehnica of Bucharest, Bucharest 060042, Romania; [email protected] (O.B.); [email protected] (A.M.) 3 Faculty of Industrial Engineering, Mechanical Engineering and Computer Science, School of Engineering and Natural Sciences, University of Iceland, Reykjavik 101, Iceland; [email protected] * Correspondence: [email protected]; Tel.: +354-868-1407 Academic Editor: Bernadette Murphy Received: 28 April 2016; Accepted: 23 June 2016; Published: 27 June 2016 Abstract: The Sound of Vision project involves developing a sensory substitution device that is aimed at creating and conveying a rich auditory representation of the surrounding environment to the visually impaired. However, the feasibility of such an approach is strongly constrained by neural flexibility, possibilities of sensory substitution and adaptation to changed sensory input. We review evidence for such flexibility from various perspectives. We discuss neuroplasticity of the adult brain with an emphasis on functional changes in the visually impaired compared to sighted people. We discuss effects of adaptation on brain activity, in particular short-term and long-term effects of repeated exposure to particular stimuli. We then discuss evidence for sensory substitution such as Sound of Vision involves, while finally discussing evidence for adaptation to changes in the auditory environment.
    [Show full text]
  • Electronic Approaches to Restoration of Sight
    Home Search Collections Journals About Contact us My IOPscience Electronic approaches to restoration of sight This content has been downloaded from IOPscience. Please scroll down to see the full text. 2016 Rep. Prog. Phys. 79 096701 (http://iopscience.iop.org/0034-4885/79/9/096701) View the table of contents for this issue, or go to the journal homepage for more Download details: IP Address: 171.64.108.170 This content was downloaded on 09/08/2016 at 17:30 Please note that terms and conditions apply. IOP Reports on Progress in Physics Reports on Progress in Physics Rep. Prog. Phys. Rep. Prog. Phys. 79 (2016) 096701 (29pp) doi:10.1088/0034-4885/79/9/096701 79 Review 2016 Electronic approaches to restoration of sight © 2016 IOP Publishing Ltd G A Goetz1,2 and D V Palanker1,3 RPPHAG 1 Hansen Experimental Physics Laboratory, Stanford University, Stanford, CA 94305, USA 2 Neurosurgery, Stanford University, Stanford, CA 94305, USA 096701 3 Ophthalmology, Stanford University, Stanford, CA 94305, USA E-mail: [email protected] and [email protected] G A Goetz and D V Palanker Received 11 November 2014, revised 11 April 2016 Accepted for publication 23 May 2016 Electronic approaches to restoration of sight Published 9 August 2016 Abstract Printed in the UK Retinal prostheses are a promising means for restoring sight to patients blinded by the gradual atrophy of photoreceptors due to retinal degeneration. They are designed to reintroduce ROP information into the visual system by electrically stimulating surviving neurons in the retina. This review outlines the concepts and technologies behind two major approaches to retinal prosthetics: epiretinal and subretinal.
    [Show full text]
  • Neural Representation of Force, Grasp, and Volitional State in Intracortical Brain-Computer Interface Uers with Tetraplegia
    NEURAL REPRESENTATION OF FORCE, GRASP, AND VOLITIONAL STATE IN INTRACORTICAL BRAIN-COMPUTER INTERFACE UERS WITH TETRAPLEGIA BY ANISHA RASTOGI Submitted in partial fulfillment of the requirements For the degree of doctor of philosophy Dissertation Advisor: Dr. A. Bolu Ajiboye Department of Biomedical Engineering CASE WESTERN RESERVE UNIVERSITY January, 2021 CASE WESTERN RESERVE UNIVERSITY SCHOOL OF GRADUATE STUDIES We hereby approve the thesis/dissertation of Anisha Rastogi candidate for the degree of Doctor of Philosophy* Committee Chair Dr. Dominique Durand Committee Member Dr. A. Bolu Ajiboye Committee Member Dr. Dawn M. Taylor Committee Member Dr. Jonathan P. Miller Date of Defense 22 June 2020 *We also certify that written approval has been obtained for any proprietary material contained therein Dedication To Bill Kochevar, for dedicating his last days to bettering the world for people with disabilities. This research is a way to honor your faith in a brighter future. i Table of Contents Dedication ........................................................................................................................ i Table of Contents ............................................................................................................ ii List of Figures ................................................................................................................. vi List of Tables ................................................................................................................ viii List of Abbreviations ......................................................................................................
    [Show full text]
  • Connecting Brain Computer Interface Technology with Ubiquitous Complex Event Processing
    Connecting Brain Computer Interface Technology with Ubiquitous Complex Event Processing – Aspects of New Applications in the Fields of Biology, Brain Research and Medicine Extended Abstract Yuri Danilov Kurt Kaczmarek Mitchell Tyler TCNL, room M1012, 1550 University Ave, Biomedical Engineering Department, UW- Madison, USA [email protected] , … Rainer von Ammon CITT GmbH, Konrad-Adenauerallee 30, D-93051 Regensburg, Germany [email protected] Outline (to be replaced by abstract in the final paper, Springer style) 1. Why connecting BCI with U-CEP? 2. BCI as interface to and from the brain 3. The role of U-CEP 4. Infrastructure for connecting BCI and U-CEP: Body Area Networks and Wearable Technologies 5. The appropriate computing power - cognitive computing chips, combining digital ‘neurons’ and ‘synapses’ 5. New applications based on connecting BCI and U-CEP 6. BrainPort technology 7. CN-NINM technology 8. Example “Depression (or another example)”: Defining biomarkers, event modelling of biosensors as event sources and process modelling for influencing the protein-machinery Why connecting BCI with U-CEP? The edBPM/U-CEP workshop series has so far focused on the topics of connecting Internet of Services and Things with the management of business processes and the Future and Emerging Technologies as addressed by the ISTAG Recommendations of the European FET-F 2020 and Beyond Initiative [1]. Such FET challenges are not longer limited to business processes, but focus on new ideas in order to connect processes on the basis of CEP with disciplines of Cell Biology, Epigenetics, Brain Research, Robotics, Emergency Management, SocioGeonomics, Bio- and Quantum Computing – summarized under the concept of U-CEP [2].
    [Show full text]
  • Is There a Future for Sensory Substitution Outside Academic Laboratories?
    Multisensory Research 27 (2014) 271–291 brill.com/msr Is There a Future for Sensory Substitution Outside Academic Laboratories? Giulia V. Elli, Stefania Benetti and Olivier Collignon ∗ Center for Mind/Brain Sciences — CIMeC, University of Trento, Via delle Regole 101, 38060 Mattarello, Italy Received 28 May 2014; accepted 10 September 2014 Abstract Sensory substitution devices (SSDs) have been developed with the ultimate purpose of supporting sensory deprived individuals in their daily activities. However, more than forty years after their first appearance in the scientific literature, SSDs still remain more common in research laboratories than in the daily life of people with sensory deprivation. Here, we seek to identify the reasons behind the limited diffusion of SSDs among the blind community by discussing the ergonomic, neurocognitive and psychosocial issues potentially associated with the use of these systems. We stress that these issues should be considered together when developing future devices or improving existing ones. We provide some examples of how to achieve this by adopting a multidisciplinary and participatory approach. These efforts would contribute not solely to address fundamental theoretical research ques- tions, but also to better understand the everyday needs of blind people and eventually promote the use of SSDs outside laboratories. Keywords Sensory substitution devices, limitation, sensory deprivation, rehabilitation, ergonomics, blindness 1. Introduction In pursuing visual rehabilitation, invasive and non-invasive solutions are ex- plored. Invasive interventions, such as implanted neuroprostheses, rely on the integrity of visual pathways and brain regions. For example, several groups are currently developing artificial retinas aimed at inducing visual phosphenes through electrical stimulation of the remaining retinal cells (e.g., Ohta, 2011; Rizzo et al.,2007).Sincesuchinvasiveapproachesattempttostimulatethe deficient sensory system directly, the natural reorganization processes that oc- * To whom correspondence should be addressed.
    [Show full text]
  • Development and Plasticity in Sensory Thalamus
    Development and Plasticity in Sensory Thalamus and Cortex 1 Pioneer Neurons and Interneurons in the Developing Subplate: Molecular Markers, Cell Birthdays, and Neurotransmitters Robert F. Hevner and Nada Zecevic 1. Abstract Subplate neurons are essential for the development of cortical axon path- ways, including thalamocortical innervation as well as the formation of some cortico-cortical and descending cortical efferent connections. Previous evidence suggests that the critical subplate neurons are early- born “pioneer’’ neurons, which extend the first axons out of the cor- tex to subcortical forebrain regions. However, pioneer neurons are not the only type of neuron in the subplate layer. The subplate contains both glutamatergic and GABAergic neurons, some of which are tran- sitory due to either ongoing cell migration or subsequent cell death. We have studied the cellular composition of the subplate in develop- ing mouse and human cortex by retrograde axon tracing, cell birth- dating, and immunohistochemical analysis of specific markers. Our re- sults indicate that pioneer neurons are early-born glutamatergic neurons that express transcription factor Tbr1, transgene golli-lacZ, and other markers. In contrast, GABAergic interneurons in the subplate do not make subcortical (pioneer) axon projections, but instead migrate tan- gentially and radially through the subplate layer, express transcription factor Dlx, and are born both early and late in corticogenesis. Subplate neurons are essential in development of the initial cortical connectiv- ity, and it is thus important to distinguish between the different cell types present in this compartment, using molecular markers. The sub- plate in humans appears to contain a similar diversity of neuron types as in mice, but is markedly thicker than in mice, as confirmed by the broad band of Tbr1 expression extending below the cortical plate in humans.
    [Show full text]