Deep Learning for Signal to Noise Enhancement of Fluorescence

Total Page:16

File Type:pdf, Size:1020Kb

Deep Learning for Signal to Noise Enhancement of Fluorescence Deep Learning for Denoising of Fluorescence Microscopy Images Tram-Anh Nguyen Guy Hagen and Jonathan Ventura George Mason University University of Colorado Colorado Springs Fairfax, VA Colorado Springs, CO [email protected] [email protected], [email protected] Abstract from microscopy images. These filtering techniques have limitations, which call for a more generalized solution. Fluorescence microscopy images are often taken at low Deep learning methods have been successful in restoring light and short exposure times to preserve the integrity corrupted images, as well as in other image processing tasks of cell samples. However, imaging under these condi- such as classification, segmentation, and object detection. A tions leads to severely degraded images with low sig- nal to noise ratios. To computationally restore these im- widely used deep learning architecture in image processing ages, we introduce novel loss functions to denoise mi- is a convolutional neural network (CNN). A CNN consists croscopy images. These loss functions will be folded of an input layer, hidden layers, and an output layer. Another into the CARE algorithm. The results produced by this popular network in image processing is an autoencoder. Typ- modification will be evaluated against traditional TV fil- ically, autoencoders are used for denoising and reducing the tering and NL means techniques. The modified model number of dimensions of input data (Xing et al. 2017). will also be compared against its CARE predecessor us- ing standard image quality metrics. Related Work Deep learning approaches to image deblurring may involve Introduction blind and non-blind image deconvolution. There are a wealth of studies devoted to the non-blind image deconvolution ap- Fluorescence microscopy is vital for understanding pro- proach, but these networks are limited, as they rely on in- cesses and structures at the cellular level. Because imaging formation about the non-blurry image beforehand. By con- at the cellular level under strong lighting conditions or long trast, blind deblurring models are more flexible since in- exposure times may damage the cell sample through pho- formation about the non-blurry image is not required for the totoxicity, fluorescence microscopy images need to be re- network to deblur an input image. stored. A safe way to image a cell is to use low light condi- In 2014, Xu et al. introduced a natural image deconvo- tions and/or low exposure times, which unfortunately lowers lution that is data-driven and does not rely on traditional the signal to noise ratio (Xing et al. 2017). assumptions. For example, generative models tend to as- sume that noise in an image is identically and independ- ently distributed, even if this assumption is not necessarily true. Instead, this CNN was trained on images that were not deblurred ahead of time and the network learned the decon- volution operation without requiring information about the original image. The main contribution of Xu et al. was devel- oping this deep convolutional neural network (DCNN) that consisted of two sub models—one for deconvolution and the other for denoising. The models perform inverse filtering us- ing large 1D kernels and the former sub model is pre-trained to mimic Weiner deconvolution (Xu et al. 2014). A major drawback of the previously discussed blind Figure 1: Fluorescence microscopy images taken of actin DCNN was that it failed if the original image was not blurry. under high light (left) and low light conditions (right) To address this shortcoming, Conti et.al introduced a con- volutional neural network that consisted of a regularization Noise in an image depends on a combination of factors, term in the cost function. This improved model was able including exposure time and physical experimental condi- to denoise a blurry image and maintain the quality of im- tions. In fluorescence microscopy, noise is typically de- ages that are not noisy. The major modifications made by scribed by a Poison-Gaussian model. There has been extens- Conti et al. were that they used single 2D convolutional lay- ive work done in image restoration through filtering noise ers rather than 1D kernels for deblurring. The regularized cost function was built using the results of a classification microscopy images. network trained to distinguish blurry and non-blurry images that had roughly 80 percent accuracy when evaluated(Conti, Research Questions and Hypothesis Minucci, and Derakhshan 2017). The questions we will address in this study are: More recently, Weigert et al. published a series of image restoration methods that succeeded in restoring seven im- 1. How well does the CARE network perform on our micro- ages of various organisms (i.e. planaria flatworm, fruit fly scopy data set? wings). The major contributions of this work include: gener- 2. How can we produce image restorations that are sharper ating training data without requiring manual labeling, replic- than those produced by CARE? ating live imaging for organisms in which live imaging had once been near-impossible, and restoring microscopy im- 3. How does the network perform when it is trained on one ages even when lighting conditions are reduced by 60-fold. kind of sample and tested on a different kind of sample? Weigert et al. demonstrated high quality results with restor- 4. Can we reliably restore live cell images? ations of images containing Tribolium castaneum (red flour We hypothesize that our improvement on the method of Wei- beetle) and Schmidtea mediterranea (a flatworm commonly gert et al. will restore microscopy images that are less blurry known as planaria). This Content-Aware Image Restoration and more detailed than the restorations of the original CARE (CARE) network is based on the U-Net network, which con- model. Thus, our model will be a more faithful solution sists of an encoder-decoder architecture. The only difference compared to the denoising approach of Weigert et al. between the CARE network and the U-net algorithm is that the former outputs a per-pixel Laplace distribution whereas Proposed Implementation the latter outputs one value per pixel. Although these res- toration methods are promising, each pair of image content We modified the loss function of the CARE network in and corruption requires a unique data set. Each model must hopes of producing restorations that are more faithful to be retrained for an image of that particular content and cor- ground truth images. Ultimately, the objective of this study ruption to be successfully restored (Weigert et al. 2017). is to restore microscopy images free of artifacts and without Modifying network architecture is a popular strategy to loss of fine details. achieve better performance. Another widely used approach to this goal is data augmentation. Data augmentation is Experimental Setup widely used for image classification(Paulin et al. 2014), as The data set we used to train the standard CARE model is applying random transformations to training data effectively High Low, which consists of over 400 fluorescent images of provides more data for a network to learn from. Not sur- actin and mitochondria, in addition to 170 images of dendra. prisingly, training with large data sets produces high-quality All images in the High Low data set were taken using an image restorations (Burger, Schuler, and Harmeling 2012). Olympus IX83 with 60X/1.3NA objective lens. The Andor Typical transformations for data augmentation include trans- Zyla CMOS camera was used to image cell organelles. lation, rotation, and scaling. Augmenting data is a manual For all of our experiments, we first observed the beha- process in which these image transformations are specified vior of the network and assessed the quality of the network by humans. To automate this process, Jain et al. developed without loss function modifications. These results were then an unsupervised learning procedure that generated training compared to results produced by training the CARE network samples using different noise models (Jain and Seung 2009). using our FFT and bandpass cost functions. These experi- Likewise, Hauberg et al. developed a learned augmentation ments include restorations of actin imaged at 1 millisecond scheme that outperforms manual augmentation of MNIST and 10 milliseconds, restorations of mitochondria imaged data when used as training for a multilayer perceptron and a at 1 millisecond, model mismatch experiments, and restora- CNN (Hauberg et al. 2016). tions of dendra imaged at 10 milliseconds. We use the term Lastly, altering the loss function is a viable strategy, model mismatch to indicate experiments in which images though this approach tends to be overlooked (Zhao et al. of one type of cell content are used for training while im- 2017). Typically, the mean absolute error and mean squared ages of another type of cell content are used for testing. For error loss functions are employed in image processing net- example, we used images of mitochondria as training data works (Zhao et al. 2017; Burger, Schuler, and Harmeling for the CARE algorithm, and subsequently tested the model 2012; Agostinelli, Anderson, and Lee 2013; Chen et al. using noisy images of actin. Our most recent experiments in- 2018). In 2017, Zhao et al. introduced a new loss function for volved restorations of dendra imaged at 10 milliseconds us- image restoration that combined the multi-scale SSIM (MS- ing dendra imaged at 10 and 400 milliseconds as the training SSIM) metric with L1 loss. Without changing network ar- set. The results of these experiments were evaluated using chitecture, Zhao et al. demonstrated that by using this mixed peak signal to noise (PSNR) and structural similarity (SSIM) loss function, their fully convolutional neural network out- image quality metrics. performed state-of-the-art networks on tasks such as joint To conduct our experiments, we used the default config- denoising and demosaicking (Zhao et al. 2017). Drawing in- urations of the standard CARE model.
Recommended publications
  • Deblured Gaussian Blurred Images
    JOURNAL OF COMPUTING, VOLUME 2, ISSUE 4, APRIL 2010, ISSN 2151-9617 HTTPS://SITES.GOOGLE.COM/SITE/JOURNALOFCOMPUTING/ 33 Deblured Gaussian Blurred Images Mr. Salem Saleh Al-amri1, Dr. N.V. Kalyankar2 and Dr. Khamitkar S.D 3 Abstract—This paper attempts to undertake the study of Restored Gaussian Blurred Images. by using four types of techniques of deblurring image as Wiener filter, Regularized filter ,Lucy Richardson deconvlutin algorithm and Blind deconvlution algorithm with an information of the Point Spread Function (PSF) corrupted blurred image with Different values of Size and Alfa and then corrupted by Gaussian noise. The same is applied to the remote sensing image and they are compared with one another, So as to choose the base technique for restored or deblurring image.This paper also attempts to undertake the study of restored Gaussian blurred image with no any information about the Point Spread Function (PSF) by using same four techniques after execute the guess of the PSF, the number of iterations and the weight threshold of it. To choose the base guesses for restored or deblurring image of this techniques. Index Terms—Blur/Types of Blur/PSF; Deblurring/Deblurring Methods. —————————— —————————— 1 INTRODUCTION he restoration image is very important process in the image entire image. Tprocessing to restor the image by using the image processing This type of blurring can be distribution in horizontal and vertical techniques to easily understand this image without any artifacts direction and can be circular averaging by radius
    [Show full text]
  • A Comparative Study of Different Noise Filtering Techniques in Digital Images
    International Journal of Engineering Research and General Science Volume 3, Issue 5, September-October, 2015 ISSN 2091-2730 A COMPARATIVE STUDY OF DIFFERENT NOISE FILTERING TECHNIQUES IN DIGITAL IMAGES Sanjib Das1, Jonti Saikia2, Soumita Das3 and Nural Goni4 1Deptt. of IT & Mathematics,ICFAI University Nagaland, Dimapur, 797112, India, [email protected] 2Service Engineer, Indian export & import office, Shillong- 793003, Meghalaya, India, [email protected] 3Department of IT, School of Technology, NEHU, Shillong-22, Meghalaya, India, [email protected] 4Department of IT, School of Technology, NEHU, Shillong-22, Meghalaya, India, [email protected] Abstract: Although various solutions are available for de-noising them, a detail study of the research is required in order to design a filter which will fulfil the desire aspects along with handling most of the image filtering issues. In this paper we want to present some of the most commonly used noise filtering techniques namely: Median filter, Gaussian filter, Kuan filter, Morphological filter, Homomorphic Filter, Bilateral Filter and wavelet filter. Median filter is used for reducing the amount of intensity variation from one pixel to another pixel. Gaussian filter is a smoothing filter in the 2D convolution operation that is used to remove noise and blur from image. Kuan filtering technique transforms multiplicative noise model into additive noise model. Morphological filter is defined as increasing idempotent operators and their laws of composition are proved. Homomorphic Filter normalizes the brightness across an image and increases contrast. Bilateral Filter is a non-linear edge preserving and noise reducing smoothing filter for images. Wavelet transform can be applied to image de-noising and it has been extremely successful.
    [Show full text]
  • A Simple Second Derivative Based Blur Estimation Technique Thesis
    A Simple Second Derivative Based Blur Estimation Technique Thesis Presented in Partial Fulfillment of the Requirements for the Degree Master of Science in the Graduate School of The Ohio State University By Gourab Ghosh Roy, B.E. Graduate Program in Computer Science & Engineering The Ohio State University 2013 Thesis Committee: Brian Kulis, Advisor Mikhail Belkin Copyright by Gourab Ghosh Roy 2013 Abstract Blur detection is a very important problem in image processing. Different sources can lead to blur in images, and much work has been done to have automated image quality assessment techniques consistent with human rating. In this work a no-reference second derivative based image metric for blur detection and estimation has been proposed. This method works by evaluating the magnitude of the second derivative at the edge points in an image, and calculating the proportion of edge points where the magnitude is greater than a certain threshold. Lower values of this proportion or the metric denote increased levels of blur in the image. Experiments show that this method can successfully differentiate between images with no blur and varying degrees of blur. Comparison with some other state-of-the-art quality assessment techniques on a standard dataset of Gaussian blur images shows that the proposed method gives moderately high performance values in terms of correspondence with human subjective scores. Coupled with the method’s primary aspect of simplicity and subsequent ease of implementation, this makes it a probable choice for mobile applications. ii Acknowledgements This work was motivated by involvement in the personal analytics research group with Dr.
    [Show full text]
  • Histopathology Learning Dataset Augmentation: a Hybrid Approach
    Proceedings of Machine Learning Research { Under Review:1{9, 2021 Full Paper { MIDL 2021 submission HISTOPATHOLOGY LEARNING DATASET AUGMENTATION: A HYBRID APPROACH Ramzi HAMDI1 [email protected] Cl´ement BOUVIER1 [email protected] Thierry DELZESCAUX2 [email protected] C´edricCLOUCHOUX1 [email protected] 1 WITSEE, Neoxia, Paris, France 2 CEA-CNRS-UMR 9199, LMN, MIRCen, Univ. Paris-Saclay, Fontenay-aux-Roses, France Editors: Under Review for MIDL 2021 Abstract In histopathology, staining quantification is a mandatory step to unveil and characterize disease progression and assess drug efficiency in preclinical and clinical settings. Super- vised Machine Learning (SML) algorithms allow the automation of such tasks but rely on large learning datasets, which are not easily available for pre-clinical settings. Such databases can be extended using traditional Data Augmentation methods, although gen- erated images diversity is heavily dependent on hyperparameter tuning. Generative Ad- versarial Networks (GAN) represent a potential efficient way to synthesize images with a parameter-independent range of staining distribution. Unfortunately, generalization of such approaches is jeopardized by the low quantity of publicly available datasets. To leverage this issue, we propose a hybrid approach, mixing traditional data augmentation and GAN to produce partially or completely synthetic learning datasets for segmentation application. The augmented datasets are validated by a two-fold cross-validation analysis using U-Net as a SML method and F-Score as a quantitative criterion. Keywords: Generative Adversarial Networks, Histology, Machine Learning, Whole Slide Imaging 1. Introduction Histology is widely used to detect biological objects with specific staining such as protein deposits, blood vessels or cells.
    [Show full text]
  • Anisotropic Diffusion in Image Processing
    Anisotropic diffusion in image processing Tomi Sariola Advisor: Petri Ola September 16, 2019 Abstract Sometimes digital images may suffer from considerable noisiness. Of course, we would like to obtain the original noiseless image. However, this may not be even possible. In this thesis we utilize diffusion equations, particularly anisotropic diffusion, to reduce the noise level of the image. Applying these kinds of methods is a trade-off between retaining information and the noise level. Diffusion equations may reduce the noise level, but they also may blur the edges and thus information is lost. We discuss the mathematics and theoretical results behind the diffusion equations. We start with continuous equations and build towards discrete equations as digital images are fully discrete. The main focus is on iterative method, that is, we diffuse the image step by step. As it occurs, we need certain assumptions for these equations to produce good results, one of which is a timestep restriction and the other is a correct choice of a diffusivity function. We construct an anisotropic diffusion algorithm to denoise images and compare it to other diffusion equations. We discuss the edge-enhancing property, the noise removal properties and the convergence of the anisotropic diffusion. Results on test images show that the anisotropic diffusion is capable of reducing the noise level of the image while retaining the edges of image and as mentioned, anisotropic diffusion may even sharpen the edges of the image. Contents 1 Diffusion filtering 3 1.1 Physical background on diffusion . 3 1.2 Diffusion filtering with heat equation . 4 1.3 Non-linear models .
    [Show full text]
  • 22Nd International Congress on Acoustics ICA 2016
    Page intentionaly left blank 22nd International Congress on Acoustics ICA 2016 PROCEEDINGS Editors: Federico Miyara Ernesto Accolti Vivian Pasch Nilda Vechiatti X Congreso Iberoamericano de Acústica XIV Congreso Argentino de Acústica XXVI Encontro da Sociedade Brasileira de Acústica 22nd International Congress on Acoustics ICA 2016 : Proceedings / Federico Miyara ... [et al.] ; compilado por Federico Miyara ; Ernesto Accolti. - 1a ed . - Gonnet : Asociación de Acústicos Argentinos, 2016. Libro digital, PDF Archivo Digital: descarga y online ISBN 978-987-24713-6-1 1. Acústica. 2. Acústica Arquitectónica. 3. Electroacústica. I. Miyara, Federico II. Miyara, Federico, comp. III. Accolti, Ernesto, comp. CDD 690.22 ISBN 978-987-24713-6-1 © Asociación de Acústicos Argentinos Hecho el depósito que marca la ley 11.723 Disclaimer: The material, information, results, opinions, and/or views in this publication, as well as the claim for authorship and originality, are the sole responsibility of the respective author(s) of each paper, not the International Commission for Acoustics, the Federación Iberoamaricana de Acústica, the Asociación de Acústicos Argentinos or any of their employees, members, authorities, or editors. Except for the cases in which it is expressly stated, the papers have not been subject to peer review. The editors have attempted to accomplish a uniform presentation for all papers and the authors have been given the opportunity to correct detected formatting non-compliances Hecho en Argentina Made in Argentina Asociación de Acústicos Argentinos, AdAA Camino Centenario y 5006, Gonnet, Buenos Aires, Argentina http://www.adaa.org.ar Proceedings of the 22th International Congress on Acoustics ICA 2016 5-9 September 2016 Catholic University of Argentina, Buenos Aires, Argentina ICA 2016 has been organised by the Ibero-american Federation of Acoustics (FIA) and the Argentinian Acousticians Association (AdAA) on behalf of the International Commission for Acoustics.
    [Show full text]
  • Removing Camera Shake from a Single Photograph
    Removing Camera Shake from a Single Photograph Rob Fergus1 Barun Singh1 Aaron Hertzmann2 Sam T. Roweis2 William T. Freeman1 1MIT CSAIL 2University of Toronto Figure 1: Left: An image spoiled by camera shake. Middle: result from Photoshop “unsharp mask”. Right: result from our algorithm. Abstract depth-of-field. A tripod, or other specialized hardware, can elim- inate camera shake, but these are bulky and most consumer pho- tographs are taken with a conventional, handheld camera. Users Camera shake during exposure leads to objectionable image blur may avoid the use of flash due to the unnatural tonescales that re- and ruins many photographs. Conventional blind deconvolution sult. In our experience, many of the otherwise favorite photographs methods typically assume frequency-domain constraints on images, of amateur photographers are spoiled by camera shake. A method or overly simplified parametric forms for the motion path during to remove that motion blur from a captured photograph would be camera shake. Real camera motions can follow convoluted paths, an important asset for digital photography. and a spatial domain prior can better maintain visually salient im- age characteristics. We introduce a method to remove the effects of Camera shake can be modeled as a blur kernel, describing the cam- camera shake from seriously blurred images. The method assumes era motion during exposure, convolved with the image intensities. a uniform camera blur over the image and negligible in-plane cam- Removing the unknown camera shake is thus a form of blind image era rotation. In order to estimate the blur from the camera shake, deconvolution, which is a problem with a long history in the im- the user must specify an image region without saturation effects.
    [Show full text]
  • Performance Analysis of Intensity Averaging by Anisotropic Diffusion Method for MRI Denoising Corrupted by Random Noise by Ms
    Global Journal of Computer Science and Technology Graphics & Vision Volume 12 Issue 12 Version 1.0 Year 2012 Type: Double Blind Peer Reviewed International Research Journal Publisher: Global Journals Inc. (USA) Online ISSN: 0975-4172 & Print ISSN: 0975-4350 Performance Analysis of Intensity Averaging by Anisotropic diffusion Method for MRI Denoising Corrupted by Random Noise By Ms. Ami vibhakar, Mukesh Tiwari, Jaikaran singh & Sanjay rathore Ragiv Gandhi Technical University Abstract - The two parameters which plays important role in MRI (magnetic resonance imaging), acquired by various imaging modalities are Feature extraction and object recognition. These operations will become difficult if the images are corrupted with noise. Noise in MR image is always random type of noise. This noise will change the value of amplitude and phase of each pixel in MR image. Due to this, MR image gets corrupted and we cannot make perfect diagnostic for a body. So noise removal is essential task for perfect diagnostic. There are different approaches for noise reduction, each of which has its own advantages and limitation. MRI denoising is a difficult task as fine details in medical image containing diagnostic information should not be removed during noise removal process. In this paper, we are representing an algorithm for MRI denoising in which we are using iterations and Gaussian blurring for amplitude reconstruction and image fusion, anisotropic diffusion and FFT for phase reconstruction. We are using the PSNR(Peak signal to noise ration), MSE(Mean square error) and RMSE(Root mean square error) as performance matrices to measure the quality of denoised MRI .The final result shows that this method is effectively removing the noise while preserving the edge and fine information in the images.
    [Show full text]
  • California State University Northridge Noise
    CALIFORNIA STATE UNIVERSITY NORTHRIDGE NOISE REDUCTION AND IMAGE SMOOTHING USING GAUSSIAN BLUR. A graduate project in fulfillment of the requirements For the degree of Masters of Science In Electrical Engineering By Sachin Eswar MAY 2015 The graduate project of Sachin Eswar is approved: _____________________________________ ___________________ Dr. Ramin Roosta, Ph.D. Date _____________________________________ ___________________ Dr. Matthew Radmanesh, Ph.D. Date _____________________________________ ___________________ Dr. Shahnam Mirzaei, Ph.D., Chair Date California State University Northridge ii ACKNOWLEDGEMENT Working on this “Noise Reduction and Image Smoothing using Gaussian Blur” was a source of immense knowledge to me. I would like to express my sincere gratitude to Dr. Shahnam Mirzaei for his encouragement and unlimited support towards completion of my dissertation. His guidance and support are the key points for the completion of my project. I would also like to thank Dr. Ramin Roosta and Dr. Mathew Radmanesh for their extended support, scholarly advice and inspiration in completion of my project successfully. I am grateful to the Department of Electrical and computer engineering for giving me this opportunity and providing me the required knowledge in completing my project. I would also like to thank my family and friends, who encouraged me to extend my reach, and for their love, care and support through all the tough times during my graduation. iii TABLE OF CONTENTS SIGNATURE PAGE: ........................................................................................................
    [Show full text]
  • Scale-Space and Edge Detection Using Anisotropic Diffusion
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 12. NO. 7. JULY 1990 629 Scale-Space and Edge Detection Using Anisotropic Diffusion PIETRO PERONA AND JITENDRA MALIK Abstracf-The scale-space technique introduced by Witkin involves generating coarser resolution images by convolving the original image with a Gaussian kernel. This approach has a major drawback: it is difficult to obtain accurately the locations of the “semantically mean- ingful” edges at coarse scales. In this paper we suggest a new definition of scale-space, and introduce a class of algorithms that realize it using a diffusion process. The diffusion coefficient is chosen to vary spatially in such a way as to encourage intraregion smoothing in preference to interregion smoothing. It is shown that the “no new maxima should be generated at coarse scales” property of conventional scale space is pre- served. As the region boundaries in our approach remain sharp, we obtain a high quality edge detector which successfully exploits global information. Experimental results are shown on a number of images. The algorithm involves elementary, local operations replicated over the Fig. 1. A family of l-D signals I(x, t)obtained by convolving the original image making parallel hardware implementations feasible. one (bottom) with Gaussian kernels whose variance increases from bot- tom to top (adapted from Witkin [21]). Zndex Terms-Adaptive filtering, analog VLSI, edge detection, edge enhancement, nonlinear diffusion, nonlinear filtering, parallel algo- rithm, scale-space. with the initial condition I(x, y, 0) = Zo(x, y), the orig- inal image. 1. INTRODUCTION Koenderink motivates the diffusion equation formula- HE importance of multiscale descriptions of images tion by stating two criteria.
    [Show full text]
  • Newport Beach, California
    47th International Congress and Exposition on Noise Control Engineering (INTERNOISE 2018) Impact of Noise Control Engineering Chicago, Illinois, USA 26 - 29 August 2018 Volume 1 of 10 ISBN: 978-1-5108-7303-2 Printed from e-media with permission by: Curran Associates, Inc. 57 Morehouse Lane Red Hook, NY 12571 Some format issues inherent in the e-media version may also appear in this print version. Copyright© (2018) by Institute of Noise Control Engineering - USA (INCE-USA) All rights reserved. Printed by Curran Associates, Inc. (2018) For permission requests, please contact Institute of Noise Control Engineering - USA (INCE-USA) at the address below. Institute of Noise Control Engineering - USA (INCE-USA) INCE-USA Business Office 11130 Sunrise Valley Drive Suite 350 Reston, VA 20190 USA Phone: +1 703 234 4124 Fax: +1 703 435 4390 [email protected] Additional copies of this publication are available from: Curran Associates, Inc. 57 Morehouse Lane Red Hook, NY 12571 USA Phone: 845-758-0400 Fax: 845-758-2633 Email: [email protected] Web: www.proceedings.com INTER-NOISE 2018 – Pleanary, Keynote and Technical Sessions Papers are listed by Day, Morning/Afternoon, Session Number and Order of Presentation Select the underlined text (in18_xxxx.pdf) to link to paper Papers without links are available from the American Society of Mechanical Engineers (ASME) - Noise Control and Acoustics Division (NCAD) Sunday – 26 August, 2018 Opening Plenary: Barry Marshall Gibbs Session Chair: Stuart Bolton, Room: Chicago D and E 16:00 in18_4002.pdf Structure-Borne Sound in Buildings: Application of Vibro-Acoustic Methods for Measurement and Prediction .....1 Barry Marshall Gibbs, University of Liverpool Monday Morning – 27 August, 2018 Keynote: Jean-Louis Guyader Session Chair: Steffen Marburg Room: Chicago D 08:00 in18_5001.pdf Building Optimal Vibro-Acoustics Models From Measured Responses .....26 Jean-Louis Guyader, LVA/INSA de Lyon; Michael Ruzek, LAMCOS INSA de Lyon; Charles Pezerat, LAUM Universite du Maine Keynote: Amiya R.
    [Show full text]
  • Detection of Anomalous Noise Events for Real-Time Road
    Noise Mapp. 2018; 5:71–85 Research Article Francesc Alías*, Rosa Ma Alsina-Pagès, Ferran Orga, and Joan Claudi Socoró Detection of Anomalous Noise Events for Real-Time Road-Traflc Noise Mapping: The DYNAMAP’s project case study https://doi.org/10.1515/noise-2018-0006 1 Introduction Received Dec 28, 2017; accepted Jun 26, 2018 Abstract: Environmental noise is increasing year after year, Environmental noise is increasing year after year and be- especially in urban and suburban areas. Besides annoy- coming a growing concern in urban and suburban areas, ance, environmental noise also causes harmful health especially in large cities, since it does not only cause an- effects on people. The Environmental Noise Directive noyance to citizens, but also harmful effects on people. 2002/49/EC (END) is the main instrument of the European Most of them focus on health-related problems [1], being of Union to identify and combat noise pollution, followed particular worry the impact of noise on children [2], whose by the CNOSSOS-EU methodological framework. In com- population group is especially vulnerable. Other investiga- pliance with the END legislation, the European Member tions have also shown the effects of noise pollution in con- States are required to publish noise maps and action plans centration, sleep and stress [3]. Finally, it is worth mention- every five years. The emergence of Wireless Acoustic Sen- ing that noise exposure does not only affect health, but can sor Networks (WASNs) have changed the paradigm to ad- also affect social and economic aspects [4]. dress the END regulatory requirements, allowing the dy- Among noise sources, road-traffic noise is one ofthe namic ubiquitous measurement of environmental noise main noise pollutants in cities.
    [Show full text]