A Short Course on Bioimage Informatics Lecture 1: Basic Concepts and Techniques

Total Page:16

File Type:pdf, Size:1020Kb

A Short Course on Bioimage Informatics Lecture 1: Basic Concepts and Techniques A Short Course on Bioimage Informatics Lecture 1: Basic Concepts and Techniques Ge Yang (杨戈) Department of Biomedical Engineering & Department of Computational Biology Carnegie Mellon University August 10, 2015 1 Acknowledgements • Dr. Chao Tang • Dr. Luhua Lai • Students, faculty, and staff of the Center for Quantitative Biology, Peking University • Dr. Jing Yuan (袁静) • Youjun Xu (徐优俊) • 111 Visiting Scholar Program, Ministry of Education • College of Chemistry and Molecular Engineering, Peking University • School of Engineering & School of Computer Science, Carnegie Mellon University 2 Outline • Background • What is bioimage informatics? • Where does it come from? • Where is it going? • Basic concepts of image formation and collection • Overview of bioimage informatics techniques • Summary 3 • Background • What is bioimage informatics? • Where does it come from? • Where is it going? • Basic concepts of image formation and collection • Overview of biological image analysis techniques • Summary 4 What information can we extract from images? 刘奇 5 More Examples of Biological Images Typical Readouts from Images • From static images, e.g. immunofluorescence images - Location - Shape/morphology - Spatial distribution: e.g. spatial patterns - Spatial relation: e.g. protein colocalization • From dynamic images, e.g. live cell time-lapse images - All the readouts above, which now may be time-dependent - In other words, spatiotemporal dynamics/processes • Computational analysis techniques are required to extract these quantitative readouts from images. 7 Relations of Imaging and Image Analysis with Other Biological Research Tools • None of these readouts can be directly provided by e.g. genetics or biochemistry. • However, understanding these readouts often requires close integration of computational image analysis with experimental analysis using genetics and/or biochemistry techniques. Alberts, MBoC, 5/e, 2007 8 Some Practical Questions • Q1: If I have an image analysis problem to solve, where do I look for information and solutions? • Q2: What techniques are available? Which technique should I choose? How can I be sure that I am getting the correct results? • Q3: How can I go from image analysis results to molecular mechanisms? • We will address these questions in this course. 9 • Background • What is bioimage informatics? • Where does it come from? • Where is it going? • Basic concepts of image formation and collection • Overview of biological image analysis techniques • Summary 10 What is Bioimage Informatics? • To give a simple definition, bioimage informatics is the computational analysis and understanding of biological images. • Bioimage informatics emerged as a new technical area primarily over the past ten years. - First meeting held in 2005 at Stanford University. https://sites.google.com/site/hanchuanpeng/bioimageinformatics_resource 11 • Background • What is bioimage informatics? • Where does it come from? • Where is it going? • Basic concepts of image formation and collection • Overview of biological image analysis techniques • Summary 12 Looking into History: Robert Hooke’s Microscope • Microscope was invented more than 300 years ago. 1) http://micro.magnet.fsu.edu/index.html Molecular expressions: microscopy world 2) http://www.history-of-the-microscope.org/ Robert Hooke (1635-1703) 13 Evolution of Microscopes http://www.microscopyu.com/museum/index.html 14 Pioneering Work of Computational Analysis of Biological Images • Shinya Inoue pioneered the use of video recording devices and digital image processing in microscopy in late 1970s to early 1980s. • It was found that digital image processing can improve image quality. • Today, we use the term “computational image analysis” interchangeably with “digital image processing”. Inoue, S. 1981. Video image processing greatly enhances contrast, quality Shinya Inoue, and speed in polarization based microscopy. J. Cell Biol. 89:346–356. Marine Biology Lab, Woods Hole, MA 15 Development of Image Processing and Computer Vision Techniques • Early work in digital image processing started in 1960s. • Computer vision techniques started as an area of robotics and artificial intelligence around the same period of time. • The work of Shinya Inoue is another example of interdisciplinary science. 16 Automation of Image Collection http://micro.magnet.fsu.edu/index.html Molecular expressions: microscopy world 17 High-Throughput Microscopy Pepperkok & Ellenberg, 2006 Multi-well plates 18 Advances in Fluorescence Microscopy Livet J, Weissman TA, Kang H, et al. Nature 450: 56–62, 2007 http://www.fda.gov/forconsumers/consumerupdates/ucm107783.htm 19 Systems Level Study of Biological Processes June 26, 2000 • International Human Genome Sequencing Consortium (2001). "Initial sequencing and analysis of the human genome" Nature 409 (6822): 860–921. • Venter, JC et al. (2001). "The sequence of the human genome" . Science 291 (5507): 1304– 1351. 20 Advances in Computer Vision Takeo Kanade lab, CMU 21 Advances in Medical Image Analysis Heimann et al, Comparison and Evaluation of Methods for Liver Segmentation From CT Datasets, IEEE Trans. Medical Imaging, vol. 28, no. 8, 1251-1265, 2009. 22 Summary • Development of bioimage informatics techniques is driven by the synergy of several factors: - The need for quantitative and systems-level study biological processes - Availability of computational image analysis techniques - Availability of automated image collection - Interdisciplinary collaboration •Q: If I have an image analysis problem to solve, where do I look for information and solutions? A: Many techniques have been developed specifically for biological image analysis. In addition, areas such as computer vision, medical image analysis are also important sources of techniques. 23 • Background • What is bioimage informatics? • Where does it come from? • Where is it going? • Basic concepts of image formation and collection • Overview of biological image analysis techniques • Summary 24 Important Trends in Fluorescence Microscopy (I): Fluorescence Biosensors Rho GTPase sensor in mouse embryonic fibroblast, Machacek et al, Nature, 2009 • Spatiotemporal cell dynamics visualized by fluorescence biosensors. Important Trends in Fluorescence Microscopy (II): Super-Resolution Microscopy Chen et al, IEEE Int. Sym. Biomed Imaging, 2014 26 Important Trends in Fluorescence Microscopy (III): Light-Sheet Microscopy B.-C. Chen et al, Science, 346: 439, 2014. 27 Summary • Bioimage informatics techniques are required for understanding complex spatiotemporal dynamics revealed by fluorescence biosensors. • Bioimage informatics techniques are required for further improve the resolution of super-resolution microscopy techniques. • Light sheet microscopy will drive the transition of cellular imaging from 2D to 3D. Bioimage informatics techniques are essential for 3D image analysis. 28 • Background • What is bioimage informatics? • Where does it come from? • Where is it going? • Basic concepts of image formation and collection • Overview of biological image analysis techniques • Summary 29 General Comments • To analyze biological images, it is often necessary to have at least some knowledge about the image formation process. • It is often feasible to significantly simplify image analysis by optimizing the imaging process. collaboration between image collection and image analysis • It is critical to choose imaging settings properly. Erroneous setting can make the collected images unusable. Nyquist/Shannon sampling criterion 30 Spectrum of Visible Light Color Wavelength violet 380–450 nm blue 450–495 nm green 495–570 nm yellow 570–590 nm orange 590–620 nm red 620–750 nm http://en.wikipedia.org/wiki/Visible_spectrum http://science.hq.nasa.gov/kids/imagers/ems/visible.html 31 Microscope as a Linear System • A light microscope can be considered as a linear system. http://micro.magnet.fsu.edu/primer/java/imageformation/airydiskformation/index.html • A linear system is characterized by its impulse response. For a microscope, this is its point spread function (PSF). 32 Airy Disk • Airy (after George Biddell Airy) disk is the diffraction pattern of a point feature under a circular aperture. • It has the following form 2 2Jx1 y x J1(x) is a Bessel function of the first kind. • Detailed derivation is given in Born & Wolf, Principles of Optics, 7th ed., pp. 439-441. 33 Microscope Image Formation • Microscope image formation can be modeled as a convolution with the PSF. I x,y O x,y psf x,y F I x,y F O x,y F psf x,y http://micro.magnet.fsu.edu/primer/java/mtf/airydisksize/index.html 34 Resolution Limit of Light Microscopy 061. • Rayleigh limit D NA • Sparrow limit 047. D NA http://www.microscopy.fsu.edu/primer/java/imageformation/rayleighdisks/index.html 35 Nyquist/Shannon Sampling Theorem • The Nyquist-Shannon sampling theorem: The sampling frequency should be at least twice the highest frequency contained in the signal. • This is true for both spatial and temporal sampling. • Spatial sampling is determined by the effective pixel size. It should be no larger than one half of the smallest feature size. • Temporal sampling is determined by frame rate. It should be at least twice as fast as the dynamic biological process to be visualized. 36 • Background • What is bioimage informatics? • Where does it come from? • Where is it going? • Basic concepts of image formation and collection • Overview of biological image analysis techniques • Summary 37 From Biological Images to Biological Knowledge 38 How are image analysis problem
Recommended publications
  • Bioimage Analysis Tools
    Bioimage Analysis Tools Kota Miura, Sébastien Tosi, Christoph Möhl, Chong Zhang, Perrine Paul-Gilloteaux, Ulrike Schulze, Simon Norrelykke, Christian Tischer, Thomas Pengo To cite this version: Kota Miura, Sébastien Tosi, Christoph Möhl, Chong Zhang, Perrine Paul-Gilloteaux, et al.. Bioimage Analysis Tools. Kota Miura. Bioimage Data Analysis, Wiley-VCH, 2016, 978-3-527-80092-6. hal- 02910986 HAL Id: hal-02910986 https://hal.archives-ouvertes.fr/hal-02910986 Submitted on 3 Aug 2020 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. 2 Bioimage Analysis Tools 1 2 3 4 5 6 Kota Miura, Sébastien Tosi, Christoph Möhl, Chong Zhang, Perrine Pau/-Gilloteaux, - Ulrike Schulze,7 Simon F. Nerrelykke,8 Christian Tischer,9 and Thomas Penqo'" 1 European Molecular Biology Laboratory, Meyerhofstraße 1, 69117 Heidelberg, Germany National Institute of Basic Biology, Okazaki, 444-8585, Japan 2/nstitute for Research in Biomedicine ORB Barcelona), Advanced Digital Microscopy, Parc Científic de Barcelona, dBaldiri Reixac 1 O, 08028 Barcelona, Spain 3German Center of Neurodegenerative
    [Show full text]
  • Other Departments and Institutes Courses 1
    Other Departments and Institutes Courses 1 Other Departments and Institutes Courses About Course Numbers: 02-250 Introduction to Computational Biology Each Carnegie Mellon course number begins with a two-digit prefix that Spring: 12 units designates the department offering the course (i.e., 76-xxx courses are This class provides a general introduction to computational tools for biology. offered by the Department of English). Although each department maintains The course is divided into two halves. The first half covers computational its own course numbering practices, typically, the first digit after the prefix molecular biology and genomics. It examines important sources of biological indicates the class level: xx-1xx courses are freshmen-level, xx-2xx courses data, how they are archived and made available to researchers, and what are sophomore level, etc. Depending on the department, xx-6xx courses computational tools are available to use them effectively in research. may be either undergraduate senior-level or graduate-level, and xx-7xx In the process, it covers basic concepts in statistics, mathematics, and courses and higher are graduate-level. Consult the Schedule of Classes computer science needed to effectively use these resources and understand (https://enr-apps.as.cmu.edu/open/SOC/SOCServlet/) each semester for their results. Specific topics covered include sequence data, searching course offerings and for any necessary pre-requisites or co-requisites. and alignment, structural data, genome sequencing, genome analysis, genetic variation, gene and protein expression, and biological networks and pathways. The second half covers computational cell biology, including biological modeling and image analysis. It includes homework requiring Computational Biology Courses modification of scripts to perform computational analyses.
    [Show full text]
  • An End-To-End System for Automatic Characterization of Iba1 Immunopositive Microglia in Whole Slide Imaging
    Neuroinformatics (2019) 17:373–389 https://doi.org/10.1007/s12021-018-9405-x ORIGINAL ARTICLE An End-to-end System for Automatic Characterization of Iba1 Immunopositive Microglia in Whole Slide Imaging Alexander D. Kyriazis1 · Shahriar Noroozizadeh1 · Amir Refaee1 · Woongcheol Choi1 · Lap-Tak Chu1 · Asma Bashir2 · Wai Hang Cheng2 · Rachel Zhao2 · Dhananjay R. Namjoshi2 · Septimiu E. Salcudean3 · Cheryl L. Wellington2 · Guy Nir4 Published online: 8 November 2018 © Springer Science+Business Media, LLC, part of Springer Nature 2018 Abstract Traumatic brain injury (TBI) is one of the leading causes of death and disability worldwide. Detailed studies of the microglial response after TBI require high throughput quantification of changes in microglial count and morphology in histological sections throughout the brain. In this paper, we present a fully automated end-to-end system that is capable of assessing microglial activation in white matter regions on whole slide images of Iba1 stained sections. Our approach involves the division of the full brain slides into smaller image patches that are subsequently automatically classified into white and grey matter sections. On the patches classified as white matter, we jointly apply functional minimization methods and deep learning classification to identify Iba1-immunopositive microglia. Detected cells are then automatically traced to preserve their complex branching structure after which fractal analysis is applied to determine the activation states of the cells. The resulting system detects white matter regions with 84% accuracy, detects microglia with a performance level of 0.70 (F1 score, the harmonic mean of precision and sensitivity) and performs binary microglia morphology classification with a 70% accuracy.
    [Show full text]
  • Digital Image Processing
    11/11/20 DIGITAL IMAGE PROCESSING Lecture 4 Frequency domain Tammy Riklin Raviv Electrical and Computer Engineering Ben-Gurion University of the Negev Fourier Domain 1 11/11/20 2D Fourier transform and its applications Jean Baptiste Joseph Fourier (1768-1830) ...the manner in which the author arrives at these A bold idea (1807): equations is not exempt of difficulties and...his Any univariate function can analysis to integrate them still leaves something be rewritten as a weighted to be desired on the score of generality and even sum of sines and cosines of rigour. different frequencies. • Don’t believe it? – Neither did Lagrange, Laplace, Poisson and Laplace other big wigs – Not translated into English until 1878! • But it’s (mostly) true! – called Fourier Series Legendre Lagrange – there are some subtle restrictions Hays 2 11/11/20 A sum of sines and cosines Our building block: Asin(wx) + B cos(wx) Add enough of them to get any signal g(x) you want! Hays Reminder: 1D Fourier Series 3 11/11/20 Fourier Series of a Square Wave Fourier Series: Just a change of basis 4 11/11/20 Inverse FT: Just a change of basis 1D Fourier Transform 5 11/11/20 1D Fourier Transform Fourier Transform 6 11/11/20 Example: Music • We think of music in terms of frequencies at different magnitudes Slide: Hoiem Fourier Analysis of a Piano https://www.youtube.com/watch?v=6SR81Wh2cqU 7 11/11/20 Fourier Analysis of a Piano https://www.youtube.com/watch?v=6SR81Wh2cqU Discrete Fourier Transform Demo http://madebyevan.com/dft/ Evan Wallace 8 11/11/20 2D Fourier Transform
    [Show full text]
  • 9. Biomedical Imaging Informatics Daniel L
    9. Biomedical Imaging Informatics Daniel L. Rubin, Hayit Greenspan, and James F. Brinkley After reading this chapter, you should know the answers to these questions: 1. What makes images a challenging type of data to be processed by computers, as opposed to non-image clinical data? 2. Why are there many different imaging modalities, and by what major two characteristics do they differ? 3. How are visual and knowledge content in images represented computationally? How are these techniques similar to representation of non-image biomedical data? 4. What sort of applications can be developed to make use of the semantic image content made accessible using the Annotation and Image Markup model? 5. Describe four different types of image processing methods. Why are such methods assembled into a pipeline when creating imaging applications? 6. Give an example of an imaging modality with high spatial resolution. Give an example of a modality that provides functional information. Why are most imaging modalities not capable of providing both? 7. What is the goal in performing segmentation in image analysis? Why is there more than one segmentation method? 8. What are two types of quantitative information in images? What are two types of semantic information in images? How might this information be used in medical applications? 9. What is the difference between image registration and image fusion? Given an example of each. 1 9.1. Introduction Imaging plays a central role in the healthcare process. Imaging is crucial not only to health care, but also to medical communication and education, as well as in research.
    [Show full text]
  • Recognition of Electrical & Electronics Components
    NATIONAL INSTITUTE OF TECHNOLOGY, ROURKELA Submission of project report for the evaluation of the final year project Titled Recognition of Electrical & Electronics Components Under the guidance of: Prof. S. Meher Dept. of Electronics & Communication Engineering NIT, Rourkela. Submitted by: Submitted by: Amiteshwar Kumar Abhijeet Pradeep Kumar Chachan 10307022 10307030 B.Tech IV B.Tech IV Dept. of Electronics & Communication Dept. of Electronics & Communication Engg. Engg. NIT, Rourkela NIT, Rourkela NATIONAL INSTITUTE OF TECHNOLOGY, ROURKELA Submission of project report for the evaluation of the final year project Titled Recognition of Electrical & Electronics Components Under the guidance of: Prof. S. Meher Dept. of Electronics & Communication Engineering NIT, Rourkela. Submitted by: Submitted by: Amiteshwar Kumar Abhijeet Pradeep Kumar Chachan 10307022 10307030 B.Tech IV B.Tech IV Dept. of Electronics & Communication Dept. of Electronics & Communication Engg. Engg. NIT, Rourkela NIT, Rourkela National Institute of Technology Rourkela CERTIFICATE This is to certify that the thesis entitled, “RECOGNITION OF ELECTRICAL & ELECTRONICS COMPONENTS ” submitted by Sri AMITESHWAR KUMAR ABHIJEET and Sri PRADEEP KUMAR CHACHAN in partial fulfillments for the requirements for the award of Bachelor of Technology Degree in Electronics & Instrumentation Engg. at National Institute of Technology, Rourkela (Deemed University) is an authentic work carried out by him under my supervision and guidance. To the best of my knowledge, the matter embodied in the thesis has not been submitted to any other University / Institute for the award of any Degree or Diploma. Date: Dr. S. Meher Dept. of Electronics & Communication Engg. NIT, Rourkela. ACKNOWLEDGEMENT I wish to express my deep sense of gratitude and indebtedness to Dr.S.Meher, Department of Electronics & Communication Engg.
    [Show full text]
  • Sampling Theory for Digital Video Acquisition: the Guide for the Perplexed User
    Sampling Theory for Digital Video Acquisition: The Guide for the Perplexed User Prof. Lenny Rudin1,2,3 Ping Yu1 Prof. Jean-Michel Morel2 Cognitech, Inc, Pasadena, California (1) Ecole Normale Superieure de Cachan, France(2) SMPTE Member (Hollywood Technical Section) (3) Abstract: Recently, the law enforcement community with professional interests in applications of image/video processing technology has been exposed to scientifically flawed assertions regarding the advantages and disadvantages of various hardware image acquisition devices (video digitizing cards). These assertions state a necessity of using SMPTE CCIR-601 standard when digitizing NTSC composite video signals from surveillance videotapes. In particular, it would imply that the pixel-sampling rate of 720*486 is absolutely required to capture all the available video information encoded in the composite video signal and recorded on (S)VHS videotapes. Fortunately, these statements can be directly analyzed within the strict mathematical context of Shannons Sampling Theory, and shown to be wrong. Here we apply the classical Shannon-Nyquist results to the process of digitizing composite analog video from videotapes to dispel the theoretically unfounded, wrong assertions. Introduction: The field of image processing is a mature interdisciplinary science that spans the fields of electrical engineering, applied mathematics, and computer science. Over the last sixty years, this field accumulated fundamental results from the above scientific disciplines. In particular, the field of signal processing has contributed to the basic understanding of the exact mathematical relationship between the world of continuous (analog) images and the digitized versions of these images, when collected by computer video acquisition systems. This theory is known as Sampling Theory, and is credited to Prof.
    [Show full text]
  • An Analysis of Aliasing and Image Restoration Performance for Digital Imaging Systems
    AN ANALYSIS OF ALIASING AND IMAGE RESTORATION PERFORMANCE FOR DIGITAL IMAGING SYSTEMS Thesis Submitted to The School of Engineering of the UNIVERSITY OF DAYTON In Partial Fulfillment of the Requirements for The Degree of Master of Science in Electrical Engineering By Iman Namroud UNIVERSITY OF DAYTON Dayton, Ohio May, 2014 AN ANALYSIS OF ALIASING AND IMAGE RESTORATION PERFORMANCE FOR DIGITAL IMAGING SYSTEMS Name: Namroud, Iman APPROVED BY: Russell C. Hardie, Ph.D. John S. Loomis, Ph.D. Advisor Committee Chairman Committee Member Professor, Department of Electrical Professor, Department of Electrical and Computer Engineering and Computer Engineering Eric J. Balster, Ph.D. Committee Member Assistant Professor, Department of Electrical and Computer Engineering John G. Weber, Ph.D. Tony E. Saliba, Ph.D. Associate Dean Dean, School of Engineering School of Engineering & Wilke Distinguished Professor ii c Copyright by Iman Namroud All rights reserved 2014 ABSTRACT AN ANALYSIS OF ALIASING AND IMAGE RESTORATION PERFORMANCE FOR DIGITAL IMAGING SYSTEMS Name: Namroud, Iman University of Dayton Advisor: Dr. Russell C. Hardie It is desirable to obtain a high image quality when designing an imaging system. The design depends on many factors such as the optics, the pitch, and the cost. The effort to enhance one aspect of the image may reduce the chances of enhancing another one, due to some tradeoffs. There is no imaging system capable of producing an ideal image, since that the system itself presents distortion in the image. When designing an imaging system, some tradeoffs favor aliasing, such as the desire for a wide field of view (FOV) and a high signal to noise ratio (SNR).
    [Show full text]
  • Bioimage Informatics
    Bioimage Informatics Lecture 22, Spring 2012 Empirical Performance Evaluation of Bioimage Informatics Algorithms Image Database; Content Based Image Retrieval Emerging Applications: Molecular Imaging Lecture 22 April 23, 2012 1 Outline • Empirical performance evaluation of bioimage Informatics algorithm • Image database; content-based image retrieval • Introduction to molecular imaging 2 • Empirical performance evaluation of bioimage Informatics algorithm • Image database; content-based image retrieval • Introduction to molecular imaging 3 Overview of Algorithm Evaluation (I) • Two sets of data are required - Input data - Corresponding correct output (ground truth) • Sources of input data - Actual/experimental data, often from experiments - Synthetic data, often from computer simulation • Actual/experimental data - Essential for performance evaluation - May be costly to acquire - May not be representative - Ground truth often is unknown Overview of Algorithm Evaluation (II) • What exactly is ground truth? Ground truth is a term used in cartography, meteorology, analysis of aerial photographs, satellite imagery and a range of other remote sensing techniques in which data are gathered at a distance. Ground truth refers to information that is collected "on location." - http://en.wikipedia.org/wiki/Ground_truth • Synthetic (simulated) data Advantages - Ground truth is known - Usually low-cost Disadvantages - Difficult to fully represent the original data 5 Overview of Algorithm Evaluation (III) • Simulated data Advantages - Ground truth is known - Usually low-cost Disadvantages - Difficult to fully represent the original data • Realistic synthetic data - To use as much information from real experimental data as possible • Quality control of manual analysis Test Protocol Development (I) • Implementation - Source codes are not always available and often are on different platforms. • Parameter setting - This is one of the most challenging issues.
    [Show full text]
  • Digital Image Processing Lectures 5 & 6
    Digital Image Processing Lectures 5 & 6 Mahmood (Mo) Azimi, Professor Colorado State University Fort Collins, Co 80523 email : [email protected] Each time you have to click on this icon wherever it appears to load the next sound file To advance click enter or page down to go back use page up 1. Parseval’s Theorem and Inner Product Preservation Another important property of FT is that the inner product of two functions is equal to the inner product of their FT’s. ZZ ∞ x(v, u)y∗(u, v)dudv −∞ ZZ ∞ 1 ∗ = 2 X(ω1, ω2)Y (ω1, ω2)dω1dω2 4π −∞ where ∗ stands for complex conjugate operation. When x = y we obtain the well-known Parseval energy conservation formula i.e. ZZ ∞ ZZ ∞ 2 1 2 |x(u, v)| dudv = 2 |X(ω1, ω2)| dω1dω2 −∞ 4π −∞ i.e. the total energy in the function is the same as in its FT. 2. Frequency Response and Eigenfunctions of 2-D LSI Systems An eigen-function of a system is defined as an input function To advance click enter or page down to go back use page up 1 that is reproduced at the output with a possible change in the amplitude. For an LSI system eigen-functions are given by f(u, v) = exp(jω1u + jω2v) Using the 2-D convolution integral ZZ ∞ 0 0 0 0 0 0 g(u, v) = h(u − u , v − v ) exp[jω1u + jω2v ]du dv −∞ change u˜ = u − u0, v˜ = v − v0, then g(u, v) = H(ω1, ω2) exp[jω1u + jω2v] where ZZ ∞ H(ω1, ω2) = h(˜u, v˜) exp[jω1u˜ + jω2v˜]dud˜ v˜ −∞ = F{h(u, v)} To advance click enter or page down to go back use page up 2 is the frequency response of the 2-D system.
    [Show full text]
  • Chapter7 Image Compression
    Chapter7 Image Compression •Preview • 71I7.1 Intro duct ion • 7.2 Fundamental concepts and theories • 7.3 Basidihdic coding methods • 7.4 Predictive coding • 7.5 Transform coding • 7.6 Introduction to international standards Digital Image Processing 1 Prof.Zhengkai Liu Dr.Rong Zhang 7.1 Introduction 7.1.1 why code data? 7.1.2 Classification • To reduce storage volume Lossless compression: reversible • To reduce transmission time Lossy compression: irreversible – One colour image • 760 by 580 pixels • 3 channels, each 8 bits • 1.3 Mbyte – Video da ta • same resolution • 25 frames per second • 33 Mbyte/second Digital Image Processing 2 Prof.Zhengkai Liu Dr.Rong Zhang 7.1 Introduction 7.1.3 Compression methods •Entropy coding: Huffman coding, arithmetic coding •Run length coding: G3, G4 •LZW coding: arj(DOS) , zi p( Wi nd ows) , compress( UNIX) •Predictive coding: linear prediction, non-linear prediction •TfTransform codi ng: DCT, KLT, W al sh -TWT(T, WT(wavel et) •Others: vector quantization (VQ), Fractal coding the second generation coding Digital Image Processing 3 Prof.Zhengkai Liu Dr.Rong Zhang 7. 2 Fundamental concepts and theories 7.2.1 Data redundancy If n1denote the original data number, and n2 denote the compressed data number, then the data redundancy can be defined as 1 RD =−1 CR where CR, commonly called the compression ratio, is n1 CR = n2 n =n FlFor example 2 1 RD=0 nn21 CRRD→∞,1 → Digital Image Processing 4 Prof.Zhengkai Liu Dr.Rong Zhang 7. 2 Fundamental concepts and theories 7.2.1 Data redundancy •Other expression of compression ratio n C = 2 n2 8/bits pixel R or L== bits/, pixel CR n1 MN× L •Three basic image redundancy Coding redundancy Inteeprpixle redu edudndancy psychovisual redundancy Digital Image Processing 5 Prof.Zhengkai Liu Dr.Rong Zhang 7.
    [Show full text]
  • A Survey of Medical Image Processing Tools
    See discussions, stats, and author profiles for this publication at: http://www.researchgate.net/publication/282610419 A Survey of Medical Image Processing Tools CONFERENCE PAPER · AUGUST 2015 DOI: 10.13140/RG.2.1.3364.4241 2 AUTHORS, INCLUDING: Siau-Chuin Liew Universiti Malaysia Pahang 24 PUBLICATIONS 30 CITATIONS SEE PROFILE Available from: Siau-Chuin Liew Retrieved on: 07 October 2015 provided by UMP Institutional Repository View metadata, citation and similar papers at core.ac.uk CORE brought to you by A Survey of Medical Image Processing Tools Lay-Khoon Lee Siau-Chuin Liew 1Faculty of Computer Systems and Software Engineering; 2Faculty of Computer Systems and Software Engineering; Universiti Malaysia Pahang, Universiti Malaysia Pahang, 26300 Gambang, Pahang, Malaysia 26300 Gambang, Pahang, Malaysia Abstract—A precise analysis of medical image is an important registration, segmentation, visualization, reconstruction, stage in the contouring phase throughout radiotherapy simulation and diffusion preparation. Medical images are mostly used as radiographic techniques in diagnosis, clinical studies and treatment planning. Medical image processing tool are also similarly as important. With a medical image processing tool, it is possible to speed up and enhance the operation of the analysis of the medical image. This paper describes medical image processing software tool which attempts to secure the same kind of programmability advantage for exploring applications of the pipelined processors. These tools simulate complete systems consisting of several of the proposed processing components, in a configuration described by a graphical schematic diagram. In this paper, fifteen different medical image processing tools will be compared in several aspects. The main objective of the comparison is to gather and analysis on the tool in order to recommend users of different operating systems on what type of medical image tools to be used when analysing different types of imaging.
    [Show full text]