An Overview of Data Science Uses in Bioimage Informatics

Total Page:16

File Type:pdf, Size:1020Kb

An Overview of Data Science Uses in Bioimage Informatics Methods 115 (2017) 110–118 Contents lists available at ScienceDirect Methods journal homepage: www.elsevier.com/locate/ymeth An Overview of data science uses in bioimage informatics Anatole Chessel LOB, Ecole Polytechnique, CNRS, INSERM, Université Paris-Saclay, 91128 Palaiseau cedex, France article info abstract Article history: This review aims at providing a practical overview of the use of statistical features and associated data Received 31 August 2016 science methods in bioimage informatics. To achieve a quantitative link between images and biological Received in revised form 9 December 2016 concepts, one typically replaces an object coming from an image (a segmented cell or intracellular object, Accepted 30 December 2016 a pattern of expression or localisation, even a whole image) by a vector of numbers. They range from Available online 3 January 2017 carefully crafted biologically relevant measurements to features learnt through deep neural networks. This replacement allows for the use of practical algorithms for visualisation, comparison and inference, Keywords: such as the ones from machine learning or multivariate statistics. While originating mainly, for biology, Data science in high content screening, those methods are integral to the use of data science for the quantitative Bioimage informatics High content screening analysis of microscopy images to gain biological insight, and they are sure to gather more interest as Image analysis the need to make sense of the increasing amount of acquired imaging data grows more pressing. Ó 2017 Elsevier Inc. All rights reserved. Contents 1. Introduction ......................................................................................................... 110 2. Numerical features computation......................................................................................... 111 2.1. Features computation strategy . ............................................................................ 112 2.2. Feature quality control and post-processing . ............................................................ 112 3. Storage, management and sharing of image-derived data. ................................................... 112 4. Features comparison . ......................................................................................... 113 5. Features analysis and interpretation. ...................................................................... 114 6. Implementations . ......................................................................................... 115 7. Conclusion . ......................................................................................................... 115 References . ......................................................................................................... 115 1. Introduction Numerous previous recent reviews, from the more specific to the more generic, have been documenting that rise and the meth- For the last couple of decades the development of increasingly ods developed and used. [1] is a fairly wide review of applications efficient fluorescence probes along with technological advances of classical computer vision techniques to biology, while [2] pro- in microscopy has led to terabytes of increasingly resolved images vide a more recent and in depth overview of bioimage informatics. being acquired across models and conditions. The discipline of [3,4] are older reviews focusing on High Content Screening (HCS); bioimage informatics (BII) is rising to develop the means to provide more recently, [5,6] are comprehensive reviews on HCS while [7] a quantitative analysis of those data, and integrate them into larger focuses on some example of phenotype analysis. [8] is an earlier biological questions and studies, inline with a more general trend review on the uses of machine learning on biological imaging of the life sciences toward more quantitative and integrative which focuses more specifically on supervised and unsupervised approaches. learning. [9] provide a more methodological review and compar- ison of segmentation methods. Here we will more specifically focus on the next steps after E-mail address: [email protected] image segmentation, i.e. the computation and use of statistical http://dx.doi.org/10.1016/j.ymeth.2016.12.014 1046-2023/Ó 2017 Elsevier Inc. All rights reserved. A. Chessel / Methods 115 (2017) 110–118 111 features, at the intersection with wider data science techniques. and technical but interested readers will be able to find more in Answering biological questions involve performing comparison depth tutorial and courses online; see also Section 6 for pointers. and inference on complex objects, be it cell shape for morphology, The plan will follow the chronology of projects using those vesicles distributions, actin or microtubule structures, etc...whose methods, beginning by how the features themselves are computed handling is not trivial: how to compute the average of a set of cell (Section 2), and handled for storage and visualisation (Section 3). shape? Or compare the intracellular organisation of segmented We will then focus specifically on comparison in Section 4 as gen- microtubules within cells? A classical path to an answer to those eric multivariate two-sample tests are an important but tricky questions is to associate n real numbers to each of them, the so- topic and look into wider statistical and machine learning tech- called features (i.e. equivalently send those objects to Rn), and niques for inference and interpretation in Section 5. We will finish use the wealth of algorithms and methods available to handle such by a few word about the practical implementations of all those numerical data. algorithms in Section 6. A lot of the examples will come from high Analysing images and image derived data in biology has a content screening (HCS) studies for historical reasons, but are now long history [10], and has grown in scope and visibility in the last increasingly of use across bioimage informatics. From Section 3 thirty years, concomitant with the increased importance of digital onward, we leave the realm of images per se and will be looking microscopy in the life sciences and the rise of quantitative biol- at rather generic biological data science questions of wider ogy. High throughput and high content microscopy, given the relevance. scale at which they do experiments, spanning 100s to 1000s of conditions or more, has always been key in moving those 2. Numerical features computation developments forward. They include full genome screens [11–13], systematic investigation of genetic interactions [14,15], As said, the aim of the computation of numerical features is to or more focused investigation of the influence of cell context send the objects of interests to Rn, i.e. associate n real number to [16] or of cell motility[17], the variability of cell shape [18],of each of them. Let us note first it’s only one solution (the most com- iPSC cell lines [19] or the investigation of small molecules effects monly used and usually the most practical), among others. For for drug developements [20–22]. example one could try and work directly in object space, by com- But similar techniques have been used outside of HCS, to anal- puting an innate distance between objects, and avoid numerical yse time lapse data [23,24], In-Situ Hybridation experiements features entirely. Examples include [33], where a representation [25,26], cell lineages [27], perform content based image retreival of drosophila embryo is built with which computation are done, [28] or build models of protein localisation [29]. Recent advance or [34] which compute kernels on graphs. Those approaches were in microscopy, leading to the acquisition of very large images particularly pursued for shape analysis, using for examples shape [30,27,31], where a single image can weight 100 Gb to several diffeomorphisms to compute geodesics in shape space [35–37] Tb, will also benefit from such large scale data analysis. but will not be presented further here as they tend to be more All those apparently very diverse applications have at their core complex and more closely tied to particular type of data or analysis the computation of numerical features on their objects of interests, frameworks. to be used as representation. The main purpose of feature representations is to quantitatively describe complex objects and concepts, essential for their further quantitative analysis. Some data science framework ‘Data science’ is a recent Conversely, one will only be able to access the aspects of an objects label that aggregates various fields and practices that aim at analysing actual data. Classical framework includes: that are described by the representation used, hence the choice of Statistics. From a statistical point of view, features are con- the right features, and the right way of visualising and comparing sidered random variables, i.e. variable whose values follow a them corresponds to hypothesis on the data at hand and their specific distribution, with the feature set being a multivariate variabilities. random variable, and each object under study to which a fea- The precise nature of the study will of course determine the ture set is associated (a individual cell, a track, a gene) can be objects for which features are computed, and how. Computing fea- seen as a particular realisation of that random variable. Sta- tures for whole images or systematic tiles of images has been done. tistical methods can be parametric, if they assume random More
Recommended publications
  • Bioimage Analysis Tools
    Bioimage Analysis Tools Kota Miura, Sébastien Tosi, Christoph Möhl, Chong Zhang, Perrine Paul-Gilloteaux, Ulrike Schulze, Simon Norrelykke, Christian Tischer, Thomas Pengo To cite this version: Kota Miura, Sébastien Tosi, Christoph Möhl, Chong Zhang, Perrine Paul-Gilloteaux, et al.. Bioimage Analysis Tools. Kota Miura. Bioimage Data Analysis, Wiley-VCH, 2016, 978-3-527-80092-6. hal- 02910986 HAL Id: hal-02910986 https://hal.archives-ouvertes.fr/hal-02910986 Submitted on 3 Aug 2020 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. 2 Bioimage Analysis Tools 1 2 3 4 5 6 Kota Miura, Sébastien Tosi, Christoph Möhl, Chong Zhang, Perrine Pau/-Gilloteaux, - Ulrike Schulze,7 Simon F. Nerrelykke,8 Christian Tischer,9 and Thomas Penqo'" 1 European Molecular Biology Laboratory, Meyerhofstraße 1, 69117 Heidelberg, Germany National Institute of Basic Biology, Okazaki, 444-8585, Japan 2/nstitute for Research in Biomedicine ORB Barcelona), Advanced Digital Microscopy, Parc Científic de Barcelona, dBaldiri Reixac 1 O, 08028 Barcelona, Spain 3German Center of Neurodegenerative
    [Show full text]
  • Other Departments and Institutes Courses 1
    Other Departments and Institutes Courses 1 Other Departments and Institutes Courses About Course Numbers: 02-250 Introduction to Computational Biology Each Carnegie Mellon course number begins with a two-digit prefix that Spring: 12 units designates the department offering the course (i.e., 76-xxx courses are This class provides a general introduction to computational tools for biology. offered by the Department of English). Although each department maintains The course is divided into two halves. The first half covers computational its own course numbering practices, typically, the first digit after the prefix molecular biology and genomics. It examines important sources of biological indicates the class level: xx-1xx courses are freshmen-level, xx-2xx courses data, how they are archived and made available to researchers, and what are sophomore level, etc. Depending on the department, xx-6xx courses computational tools are available to use them effectively in research. may be either undergraduate senior-level or graduate-level, and xx-7xx In the process, it covers basic concepts in statistics, mathematics, and courses and higher are graduate-level. Consult the Schedule of Classes computer science needed to effectively use these resources and understand (https://enr-apps.as.cmu.edu/open/SOC/SOCServlet/) each semester for their results. Specific topics covered include sequence data, searching course offerings and for any necessary pre-requisites or co-requisites. and alignment, structural data, genome sequencing, genome analysis, genetic variation, gene and protein expression, and biological networks and pathways. The second half covers computational cell biology, including biological modeling and image analysis. It includes homework requiring Computational Biology Courses modification of scripts to perform computational analyses.
    [Show full text]
  • Bioimage Informatics
    Bioimage Informatics Lecture 22, Spring 2012 Empirical Performance Evaluation of Bioimage Informatics Algorithms Image Database; Content Based Image Retrieval Emerging Applications: Molecular Imaging Lecture 22 April 23, 2012 1 Outline • Empirical performance evaluation of bioimage Informatics algorithm • Image database; content-based image retrieval • Introduction to molecular imaging 2 • Empirical performance evaluation of bioimage Informatics algorithm • Image database; content-based image retrieval • Introduction to molecular imaging 3 Overview of Algorithm Evaluation (I) • Two sets of data are required - Input data - Corresponding correct output (ground truth) • Sources of input data - Actual/experimental data, often from experiments - Synthetic data, often from computer simulation • Actual/experimental data - Essential for performance evaluation - May be costly to acquire - May not be representative - Ground truth often is unknown Overview of Algorithm Evaluation (II) • What exactly is ground truth? Ground truth is a term used in cartography, meteorology, analysis of aerial photographs, satellite imagery and a range of other remote sensing techniques in which data are gathered at a distance. Ground truth refers to information that is collected "on location." - http://en.wikipedia.org/wiki/Ground_truth • Synthetic (simulated) data Advantages - Ground truth is known - Usually low-cost Disadvantages - Difficult to fully represent the original data 5 Overview of Algorithm Evaluation (III) • Simulated data Advantages - Ground truth is known - Usually low-cost Disadvantages - Difficult to fully represent the original data • Realistic synthetic data - To use as much information from real experimental data as possible • Quality control of manual analysis Test Protocol Development (I) • Implementation - Source codes are not always available and often are on different platforms. • Parameter setting - This is one of the most challenging issues.
    [Show full text]
  • Bioimage Informatics 2019
    BioImage Informatics 2019 Co-hosted by: October 2-4, 2019 Allen Institute 615 Westlake Avenue N Seattle, WA 98109 alleninstitute.org/bioimage-informatics twitter #bioimage19 THANKS TO OUR SPONSORS BioImage Informatics 2019 Wednesday, October 2 8:00-9:00am Registration and continental breakfast 9:00-9:15am Welcome and introductory remarks, BioImage Informatics Local Planning Committee • Michael Hawrylycz, Ph.D., Allen Institute for Brain Science (co-chair) • Winfried Wiegraebe, Ph.D., Allen Institute for Cell Science (co-chair) • Forrest Collman, Ph.D., Allen Institute for Brain Science • Greg Johnson, Ph.D., Allen Institute for Cell Science • Lydia Ng, Ph.D., Allen Institute for Brain Science • Kaitlyn Casimo, Ph.D., Allen Institute 9:15-10:15am Opening keynote Gaudenz Danuser, UT Southwestern Medical Center “Inference of causality in molecular pathways from live cell images” Defining cause-effect relations among molecular components is the holy grail of every cell biological study. It turns out that for many of the molecular pathways this task is far from trivial. Many of the pathways are characterized by functional overlap and nonlinear interactions between components. In this configuration, perturbation of one component may result in a measureable shift of the pathway output – commonly referred to as a phenotype – but it is strictly impossible to interpret the phenotype in terms of the roles the targeted component plays in the unperturbed system. This caveat of perturbation approaches applies equally to genetic perturbations, which lead to long-term adaptation, and more acute pharmacological and optogenetic approaches, which often induce short-term adaptation. For nearly two decades my lab has been devoted to circumventing this problem by exploiting basal fluctuations of unperturbed systems to establish cause-and-effect cascades between pathway components.
    [Show full text]
  • A Bioimage Informatics Platform for High-Throughput Embryo Phenotyping James M
    Briefings in Bioinformatics, 19(1), 2018, 41–51 doi: 10.1093/bib/bbw101 Advance Access Publication Date: 14 October 2016 Paper A bioimage informatics platform for high-throughput embryo phenotyping James M. Brown, Neil R. Horner, Thomas N. Lawson, Tanja Fiegel, Simon Greenaway, Hugh Morgan, Natalie Ring, Luis Santos, Duncan Sneddon, Lydia Teboul, Jennifer Vibert, Gagarine Yaikhom, Henrik Westerberg and Ann-Marie Mallon Corresponding author: James Brown, MRC Harwell Institute, Harwell Campus, Oxfordshire, OX11 0RD. Tel. þ44-0-1235-841237; Fax: +44-0-1235-841172; E-mail: [email protected] Abstract High-throughput phenotyping is a cornerstone of numerous functional genomics projects. In recent years, imaging screens have become increasingly important in understanding gene–phenotype relationships in studies of cells, tissues and whole organisms. Three-dimensional (3D) imaging has risen to prominence in the field of developmental biology for its ability to capture whole embryo morphology and gene expression, as exemplified by the International Mouse Phenotyping Consortium (IMPC). Large volumes of image data are being acquired by multiple institutions around the world that encom- pass a range of modalities, proprietary software and metadata. To facilitate robust downstream analysis, images and metadata must be standardized to account for these differences. As an open scientific enterprise, making the data readily accessible is essential so that members of biomedical and clinical research communities can study the images for them- selves without the need for highly specialized software or technical expertise. In this article, we present a platform of soft- ware tools that facilitate the upload, analysis and dissemination of 3D images for the IMPC.
    [Show full text]
  • View of Above Observations and Arguments, the Fol- Scopy Environment) [9]
    Loyek et al. BMC Bioinformatics 2011, 12:297 http://www.biomedcentral.com/1471-2105/12/297 SOFTWARE Open Access BioIMAX: A Web 2.0 approach for easy exploratory and collaborative access to multivariate bioimage data Christian Loyek1*, Nasir M Rajpoot2, Michael Khan3 and Tim W Nattkemper1 Abstract Background: Innovations in biological and biomedical imaging produce complex high-content and multivariate image data. For decision-making and generation of hypotheses, scientists need novel information technology tools that enable them to visually explore and analyze the data and to discuss and communicate results or findings with collaborating experts from various places. Results: In this paper, we present a novel Web2.0 approach, BioIMAX, for the collaborative exploration and analysis of multivariate image data by combining the webs collaboration and distribution architecture with the interface interactivity and computation power of desktop applications, recently called rich internet application. Conclusions: BioIMAX allows scientists to discuss and share data or results with collaborating experts and to visualize, annotate, and explore multivariate image data within one web-based platform from any location via a standard web browser requiring only a username and a password. BioIMAX can be accessed at http://ani.cebitec. uni-bielefeld.de/BioIMAX with the username “test” and the password “test1” for testing purposes. Background key challenge of bioimage informatics is to present the In the last two decades, innovative biological and biome- large amount of image data in a way that enables them dical imaging technologies greatly improved our under- to be browsed, analyzed, queried or compared with standing of structures of molecules, cells and tissue and other resources [2].
    [Show full text]
  • Principles of Bioimage Informatics: Focus on Machine Learning of Cell Patterns
    Principles of Bioimage Informatics: Focus on Machine Learning of Cell Patterns Luis Pedro Coelho1,2,3, Estelle Glory-Afshar3,6, Joshua Kangas1,2,3, Shannon Quinn2,3,4, Aabid Shariff1,2,3, and Robert F. Murphy1,2,3,4,5,6 1 Joint Carnegie Mellon University–University of Pittsburgh Ph.D. Program in Computational Biology 2 Lane Center for Computational Biology, Carnegie Mellon University 3 Center for Bioimage Informatics, Carnegie Mellon University 4 Department of Biological Sciences, Carnegie Mellon University 5 Machine Learning Department, Carnegie Mellon University 6 Department of Biomedical Engineering, Carnegie Mellon University Abstract. The field of bioimage informatics concerns the development and use of methods for computational analysis of biological images. Tra- ditionally, analysis of such images has been done manually. Manual an- notation is, however, slow, expensive, and often highly variable from one expert to another. Furthermore, with modern automated microscopes, hundreds to thousands of images can be collected per hour, making man- ual analysis infeasible. This field borrows from the pattern recognition and computer vi- sion literature (which contain many techniques for image processing and recognition), but has its own unique challenges and tradeoffs. Fluorescence microscopy images represent perhaps the largest class of biological images for which automation is needed. For this modality, typical problems include cell segmentation, classification of phenotypical response, or decisions regarding differentiated responses (treatment vs. control setting). This overview focuses on the problem of subcellular lo- cation determination as a running example, but the techniques discussed are often applicable to other problems. 1 Introduction Bioimage informatics employs computational and statistical techniques to ana- lyze images and related metadata.
    [Show full text]
  • Introduction to Bio(Medical)- Informatics
    Introduction to Bio(medical)- Informatics Kun Huang, PhD Raghu Machiraju, PhD Department of Biomedical Informatics Department of Computer Science and Engineering The Ohio State University Outline • Overview of bioinformatics • Other people’s (public) data • Basic biology review • Basic bioinformatics techniques • Logistics 2 Technical areas of bioinformatics IMMPORT https://immport.niaid.nih.gov/immportWeb/home/home.do?loginType=full# Databases and Query Databases and Query 547 5 Visualization 6 Visualization http://supramap.osu.edu 7 Network 8 Visualization and Visual Analytics § Visualizing genome data annotation 9 Network Mining Hopkins 2007, Nature Biotechnology, Network pharmacology 10 Role of Bioinformatics – Hypothesis Generation Hypothesis generation Bioinformatics Hypothesis testing 11 Training for Bioinformatics § Biology / medicine § Computing – programming, algorithm, pattern recognition, machine learning, data mining, network analysis, visualization, software engineering, database, etc § Statistics – hypothesis testing, permutation method, multivariate analysis, Bayesian method, etc Beyond Bioinformatics • Translational Bioinformatics • Biomedical Informatics • Medical Informatics • Nursing Informatics • Imaging Informatics • Bioimage Informatics • Clinical Research Informatics • Pathology Informatics • … • Health Analytics • … • Legal Informatics • Business Informatics Source: Department of Biomedical Informatics 13 Beyond Bioinformatics Basic Science Clinical Research Informatics Biomedical Informatics Theories & Methods
    [Show full text]
  • Bioimagexd: an Open, General-Purpose and High-Throughput Image-Processing Platform
    FOCUS ON BIOIMAGE INFORMATICS PERSPECTIVE BioImageXD: an open, general-purpose and high-throughput image-processing platform Pasi Kankaanpää1, Lassi Paavolainen2, Silja Tiitta1, Mikko Karjalainen2, Joacim Päivärinne1, Jonna Nieminen1, Varpu Marjomäki2, Jyrki Heino1 & Daniel J White2,3 BioImageXD puts open-source computer science BioImageXD in practice with user-friendly, auto- tools for three-dimensional visualization and mated, multiparametric image analyses of the move- analysis into the hands of all researchers, through a ment, clustering and internalization of integrins. user-friendly graphical interface tuned to the needs of biologists. BioImageXD has no restrictive licenses Criteria in BioImageXD development or undisclosed algorithms and enables publication Open. Our goal was to develop software that is trans- of precise, reproducible and modifiable workflows. parent, with open and documented source code and It allows simple construction of processing pipelines no undisclosed algorithms or hidden processing the and should enable biologists to perform challenging user is unaware of or cannot control. It is our view that analyses of complex processes. We demonstrate its software for scientific data analysis, in contrast to engi- performance in a study of integrin clustering in neering, is best when it is not a ‘black box’. Different response to selected inhibitors. implementations of the same method can often give dif- Bioimaging has emerged as one of the key tools in bio- ferent results, the reasons for which only become clear medical research. New imaging devices are appearing by comparing the source codes. Increasingly (and cor- rapidly, producing large multidimensional images with rectly in our view), disclosure of source code for image- several data channels in medium-to-high throughput.
    [Show full text]
  • Bioimage Informatics for Plant Sciences Blobs and Curves: Object-Based Colocalisation for Plant Cells
    ORE Open Research Exeter TITLE Blobs and curves: object-based colocalisation for plant cells AUTHORS Nelson, CJ; Duckney, P; Hawkins, TJ; et al. JOURNAL Functional Plant Biology DEPOSITED IN ORE 10 March 2017 This version available at http://hdl.handle.net/10871/26430 COPYRIGHT AND REUSE Open Research Exeter makes this work available in accordance with publisher policies. A NOTE ON VERSIONS The version presented here may differ from the published version. If citing, you are advised to consult the published version for pagination, volume/issue and date of publication Bioimage Informatics For Plant Sciences Blobs and Curves: Object-Based Colocalisation for Plant Cells Carl J. Nelson (A), Patrick Duckney, (B), Timothy J. Hawkins (B), Michael J. Deeks (C), Philippe Laissue (D), Patrick J. Hussey (B) and Boguslaw Obara (A)(*) (A) School of Engineering and Computing Sciences, Durham University, Durham, UK (B) School of Biological and Biomedical Sciences, Durham University, Durham, UK (C) College of Life and Environmental Sciences, University of Exeter, Exeter, UK (D) School of Biological Sciences, University of Essex, Colchester, UK (*) Corresponding author (email: [email protected]) Abstract: Quantifying the colocalisation of labels is a major application of fluorescent microscopy in plant biology. Pixel-based, i.e. Pearson's coefficient, quantification of colocalisation is a limited tool that gives limited information for further analysis. We show how applying bioimage informatics tools to a commonplace experiment allows further quantifiable results to be extracted. We use our object-based colocalisation technique to extract distance information, show temporal changes and demonstrate the advantages, and pitfalls, of using bioimage informatics for plant science.
    [Show full text]
  • A Short Course on Bioimage Informatics Lecture 1: Basic Concepts and Techniques
    A Short Course on Bioimage Informatics Lecture 1: Basic Concepts and Techniques Ge Yang (杨戈) Department of Biomedical Engineering & Department of Computational Biology Carnegie Mellon University August 10, 2015 1 Acknowledgements • Dr. Chao Tang • Dr. Luhua Lai • Students, faculty, and staff of the Center for Quantitative Biology, Peking University • Dr. Jing Yuan (袁静) • Youjun Xu (徐优俊) • 111 Visiting Scholar Program, Ministry of Education • College of Chemistry and Molecular Engineering, Peking University • School of Engineering & School of Computer Science, Carnegie Mellon University 2 Outline • Background • What is bioimage informatics? • Where does it come from? • Where is it going? • Basic concepts of image formation and collection • Overview of bioimage informatics techniques • Summary 3 • Background • What is bioimage informatics? • Where does it come from? • Where is it going? • Basic concepts of image formation and collection • Overview of biological image analysis techniques • Summary 4 What information can we extract from images? 刘奇 5 More Examples of Biological Images Typical Readouts from Images • From static images, e.g. immunofluorescence images - Location - Shape/morphology - Spatial distribution: e.g. spatial patterns - Spatial relation: e.g. protein colocalization • From dynamic images, e.g. live cell time-lapse images - All the readouts above, which now may be time-dependent - In other words, spatiotemporal dynamics/processes • Computational analysis techniques are required to extract these quantitative readouts from images. 7 Relations of Imaging and Image Analysis with Other Biological Research Tools • None of these readouts can be directly provided by e.g. genetics or biochemistry. • However, understanding these readouts often requires close integration of computational image analysis with experimental analysis using genetics and/or biochemistry techniques.
    [Show full text]
  • Phenotypic Image Analysis Software Tools for Exploring and Understanding Big Image Data from Cell-Based Assays
    Cell Systems Review Phenotypic Image Analysis Software Tools for Exploring and Understanding Big Image Data from Cell-Based Assays Kevin Smith,1,2 Filippo Piccinini,3 Tamas Balassa,4 Krisztian Koos,4 Tivadar Danka,4 Hossein Azizpour,1,2 and Peter Horvath4,5,* 1KTH Royal Institute of Technology, School of Electrical Engineering and Computer Science, Lindstedtsvagen€ 3, 10044 Stockholm, Sweden 2Science for Life Laboratory, Tomtebodavagen€ 23A, 17165 Solna, Sweden 3Istituto Scientifico Romagnolo per lo Studio e la Cura dei Tumori (IRST) IRCCS, Via P. Maroncelli 40, Meldola, FC 47014, Italy 4Synthetic and Systems Biology Unit, Hungarian Academy of Sciences, Biological Research Center (BRC), Temesva´ ri krt. 62, 6726 Szeged, Hungary 5Institute for Molecular Medicine Finland, University of Helsinki, Tukholmankatu 8, 00014 Helsinki, Finland *Correspondence: [email protected] https://doi.org/10.1016/j.cels.2018.06.001 Phenotypic image analysis is the task of recognizing variations in cell properties using microscopic image data. These variations, produced through a complex web of interactions between genes and the environ- ment, may hold the key to uncover important biological phenomena or to understand the response to a drug candidate. Today, phenotypic analysis is rarely performed completely by hand. The abundance of high-dimensional image data produced by modern high-throughput microscopes necessitates computa- tional solutions. Over the past decade, a number of software tools have been developed to address this need. They use statistical learning methods to infer relationships between a cell’s phenotype and data from the image. In this review, we examine the strengths and weaknesses of non-commercial phenotypic image analysis software, cover recent developments in the field, identify challenges, and give a perspective on future possibilities.
    [Show full text]