Physical Mapping by STS Hybridization: Algorithmic

Total Page:16

File Type:pdf, Size:1020Kb

Physical Mapping by STS Hybridization: Algorithmic Physical Mapping by STS Hybridization Algorithmic Strategies and the Challenge of Software Evaluation David S Greenb erg Sorin Istrail Sandia National Lab oratories Sandia National Lab oratories Mail Stop Mail Stop PO Box PO Box Albuquerque NM Albuquerque NM dsgreencssandiagov scistracssandiagov March To app ear in the Journal of Computational Biology summer Abstract An imp ortant to ol in the analysis of genomic sequences is the physical map In this pap er we examine the construction of physical maps from hybridization data b etween STS sequence tag sites prob es and clones of genomic fragments An algorithmic theory of the mapping pro cess a prop osed p erformance evaluation pro cedure and several new algorithmic strategies for mapping are given A unifying theme for these developments is the idea of a conservative extension An algorithm measure of algorithm quality or description of physical map is a conservative extension if it is a generalization for data with errors of a corresp onding concept in the errorfree case In our algorithmic theory we show that the nature of hybridization exp eriments imp oses inherent limitations on the mapping information recorded in the exp erimental data We prove that only certain typ es of mapping information can b e reliably cal culated by any algorithm A test generator is then presented along with quantitative measures for determining how much of the p ossible information is b eing computed by a given algorithm Weaknesses and strengths of these measures are discussed Each of the new algorithms presented in this pap er is based on combinatorial op timizations Despite the fact that all the optimizations are NPcomplete we have de velop ed algorithmic to ols for the design of comp etitive approximation algorithms We apply our p erformance evaluation program to our algorithms and obtain solid evidence that the algorithms are capable of retrieving highlevel reliable mapping information This work supp orted in part by the US Department of Energy under contract DEACDP Intro duction A central question for the Human Genome Program is how to bridge the gap b etween the size of DNA fragments which can b e directly sequenced and the size of the human genome Researchers continue to lo ok for ways of extending the size of fragments which can b e directly sequenced and ways of picking out imp ortant pieces to sequence However at present and in the near term large scale sequencing pro jects dep end on some pro cess which involves dividing a large segment of DNA into overlapping pieces analyzing the smaller pieces separately and determining the order of the pieces so as to combine information ab out the pieces into information ab out the whole It is this last step reordering fragments which is the fo cus of this pap er Many metho ds have b een prop osed for reordering fragments ranging from to ols for helping exp erts reorder the fragments by hand to automatic programs employing maximum likeliho o d analysis combinatorial optimizations or a variety of heuristics These to ols have b een aimed at data which includes STS prob eclone interactions FISH data radiation hybridization data genetic map orderings break p oint p ositions etc Typically a given to ol or algorithm is evaluated by applying it to actual data and displaying the resulting map In this pap er we attempt to provide a framework for a more rigorous analysis of mapping algorithms We aim to answer the following questions For a given typ e of data what typ e of information is reasonable to exp ect from an algorithm For a given algorithm how well do es it work on dierent typ es of data Are there go o d algorithms for particular typ es of data and if so what are they Since these questions are clearly broad and dicult to answer we have b egun by concen trating on a particular typ e of data hybridizations of sequencetagsite STS prob es with clones of genomic fragments and a particular class of algorithms those based on combina torial optimizations We pay esp ecial attention to data in which there are errors of various typ es clones which contain multiple unrelated fragments ie chimera clones with internal deletions and b oth falsep ositive and falsenegative errors in the hybridization data In this rst study we were able to show The information available in a hybridization matrix of even error free exp eriments is limited The amount of available information is directly related to the information in the PQtree of the matrix The presence of errors further degrades the information available Furthermore the errors may obscure redundancies and singularities which could easily have b een ltered in the errorfree case Despite these limitations on information available in hybridization matrices it is p ossi ble even data containing errors to determine relevant facts ab out the target genome 1 Some terms here such as PQtree may b e unfamiliar to readers new to the sub ject They are dened in later sections In particular we identify lo cal prop erties which can b e determined such as the fact that certain prob e pairs are adjacent Given the correct ordering for a matrix it is p ossible to divide the adjacent prob e pairs into a weak and a strong set where the weak set cannot b e reliably determined by any algorithm and the strong set is p otentially determinable The denitions of these sets is formal and precise for errorfree data and heuristic for data containing errors The p ercentage of strong adjacent pairs identied by an algorithm is a reasonable measure of algorithm success and can b e used to compare algorithms However this measure is lo cal in nature and more global measures are still needed We formulate a noise mo del which allows us to unify the seemingly dissimilar chal lenges of nding the correct order of prob es and of dealing with the errors in the hybridization matrix We show that combinatorial optimization functions can b e used to search for solutions which require the p ostulation of as few errors as p ossible We b elieve that the success of these functions is due to their b eing conservative ex tensions of the errorfree case and their having monotonicity with resp ect to common error typ es Figure gives a table of contents for this pap er The remainder of the pap er is divided up as follows Section presents the basic biology required for the rest of the pap er Section rigorously denes the physical mapping problem and shows that the information ab out a genomic target which can b e reliably inferred from an errorfree hybridization matrix is related to the PQtree of the matrix In Section we discuss p ossible sources of errors in the hybridization matrix and extend our theory to include the errors In Sections and we discuss how to generate synthetic data for testing mapping algorithms and show exp erimentally that hybridization matrices will tend to have ambiguities which algorithms cannot resolve In Sections and we describ e how combinatorial optimizations can b e used to search for go o d maps and present several algorithms based on the optimizations In Sections and we describ e a pilot exp eriment in which we examine the p erformance of our algorithms on varying amounts of errors In Sections and we contrast our work with other studies in the eld and describ e the large amount of work which still remains to b e done A biology primer Readers who are already familiar with the pro cess of physical mapping can skip this section Readers who desire more details than present in this section are encouraged to read Browns or Nelson and Brownsteins b o oks The essence of the physical mapping pro cess is as follows The exp eriment b egins with a sample of target DNA recall that DNA is a linear sequence of basepairs ACG and T Pure samples of the target DNA are cut at sp ecic p oints and then each fragment of DNA is inserted into a circular DNA molecule called a vector to pro duce a recombinant DNA molecule The DNA fragment incorp orated into the vector is called an insert The vector Contents Intro duction A biology primer Exp erimental errors A theory of physical mapping Errorfree hybridization exp eriments Mapping Information Algorithmically retrievable mapping information Prob e restrictions Consecutive ones matrices Bo oth and Luekers PQtree algorithm The true map is almost never algorithmically retrievable PQtrees enco de information in common to all maps Exp erimental errors Extending the mo del to data with errors NPcompleteness barriers Reasonable goals Evaluating Algorithms on Hybridization Data Generator Mimicking real data Exploring a range of data The Quality of Data Connected Comp onents Redundant prob es Weak adjacencies Summary of data quality Combinatorial Optimizations as Mo
Recommended publications
  • Chapter 1 Preliminaries
    Chapter 1 Preliminaries 1.1 First-order theories In this section we recall some basic definitions and facts about first-order theories. Most proofs are postponed to the end of the chapter, as exercises for the reader. 1.1.1 Definitions Definition 1.1 (First-order theory) —A first-order theory T is defined by: A first-order language L , whose terms (notation: t, u, etc.) and formulas (notation: φ, • ψ, etc.) are constructed from a fixed set of function symbols (notation: f , g, h, etc.) and of predicate symbols (notation: P, Q, R, etc.) using the following grammar: Terms t ::= x f (t1,..., tk) | Formulas φ, ψ ::= t1 = t2 P(t1,..., tk) φ | | > | ⊥ | ¬ φ ψ φ ψ φ ψ x φ x φ | ⇒ | ∧ | ∨ | ∀ | ∃ (assuming that each use of a function symbol f or of a predicate symbol P matches its arity). As usual, we call a constant symbol any function symbol of arity 0. A set of closed formulas of the language L , written Ax(T ), whose elements are called • the axioms of the theory T . Given a closed formula φ of the language of T , we say that φ is derivable in T —or that φ is a theorem of T —and write T φ when there are finitely many axioms φ , . , φ Ax(T ) such ` 1 n ∈ that the sequent φ , . , φ φ (or the formula φ φ φ) is derivable in a deduction 1 n ` 1 ∧ · · · ∧ n ⇒ system for classical logic. The set of all theorems of T is written Th(T ). Conventions 1.2 (1) In this course, we only work with first-order theories with equality.
    [Show full text]
  • Automated ZFC Theorem Proving with E
    Automated ZFC Theorem Proving with E John Hester Department of Mathematics, University of Florida, Gainesville, Florida, USA https://people.clas.ufl.edu/hesterj/ [email protected] Abstract. I introduce an approach for automated reasoning in first or- der set theories that are not finitely axiomatizable, such as ZFC, and describe its implementation alongside the automated theorem proving software E. I then compare the results of proof search in the class based set theory NBG with those of ZFC. Keywords: ZFC · ATP · E. 1 Introduction Historically, automated reasoning in first order set theories has faced a fun- damental problem in the axiomatizations. Some theories such as ZFC widely considered as candidates for the foundations of mathematics are not finitely axiomatizable. Axiom schemas such as the schema of comprehension and the schema of replacement in ZFC are infinite, and so cannot be entirely incorpo- rated in to the prover at the beginning of a proof search. Indeed, there is no finite axiomatization of ZFC [1]. As a result, when reasoning about sufficiently strong set theories that could no longer be considered naive, some have taken the alternative approach of using extensions of ZFC that admit objects such as proper classes as first order objects, but this is not without its problems for the individual interested in proving set-theoretic propositions. As an alternative, I have programmed an extension to the automated the- orem prover E [2] that generates instances of parameter free replacement and comprehension from well formuled formulas of ZFC that are passed to it, when eligible, and adds them to the proof state while the prover is running.
    [Show full text]
  • (Aka “Geometric”) Iff It Is Axiomatised by “Coherent Implications”
    GEOMETRISATION OF FIRST-ORDER LOGIC ROY DYCKHOFF AND SARA NEGRI Abstract. That every first-order theory has a coherent conservative extension is regarded by some as obvious, even trivial, and by others as not at all obvious, but instead remarkable and valuable; the result is in any case neither sufficiently well-known nor easily found in the literature. Various approaches to the result are presented and discussed in detail, including one inspired by a problem in the proof theory of intermediate logics that led us to the proof of the present paper. It can be seen as a modification of Skolem's argument from 1920 for his \Normal Form" theorem. \Geometric" being the infinitary version of \coherent", it is further shown that every infinitary first-order theory, suitably restricted, has a geometric conservative extension, hence the title. The results are applied to simplify methods used in reasoning in and about modal and intermediate logics. We include also a new algorithm to generate special coherent implications from an axiom, designed to preserve the structure of formulae with relatively little use of normal forms. x1. Introduction. A theory is \coherent" (aka \geometric") iff it is axiomatised by \coherent implications", i.e. formulae of a certain simple syntactic form (given in Defini- tion 2.4). That every first-order theory has a coherent conservative extension is regarded by some as obvious (and trivial) and by others (including ourselves) as non- obvious, remarkable and valuable; it is neither well-enough known nor easily found in the literature. We came upon the result (and our first proof) while clarifying an argument from our paper [17].
    [Show full text]
  • 1.4 Theories
    1.4 Theories Before we start to work with one of the most central notions of mathematical logic – that of a theory – it will be useful to make some observations about countable sets. Lemma 1.3. 1. A set X is countable iff there is a surjection ψ : N X. 2. If X is countable and Y X then Y is countable. 3. If X and Y are countable, then X Y is countable. 4. If X and Y are countable, then X ¢ Y is countable. k 5. If X is countable and k 1 then X is countable. ä 6. If Xi is countable for each i È N, then Xi is countable. iÈN 7. If X is countable, then ä Xk is countable. k ÈN Ø Ù Proof. 1. If X is countably infinite note that every bijection is a surjection. If X x0,...,xn Õ Ô Õ is finite, let ψ Ôi xi for 0 i n and ψ j xn for j n. For the other direction, Ô Õ Ô Õ Ô Õ Ù let ψ be a surjection, then X is Øψ 0 , ψ 1 , ψ 2 ,... and removing duplicates from the Ô ÕÕ Ô N sequence ψ i i¥0 yields a bijective ϕ : X. È 2. Obvious if Y is finite. Otherwise, let ϕ : N X be a bijection and let y0 Y . Define " Õ Ô Õ È ϕÔn if ϕ n Y ψ : N Y,n y0 otherwise which is a surjection. 3. Let ϕ : N X and ψ : N Y be bijections.
    [Show full text]
  • Lipics-FSCD-2021-29.Pdf (0.9
    On the Logical Strength of Confluence and Normalisation for Cyclic Proofs Anupam Das ! Ï University of Birmingham, UK Abstract In this work we address the logical strength of confluence and normalisation for non-wellfounded typing derivations, in the tradition of “cyclic proof theory” . We present a circular version CT of Gödel’ s system T, with the aim of comparing the relative expressivity of the theories CT and T. We approach this problem by formalising rewriting-theoretic results such as confluence and normalisation for the underlying “coterm” rewriting system of CT within fragments of second-order arithmetic. We establish confluence of CT within the theory RCA0, a conservative extension of primitive recursive arithmetic and IΣ1. This allows us to recast type structures of hereditarily recursive operations as “coterm” models of T. We show that these also form models of CT, by formalising a totality argument for circular typing derivations within suitable fragments of second-order arithmetic. Relying on well-known proof mining results, we thus obtain an interpretation of CT into T; in fact, more precisely, we interpret level-n-CT into level-(n + 1)-T, an optimum result in terms of abstraction complexity. A direct consequence of these model-theoretic results is weak normalisation for CT. As further results, we also show strong normalisation for CT and continuity of functionals computed by its type 2 coterms. 2012 ACM Subject Classification Theory of computation → Equational logic and rewriting; Theory of computation → Proof theory; Theory of computation → Higher order logic; Theory of computation → Lambda calculus Keywords and phrases confluence, normalisation, system T, circular proofs, reverse mathematics, type structures Digital Object Identifier 10.4230/LIPIcs.FSCD.2021.29 Related Version This work is based on part of the following preprint, where related results, proofs and examples may be found.
    [Show full text]
  • Inseparability and Conservative Extensions of Description Logic Ontologies: a Survey
    Inseparability and Conservative Extensions of Description Logic Ontologies: A Survey Elena Botoeva1, Boris Konev2, Carsten Lutz3, Vladislav Ryzhikov1, Frank Wolter2, and Michael Zakharyaschev4 1 Free University of Bozen-Bolzano, Italy fbotoeva,[email protected] 2 University of Liverpool, UK fkonev,[email protected] 3 University of Bremen, Germany [email protected] 4 Birkbeck, University of London, UK [email protected] Abstract. The question whether an ontology can safely be replaced by another, possibly simpler, one is fundamental for many ontology engi- neering and maintenance tasks. It underpins, for example, ontology ver- sioning, ontology modularization, forgetting, and knowledge exchange. What `safe replacement' means depends on the intended application of the ontology. If, for example, it is used to query data, then the answers to any relevant ontology-mediated query should be the same over any relevant data set; if, in contrast, the ontology is used for conceptual reasoning, then the entailed subsumptions between concept expressions should coincide. This gives rise to different notions of ontology insep- arability such as query inseparability and concept inseparability, which generalize corresponding notions of conservative extensions. We survey results on various notions of inseparability in the context of descrip- tion logic ontologies, discussing their applications, useful model-theoretic characterizations, algorithms for determining whether two ontologies are inseparable (and, sometimes, for computing the difference between them if they are not), and the computational complexity of this problem. 1 Introduction Description logic (DL) ontologies provide a common vocabulary for a domain of interest together with a formal modeling of the semantics of the vocabulary items (concept names and role names).
    [Show full text]
  • The Strength of Mac Lane Set Theory
    The Strength of Mac Lane Set Theory A. R. D. MATHIAS D´epartement de Math´ematiques et Informatique Universit´e de la R´eunion To Saunders Mac Lane on his ninetieth birthday Abstract AUNDERS MAC LANE has drawn attention many times, particularly in his book Mathematics: Form and S Function, to the system ZBQC of set theory of which the axioms are Extensionality, Null Set, Pairing, Union, Infinity, Power Set, Restricted Separation, Foundation, and Choice, to which system, afforced by the principle, TCo, of Transitive Containment, we shall refer as MAC. His system is naturally related to systems derived from topos-theoretic notions concerning the category of sets, and is, as Mac Lane emphasizes, one that is adequate for much of mathematics. In this paper we show that the consistency strength of Mac Lane's system is not increased by adding the axioms of Kripke{Platek set theory and even the Axiom of Constructibility to Mac Lane's axioms; our method requires a close study of Axiom H, which was proposed by Mitchell; we digress to apply these methods to subsystems of Zermelo set theory Z, and obtain an apparently new proof that Z is not finitely axiomatisable; we study Friedman's strengthening KPP + AC of KP + MAC, and the Forster{Kaye subsystem KF of MAC, and use forcing over ill-founded models and forcing to establish independence results concerning MAC and KPP ; we show, again using ill-founded models, that KPP + V = L proves the consistency of KPP ; turning to systems that are type-theoretic in spirit or in fact, we show by arguments of Coret
    [Show full text]
  • More Model Theory Notes Miscellaneous Information, Loosely Organized
    More Model Theory Notes Miscellaneous information, loosely organized. 1. Kinds of Models A countable homogeneous model M is one such that, for any partial elementary map f : A ! M with A ⊆ M finite, and any a 2 M, f extends to a partial elementary map A [ fag ! M. As a consequence, any partial elementary map to M is extendible to an automorphism of M. Atomic models (see below) are homogeneous. A prime model of T is one that elementarily embeds into every other model of T of the same cardinality. Any theory with fewer than continuum-many types has a prime model, and if a theory has a prime model, it is unique up to isomorphism. Prime models are homogeneous. On the other end, a model is universal if every other model of its size elementarily embeds into it. Recall a type is a set of formulas with the same tuple of free variables; generally to be called a type we require consistency. The type of an element or tuple from a model is all the formulas it satisfies. One might think of the type of an element as a sort of identity card for automorphisms: automorphisms of a model preserve types. A complete type is the analogue of a complete theory, one where every formula of the appropriate free variables or its negation appears. Types of elements and tuples are always complete. A type is principal if there is one formula in the type that implies all the rest; principal complete types are often called isolated. A trivial example of an isolated type is that generated by the formula x = c where c is any constant in the language, or x = t(¯c) where t is any composition of appropriate-arity functions andc ¯ is a tuple of constants.
    [Show full text]
  • Warren Goldfarb, Notes on Metamathematics
    Notes on Metamathematics Warren Goldfarb W.B. Pearson Professor of Modern Mathematics and Mathematical Logic Department of Philosophy Harvard University DRAFT: January 1, 2018 In Memory of Burton Dreben (1927{1999), whose spirited teaching on G¨odeliantopics provided the original inspiration for these Notes. Contents 1 Axiomatics 1 1.1 Formal languages . 1 1.2 Axioms and rules of inference . 5 1.3 Natural numbers: the successor function . 9 1.4 General notions . 13 1.5 Peano Arithmetic. 15 1.6 Basic laws of arithmetic . 18 2 G¨odel'sProof 23 2.1 G¨odelnumbering . 23 2.2 Primitive recursive functions and relations . 25 2.3 Arithmetization of syntax . 30 2.4 Numeralwise representability . 35 2.5 Proof of incompleteness . 37 2.6 `I am not derivable' . 40 3 Formalized Metamathematics 43 3.1 The Fixed Point Lemma . 43 3.2 G¨odel'sSecond Incompleteness Theorem . 47 3.3 The First Incompleteness Theorem Sharpened . 52 3.4 L¨ob'sTheorem . 55 4 Formalizing Primitive Recursion 59 4.1 ∆0,Σ1, and Π1 formulas . 59 4.2 Σ1-completeness and Σ1-soundness . 61 4.3 Proof of Representability . 63 3 5 Formalized Semantics 69 5.1 Tarski's Theorem . 69 5.2 Defining truth for LPA .......................... 72 5.3 Uses of the truth-definition . 74 5.4 Second-order Arithmetic . 76 5.5 Partial truth predicates . 79 5.6 Truth for other languages . 81 6 Computability 85 6.1 Computability . 85 6.2 Recursive and partial recursive functions . 87 6.3 The Normal Form Theorem and the Halting Problem . 91 6.4 Turing Machines .
    [Show full text]
  • Choice, Extension, Conservation. from Transfinite to finite Methods in Abstract Algebra
    Choice, extension, conservation. From transfinite to finite methods in abstract algebra Daniel Wessel Universita` degli Studi di Trento Universita` degli Studi di Verona December 3, 2017 Doctoral thesis in Mathematics Joint doctoral programme in Mathematics, 30th cycle Department of Mathematics, University of Trento Department of Computer Science, University of Verona Academic year 2017/18 Supervisor: Peter Schuster, University of Verona Trento, Italy December 3, 2017 Abstract Maximality principles such as the ones going back to Kuratowski and Zorn ensure the existence of higher type ideal objects the use of which is commonly held indispensable for mathematical practice. The modern turn towards computational methods, which can be witnessed to have a strong influence on contemporary foundational studies, encourages a reassessment within a constructive framework of the methodological intricacies that go along with invocations of maximality principles. The common thread that can be followed through the chapters of this thesis is explained by the attempt to put the widespread use of ideal objects under constructive scrutiny. It thus walks the tracks of a revised Hilbert's programme which has inspired a reapproach to constructive algebra by finitary means, and for which Scott's entailment relations have already shown to provide a vital and utmost versatile tool. In this thesis several forms of the Kuratowski-Zorn Lemma are introduced and proved equivalent over constructive set theory; the notion of Jacobson radical is brought from com- mutative
    [Show full text]
  • Constructive Set Theory and Brouwerian Principles1
    Constructive Set Theory and Brouwerian Principles1 Michael Rathjen (Department of Mathematics, The Ohio State University,2 Email: [email protected]) Abstract: The paper furnishes realizability models of constructive Zermelo-Fraenkel set theory, CZF, which also validate Brouwerian principles such as the axiom of con- tinuous choice (CC), the fan theorem (FT), and monotone bar induction (BIM), and thereby determines the proof-theoretic strength of CZF augmented by these principles. The upshot is that CZF+CC+FT possesses the same strength as CZF, or more pre- 0 cisely, that CZF+CC+FT is conservative over CZF for Π2 statements of arithmetic, whereas the addition of a restricted version of bar induction to CZF (called decidable bar induction, BID) leads to greater proof-theoretic strength in that CZF+BID proves the consistency of CZF. Key Words: Constructive set theory, Brouwerian principles, partial combinatory al- gebra, realizability Category: F.4.1 1 Introduction Constructive Zermelo-Fraenkel Set Theory has emerged as a standard reference theory that relates to constructive predicative mathematics as ZFC relates to classical Cantorian mathematics. The general topic of Constructive Set Theory originated in the seminal 1975 paper of John Myhill (cf. [16]), where a specific ax- iom system CST was introduced. Myhill developed constructive set theory with the aim of isolating the principles underlying Bishop's conception of what sets and functions are. Moreover, he wanted \these principles to be such as to make the process of formalization completely trivial, as it is in the classical case" ([16], p. 347). Indeed, while he uses other primitives in his set theory CST besides the notion of set, it can be viewed as a subsystem of ZF.
    [Show full text]
  • Towards Flexiformal Mathematics By
    Towards Flexiformal Mathematics by Mihnea Iancu A thesis submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Computer Science Approved Dissertation Committee Prof. Dr. Michael Kohlhase Jacobs University Bremen, Germany Prof. Dr. Herbert Jaeger Jacobs University Bremen, Germany Prof. Dr. William M. Farmer McMaster University, Canada Priv.-Doz. Dr. Florian Rabe Jacobs University Bremen, Germany Date of defense: Jan 5, 2017 Computer Science & Electrical Engineering Statutory Declaration (Declaration on Authorship of a Dissertation) I, Mihnea Iancu, hereby declare, under penalty of perjury, that I am aware of the consequences of a deliberately or negligently wrongly submitted affidavit, in particular the punitive provisions of x156 and x161 of the Criminal Code (up to 1 year imprisonment or a fine at delivering a negligent or 3 years or a fine at a knowingly false affidavit). Furthermore I declare that I have written this PhD thesis independently, unless where clearly stated otherwise. I have used only the sources, the data and the support that I have clearly mentioned. This PhD thesis has not been submitted for the conferral of a degree elsewhere. Place Date Signature ii Parts of this thesis are based on or closely related to previously published material or material that is prepared for publication at the time of this writing. These are [IKR], [IK15b], [IK15a], [KI14], [IKP14], [IJKW14], [GIJ+16], [IKRU13] and [IKR11]. Parts of this work are collaborations with other researchers: For Part I with Michael Kohlhase, for Part II with Michael Kohlhase and Florian Rabe and for Part III with Constantin Jucovschi, Michael Kohlhase, Enxhell Luzhnica, Akbar Oripov, Corneliu Prodescu, Florian Rabe and Tom Wiesing.
    [Show full text]