Isomorphisms Math 130 Linear Algebra

Total Page:16

File Type:pdf, Size:1020Kb

Isomorphisms Math 130 Linear Algebra the identity function on B. The usual notation for the function inverse to f is f −1. If f and g are inverse to each other, that is, if g is the inverse of f, g = f −1, then f is the inverse of Isomorphisms g, f = g−1. Thus, (f −1)−1 = f. Math 130 Linear Algebra An important property of bijections is that you D Joyce, Fall 2013 can convert equations involving f to equations in- volving f −1: Frequently in mathematics we look at two alge- −1 braic structures A and B of the same kind and want f(x) = y if and only if x = f (y) to compare them. For instance, we might think they're really the same thing, but they have dif- . ferent names for their elements. That leads to the concept of isomorphism f : A !' B, and we'll talk Isomorphisms of algebraic structures. There about that first. Other times we'll know they're are lots of different kinds of algebraic structures. not the same thing, but there is a relation between We've already studied two of them, namely, fields them, and that will lead to the next concept, ho- and vector spaces. momorphism, f : A ! B. We'll then look as some We'll say two algebraic structures A and B are special homomorphisms such as monomorphisms. isomorphic if they have exactly the same structure, When we have a homomorphism f : A ! A, we'll but their elements may be different. For instance, call it an endomorphism, and when an isomorphism let A be the vector space R[x] of polynomials in f : A !' A, we'll call it an automorphism. We'll the variable x, and let B be the vector space R[y] take each of these variants in turn. of polynomials in y. They're both just polynomials in one variable, it's just that the choice of variable Injections, surjections, and bijections of is different in the two rings. functions between sets. These are words that We're studying vector spaces, so we need a pre- describe certain functions f : A ! B from one set cise definition of isomorphism for them. to another. Definition 1 (Isomorphism of vector spaces). Two An injection, also called a one-to-one function vector spaces V and W over the same field F are is a function that maps distinct elements to dis- isomorphic if there is a bijection T : V ! W which tinct elements, that is, if x 6= y, then f(x) 6= f(y). preserves addition and scalar multiplication, that Equivalently, if f(x) = f(y) then x = y. If A is, for all vectors u and v in V , and all scalars is a subset of B, then there is a natural injection c 2 F , ι : A ! B, called the inclusion function, defined by ι(x) = x. T (u + v) = T (u) + T (v) and T (cv) = cT (v): A surjection, also called an onto function is one that includes all of B in its image, that is, if y 2 B, The correspondence T is called an isomorphism of then there is an x 2 A such that f(x) = y. vector spaces. A bijection, also called a one-to-one correspon- dence, is a function that is simultaneously injective When T : V ! W is an isomorphism we'll write and bijective. Another way to describe a bijection T : V !' W if we want to emphasize that it is an f : A ! B is to say that there is an inverse function isomorphism. When V and W are isomorphic, but g : B ! A so that the composition g ◦ f : A ! A the specific isomorphism is not named, we'll just is the identity function on A while f ◦ g : B ! B is write V ∼= W . 1 ' Of course, the identity function IV : V ! V is Example 4. Consider P3, the vector space of poly- ' an isomorphism. nomials over R of degree 3 or less. Define T : P3 ! 4 3 2 After we introduce linear transformations (which R by T (a1x + a2x + a3x + a4) = (a1; a2; a3; a4). is what homomorphisms of vector spaces are It just associates to a polynomial its 4-tuple of co- called), we'll have another way to describe isomor- efficients starting with the coefficient of x3 and go- phisms. ing down in degree. This T preserves addition and You can prove various properties of vector space scalar multiplication, it is one-to-one, and it is onto. isomorphisms from this definition. (Those statements are easy to verify.) ' 4 Since the structure of vector spaces is defined in This is not the only isomorphism P3 ! R .A terms of addition and scalar multiplication, if T cubic polynomial is determined by its value at any preserves them, it will preserve structure defined in four points. The association f(x) to the 4-tuple terms of them. For instance, T preserves 0, nega- (f(1); f(2); f(3); f(4)) is also an isomorphism. tion, subtraction, and linear transformations. Theorem 5. If T : V ! W is an isomorphism, Theorem 2. If T : V !' W is an isomorphism of then T carries linearly independent sets to linearly vector spaces, then its inverse T −1 : W !' V is also independent sets, spanning sets to spanning sets, an isomorphism. and bases to bases. Proof. Since T is a bijection, T −1 exists as a func- Proof. For the first statement, let S be a set of lin- tion W ! V . We have to show T −1 preserves ad- early independent vectors in V . We'll show that its dition and scalar multiplication. image T (S) is a set of linearly independent vectors in W . If 0 were a nontrivial linear combination of First, we'll do addition. Let w and x be elements −1 of W . We have to show that vectors in T (S), then an application of T would yield a nontrivial linear combination of vectors in T −1(w + x) = T −1(w) + T −1(x): S, but there is none since S is independent. There- fore, T (S) is linear independent. We'll show that by simplifying it to logically equiv- For the second statement, let w be any vector in alent statements until we reach one which we know W , then T −1(w) is a linear combination of vectors is true. Since T and T −1 are inverse functions, that in V . Apply T to that linear combination to see equation holds if and only if that w is a linear combination of vectors in W . Since T carries both independent and spanning w + x = T (T −1(w) + T −1(x)): sets from V to W , it carries bases to bases. q.e.d. Since T is an isomorphism, we can rewrite that as More generally, any property of vector spaces de- fined in terms of the structure of vector spaces (ad- w + x = T (T −1(w)) + T (T −1(x)) dition and scalar multiplication) is preserved by iso- morphisms. which simplifies to w + x = w + x which is true. Scalar multiplication is left to you. Show Coordinates with respect to a basis deter- T −1(cw) = cT −1(w). q.e.d. mine an isomorphism. One of the main uses of a basis β = (b1; b2;:::; bn) for a vector space V We'll omit the proof of the next theorem. over a field is to impose coordinates on V . Each ' ' vector v in V is a unique linear combination of of Theorem 3. If S : V ! W and T : W ! X are the basis vectors both isomorphisms of vector spaces, then so is their composition (T ◦ S): V !' X. v = v1b1 + v2b2 + ··· + vnbn: 2 The coefficients are used as coordinates for v with the respect to the basis β 2 3 v1 6v 7 6 27 [v]β = 6 . 7 : 4 . 5 vn Let's denote the function that assigns these co- ordinates φβ. Theorem 6. The correspondence v to [v]β is an ' n isomorphism φβ : V ! F . To prove that theorem, you'll need to note that this is a bijection, prove that [u + v]β = [u]β + [v]β, and prove that [cv]β = c[v]β. Since the correspondence φβ is an isomorphism, it means we can work with coordinates with respect to a basis β of V just like ordinary coordinates. Corollary 7. Two finite dimensional vector spaces are isomorphic if and only if they have the same dimension. Proof. If they're isomorphic, then there's an iso- morphism T from one to the other, and it carries a basis of the first to a basis of the second. Therefore they have the same dimension. On the other hand, if they have the same dimen- sion n, then they're each isomorphic to F n, and therefore they're isomorphic to each other. q.e.d. Linear transformations. Next we'll look at lin- ear transformations of vector spaces. Whereas isomorphisms are bijections that pre- serve the algebraic structure, homomorphisms are simply functions that preserve the algebraic struc- ture. In the case of vector spaces, the term linear transformation is used in preference to homomor- phism. Math 130 Home Page at http://math.clarku.edu/~djoyce/ma130/ 3.
Recommended publications
  • Chapter 4. Homomorphisms and Isomorphisms of Groups
    Chapter 4. Homomorphisms and Isomorphisms of Groups 4.1 Note: We recall the following terminology. Let X and Y be sets. When we say that f is a function or a map from X to Y , written f : X ! Y , we mean that for every x 2 X there exists a unique corresponding element y = f(x) 2 Y . The set X is called the domain of f and the range or image of f is the set Image(f) = f(X) = f(x) x 2 X . For a set A ⊆ X, the image of A under f is the set f(A) = f(a) a 2 A and for a set −1 B ⊆ Y , the inverse image of B under f is the set f (B) = x 2 X f(x) 2 B . For a function f : X ! Y , we say f is one-to-one (written 1 : 1) or injective when for every y 2 Y there exists at most one x 2 X such that y = f(x), we say f is onto or surjective when for every y 2 Y there exists at least one x 2 X such that y = f(x), and we say f is invertible or bijective when f is 1:1 and onto, that is for every y 2 Y there exists a unique x 2 X such that y = f(x). When f is invertible, the inverse of f is the function f −1 : Y ! X defined by f −1(y) = x () y = f(x). For f : X ! Y and g : Y ! Z, the composite g ◦ f : X ! Z is given by (g ◦ f)(x) = g(f(x)).
    [Show full text]
  • The General Linear Group
    18.704 Gabe Cunningham 2/18/05 [email protected] The General Linear Group Definition: Let F be a field. Then the general linear group GLn(F ) is the group of invert- ible n × n matrices with entries in F under matrix multiplication. It is easy to see that GLn(F ) is, in fact, a group: matrix multiplication is associative; the identity element is In, the n × n matrix with 1’s along the main diagonal and 0’s everywhere else; and the matrices are invertible by choice. It’s not immediately clear whether GLn(F ) has infinitely many elements when F does. However, such is the case. Let a ∈ F , a 6= 0. −1 Then a · In is an invertible n × n matrix with inverse a · In. In fact, the set of all such × matrices forms a subgroup of GLn(F ) that is isomorphic to F = F \{0}. It is clear that if F is a finite field, then GLn(F ) has only finitely many elements. An interesting question to ask is how many elements it has. Before addressing that question fully, let’s look at some examples. ∼ × Example 1: Let n = 1. Then GLn(Fq) = Fq , which has q − 1 elements. a b Example 2: Let n = 2; let M = ( c d ). Then for M to be invertible, it is necessary and sufficient that ad 6= bc. If a, b, c, and d are all nonzero, then we can fix a, b, and c arbitrarily, and d can be anything but a−1bc. This gives us (q − 1)3(q − 2) matrices.
    [Show full text]
  • Useful Relations in Permutations and Combination 1. Useful Relations
    Useful Relations in Permutations and Combination 1. Useful Relations - Factorial n! = n.(n-1)! 2. n퐶푟= n푃푟/r! n n-1 3. Pr = n( Pr-1) 4. Useful Relations - Combinations n n 1. Cr = C(n - r) Example 8 8 C6 = C2 = 8×72×1 = 28 n 2. Cn = 1 n 3. C0 = 1 n n n n n 4. C0 + C1 + C2 + ... + Cn = 2 Example 4 4 4 4 4 4 C0 + C1 + C2 + C3+ C4 = (1 + 4 + 6 + 4 + 1) = 16 = 2 n n (n+1) Cr-1 + Cr = Cr (Pascal's Law) n퐶푟 =n/퐶푟−1=n-r+1/r n n If Cx = Cy then either x = y or (n-x) = y. 5. Selection from identical objects: Some Basic Facts The number of selections of r objects out of n identical objects is 1. Total number of selections of zero or more objects from n identical objects is n+1. 6. Permutations of Objects when All Objects are Not Distinct The number of ways in which n things can be arranged taking them all at a time, when st nd 푃1 of the things are exactly alike of 1 type, 푃2 of them are exactly alike of a 2 type, and th 푃푟of them are exactly alike of r type and the rest of all are distinct is n!/ 푃1! 푃2! ... 푃푟! 1 Example: how many ways can you arrange the letters in the word THESE? 5!/2!=120/2=60 Example: how many ways can you arrange the letters in the word REFERENCE? 9!/2!.4!=362880/2*24=7560 7.Circular Permutations: Case 1: when clockwise and anticlockwise arrangements are different Number of circular permutations (arrangements) of n different things is (n-1)! 1.
    [Show full text]
  • Categories, Functors, and Natural Transformations I∗
    Lecture 2: Categories, functors, and natural transformations I∗ Nilay Kumar June 4, 2014 (Meta)categories We begin, for the moment, with rather loose definitions, free from the technicalities of set theory. Definition 1. A metagraph consists of objects a; b; c; : : :, arrows f; g; h; : : :, and two operations, as follows. The first is the domain, which assigns to each arrow f an object a = dom f, and the second is the codomain, which assigns to each arrow f an object b = cod f. This is visually indicated by f : a ! b. Definition 2. A metacategory is a metagraph with two additional operations. The first is the identity, which assigns to each object a an arrow Ida = 1a : a ! a. The second is the composition, which assigns to each pair g; f of arrows with dom g = cod f an arrow g ◦ f called their composition, with g ◦ f : dom f ! cod g. This operation may be pictured as b f g a c g◦f We require further that: composition is associative, k ◦ (g ◦ f) = (k ◦ g) ◦ f; (whenever this composition makese sense) or diagrammatically that the diagram k◦(g◦f)=(k◦g)◦f a d k◦g f k g◦f b g c commutes, and that for all arrows f : a ! b and g : b ! c, we have 1b ◦ f = f and g ◦ 1b = g; or diagrammatically that the diagram f a b f g 1b g b c commutes. ∗This talk follows [1] I.1-4 very closely. 1 Recall that a diagram is commutative when, for each pair of vertices c and c0, any two paths formed from direct edges leading from c to c0 yield, by composition of labels, equal arrows from c to c0.
    [Show full text]
  • Irreducible Representations of Finite Monoids
    U.U.D.M. Project Report 2019:11 Irreducible representations of finite monoids Christoffer Hindlycke Examensarbete i matematik, 30 hp Handledare: Volodymyr Mazorchuk Examinator: Denis Gaidashev Mars 2019 Department of Mathematics Uppsala University Irreducible representations of finite monoids Christoffer Hindlycke Contents Introduction 2 Theory 3 Finite monoids and their structure . .3 Introductory notions . .3 Cyclic semigroups . .6 Green’s relations . .7 von Neumann regularity . 10 The theory of an idempotent . 11 The five functors Inde, Coinde, Rese,Te and Ne ..................... 11 Idempotents and simple modules . 14 Irreducible representations of a finite monoid . 17 Monoid algebras . 17 Clifford-Munn-Ponizovski˘ıtheory . 20 Application 24 The symmetric inverse monoid . 24 Calculating the irreducible representations of I3 ........................ 25 Appendix: Prerequisite theory 37 Basic definitions . 37 Finite dimensional algebras . 41 Semisimple modules and algebras . 41 Indecomposable modules . 42 An introduction to idempotents . 42 1 Irreducible representations of finite monoids Christoffer Hindlycke Introduction This paper is a literature study of the 2016 book Representation Theory of Finite Monoids by Benjamin Steinberg [3]. As this book contains too much interesting material for a simple master thesis, we have narrowed our attention to chapters 1, 4 and 5. This thesis is divided into three main parts: Theory, Application and Appendix. Within the Theory chapter, we (as the name might suggest) develop the necessary theory to assist with finding irreducible representations of finite monoids. Finite monoids and their structure gives elementary definitions as regards to finite monoids, and expands on the basic theory of their structure. This part corresponds to chapter 1 in [3]. The theory of an idempotent develops just enough theory regarding idempotents to enable us to state a key result, from which the principal result later follows almost immediately.
    [Show full text]
  • SOME ALGEBRAIC DEFINITIONS and CONSTRUCTIONS Definition
    SOME ALGEBRAIC DEFINITIONS AND CONSTRUCTIONS Definition 1. A monoid is a set M with an element e and an associative multipli- cation M M M for which e is a two-sided identity element: em = m = me for all m M×. A−→group is a monoid in which each element m has an inverse element m−1, so∈ that mm−1 = e = m−1m. A homomorphism f : M N of monoids is a function f such that f(mn) = −→ f(m)f(n) and f(eM )= eN . A “homomorphism” of any kind of algebraic structure is a function that preserves all of the structure that goes into the definition. When M is commutative, mn = nm for all m,n M, we often write the product as +, the identity element as 0, and the inverse of∈m as m. As a convention, it is convenient to say that a commutative monoid is “Abelian”− when we choose to think of its product as “addition”, but to use the word “commutative” when we choose to think of its product as “multiplication”; in the latter case, we write the identity element as 1. Definition 2. The Grothendieck construction on an Abelian monoid is an Abelian group G(M) together with a homomorphism of Abelian monoids i : M G(M) such that, for any Abelian group A and homomorphism of Abelian monoids−→ f : M A, there exists a unique homomorphism of Abelian groups f˜ : G(M) A −→ −→ such that f˜ i = f. ◦ We construct G(M) explicitly by taking equivalence classes of ordered pairs (m,n) of elements of M, thought of as “m n”, under the equivalence relation generated by (m,n) (m′,n′) if m + n′ = −n + m′.
    [Show full text]
  • Limits Commutative Algebra May 11 2020 1. Direct Limits Definition 1
    Limits Commutative Algebra May 11 2020 1. Direct Limits Definition 1: A directed set I is a set with a partial order ≤ such that for every i; j 2 I there is k 2 I such that i ≤ k and j ≤ k. Let R be a ring. A directed system of R-modules indexed by I is a collection of R modules fMi j i 2 Ig with a R module homomorphisms µi;j : Mi ! Mj for each pair i; j 2 I where i ≤ j, such that (i) for any i 2 I, µi;i = IdMi and (ii) for any i ≤ j ≤ k in I, µi;j ◦ µj;k = µi;k. We shall denote a directed system by a tuple (Mi; µi;j). The direct limit of a directed system is defined using a universal property. It exists and is unique up to a unique isomorphism. Theorem 2 (Direct limits). Let fMi j i 2 Ig be a directed system of R modules then there exists an R module M with the following properties: (i) There are R module homomorphisms µi : Mi ! M for each i 2 I, satisfying µi = µj ◦ µi;j whenever i < j. (ii) If there is an R module N such that there are R module homomorphisms νi : Mi ! N for each i and νi = νj ◦µi;j whenever i < j; then there exists a unique R module homomorphism ν : M ! N, such that νi = ν ◦ µi. The module M is unique in the sense that if there is any other R module M 0 satisfying properties (i) and (ii) then there is a unique R module isomorphism µ0 : M ! M 0.
    [Show full text]
  • A Review of Commutative Ring Theory Mathematics Undergraduate Seminar: Toric Varieties
    A REVIEW OF COMMUTATIVE RING THEORY MATHEMATICS UNDERGRADUATE SEMINAR: TORIC VARIETIES ADRIANO FERNANDES Contents 1. Basic Definitions and Examples 1 2. Ideals and Quotient Rings 3 3. Properties and Types of Ideals 5 4. C-algebras 7 References 7 1. Basic Definitions and Examples In this first section, I define a ring and give some relevant examples of rings we have encountered before (and might have not thought of as abstract algebraic structures.) I will not cover many of the intermediate structures arising between rings and fields (e.g. integral domains, unique factorization domains, etc.) The interested reader is referred to Dummit and Foote. Definition 1.1 (Rings). The algebraic structure “ring” R is a set with two binary opera- tions + and , respectively named addition and multiplication, satisfying · (R, +) is an abelian group (i.e. a group with commutative addition), • is associative (i.e. a, b, c R, (a b) c = a (b c)) , • and the distributive8 law holds2 (i.e.· a,· b, c ·R, (·a + b) c = a c + b c, a (b + c)= • a b + a c.) 8 2 · · · · · · Moreover, the ring is commutative if multiplication is commutative. The ring has an identity, conventionally denoted 1, if there exists an element 1 R s.t. a R, 1 a = a 1=a. 2 8 2 · ·From now on, all rings considered will be commutative rings (after all, this is a review of commutative ring theory...) Since we will be talking substantially about the complex field C, let us recall the definition of such structure. Definition 1.2 (Fields).
    [Show full text]
  • Algebraic Structures Lecture 18 Thursday, April 4, 2019 1 Type
    Harvard School of Engineering and Applied Sciences — CS 152: Programming Languages Algebraic structures Lecture 18 Thursday, April 4, 2019 In abstract algebra, algebraic structures are defined by a set of elements and operations on those ele- ments that satisfy certain laws. Some of these algebraic structures have interesting and useful computa- tional interpretations. In this lecture we will consider several algebraic structures (monoids, functors, and monads), and consider the computational patterns that these algebraic structures capture. We will look at Haskell, a functional programming language named after Haskell Curry, which provides support for defin- ing and using such algebraic structures. Indeed, monads are central to practical programming in Haskell. First, however, we consider type constructors, and see two new type constructors. 1 Type constructors A type constructor allows us to create new types from existing types. We have already seen several different type constructors, including product types, sum types, reference types, and parametric types. The product type constructor × takes existing types τ1 and τ2 and constructs the product type τ1 × τ2 from them. Similarly, the sum type constructor + takes existing types τ1 and τ2 and constructs the product type τ1 + τ2 from them. We will briefly introduce list types and option types as more examples of type constructors. 1.1 Lists A list type τ list is the type of lists with elements of type τ. We write [] for the empty list, and v1 :: v2 for the list that contains value v1 as the first element, and v2 is the rest of the list. We also provide a way to check whether a list is empty (isempty? e) and to get the head and the tail of a list (head e and tail e).
    [Show full text]
  • Combinatorics
    Combinatorics Problem: How to count without counting. I How do you figure out how many things there are with a certain property without actually enumerating all of them. Sometimes this requires a lot of cleverness and deep mathematical insights. But there are some standard techniques. I That's what we'll be studying. We sometimes use the bijection rule without even realizing it: I count how many people voted are in favor of something by counting the number of hands raised: I I'm hoping that there's a bijection between the people in favor and the hands raised! Bijection Rule The Bijection Rule: If f : A ! B is a bijection, then jAj = jBj. I We used this rule in defining cardinality for infinite sets. I Now we'll focus on finite sets. Bijection Rule The Bijection Rule: If f : A ! B is a bijection, then jAj = jBj. I We used this rule in defining cardinality for infinite sets. I Now we'll focus on finite sets. We sometimes use the bijection rule without even realizing it: I count how many people voted are in favor of something by counting the number of hands raised: I I'm hoping that there's a bijection between the people in favor and the hands raised! Answer: 26 choices for the first letter, 26 for the second, 10 choices for the first number, the second number, and the third number: 262 × 103 = 676; 000 Example 2: A traveling salesman wants to do a tour of all 50 state capitals. How many ways can he do this? Answer: 50 choices for the first place to visit, 49 for the second, .
    [Show full text]
  • Inclusion‒Exclusion Principle
    Inclusionexclusion principle 1 Inclusion–exclusion principle In combinatorics, the inclusion–exclusion principle (also known as the sieve principle) is an equation relating the sizes of two sets and their union. It states that if A and B are two (finite) sets, then The meaning of the statement is that the number of elements in the union of the two sets is the sum of the elements in each set, respectively, minus the number of elements that are in both. Similarly, for three sets A, B and C, This can be seen by counting how many times each region in the figure to the right is included in the right hand side. More generally, for finite sets A , ..., A , one has the identity 1 n This can be compactly written as The name comes from the idea that the principle is based on over-generous inclusion, followed by compensating exclusion. When n > 2 the exclusion of the pairwise intersections is (possibly) too severe, and the correct formula is as shown with alternating signs. This formula is attributed to Abraham de Moivre; it is sometimes also named for Daniel da Silva, Joseph Sylvester or Henri Poincaré. Inclusion–exclusion illustrated for three sets For the case of three sets A, B, C the inclusion–exclusion principle is illustrated in the graphic on the right. Proof Let A denote the union of the sets A , ..., A . To prove the 1 n Each term of the inclusion-exclusion formula inclusion–exclusion principle in general, we first have to verify the gradually corrects the count until finally each identity portion of the Venn Diagram is counted exactly once.
    [Show full text]
  • STRUCTURE ENUMERATION and SAMPLING Chemical Structure Enumeration and Sampling Have Been Studied by Mathematicians, Computer
    STRUCTURE ENUMERATION AND SAMPLING MARKUS MERINGER To appear in Handbook of Chemoinformatics Algorithms Chemical structure enumeration and sampling have been studied by 5 mathematicians, computer scientists and chemists for quite a long time. Given a molecular formula plus, optionally, a list of structural con- straints, the typical questions are: (1) How many isomers exist? (2) Which are they? And, especially if (2) cannot be answered completely: (3) How to get a sample? 10 In this chapter we describe algorithms for solving these problems. The techniques are based on the representation of chemical compounds as molecular graphs (see Chapter 2), i.e. they are mainly applied to constitutional isomers. The major problem is that in silico molecular graphs have to be represented as labeled structures, while in chemical 15 compounds, the atoms are not labeled. The mathematical concept for approaching this problem is to consider orbits of labeled molecular graphs under the operation of the symmetric group. We have to solve the so–called isomorphism problem. According to our introductory questions, we distinguish several dis- 20 ciplines: counting, enumerating and sampling isomers. While counting only delivers the number of isomers, the remaining disciplines refer to constructive methods. Enumeration typically encompasses exhaustive and non–redundant methods, while sampling typically lacks these char- acteristics. However, sampling methods are sometimes better suited to 25 solve real–world problems. There is a wide range of applications where counting, enumeration and sampling techniques are helpful or even essential. Some of these applications are closely linked to other chapters of this book. Counting techniques deliver pure chemical information, they can help to estimate 30 or even determine sizes of chemical databases or compound libraries obtained from combinatorial chemistry.
    [Show full text]