On the Equivalence of Tries and Dendrograms - E€Icient Hierarchical Clustering of Tra€Ic Data

Total Page:16

File Type:pdf, Size:1020Kb

On the Equivalence of Tries and Dendrograms - E€Icient Hierarchical Clustering of Tra€Ic Data On The Equivalence of Tries and Dendrograms - Eicient Hierarchical Clustering of Traic Data Chia-Tung Kuo Ian Davidson University of California, Davis University of California, Davis [email protected] [email protected] ABSTRACT clustering is its O¹n2 lognº run time, which is prohibitive when the e widespread use of GPS-enabled devices generates voluminous number of instances, n is large. For example, in our experiments and continuous amounts of trac data but analyzing such data for our data set has nearly half a million trips. Classic agglomerative interpretable and actionable insights poses challenges. A hierar- algorithms would take days or even weeks to build a dendrogram chical clustering of the trips has many uses such as discovering on such large data set. shortest paths, common routes and oen traversed areas. However, In this paper we consider an alternative way to eciently con- hierarchical clustering typically has time complexity of O¹n2 lognº struct a dendrogram of the trip data without the long computation where n is the number of instances, and is dicult to scale to large time. We achieve this by rst converting each trip to a trajectory data sets associated with GPS data. Furthermore, incremental hier- string and then building a prex tree from the trip-strings. We then archical clustering is still a developing area. Prex trees (also called formally show the equivalence between a prex tree and a dendro- tries) can be eciently constructed and updated in linear time (in gram by showing that the prex tree we create would have been n). We show how a specially constructed trie can compactly store built by a classic agglomerative method using a specic distance the trips and further show this trie is equivalent to a dendrogram metric on the strings. is result is not trivial as the prex tree is that would have been built by classic agglomerative hierarchical created top-down and the dendrogram boom-up. algorithms using a specic distance metric. is allows creating Creating Trajectory Strings from GPS Data. We discretize hierarchical clusterings of GPS trip data and updating this hierar- both the spatial and temporal dimensions with respective pre- chy in linear time. We demonstrate the usefulness of our proposed dened resolutions, eectively converting each spatial location approach on a real world data set of half a million taxis’ GPS traces, to a unique symbol. For example, in our experiments we disceretize well beyond the capabilities of agglomerative clustering methods. the San Francisco Bay area into a 100 × 100 grid so our alphabet Our work is not limited to trip data and can be used with other data contains 10,000 symbols. We can then naturally represent a trip with a string representation. as a sequence (i.e. string) of the discretized regions (symbols). e symbol at position i in the string represents the location of the trip ACM Reference format: at time step i as shown in Figure 1. Chia-Tung Kuo and Ian Davidson. 2016. On e Equivalence of Tries and Dendrograms - Ecient Hierarchical Clustering of Trac Data. In Proceed- Creating Trip Tries. A prex tree is a tree structure built from ings of ACM Conference, Washington, DC, USA, July 2017 (Conference’17), strings where each path from the root to any node corresponds 9 pages. to a unique string prex and is commonly used for indexing and DOI: 10.1145/nnnnnnn.nnnnnnn retrievals of text and symbolic data. A prex tree (trie) can be constructed in linear time to both number of trips, n and maximum 1 INTRODUCTION number of discretized time steps, l. An example of such a tree is shown in Figure 2. Location tracking devices have become widely popular over the Uses of Trip Tries. A trip trie is not only a hierarchical clus- last decade and this has enabled the collection of large amounts tering (as we shall see) but has other uses. For example, easy to of spatial temporal trip data. Given a collection of such trip data, understand visualizations of a collection of trips such as heat maps clustering is oen a natural start to explore the general properties (see Figures 4 and ) can be created from a trip trie; and trip tries of the data and among clustering methods, hierarchical clustering constructed from dierent collections of trips can be compared (i.e. arXiv:1810.05357v1 [cs.DB] 12 Oct 2018 is well suited as it provides a set of groupings at dierent levels Table 1). Tries have many useful properties such as the ability to ef- where each grouping at one level is a renement of the groupings ciently compute Levenshtein distances and we describe uses such at the previous levels. is dendrogram structure can also naturally as creating more robust clusters using these properties. ough represent the evolutionary/temporal nature of trip data where each tries are commonly used in the database literature for tasks such as level in the dendrogram corresponds to a particular time. However, retrieval and indexing, to our knowledge they have not been used for one signicant drawback of the standard agglomerative hierarchical the purposes we outlined in this paper. Permission to make digital or hard copies of all or part of this work for personal or Uses Beyond GPS Trip Data. In this paper we have focused classroom use is granted without fee provided that copies are not made or distributed on GPS trip data as the application domain is important and has for prot or commercial advantage and that copies bear this notice and the full citation on the rst page. Copyrights for components of this work owned by others than ACM readily available public data. However, our work is applicable in must be honored. Abstracting with credit is permied. To copy otherwise, or republish, other domains where the data represents behavior over time such as to post on servers or to redistribute to lists, requires prior specic permission and/or a seings where some categorical event (a symbol) occurs over time fee. Request permissions from [email protected]. Conference’17, Washington, DC, USA (the position of the symbol). In our earlier work [8] we modeled © 2016 ACM. 978-x-xxxx-xxxx-x/YY/MM...$15.00 behavioral data as these event strings but other applications exist DOI: 10.1145/nnnnnnn.nnnnnnn Conference’17, July 2017, Washington, DC, USA Chia-Tung Kuo and Ian Davidson x 104 12 in areas such as computer network trac where each location is ????? 10 8 an IP address. z z 6 1 9 4 2 Our contributions can be summarized as follows. 0 0 5 10 15 20 25 30 35 • We provide a novel way to organize trip data into symbolic z1???? ::: z9???? data and then a prex tree/trie (see section 2). Tries can be z2 z5 built and updated in linear time to the number of instances : and alphabet size. z1z2??? ::: z1z5??? : • We derive the equivalence between a prex tree and a : dendrogram and verify it empirically (see eorem 3.1 and derivation in section 3). Figure 2: An example of a trip trie. Each node corresponds to • We discuss extensions of our work including uses beyond a prex which matches a subset of the trips with the symbol, hierarchical clustering such as outlier detection (see section “?” meaning unknown as yet. e root of the tree contains 4). all unknown and denotes the beginning of a trip, the next • We demonstrate the usefulness of our dendrogram in illus- level time step 1 and so on. e user can store additional trating interesting insights of trip data on a real data set of relevant data about the trips at each node, such as the trip GPS traces of taxis (see section 5). durations, as shown in the histogram. Our paper is organized as follows. We describe the steps to create string representations and a trip trie in section 2, which is then followed by a proof of its equivalence to standard agglomera- irregular time intervals and later in experiments we will use inter- tive hierarchical clustering in section 3. We further discuss other polation/extrapolation. Our approach consists of three major steps, dierent ways this trip trie can be used in section 4. In section shown in corresponding order in Figure 3(a), 3(b) and 3(c). 5 we evaluate our approach on a real world dataset of taxis’ GPS traces and demonstrate the usefulness of our approach in obtaining (1) Discretization of the geographic space: is step breaks insights on the trac dynamics. We discuss related work in section the modeled space into a nite set of distinct non-overlapping f ; ;:::; g 6 and conclude our paper in section 7. regions Z = z1 z2 zs which we will use as symbols in an alphabet (see Figure 1). is preprocessing is carried out before the major algorithm is applied, just like most work in trajectory mining [11][4]. In our experiment we use equal-sized rectangular grids over the modeled space; this allows straightforward mapping between actual spatial coordinates to the symbols; however our method can be used with any shaped regions. It is worth noting that the actual number of regions with activity is typically much smaller than the possible number of grid cells due to physi- cal presence of roads. is is a desirable property allowing large geographic areas to be eciently represented. For example, in our experiments, though we discretize the San Francisco Bay area into 10,000 cells, less than 20% of them Figure 1: An example that shows the construction of a tra- see any activity. jectory string from the given spatial grids (in black) and tem- (2) Build trajectory strings: In this step we build a trajec- poral resolution.
Recommended publications
  • Headprint: Detecting Anomalous Communications Through Header-Based Application Fingerprinting
    HeadPrint: Detecting Anomalous Communications through Header-based Application Fingerprinting Riccardo Bortolameotti Thijs van Ede Andrea Continella University of Twente University of Twente UC Santa Barbara [email protected] [email protected] [email protected] Thomas Hupperich Maarten H. Everts Reza Rafati University of Muenster University of Twente Bitdefender [email protected] [email protected] [email protected] Willem Jonker Pieter Hartel Andreas Peter University of Twente Delft University of Technology University of Twente [email protected] [email protected] [email protected] ABSTRACT 1 INTRODUCTION Passive application fingerprinting is a technique to detect anoma- Data breaches are a major concern for enterprises across all indus- lous outgoing connections. By monitoring the network traffic, a tries due to the severe financial damage they cause [25]. Although security monitor passively learns the network characteristics of the many enterprises have full visibility of their network traffic (e.g., applications installed on each machine, and uses them to detect the using TLS-proxies) and they stop many cyber attacks by inspect- presence of new applications (e.g., malware infection). ing traffic with different security solutions (e.g., IDS, IPS, Next- In this work, we propose HeadPrint, a novel passive fingerprint- Generation Firewalls), attackers are still capable of successfully ing approach that relies only on two orthogonal network header exfiltrating data. Data exfiltration often occurs over network covert characteristics to distinguish applications, namely the order of the channels, which are communication channels established by attack- headers and their associated values. Our approach automatically ers to hide the transfer of data from a security monitor.
    [Show full text]
  • Author Guidelines for 8
    I.J. Information Technology and Computer Science, 2014, 01, 68-75 Published Online December 2013 in MECS (http://www.mecs-press.org/) DOI: 10.5815/ijitcs.2014.01.08 FSM Circuits Design for Approximate String Matching in Hardware Based Network Intrusion Detection Systems Dejan Georgiev Faculty of Electrical Engineering and Information Technologies , Skopje, Macedonia E-mail: [email protected] Aristotel Tentov Faculty of Electrical Engineering and Information Technologies, Skopje, Macedonia E-mail: [email protected] Abstract— In this paper we present a logical circuits to perform a check to a possible variations of the design for approximate content matching implemented content because of so named "evasion" techniques. as finite state machines (FSM). As network speed Software systems have processing limitations in that increases the software based network intrusion regards. On the other hand hardware platforms, like detection and prevention systems (NIDPS) are lagging FPGA offer efficient implementation for high speed behind requirements in throughput of so called deep network traffic. package inspection - the most exhaustive process of Excluding the regular expressions as special case ,the finding a pattern in package payloads. Therefore, there problem of approximate content matching basically is a is a demand for hardware implementation. k-differentiate problem of finding a content Approximate content matching is a special case of content finding and variations detection used by C c1c2.....cm in a string T t1t2.......tn such as the "evasion" techniques. In this research we will enhance distance D(C,X) of content C and all the occurrences of the k-differentiate problem with "ability" to detect a a substring X is less or equal to k , where k m n .
    [Show full text]
  • Comparison of Levenshtein Distance Algorithm and Needleman-Wunsch Distance Algorithm for String Matching
    COMPARISON OF LEVENSHTEIN DISTANCE ALGORITHM AND NEEDLEMAN-WUNSCH DISTANCE ALGORITHM FOR STRING MATCHING KHIN MOE MYINT AUNG M.C.Sc. JANUARY2019 COMPARISON OF LEVENSHTEIN DISTANCE ALGORITHM AND NEEDLEMAN-WUNSCH DISTANCE ALGORITHM FOR STRING MATCHING BY KHIN MOE MYINT AUNG B.C.Sc. (Hons:) A Dissertation Submitted in Partial Fulfillment of the Requirements for the Degree of Master of Computer Science (M.C.Sc.) University of Computer Studies, Yangon JANUARY2019 ii ACKNOWLEDGEMENTS I would like to take this opportunity to express my sincere thanks to those who helped me with various aspects of conducting research and writing this thesis. Firstly, I would like to express my gratitude and my sincere thanks to Prof. Dr. Mie Mie Thet Thwin, Rector of the University of Computer Studies, Yangon, for allowing me to develop this thesis. Secondly, my heartfelt thanks and respect go to Dr. Thi Thi Soe Nyunt, Professor and Head of Faculty of Computer Science, University of Computer Studies, Yangon, for her invaluable guidance and administrative support, throughout the development of the thesis. I am deeply thankful to my supervisor, Dr. Ah Nge Htwe, Associate Professor, Faculty of Computer Science, University of Computer Studies, Yangon, for her invaluable guidance, encouragement, superior suggestion and supervision on the accomplishment of this thesis. I would like to thank Daw Ni Ni San, Lecturer, Language Department of the University of Computer Studies, Yangon, for editing my thesis from the language point of view. I thank all my teachers who taught and helped me during the period of studies in the University of Computer Studies, Yangon. Finally, I am grateful for all my teachers who have kindly taught me and my friends from the University of Computer Studies, Yangon, for their advice, their co- operation and help for the completion of thesis.
    [Show full text]
  • Automating Error Frequency Analysis Via the Phonemic Edit Distance Ratio
    JSLHR Research Note Automating Error Frequency Analysis via the Phonemic Edit Distance Ratio Michael Smith,a Kevin T. Cunningham,a and Katarina L. Haleya Purpose: Many communication disorders result in speech corresponds to percent segments with these phonemic sound errors that listeners perceive as phonemic errors. errors. Unfortunately, manual methods for calculating phonemic Results: Convergent validity between the manually calculated error frequency are prohibitively time consuming to use in error frequency and the automated edit distance ratio was large-scale research and busy clinical settings. The purpose excellent, as demonstrated by nearly perfect correlations of this study was to validate an automated analysis based and negligible mean differences. The results were replicated on a string metric—the unweighted Levenshtein edit across 2 speech samples and 2 computation applications. distance—to express phonemic error frequency after left Conclusions: The phonemic edit distance ratio is well hemisphere stroke. suited to estimate phonemic error rate and proves itself Method: Audio-recorded speech samples from 65 speakers for immediate application to research and clinical practice. who repeated single words after a clinician were transcribed It can be calculated from any paired strings of transcription phonetically. By comparing these transcriptions to the symbols and requires no additional steps, algorithms, or target, we calculated the percent segments with a combination judgment concerning alignment between target and production. of phonemic substitutions, additions, and omissions and We recommend it as a valid, simple, and efficient substitute derived the phonemic edit distance ratio, which theoretically for manual calculation of error frequency. n the context of communication disorders that affect difference between a series of targets and a series of pro- speech production, sound error frequency is often ductions.
    [Show full text]
  • APPROXIMATE STRING MATCHING METHODS for DUPLICATE DETECTION and CLUSTERING TASKS by Oleksandr Rudniy
    Copyright Warning & Restrictions The copyright law of the United States (Title 17, United States Code) governs the making of photocopies or other reproductions of copyrighted material. Under certain conditions specified in the law, libraries and archives are authorized to furnish a photocopy or other reproduction. One of these specified conditions is that the photocopy or reproduction is not to be “used for any purpose other than private study, scholarship, or research.” If a, user makes a request for, or later uses, a photocopy or reproduction for purposes in excess of “fair use” that user may be liable for copyright infringement, This institution reserves the right to refuse to accept a copying order if, in its judgment, fulfillment of the order would involve violation of copyright law. Please Note: The author retains the copyright while the New Jersey Institute of Technology reserves the right to distribute this thesis or dissertation Printing note: If you do not wish to print this page, then select “Pages from: first page # to: last page #” on the print dialog screen The Van Houten library has removed some of the personal information and all signatures from the approval page and biographical sketches of theses and dissertations in order to protect the identity of NJIT graduates and faculty. ABSTRACT APPROXIMATE STRING MATCHING METHODS FOR DUPLICATE DETECTION AND CLUSTERING TASKS by Oleksandr Rudniy Approximate string matching methods are utilized by a vast number of duplicate detection and clustering applications in various knowledge domains. The application area is expected to grow due to the recent significant increase in the amount of digital data and knowledge sources.
    [Show full text]
  • Finite State Recognizer and String Similarity Based Spelling
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by BRAC University Institutional Repository FINITE STATE RECOGNIZER AND STRING SIMILARITY BASED SPELLING CHECKER FOR BANGLA Submitted to A Thesis The Department of Computer Science and Engineering of BRAC University by Munshi Asadullah In Partial Fulfillment of the Requirements for the Degree of Bachelor of Science in Computer Science and Engineering Fall 2007 BRAC University, Dhaka, Bangladesh 1 DECLARATION I hereby declare that this thesis is based on the results found by me. Materials of work found by other researcher are mentioned by reference. This thesis, neither in whole nor in part, has been previously submitted for any degree. Signature of Supervisor Signature of Author 2 ACKNOWLEDGEMENTS Special thanks to my supervisor Mumit Khan without whom this work would have been very difficult. Thanks to Zahurul Islam for providing all the support that was required for this work. Also special thanks to the members of CRBLP at BRAC University, who has managed to take out some time from their busy schedule to support, help and give feedback on the implementation of this work. 3 Abstract A crucial figure of merit for a spelling checker is not just whether it can detect misspelled words, but also in how it ranks the suggestions for the word. Spelling checker algorithms using edit distance methods tend to produce a large number of possibilities for misspelled words. We propose an alternative approach to checking the spelling of Bangla text that uses a finite state automaton (FSA) to probabilistically create the suggestion list for a misspelled word.
    [Show full text]
  • Adaptive Name Matching in Information Integration
    Information Integration on the Web Adaptive Name Matching in Information Integration Mikhail Bilenko and Raymond Mooney, University of Texas at Austin William Cohen, Pradeep Ravikumar, and Stephen Fienberg, Carnegie Mellon University hen you combine information from heterogeneous information sources, you W must identify data records that refer to equivalent entities. However, records that describe the same object might differ syntactically—for example, the same person can be referred to as “William Jefferson Clinton” and “bill clinton.” Figure 1 presents Identifying more complex examples of duplicate records that are tant in duplicate identification. We examine a few effec- not identical. tive and widely used metrics for measuring similarity. approximately Variations in representation across information sources can arise from differences in formats that Edit distance duplicate database store data, typographical and optical character recog- An important class of such metrics are edit dis- nition (OCR) errors, and abbreviations. Variations tances. Here, the distance between strings s and t is records that refer to are particularly pronounced in data that’s automati- the cost of the best sequence of edit operations that cally extracted from Web pages and unstructured or converts s to t. For example, consider mapping the the same entity is semistructured documents, making the matching task string s = “Willlaim” to t = “William” using these essential for information integration on the Web. edit operations: essential for Researchers have investigated the problem of identifying duplicate objects under several monikers, • Copy the next letter in s to the next position in t. information including record linkage, merge-purge, duplicate • Insert a new letter in t that does not appear in s.
    [Show full text]
  • Form Similarity Via Levenshtein Distance Between Ortho-Filtered Logarithmic Ruling-Gap Ratios, Procs
    G. Nagy, D. Lopresti, Form similarity via Levenshtein distance between ortho-filtered logarithmic ruling-gap ratios, Procs. SPIE/IST Document Recognition and Retrieval, Feb. 201 Form similarity via Levenshtein distance between ortho-filtered logarithmic ruling-gap ratios George Nagya and Daniel Loprestib* a RPI, Troy, NY USA; b Lehigh University, Bethlehem, PA USA ABSTRACT Geometric invariants are combined with edit distance to compare the ruling configuration of noisy filled-out forms. It is shown that gap-ratios used as features capture most of the ruling information of even low-resolution and poorly scanned form images, and that the edit distance is tolerant of missed and spurious rulings. No preprocessing is required and the potentially time-consuming string operations are performed on a sparse representation of the detected rulings. Based on edit distance, 158 Arabic forms are classified into 15 groups with 89% accuracy. Since the method was developed for an application that precludes public dissemination of the data, it is illustrated on public-domain death certificates. Keywords: Office forms, form identification, rulings, edit distance 1. INTRODUCTION We propose a translation-, scale- and rotation-invariant metric for classifying forms. The method is independent of language and script. It is remarkably immune to common types of document degradation, especially those incurred in copying and scanning. Even the Matlab implementation is fast enough for production runs. Our program accepts a set of scanned form images and produces a matrix of pairwise similarity measures based on the configuration of detected rulings. Geometric invariance is achieved by using ratios of the horizontal and vertical distances between rulings.
    [Show full text]
  • Using String Metrics to Improve the Design of Virtual Conversational Characters: Behavior Simulator Development Study
    JMIR SERIOUS GAMES García-Carbajal et al Original Paper Using String Metrics to Improve the Design of Virtual Conversational Characters: Behavior Simulator Development Study Santiago García-Carbajal1*, PhD; María Pipa-Muniz2*, MSc; Jose Luis Múgica3*, MSc 1Computer Science Department, Universidad de Oviedo, Gijón, Spain 2Cabueñes Hospital, Gijón, Spain 3Signal Software SL, Parque Científico Tecnológico de Gijón, Gijón, Asturias, Spain *all authors contributed equally Corresponding Author: Santiago García-Carbajal, PhD Computer Science Department Universidad de Oviedo Campus de Viesques Office 1 b 15 Gijón, 33203 Spain Phone: 34 985182487 Email: [email protected] Abstract Background: An emergency waiting room is a place where conflicts often arise. Nervous relatives in a hostile, unknown environment force security and medical staff to be ready to deal with some awkward situations. Additionally, it has been said that the medical interview is the first diagnostic and therapeutic tool, involving both intellectual and emotional skills on the part of the doctor. At the same time, it seems that there is something mysterious about interviewing that cannot be formalized or taught. In this context, virtual conversational characters (VCCs) are progressively present in most e-learning environments. Objective: In this study, we propose and develop a modular architecture for a VCC-based behavior simulator to be used as a tool for conflict avoidance training. Our behavior simulators are now being used in hospital environments, where training exercises must be easily designed and tested. Methods: We define training exercises as labeled, directed graphs that help an instructor in the design of complex training situations. In order to increase the perception of talking to a real person, the simulator must deal with a huge number of sentences that a VCC must understand and react to.
    [Show full text]
  • MS Word Template for Final Thesis Report
    Offline Approximate String Matching for Information Retrieval: An experiment on technical documentation Simon Dubois MASTER THESIS 2013 INFORMATICS Postadress: Besöksadress: Telefon: Box 1026 Gjuterigatan 5 036-10 10 00 (vx) 551 11 Jönköping EXAMENSARBETETS TITEL Offline Approximate String Matching for Information Retrieval:An experiment on technical documentation Simon Dubois Detta examensarbete är utfört vid Tekniska Högskolan i Jönköping inom ämnesområdet informatik. Arbetet är ett led i masterutbildningen med inriktning informationsteknik och management. Författarna svarar själva för framförda åsikter, slutsatser och resultat. Handledare: He Tan Examinator: Vladimir Tarasov Omfattning: 30 hp (D-nivå) Datum: 2013 Arkiveringsnummer: Postadress: Besöksadress: Telefon: Box 1026 Gjuterigatan 5 036-10 10 00 (vx) 551 11 Jönköping Abstract Abstract Approximate string matching consists in identifying strings as similar even if there is a number of mismatch between them. This technique is one of the solutions to reduce the exact matching strictness in data comparison. In many cases it is useful to identify stream variation (e.g. audio) or word declension (e.g. prefix, suffix, plural). Approximate string matching can be used to score terms in Information Retrieval (IR) systems. The benefit is to return results even if query terms does not exactly match indexed terms. However, as approximate string matching algorithms only consider characters (nor context neither meaning), there is no guarantee that additional matches are relevant matches. This paper presents the effects of some approximate string matching algorithms on search results in IR systems. An experimental research design has been conducting to evaluate such effects from two perspectives. First, result relevance is analysed with precision and recall.
    [Show full text]
  • Ontology Alignment Based on Word Embedding and Random Forest Classification
    NKISI-ORJI, I., WIRATUNGA, N., MASSIE, S., HUI, K.-Y. and HEAVEN, R. 2019. Ontology alignment based on word embedding and random forest classification. In Berlingerio, M., Bonchi, F., Gärtner, T., Hurley, N. and Ifrim, G. (eds.) Machine learning and knowledge discovery in databases: proceedings of the 2018 European conference on machine learning and principles and practice of knowledge discovery in databases (ECML PKDD 2018), 10-14 September 2018, Dublin, Ireland. Lecture notes in computer science, 11051. Cham: Springer [online], part I, pages 557-572. Available from: https://doi.org/10.1007/978-3-030-10925-7_34 Ontology alignment based on word embedding and random forest classification. NKISI-ORJI, I., WIRATUNGA, N., MASSIE, S., HUI, K.-Y., HEAVEN, R. 2018 This document was downloaded from https://openair.rgu.ac.uk Ontology alignment based on word embedding and random forest classification Ikechukwu Nkisi-Orji1[0000−0001−9734−9978], Nirmalie Wiratunga1, Stewart Massie1, Kit-Ying Hui1, and Rachel Heaven2 1 Robert Gordon University, Aberdeen, UK {i.o.nkisi-orji , n.wiratunga, s.massie, k.hui}@rgu.ac.uk 2 British Geological Survey, Nottingham, UK [email protected] Abstract. Ontology alignment is crucial for integrating heterogeneous data sources and forms an important component for realising the goals of the semantic web. Accordingly, several ontology alignment techniques have been proposed and used for discovering correspondences between the concepts (or entities) of different ontologies. However, these tech- niques mostly depend on string-based similarities which are unable to handle the vocabulary mismatch problem. Also, determining which simi- larity measures to use and how to effectively combine them in alignment systems are challenges that have persisted in this area.
    [Show full text]
  • Dedupe Documentation Release 2.0.0
    dedupe Documentation Release 2.0.0 Forest Gregg, Derek Eder, and contributors Sep 03, 2021 CONTENTS 1 Important links 3 2 Tools built with dedupe 5 3 Contents 7 4 Features 41 5 Installation 43 6 Using dedupe 45 7 Errors / Bugs 47 8 Contributing to dedupe 49 9 Citing dedupe 51 10 Indices and tables 53 Index 55 i ii dedupe Documentation, Release 2.0.0 dedupe is a library that uses machine learning to perform de-duplication and entity resolution quickly on structured data. If you’re looking for the documentation for the Dedupe.io Web API, you can find that here: https://apidocs.dedupe.io/ dedupe will help you: • remove duplicate entries from a spreadsheet of names and addresses • link a list with customer information to another with order history, even without unique customer id’s • take a database of campaign contributions and figure out which ones were made by the same, person even if the names were entered slightly differently for each record dedupe takes in human training data and comes up with the best rules for your dataset to quickly and automatically find similar records, even with very large databases. CONTENTS 1 dedupe Documentation, Release 2.0.0 2 CONTENTS CHAPTER ONE IMPORTANT LINKS • Documentation: https://docs.dedupe.io/ • Repository: https://github.com/dedupeio/dedupe • Issues: https://github.com/dedupeio/dedupe/issues • Mailing list: https://groups.google.com/forum/#!forum/open-source-deduplication • Examples: https://github.com/dedupeio/dedupe-examples • IRC channel, #dedupe on irc.freenode.net 3 dedupe Documentation, Release 2.0.0 4 Chapter 1.
    [Show full text]