A Unified Model for Document-Based Question Answering Based On

Total Page:16

File Type:pdf, Size:1020Kb

A Unified Model for Document-Based Question Answering Based On The Thirty-Second AAAI Conference on Artificial Intelligence (AAAI-18) A Unified Model for Document-Based Question Answering Based on Human-Like Reading Strategy Weikang Li, Wei Li, Yunfang Wu∗ Key Laboratory of Computational Linguistics, Peking University, MOE, China {wavejkd,liweitj47,wuyf}@pku.edu.cn Abstract sentences from document as document summary for the an- swer selection task. Document-based Question Answering (DBQA) in Natural However, the simple transfer between deep learning and Language Processing (NLP) is important but difficult because DBQA is not so logical. In our opinion, DBQA is similar to of the long document and the complex question. Most of pre- vious deep learning methods mainly focus on the similarity the reading comprehension test, which is defined to test peo- computation between two sentences. However, DBQA stems ple’s comprehension of a document. In their school years, from the reading comprehension in some degree, which is students would do lots of reading comprehension tests. In originally used to train and test people’s ability of reading this paper, we provide a solution for the problem to stimu- and logical thinking. Inspired by the strategy of doing read- late men’s reading strategy of doing the tests. With the as- ing comprehension tests, we propose a unified model based sumption, the detailed reading strategy is as follow: on the human-like reading strategy. The unified model con- tains three major encoding layers that are consistent to differ- • 1. Go over the document quickly to get a general under- ent steps of the reading strategy, including the basic encoder, standing of the document; combined encoder and hierarchical encoder. We conduct ex- tensive experiments on both the English WikiQA dataset and • 2. Read the question carefully equipped with the general the Chinese dataset, and the experimental results show that understanding of the document; our unified model is effective and yields state-of-the-art re- • sults on WikiQA dataset. 3. Go back to the document with the prior knowledge of question and get the right answer. Introduction Such a reading strategy could be implemented by neural net- work models. Document-based Question Answering (DBQA) is an impor- As we know, the document in reading comprehension tant issue in natural language processing(NLP). Given a doc- tests usually has a title, which has an important impact ument and a question related to the document, the system on doing reading comprehension tests for people. Unfortu- is required to give an answer for the question. The answer nately, the title information is neglected by most researches could be a word, a text span or a sentence extracted from the on DBQA. In this paper, we use the title information (a nat- document. Table 1 gives an example of DBQA. Recently, ural document summary) as the general understanding of a more and more researches have focused on this challenging document. As for the document without title, we make many problem. attempts to get the general understanding of the document, A lot of achievements have been achieved via deep learn- by using the first sentence, the last sentence and training a ing models, which obtain better performances than tradi- LDA or LSA model to get the topic of a document. In addi- tional machine learning methods. Inspired by the great suc- tion, we have tried many ways to understand questions well cess of deep learning methods in voice and image recog- given the general understanding of the document. nition, researchers have adopted various ways to solve the At the end, we propose a unified neural network model problem of DBQA, including convolutional neural network according to the human-like reading strategy above. (CNN) (Feng et al. 2015), recurrent neural network (RNN) Our contributions in this paper can be summarized as fol- (Tan et al. 2015), Attention-Way (Seo et al. 2016) and gener- lows: ative adversarial networks (GAN) (Wang et al. 2017). Many other ways have emerged to dig out more information to • We propose a human-like reading strategy for DBQA task solve the problem of DBQA. Document summary could also which is similar to the logic of students when they do the be seen as an effective information in many NLP tasks. Choi test of reading comprehension. et al. (2017) and Miller et al. (2016) used the most related • Based on the reading strategy, we make a good combina- Copyright c 2018, Association for the Advancement of Artificial tion of general understanding of both document and ques- Intelligence (www.aaai.org). All rights reserved. tion. 604 Document Title Uncle Sam Document J. M. Flagg ’s 1917 poster, based on the original British Lord Kitchener poster of three years earlier, was used to recruit soldiers for both World War I and World War II. ...... Un- cle Sam (initials U.S.) is a common national personification of the American government that, according to legend, came into use during the War of 1812 and was supposedly named for Samuel Wilson. It is not clear whether this reference is to Uncle Sam as a metaphor for the United States. ...... Question what does uncle sam represent to the American people ? Answer(Sentence) Uncle Sam (initials U.S.) is a common national personification of the American govern- ment that, according to legend, came into use during the War of 1812 and was supposedly named for Samuel Wilson. Answer(Span) a common national personification of the American government. Answer(Word) national personification. Table 1: An Outline of DBQA • We propose a unified neural network model which is suit- Pang et al. (2016) built hierarchical convolution layers on able for our reading strategy to tackle the problem of the word similarity matrix between sentences, and Yin and DBQA. Schutze¨ (2015) proposed MultiGranCNN to integrate mul- tiple granularity levels of matching models. The Represen- We conduct experiments on the English WikiQA dataset tation (MPSR) model (Wan et al. 2016) employed LSTM (Yang, Yih, and Meek 2015) and the Chinese DBQA dataset and interactive tensor to capture matching features with po- (Duan 2016). On the WikiQA dataset, our model obtains a sitional local context. For both levels of matching strategies, MAP of 0.754, which outperforms the best previous method the ways of computing similarity between two sentences are by 1.1 MAP points. On the Chinese DBQA dataset, our similar. The most popular methods are cosine similarity (Tan model gets comparable results without using any features. et al. 2016), element-wise product (Seo et al. 2016) and ten- sor computation (Bowman et al. 2015). Related Work As a task to train people’s reading and understanding Reading documents and being able to answer related ques- skills, DBQA is more complex, logical and skillful than a tions by machines is a useful and meaningful issue. How- simple comparison of the similarity between two sentences. ever, it is still an unsolved challenge. DBQA has several dif- We will imitate people’s reading strategy of doing reading ferent answer types, as outlined in Table 1. Our work mainly comprehension tests via the neural network. focuses on the form in which answer is a whole sentence. Attempts have also been made to study how people read. As we know, many NLP problems involve matching two Masson (1983) conducted studies on how people answer or more sequences to make a decision. For DBQA, some re- questions by first skimming the document, identifying rel- searches also see this problem as matching two sequences evant parts, and carefully reading these parts to obtain an (question and candidate answer) to decide whether a sen- answer. Inspired by this observation, Golub et al. (2017) pro- tence from the document could answer the question. posed a coarse-to-fine model for question answering. It first In the field of sentence pairs matching, there have been selects relevant sentences and then generates an answer. Dif- various deep neural network models proposed. Two levels ferent from their method, our work mainly focuses on peo- of matching strategies are considered: the first is converting ple’s reading strategy when doing reading comprehension the whole source and target sentence into embedding vectors tests. of latent semantic spaces respectively, and then calculating Titles could be naturally used to obtain a general under- similarity score between them; the second is calculating the standing of documents in people’s reading, and summariza- similarity score among all possible local positions of source tion offers another way to get the general meaning of a doc- and target sentences, and then summarizing the local scores ument in the absence of a title. Usually, there are two ways into the final similarity score. to automatically summarize a document, including extrac- Works using the first strategy include bag of words based tive summarization and abstractive summarization. There methods (Wang et al. 2011) and CNN model (Arc-I) (Hu et has been an increase interest in document summarization al. 2014). Qiu and Huang (2015) applied a tensor transfor- over the years. mation layer on CNN based embeddings to capture the inter- Extractive summarization works with the method of find- actions between question and answer more effectively. Long ing the salient sentences in a document. The research of IBM short-term memory (LSTM) network model (Palangi et al. laboratory (Luhn 1958) worked on the frequency of words 2016) are also explored in this problem. Works using the in the text. Edmundson (1969) used the title words, core second strategy include DeepMatch (Lu and Li 2013) which phrases, key concepts, position method, which are the sur- incorporated latent topics to make the local matching struc- face level information. Gillick (2011) employed a classifica- ture sparse, Arc-II (Hu et al.
Recommended publications
  • Shakespeare in the Eighteenth Century: Algorithm for Quotation Identification
    University of Arkansas, Fayetteville ScholarWorks@UARK Theses and Dissertations 5-2020 Shakespeare in the Eighteenth Century: Algorithm for Quotation Identification Marion Pauline Chiariglione University of Arkansas, Fayetteville Follow this and additional works at: https://scholarworks.uark.edu/etd Part of the Numerical Analysis and Scientific Computing Commons, and the Theory and Algorithms Commons Citation Chiariglione, M. P. (2020). Shakespeare in the Eighteenth Century: Algorithm for Quotation Identification. Theses and Dissertations Retrieved from https://scholarworks.uark.edu/etd/3580 This Thesis is brought to you for free and open access by ScholarWorks@UARK. It has been accepted for inclusion in Theses and Dissertations by an authorized administrator of ScholarWorks@UARK. For more information, please contact [email protected]. Shakespeare in the Eighteenth Century: Algorithm for Quotation Identification A thesis submitted in partial fulfillment of the requirements for the degree of Master of Science in Computer Science by Marion Pauline Chiariglione IUT Dijon, University of Burgundy Bachelor of Science in Computer Science, 2017 May 2020 University of Arkansas This thesis is approved for recommendation to the Graduate Council Susan Gauch, Ph.D. Thesis Director Qinghua Li, Ph.D. Committee member Khoa Luu, Ph.D. Committee member Abstract Quoting a borrowed excerpt of text within another literary work was infrequently done prior to the beginning of the eighteenth century. However, quoting other texts, particularly Shakespeare, became quite common after that. Our work develops automatic approaches to identify that trend. Initial work focuses on identifying exact and modified sections of texts taken from works of Shakespeare in novels spanning the eighteenth century. We then introduce a novel approach to identifying modified quotes by adapting the Edit Distance metric, which is character based, to a word based approach.
    [Show full text]
  • Using N-Grams to Understand the Nature of Summaries
    Using N-Grams to Understand the Nature of Summaries Michele Banko and Lucy Vanderwende One Microsoft Way Redmond, WA 98052 {mbanko, lucyv}@microsoft.com views of the event being described over different Abstract documents, or present a high-level view of an event that is not explicitly reflected in any single document. A Although single-document summarization is a useful multi-document summary may also indicate the well-studied task, the nature of multi- presence of new or distinct information contained within document summarization is only beginning to a set of documents describing the same topic (McKeown be studied in detail. While close attention has et. al., 1999, Mani and Bloedorn, 1999). To meet these been paid to what technologies are necessary expectations, a multi-document summary is required to when moving from single to multi-document generalize, condense and merge information coming summarization, the properties of human- from multiple sources. written multi-document summaries have not Although single-document summarization is a well- been quantified. In this paper, we empirically studied task (see Mani and Maybury, 1999 for an characterize human-written summaries overview), multi-document summarization is only provided in a widely used summarization recently being studied closely (Marcu & Gerber 2001). corpus by attempting to answer the questions: While close attention has been paid to multi-document Can multi-document summaries that are summarization technologies (Barzilay et al. 2002, written by humans be characterized as Goldstein et al 2000), the inherent properties of human- extractive or generative? Are multi-document written multi-document summaries have not yet been summaries less extractive than single- quantified.
    [Show full text]
  • Multi-Document Biography Summarization
    Multi-document Biography Summarization Liang Zhou, Miruna Ticrea, Eduard Hovy University of Southern California Information Sciences Institute 4676 Admiralty Way Marina del Rey, CA 90292-6695 {liangz, miruna, hovy} @isi.edu Abstract In this paper we describe a biography summarization system using sentence classification and ideas from information retrieval. Although the individual techniques are not new, assembling and applying them to generate multi-document biographies is new. Our system was evaluated in DUC2004. It is among the top performers in task 5–short summaries focused by person questions. 1 Introduction Automatic text summarization is one form of information management. It is described as selecting a subset of sentences from a document that is in size a small percentage of the original and Figure 1. Overall design of the biography yet is just as informative. Summaries can serve as summarization system. surrogates of the full texts in the context of To determine what and how sentences are Information Retrieval (IR). Summaries are created selected and ranked, a simple IR method and from two types of text sources, a single document experimental classification methods both or a set of documents. Multi-document contributed. The set of top-scoring sentences, after summarization (MDS) is a natural and more redundancy removal, is the resulting biography. elaborative extension of the single-document As yet, the system contains no inter-sentence summarization, and poses additional difficulties on ‘smoothing’ stage. algorithm design. Various kinds of summaries fall In this paper, work in related areas is discussed into two broad categories: generic summaries are in Section 2; a description of our biography corpus the direct derivatives of the source texts; special- used for training and testing the classification interest summaries are generated in response to component is in Section 3; Section 4 explains the queries or topic-oriented questions.
    [Show full text]
  • Quesgen Using Nlp 01
    Quesgen Using Nlp 01 International Journal of Latest Trends in Engineering and Technology Vol.(13)Issue(2), pp.009-014 DOI: http://dx.doi.org/10.21172/1.132.02 e-ISSN:2278-621X QUESGEN USING NLP Pawan NGP1, Pooja Bahuguni2, Pooja Dattatri3, Shilpi Kumari4, Vikranth B.M5 Abstract— when people read for long hours, they seldom are able to grasp concepts and it gives them false sense of understanding it. The aim of this project is to tackle this problem by processing given text and generating applicable questions and answer. The steps followed are: 1. Candidate key sentences are selected (using Text Rank). 2. Candidate key words are selected from candidate key sentences (RAKE). 3. These selected key sentences and words are stored in the database (MongoDB) and presented to the user through chatbot interface. Keywords— NLP, NLP toolkit, Sentence extraction, Keyword extraction, ChatBot, RAKE, TextRank 1. INTRODUCTION Humans are the most curious by nature. Asking Questions to meet their never-ending quest for information and knowledge. For Example,teachers ask students, questions to evaluate performance of the students, pupils learn by asking questions to teachers,and even our normal life conversation consists of asking questions. Questions are the major part of countless learning interactions. However, with the advent of technology, attention spans of individuals have significantly gone down and they are not able to ask good questions. It has been noticed that when people try to read for long hours, they seldom are able to grasp concepts. But having spent some time reading gives people a false sense of understanding it.
    [Show full text]
  • An Automatic Text Summarization for Malayalam Using Sentence Extraction
    International Journal of Advanced Computational Engineering and Networking, ISSN: 2320-2106, Volume-3, Issue-8, Aug.-2015 AN AUTOMATIC TEXT SUMMARIZATION FOR MALAYALAM USING SENTENCE EXTRACTION 1RENJITH S R, 2SONY P 1M.Tech Computer and Information Science, Dept.of Computer Science, College of Engineering Cherthala Kerala, India-688541 2Assistant Professor, Dept. of Computer Science, College of Engineering Cherthala, Kerala, India-688541 Abstract—Text Summarization is the process of generating a short summary for the document that contains the significant portion of information. In an automatic text summarization process, a text is given to the computer and the computer returns a shorter less redundant extract of the original text. The proposed method is a sentence extraction based single document text summarization which produces a generic summary for a Malayalam document. Sentences are ranked based on feature scores and Googles PageRank formula. Top k ranked sentences will be included in summary where k depends on the compression ratio between original text and summary. Performance evaluation will be done by comparing the summarization outputs with manual summaries generated by human evaluators. Keywords—Text summarization, Sentence Extraction, Stemming, TF-ISF score, Sentence similarity, PageRank formula, Summary generation. I. INTRODUCTION a summary, which represents the subject matter of an article by understanding the whole meaning, which With enormous growth of information on cyberspace, are generated by reformulating the salient unit conventional Information Retrieval techniques have selected from an input sentences. It may contain some become inefficient for finding relevant information text units which are not present in the input text. An effectively. When we give a keyword to be searched extract is a summary consisting of a number of on the internet, it returns thousands of documents sentences selected from the input text.Sentence overwhelming the user.
    [Show full text]
  • History of Fairfax City
    HISTORY OF FAIRFAX CITY The City of Fairfax began as the Town of Providence in 1805, a community built around the Fairfax County Courthouse. Completed in 1800 at the corner of Little River Turnpike and Ox Road, the area was a crossroads of conflict during the American Civil War with hardships and disrupted lives for everyone. From a crossroads of conflict, the area became a crossroads of commerce in the late nineteenth century when the dairy industry propelled economic rebirth and the building of schools, churches, homes, barns, and businesses and in 1874 the Town of Providence officially became the Town of Fairfax. The early 20th century ushered in a myriad of technological and transportation changes and the emergence of civic organizations, sports clubs, a Town police unit, and a volunteer fire company. World War II spurred rapid growth across the region in housing, business ventures, and population and Fairfax quickly changed from a rural to a suburban community. The Town of Fairfax deeded a 150-acre tract of land in 1959 to the University of Virginia to establish a permanent home for what is now George Mason University. In 1961, the Town of Fairfax was incorporated as the independent City of Fairfax and in 1962 a new City Hall was completed. Rich in history and heritage, residents and visitors enjoy a small-town atmosphere and an abundance of cultural and recreational pursuits in the midst of a bustling metropolitan area. As the City's first mayor, John C. Wood said in 1962 - "Fairfax has a wonderful past and present and an even greater future." HOW DID THE JULY 4TH CELEBRATION BEGIN IN FAIRFAX CITY? Fairfax City’s Independence Day Parade and Fireworks began in 1967 and was organized by the Delta Alpha Chapter of Beta Sigma Phi Sorority.
    [Show full text]
  • Automatic Document Summarization by Sentence Extraction
    Вычислительные технологии Том 12, № 5, 2007 AUTOMATIC DOCUMENT SUMMARIZATION BY SENTENCE EXTRACTION R. M. Aliguliyev Institute of Information Technology of the National Academy of Sciences of Azerbaijan, Baku e-mail: [email protected] Представлен метод автоматического реферирования документов, который ге- нерирует резюме документа путем кластеризации и извлечения предложений из исходного документа. Преимущество предложенного подхода в том, что сгенери- рованное резюме документа может включать основное содержание практически всех тем, представленных в документе. Для определения оптимального числа кла- стеров введен критерий оценки качества кластеризации. Introduction Automatic document processing is a research field that is currently extremely active. One important task in this field is automatic document summarization, which preserves its information content [1]. With a large volume of text documents, presenting the user with a summary of each document greatly facilitates the task of finding the desired documents. Text search and summarization are the two essential technologies that complement each other. Text search engines return a set of documents that seem to be relevant to the user’s query, and text enable quick examinations through the returned documents. In general, automatic document summarization takes a source document (or source documents) as input, extracts the essence of the source(s), and presents a well-formed summary to the user. Mani and Maybury [1] formally defined automatic document summari- zation as the process of distilling the most important information from a source(s) to produce an abridged version for a particular user (or users) and task (or tasks). The process can be decomposed into three phases: analysis, transformation and synthesis. The analysis phase analyzes the input document and selects a few salient features.
    [Show full text]
  • The Elements of Automatic Summarization
    The Elements of Automatic Summarization by Daniel Jacob Gillick A dissertation submitted in partial satisfaction of the requirements for the degree of in Computer Science in the GRADUATE DIVISION of the UNIVERSITY OF CALIFORNIA, BERKELEY Committee in charge: Professor Nelson Morgan, Chair Professor Daniel Klein Professor Thomas Griffiths Spring 2011 The Elements of Automatic Summarization Copyright © 2011 by Daniel Jacob Gillick Abstract The Elements of Automatic Summarization by Daniel Jacob Gillick Doctor of Philosophy in Computer Science University of California, Berkeley Professor Nelson Morgan, Chair This thesis is about automatic summarization, with experimental results on multi- document news topics: how to choose a series of sentences that best represents a col- lection of articles about one topic. I describe prior work and my own improvements on each component of a summarization system, including preprocessing, sentence valuation, sentence selection and compression, sentence ordering, and evaluation of summaries. The centerpiece of this work is an objective function for summariza- tion that I call "maximum coverage". The intuition is that a good summary covers as many possible important facts or concepts in the original documents. It turns out that this objective, while computationally intractable in general, can be solved efficiently for medium-sized problems and has reasonably good fast approximate so- lutions. Most importantly, the use of an objective function marks a departure from previous algorithmic approaches to summarization. 1 Acknowledgements Getting a Ph.D. is hard. Not really hard in the day-to-day sense, but more because I spent a lot of time working on a few small problems, and at the end of six years, I have only made a few small contributions to the world.
    [Show full text]
  • A Query Focused Multi Document Automatic Summarization
    PACLIC 24 Proceedings 545 A Query Focused Multi Document Automatic Summarization Pinaki Bhaskar and Sivaji Bandyopadhyay Department of Computer Science & Engineering, Jadavpur University, Kolkata – 700032, India [email protected], [email protected] Abstract. The present paper describes the development of a query focused multi-document automatic summarization. A graph is constructed, where the nodes are sentences of the documents and edge scores reflect the correlation measure between the nodes. The system clusters similar texts having related topical features from the graph using edge scores. Next, query dependent weights for each sentence are added to the edge score of the sentence and accumulated with the corresponding cluster score. Top ranked sentence of each cluster is identified and compressed using a dependency parser. The compressed sentences are included in the output summary. The inter-document cluster is revisited in order until the length of the summary is less than the maximum limit. The summarizer has been tested on the standard TAC 2008 test data sets of the Update Summarization Track. Evaluation of the summarizer yielded accuracy scores of 0.10317 (ROUGE-2) and 0.13998 (ROUGE–SU-4). Keywords: Multi Document Summarizer, Query Focused, Cluster based approach, Parsed and Compressed Sentences, ROUGE Evaluation. 1 Introduction Text Summarization, as the process of identifying the most salient information in a document or set of documents (for multi document summarization) and conveying it in less space, became an active field of research in both Information Retrieval (IR) and Natural Language Processing (NLP) communities. Summarization shares some basic techniques with indexing as both are concerned with identification of the essence of a document.
    [Show full text]
  • Automatic Summarization of Medical Conversations, a Review
    Jessica López Espejel Automatic summarization of medical conversations, a review Jessica López Espejel 1, 2 (1) CEA, LIST, DIASI, F-91191 Gif-sur-Yvette, France. (2) Paris 13 University, LIPN, 93430 Villateneuse, France. [email protected] RÉSUMÉ L’analyse de la conversation joue un rôle important dans le développement d’appareils de simulation pour la formation des professionnels de la santé (médecins, infirmières). Notre objectif est de développer une méthode de synthèse automatique originale pour les conversations médicales entre un patient et un professionnel de la santé, basée sur les avancées récentes en matière de synthèse à l’aide de réseaux de neurones convolutionnels et récurrents. La méthode proposée doit être adaptée aux problèmes spécifiques liés à la synthèse des dialogues. Cet article présente une revue des différentes méthodes pour les résumés par extraction et par abstraction et pour l’analyse du dialogue. Nous décrivons aussi les utilisation du Traitement Automatique des Langues dans le domaine médical. ABSTRACT Conversational analysis plays an important role in the development of simulation devices for the training of health professionals (doctors, nurses). Our goal is to develop an original automatic synthesis method for medical conversations between a patient and a healthcare professional, based on recent advances in summarization using convolutional and recurrent neural networks. The proposed method must be adapted to the specific problems related to the synthesis of dialogues. This article presents a review of the different methods for extractive and abstractive summarization, and for dialogue analysis. We also describe the use of Natural Language Processing in the medical field. MOTS-CLÉS : Résumé automatique, domaine médical, dialogues, revue.
    [Show full text]
  • A Hybrid Approach for Multi-Document Text Summarization
    San Jose State University SJSU ScholarWorks Master's Projects Master's Theses and Graduate Research Fall 12-11-2019 A Hybrid Approach for Multi-document Text Summarization Rashmi Varma Follow this and additional works at: https://scholarworks.sjsu.edu/etd_projects Part of the Artificial Intelligence and Robotics Commons, and the Databases and Information Systems Commons A Hybrid Approach for Multi-document Text Summarization A Project Presented to The Faculty of the Department of Computer Science San Jose State University In Partial Fulfillment of the Requirements for the Degree Master of Science by Rashmi Varma December 2019 ○c 2019 Rashmi Varma ALL RIGHTS RESERVED The Designated Project Committee Approves the Project Titled A Hybrid Approach for Multi-document Text Summarization by Rashmi Varma APPROVED FOR THE DEPARTMENT OF COMPUTER SCIENCE SAN JOSE STATE UNIVERSITY December 2019 Dr. Robert Chun Department of Computer Science Dr. Katerina Potika Department of Computer Science Ms. Manasi Thakur Tuutkia Inc. ABSTRACT A Hybrid Approach for Multi-document Text Summarization by Rashmi Varma Text summarization has been a long studied topic in the field of natural language processing. There have been various approaches for both extractive text summarization as well as abstractive text summarization. Summarizing texts for a single document is a methodical task. But summarizing multiple documents poses as a greater challenge. This thesis explores the application of Latent Semantic Analysis, Text-Rank, Lex-Rank and Reduction algorithms for single document text summarization and compares it with the proposed approach of creating a hybrid system combining each of the above algorithms, individually, with Restricted Boltzmann Machines for multi-document text summarization and analyzing how all the approaches perform.
    [Show full text]
  • Cultural Icons
    http://en.wikipedia.org/wiki/UK_topics#Cultural_icons Cultural icons Bagpipes Bagpipes are a class of musical instrument, aerophones using enclosed reeds. The term is equally correct in the singular or plural, although pipers most commonly talk of "pipes" and "the bagpipe". Skirl is a term used by pipers to describe an unintended shrill sound made by the chanter, and is usually produced when the chanter reed is too easy and thus the chanter is overblown. Sometimes the term is also somewhat mistakenly used to describe the general sound produced by a bagpipe. The history of the bagpipe is very unclear. However, it seems likely they were first invented in pre-Christian times. Nero is generally accepted to have been a player; there are Greek depictions of pipers, and the Roman legions are thought to have marched to bagpipes. The idea of taking a leather bag and combining it with a chanter and inflation device seems to have originated with various ethnic groups in the Roman empire. In the modern era the use of bagpipes has become a common tradition for military funerals and memorials in the anglophone world, and they are often used at the funerals of high-ranking civilian public officials as well. http://en.wikipedia.org/wiki/UK_topics#Cultural_icons Big Ben Big Ben is the colloquial name of the Clock Tower of the Palace of Westminster in London and an informal name for the Great Bell of Westminster, part of the Great Clock of Westminster. The clock tower is located at the northwestern end of the building, the home of the Houses of Parliament, and contains the famous striking clock and bell.
    [Show full text]