<<

Toward Automatic Processing of English Metalanguage

Shomir Wilson* School of Informatics School of Computer Science University of Edinburgh Carnegie Mellon University 10 Crichton Street 5000 Forbes Avenue Edinburgh EH8 9AB, UK Pittsburgh, PA 15213 USA [email protected]

Conventional stylistic cues, such as italics in (1), Abstract (2), and (3) and quotation marks in (4), some- times help the audience to recognize metalinguis- The metalinguistic facilities of natural lan- tic statements in written . In spoken lan- guage are crucial to our ability to communi- guage or in written contexts where stylistic cues cate, but the patterns behind the appearance of are not used, the audience is expected to identify metalanguage—and thus the clues for how we metalinguistic statements using paralinguistic may instruct computers to detect it—have re- cues (such as intonation, when speaking) or con- mained relatively unknown. This paper de- scribes the first results on the feasibility of au- text and . tomatically identifying metalanguage in Eng- Metalanguage is both pervasive and, paradox- lish text. A core metalinguistic vocabulary has ically, the subject of limited attention in research been identified, supporting intuitions about the on language technologies. The ability to produce phenomenon and aiding in its detection and and understand metalanguage is a core linguistic delineation. These results open the door to ap- competence that allows humans to converse flex- plications that can extract the direct, salient in- ibly, unrestricted by domain (Anderson et al., formation that metalanguage encodes. 2002). Humans use it to establish grounding, ver- ify audience understanding, and maintain com- 1 Introduction munication channels (Anderson et al., 2004). Metalanguage encodes direct and salient infor- In linguistic communication it is sometimes nec- mation about language, but many typical exam- essary to refer to features of language, such as ples thwart parsers with novel word usage or ar- orthography, vocabulary, structure, pragmatics, rangement (Wilson, 2011a). Metalanguage is or meaning. Metalanguage enables a speaker to difficult to classify through the interpretive lens select a linguistically-relevant referent over (or in of word senses, given that conventional word addition to) other typical referents (Audi, 1995). senses have little relevance when a word appears Metalanguage is illustrated in sentences such as chiefly “as a word”. The roles of metalanguage (1) Graupel refers to a kind of precipitation. in L2 language acquisition (Hu, 2010), expres- (2) The name is actually Rolla. sion of sentiment toward others’ utterances (3) Keep tabs on is a colloquial phrase. (Jaworski et al., 2004), and irony (Sperber and (4) He wrote “All gone” and nothing more. Wilson, 1981) have also been noted. This paper describes the results of the first ef- The roles of the bold substrings in the above sen- fort to automatically identify instances of - tences contrast with those in (5)-(8) below: language in English text. Mentioned language, a (5) Graupel fell on the weary hikers. common variety of metalanguage, is focused up- (6) Rolla is a small town. on for its explicit, direct nature, which makes its (7) Keep tabs on him, will you? structure and meaning easily accessible once an (8) They were all gone when I returned. instance is identified. Section 2 reviews a prior project by Wilson (2012) to create a corpus of instances of metalanguage, a necessary resource * for the present effort. Section 3 describes an ap- This research was performed during a prior affiliation with the University of Maryland at College Park. proach to distinguishing sentences that contain

760 International Joint Conference on Natural Language Processing, pages 760–766, Nagoya, Japan, 14-18 October 2013. metalanguage from those that do not, a task re- language tasks discussed in the introduction. ferred to as detection for brevity. Results show However, other metalinguistic constructions that the performance of this approach roughly draw attention to tokens outside of the referring matches an implied performance ceiling of inter- sentence. Some examples of this are (10)-(12) annotator agreement. Section 4 describes an ap- below. Supporting contexts are not shown for proach to delineate sequences of words that are these sentences, though such contexts are easily directly mentioned by a metalinguistic ; imagined: although the results are preliminary, its accuracy (10) Disregard the last thing I said. shows promise for future development. Together, (11) That spelling, is it correct? these results on detection and delineation show (12) People don’t use those words lightly. the feasibility of enabling language technologies to extract the salient information about language In each of the above three sentences, a linguistic that metalanguage contains. entity (an utterance, a sequence of letters, and a sequence of words, respectively) is referred to, 2 Background but the referent is contained in a separate sen- tence. The referent may have been produced by a The reader is likely to be familiar with the con- different utterer or appeared in a different medi- cept of metalanguage, but a discussion is appro- um (e.g., speaking aloud while referring to writ- priate to ground the and connect to pre- ten text). These “extra-sentential” forms of meta- vious work. Section 2.1 summarizes a prior study language have clear value to understanding dis- (Wilson, 2012) to collect instances of metalan- course and coreference. The focus on mentioned guage, and 2.2 reviews some related efforts. language is a limitation to the present work, to 2.1 Prior Work utilize an existing corpus and to apply tractable boundaries to the identification tasks. A diverse variety of phenomena in natural lan- The mentioned language corpus of the prior guage satisfy the intuitive criteria that we associ- study2 was constructed by filtering a large vol- ate with metalanguage. The prior study focused ume of sentences with a heuristic, followed by on identifying sentences that contained men- annotation by a human reader. A randomly se- tioned language, a phenomenon defined below: lected subset of articles from English Wikipedia Definition: For T a token or a set of tokens in a was chosen as a source for text because of its sentence, if T is produced to draw attention to a representation of a large sample of English writ- of the token T or the type of T, then T is ers (Adler et al., 2008), the rich frequency of an instance of mentioned language.1 mentioned language in its text, and the frequent use of stylistic cues in its text that delimit men- Here, a token is an instantiation of a linguistic tioned language (i.e., bold text, italic text, and entity (e.g., a letter, , sound, word, phrase, quotation marks). Sentences were sought that or other related entity), and a property is an os- contained at least one of these stylistic cues and a tension of language (García-Carpintero, 2004; mention-significant word in close proximity. Saka, 2006), such as spelling, pronunciation, Mention-significant words were a set of 8,735 meaning (for a variety of interpretations of words and collocations with potential metalin- meaning), structure, connotation, or quotative guistic significance (e.g., word, symbol, call), source. Generally attention is drawn to the type extracted from the WordNet lexical ontology of T (for example, in Sentences (1)-(4)), but it (Fellbaum, 1998). Phrases highlighted by the can be drawn to the token of T for self-reference, stylistic cues were considered candidate instanc- as in Sentence (9): es, and these were labeled by a human reader, (9) “The” appears between quote marks. who determined that 629 sentences were mention sentences (i.e., containing instances of mentioned Although constructions like (9) are unusual and language) and the remaining 1,764 were not. carry less practical value, the definition accom- Mention sentences were categorized based on modates them for completeness. functional properties that emerged during catego- Mentioned language is a common form of rization. Table 1 shows some examples of col- metalanguage, used to perform the full variety of lected mention sentences in each category.

1 This definition was introduced by Wilson (2011a) along 2 with a practical rubric for evaluating candidate sentences. The corpus is available at For brevity, its full justification is not reproduced here. http://www.cs.cmu.edu/~shomir/um_corpus.html.

761 Category Examples es in conversational English. A lack of phrase- Words as The IP Multimedia Subsystem level annotations in their corpus as well as sub- Words architecture uses the term stantial noise made it suboptimal for the present transport plane to describe a effort. However, it is possible (if not likely) that function roughly equivalent to the indicators of metalanguage differ between writ- routing control plane. ten and spoken English, lending importance to The material was a heavy canvas the Anderson corpus as a resource. known as duck, and the brothers Metalanguage has a long history of theoretical began making work pants and treatments, which chiefly explained the mechan- shirts out of the strong material. ics of selected examples of the phenomenon. Names as Digeri is the name of a Thracian Many addressed it through the related topic of Names tribe mentioned by Pliny the El- quotation (Cappelen and Lepore, 1997; Da- der, in The Natural History. vidson, 1979; Maier, 2007; Quine, 1940; Tarski, Hazrat Syed Jalaluddin Bukhari's 1933), and others previously cited in this paper descendants are also called Naqvi discussed it directly as metalanguage or the use- al-Bukhari. mention distinction. The definition of mentioned Spelling or The French changed the spelling language in Section 2.1 was a synthesis of the Pronunciation to bataillon, whereupon it direct- most empirically-compatible theoretical treat- ly entered into German. ments, and the present effort to automatically Welles insisted on pronouncing identify metalanguage builds on that synthesis. the word apostles with a hard t. Other Men- He kneels over Fil, and seeing 3 Detection of Mentioned Language tioned Lan- that his eyes are open whispers: The corpus-building effort used a heuristic to guage brother. accelerate the collection of mentioned language, During Christmas 1941, she typed but its low precision is impractical for automatic The end on the last page of identification. Moreover, the stylistic cues that Laura. the heuristic relied upon are often inconsistently applied (or entirely absent in informal contexts), Table 1: Examples of mentioned language from and they are sometimes unavailable for the writer the corpus. Instances of the phenomenon appear to use or for the audience to extract. This section in bold, with the original stylistic cues removed. presents an approach to the detection task, to dis- criminate between mention and non-mention sentences. Early examination of the corpus sug- To verify the reliability of the corpus and the gested that mention sentences tend not to have definition of mentioned language, three addition- distinct structural differences from non-mention al expert annotators independently labeled a sentences, so a lexical approach was first taken, shuffled set of 100 sentences, consisting of 54 although combinations of lexical and structural randomly selected mention sentences and 46 approaches are later explored indirectly through randomly selected non-mention sentences. All the delineation task. In this section a sentence is three agreed with the primary annotator on 46 assumed to be a sequence of words without sty- mention sentences and 30 non-mention sentenc- listic cues for mentioned language. es, with an average pairwise Kappa of 0.74. Kappa between the primary annotator and a hy- 3.1 Approach pothetical “majority voter” of the additional an- notators was 0.90. These results were seen as a To establish performance baselines, a matrix moderate indication of reliability and a potential of feature sets and classifiers was run on the cor- performance ceiling for automatic identification. pus with ten-fold cross validation. The feature sets were bags of the following: stemmed words 2.2 Related Work (SW), unstemmed words (UW), stemmed words plus stemmed bigrams (SWSB), and unstemmed The present effort is believed to be the first to words plus unstemmed bigrams (UWUB). Clas- automatically identify a natural variety of meta- sifiers were chosen to reflect a variety of ap- language in English text. Aside from the corpus proaches to supervised learning; as implemented described above, the only other significant cor- in Weka (Hall et al., 2009), these were Naive pus of metalanguage was created by Anderson et Bayes (John and Langley, 1995), SMO (Keerthi al. (2004), who collected metalinguistic utteranc-

762 mention and non-mention sentences, informed the approach taken to the detection task. To at- tempt to improve over the baseline, the SW fea- ture set was ranked by information gain, and all features except the top ten were discarded to cre- ate the metawords feature set (MW). Feature se- lection was done using the training set for each cross-validation fold, and the testing data for each fold was pruned correspondingly. The five selected classifiers were then applied to the data.

3.2 Results and Discussion The combination of five feature sets and five classifiers produced 25 sets of annotations, for which precision, recall and F1 were calculated for detecting mentioned language. For brevity, we present the highlights and contrast the meta- words approach with baseline performances. Table 2 compares classifier performances us- ing MW with SW, its closest relation. MW pro- duced improvement for all classifiers except Na- 3 ive Bayes . The J48-MW combination had the Figure 1. Cumulative coverage over sentences by highest F1 and recall of any feature set-classifier the most common words before (top) and after combination, though some combinations exceed- (bottom) instances of mentioned language. ed its precision. For all feature set-classifier pairs, precision was higher than recall, by as little as 0.024 (IBK-MW) and as much as 0.22 (Deci- et al., 2001), J48 (Quinlan, 1993), IBk (Aha and sion Table-UW). For the baseline feature sets, Kibler, 1991), and Decision Table (Kohavi, the best classifier was consistently SMO, with F1 1995). scores of 0.70, 0.70, 0.73, and 0.71 for SW, UW, Prior observations suggested that a small set SWSB, and UWUB, respectively. J48 was con- of approximately ten words significant to meta- sistently the second best, with F1 scores within language (“metawords”, informally) appear near 0.01 of SMO for each feature set. most instances of mentioned language (Wilson, Table 3 lists differences between F1 scores us- 2011b). The metalanguage corpus described in ing the MW feature set and each baseline feature Section 2 provided an opportunity to explore this set. MW resulted in improvements over the base- observation. The sentences in the corpus were line feature sets for nearly all classifiers, and sta- part-of-speech tagged and stemmed (using tistically significant improvements (using one- NLTK (Bird, 2006)). Sets were collected of all tailed T-tests across the populations of validation unique (stemmed) words in the three-word folds, p<0.05) were observed for eleven of the phrases directly preceding and following candi- sixteen combinations. IBk appeared to benefit date instances, and (respective of position) these the most, with significant improvements over all were ranked by frequency. The appearance or baseline feature sets, and Naive Bayes the least. non-appearance of these words was then deter- In general, recall benefited more than precision. mined over all mention and non-mention sen- Examining the MW features confirmed that tences. Figure 1 shows the cumulative coverage most were intuitive metawords. Nine words ap- (i.e., appearance at least once) over sentences for peared in all ten folds of MW: name, word, call, the top ten words appearing before and after can- term, mean, refer, use, derive, and Latin. The last didate instances. For example, call, word, or two words are perhaps artifacts of the encyclo- term were the three most common words before pedic nature of the source text, but the rest gen- candidate instances; they appear at least once in eralize easily. Future research using additional 36% of mention sentences, but they appear in only 1.6% of non-mention sentences. 3 The high frequencies of intuitive metawords, It seems likely that the method used to create the MW feature set aggravated the Naive Bayes assumption of fea- combined with their in coverage over ture independence.

763 Classifier Precision Recall F1 words subject to direct reference (e.g., the bold Naive Bayes .76 / .75 .63 / .60 .69 / .66 words in Sentences (1) through (4) and in other SMO .74 / .75 .67 / .70 .70 / .73 examples throughout this paper). This task, in IBk .69 / .74 .64 / .72 .66 / .73 addition to detection, is necessary to ascribe the Decision Table .76 / .74 .61 / .68 .67 / .71 information encoded in a metalinguistic state- J48 .72 / .75 .69 / .73 .70 / .74 ment to a specific linguistic entity.

4.1 Approach Table 2. The performances of classifiers using the SW and MW (in bold) feature sets. Manual examination of the corpus showed two frequent relationship patterns between meta- words and mentioned language. The first was Classifier SW UW SWSB UWUB noun apposition, in constructions like (15) and Naive Bayes -.024 -.018 .005 .007 (16), where the metaword-noun appears in italics SMO .023* .026* .000 .015 and the mentioned word in bold: IBk .067* .088* .07* .108* Decision Table .038* .047* .027 .052* (15) The term auntie was used depreciatively. J48 .037* .046* .025 .034 (16) It comes from the root word conficere. The second pattern was the appearance of men- Table 3. Differences between F1 scores from us- tioned language in the semantic role of a meta- ing the MW feature set and baseline feature sets. word-verb, as in (17) below: Statistically significant improvements are starred. (17) We sometimes call it the alpha profile. text sources will be necessary to fully verify Notably these patterns do not guarantee the cor- whether the MW approach and these specific rect delineation of mentioned language, but their metalinguistic terms are widely applicable. applicability made them suitable for the task. It also appears that 20% to 30% of instances To assess the applicability of phrase structures of mentioned language resist identification using and semantic roles to the automatic delineation word and bigram-based features alone. Many of of mentioned language, case studies were per- the false negatives from this experiment ap- formed on the sets of sentences in the corpus peared to lack the common metawords that the containing the nouns term and word and the verb detection approach relied upon. The sentences call. All sentences containing these three meta- below (taken from the corpus) illustrate this lack: words (appearing as their respective targeted parts of speech) were examined, including those (13) Other common insulting modifiers that did not contain mentioned language, since it include “dog”, “filthy”, etc. was believed that methods of delineation could (14) To note, in the original version the indirectly perform detection as well. Because of lyrics read “Jim crack corn”. the limited data available, formal experiments While modifier in (13) and read in (14) have in- were not possible, although the results still have tuitive metalinguistic value, they also have illustrative value. common non-metalinguistic senses. This sug- The noun apposition pattern described above gests that an approach incorporating word senses was formalized for term and word using TRegex may further improve upon the MW performanc- search strings (Levy and Andrew, 2006). The 91 es, and such an approach is preliminarily ex- sentences in the corpus containing term and the plored in the following section. 107 containing word were parsed using the Stan- Finally, it is notable that the best MW perfor- ford Parser (Marneffe et al., 2006), and the mances approach the Kappa score observed be- TRegex strings were applied to each sentence; tween the additional annotators. Although this is when a match occurred, the result was a predic- an indication of some success, the higher “major- tion that a specific sequence of words was men- ity vote” Kappa score of 0.90 remains a mean- tioned language (delineation), as well as a pre- ingful goal for future research efforts. diction that the sentence contained mentioned language (detection). The semantic role pattern 4 Toward Delineation for call was explored similarly using the Illinois Semantic Role Labeler (SRL) (Punyakanok et After identifying a mention sentence, the task al., 2008). Each of the 158 sentences in the cor- remains to determine the specific sequence of pus containing call as a verb was processed by

764 SRL, and when the output contained the appro- Label Scope Metaword priate semantic role (i.e., SRL’s “attribute of Overlabeled Underlabeled Exact arg1”) with respect to the metaword, the phrase term (n) 0 2 57 fulfilling that role was considered a predicted word (n) 3 4 57 delineation of mentioned language. By proxy, call (v) 16 1 68 such matching also implied a prediction that the phenomenon was present in the sentence. Pattern Applicability Metaword 4.2 Results and Discussion Precision Recall F1 term (n) 1.0 0.89 0.90 Delineation was evaluated with respect to the word (n) 1.0 0.94 0.97 correctness of label scope: that is, for a sentence call (v) 0.87 0.76 0.81 that contained an instance of mentioned lan- guage, whether the predicted word sequence ex- Table 4: Performance statistics for delineation actly matched the sequence labeled in the corpus, (in the form of label scope) and detection (pat- overlabeled it (i.e., included the instance of men- tern applicability) for the case studies. tioned language plus additional words), or un- derlabeled it (i.e., did not include the entire in- stance). To avoid confounding detection and de- for call suffered from two sources of errors: in- lineation, the statistics on label scope do not in- correct applications of the semantic role and ap- clude instances when the appropriate pattern plications of it that, while valid, did not involve failed to annotate any phrase in a sentence that mentioned language. contained mentioned language, or annotated a For the selected metawords, it appeared pat- phrase when no mentioned language was present. terns in noun apposition and semantic roles were Such instances are instead represented through moderately effective at delineating as well as pattern applicability statistics: when one of the detecting mentioned language. However, the ac- sought relationships between a chosen metaword curacy of these patterns was a reflection of the and a phrase appeared in a sentence, it was con- dependability of the underlying language tools, sidered a positive prediction of the presence of and the case studies in aggregate covered only mentioned language. Table 4 shows performance 33% of the sentences containing mentioned lan- metrics from each of the three case studies. guage in the corpus. To create a comprehensive Noun apposition with either term or word ap- method for delineation, more relationships must peared to be adept at predicting scope, with per- be identified between metawords and mentioned fect labels for 97% and 89% of instances, respec- language. A perusal of the corpus suggests that tively. The instances of overlabeling and un- these patterns are small in variety but large in derlabeling for these two were mostly due to quantity: metawords are diverse, and some have parsing errors, which occurred prior to applying non-metalinguistic senses that must be accounted the TRegex pattern. Overlabeling was a greater for, as shown by Sentences (13) and (14) and problem for call, for which 80% of labels were others that resisted detection. perfect and nearly the rest were overlabeled. Manual examination revealed that the prediction 5 Conclusion often would “spill” far past the actual end of The detection and delineation methods presented mentioned language, due to the boundaries of the in this paper demonstrate the feasibility of identi- semantic role in SRL’s output. For example, the fying metalanguage in English text. The next entire phrase in bold in (18) below was errone- goals of this project will be to assimilate meta- ously predicted to be mentioned language, in- language from additional text sources and inte- stead of simply snow-eaters: grate the detection and delineation tasks. This (18) Winds of this type are called snow-eaters will improve performance and provide a richer for their ability to make snow melt or subli- structural knowledge of metalanguage, which mate rapidly. will enable practical systems to incorporate pro- cessing of the phenomenon and exploit the lin- Re-examining the detection task through pat- guistic information that it encodes. tern applicability, noun appositions with term and word exhibited perfect precision. The false negatives that lowered the recall were again mostly due to parse errors. Precision and recall

765

References Keerthi, S. S., Shevade, S. K., Bhattacharyya, C., & Adler, B. T., de Alfaro, L., Pye, I., & Raman, V. Murthy, K. R. K. (2001). Improvements to Platt’s (2008). Measuring author contributions to the Wik- SMO Algorithm for SVM Classifier Design. Neural ipedia. In Proceedings of the 4th International Sym- Computation, 13, 637–649. doi:10.1162/ posium on Wikis (pp. 15:1–15:10). New York, NY, 089976601300014493 USA: ACM. doi:10.1145/1822258.1822279 Kohavi, R. (1995). The Power of Decision Tables. In Aha, D. W., & Kibler, D. (1991). Instance-based Proceedings of the European Conference on Machine learning algorithms. In Machine Learning (pp. 37– Learning (pp. 174–189). Springer Verlag. 66). Levy, R., & Andrew, G. (2006). TRegex and TSur- Anderson, M. L., Fister, A., Lee, B., & Wang, D. geon: tools for querying and manipulating tree data (2004). On the frequency and types of meta-language structures. In Proceedings of the Fifth International in conversation: A preliminary report. In 14th Annual Conference on Language Resources and Evaluation. Conference of the Society for Text and Discourse. Maier, E. (2007). Mixed quotation: between use and Anderson, M. L., Okamoto, Y. A., Josyula, D., & mention. In and Engineering of Natural Lan- Perlis, D. (2002). The Use-Mention Distinction and guage Workshop. Retrieved from Its Importance to HCI. In Proceedings of the Sixth http://ncs.ruhosting.nl/emar/em_lenls_quot.pdf Workshop on the Semantics and Pragmatics of Dia- log, 21–28. Marneffe, M. D., MacCartney, B., & Manning, C. D. (2006). Generating typed dependency parses from Audi, R. (1995). The Cambridge Dictionary of Phi- phrase structure parses. In Proceedings of the Interna- losophy. Cambridge University Press. tional Conference on Language Resources and Eval- Bird, S. (2006). NLTK: The natural language toolkit. uation, pp. 449–454. In Proceedings of the COLING/ACL on Interactive Punyakanok, V., Roth, D., & Yih, W. (2008). The Presentation Sessions (pp. 69–72). Stroudsburg, PA, importance of syntactic parsing and inference in se- USA: Association for Computational . mantic role labeling. Computational Linguistics, Cappelen, H., & Lepore, E. (1997). Varieties of quo- 34(2), 257–287. doi:10.1162/coli.2008.34.2.257 tation. Mind, 106(423), 429–450. Quine, W. V. O. (1940). Mathematical logic. Cam- doi:10.1093/mind/106.423.429 bridge, MA: Harvard University Press. Davidson, D. (1979). Quotation. and Deci- Quinlan, J. (1993). C4.5: Programs for Machine sion, 11(1), 27–40. doi:10.1007/BF00126690 Learning. San Mateo, Calif.: Morgan Kaufmann Pub- Fellbaum, C. (1998). WordNet: An electronic lexical lishers. database. Cambridge: MIT Press. Saka, P. (2006). The Demonstrative and Identity The- García-Carpintero, M. (2004). The deferred ostension ories of Quotation. Journal of Philosophy, 103(9), theory of quotation. Noûs, 38(4), 674–692. 452–471. Hall, M., Frank, E., Holmes, G., Pfahringer, B., Sperber, D., & Wilson, D. (1981). Irony and the Use- Reutemann, P., & Witten, I. H. (2009). The WEKA Mention Distinction. In Radical Pragmatics (pp. 295– data mining software: An update. ACM SIGKDD Ex- 318). New York. plorations Newsletter, 11, 10–18. doi:10.1145/ Tarski, A. (1933). The concept of in formalized 1656274.1656278 . In J. H. Woodger (Ed.), Logic, Semantics, Hu, G. (2010). A place for metalanguage in the L2 Mathematics. Oxford: Oxford University Press. classroom. ELT Journal. doi:10.1093/elt/ccq037 Wilson, S. (2011a). A computational theory of the Jaworski, A., Coupland, N., & Galasinski, D. (Eds.). use-mention distinction in natural language. Univer- (2004). Metalanguage: Language, Power, and Social sity of Maryland at College Park. PhD Thesis, Col- Process. De Gruyter. lege Park, MD, USA. John, G., & Langley, P. (1995). Estimating Continu- Wilson, S. (2011b). In search of the use-mention dis- ous Distributions in Bayesian Classifiers. In Proceed- tinction and its impact on language processing tasks. ings of the Eleventh Conference on Uncertainty in International Journal of Computational Linguistics Artificial Intelligence (pp. 338–345). Morgan Kauf- and Applications, 2(1-2), 139–154. mann. Wilson, S. (2012). The Creation of a Corpus of Eng- lish Metalanguage. In Proceedings of the 50th Annual Meeting of the Association for Computational Lin- guistics, (pp. 638–646).

766