Causal Inference of Script Knowledge

Noah Weber 1, Rachel Rudinger 2,3, Benjamin Van Durme 1 1Johns Hopkins University 2University of Maryland, College Park 3Allen Institute for AI

Abstract When does a sequence of events define an ev- eryday scenario and how can this knowledge be induced from text? Prior works in induc- ing such scripts have relied on, in one form or another, measures of correlation between instances of events in a corpus. We argue from both a conceptual and practical sense that a purely correlation-based approach is insuf- ficient, and instead propose an approach to script induction based on the causal effect be- Figure 1: The events of Watching a sad movie, Eat- tween events, formally defined via interven- ing popcorn, and Crying, may highly co-occur in a hy- tions. Through both human and automatic pothetical corpus. What distinguishes valid event pair evaluations, we show that the output of our inferences (event pairs linked in a commensense sce- method based on causal effects better matches nario; noted by checkmarks above) versus invalid infer- the intuition of what a script represents. ences (noted by a ‘X’)? 1 Introduction Commonsense knowledge of everyday situations, Hanks, 1990) between event mentions. Others em- as defined in terms of prototypical sequences of ployed probabilities from a model over events, has long been held to play a major role in event sequences (Jans et al., 2012; Rudinger et al., text comprehension and understanding (Minsky, 2015; Pichotta and Mooney, 2016; Peng and Roth, 1974; Schank and Abelson, 1975, 1977; Bower 2016; Weber et al., 2018b), or other measures of et al., 1979; Abbott et al., 1985). Naturally, this has event co-occurrence (Balasubramanian et al., 2013; motivated a large body of work looking to learn Modi and Titov, 2014). 1 such knowledge, such scripts, from text corpora In this work we ask: do measures rooted in co- through data-driven approaches. occurrence best capture the notion of whether one A minimal and often implicit requirement for event should follow another in a script? We posit any such approach is to resolve for any pair of that it does not, that while observed correlations events e1 and e2 what quantitative measure should between events indicate relatedness, relatedness is be used to determine whether e2 should ”follow” e1 not the only factor in determining whether events in script. That is, documents may serve as descrip- form a meaningful script. tions of events that occur in the same situation as Consider the example of Ge et al.(2016): hur- other events: what function may we compute over ricane events are prototypically connected with the raw presence or absence of events in documents events of donations coming in. Likewise, hurri- that is most useful for script induction? cane events are connected to evacuation events. Chambers and Jurafsky (2008; 2009) adopted However, while donation and evacuation events are point-wise mutual information (PMI) (Church and not conceptually connected in the same sense, there 1For simplicity we will refer to these ‘prototypical event will exist strong statistical associations between the sequences’ as scripts throughout the paper, though it should be noted scripts as originally proposed contain further structure two. Figure1 provides a second example: eating not captured in this definition. popcorn is not conceptually associated with crying,

7583 Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, pages 7583–7596, November 16–20, 2020. c 2020 Association for Computational but they might co-occur in a hypothetical corpus 1985), as do related works in psychology (Black describing situations of watching a sad movie. and Bower, 1980; Trabasso and Sperry, 1985). What do strict co-occurrence measures miss? In But any measure based solely on p(e2|e1) is ag- both examples the ‘invalid’ inferences arise from nostic to notions of causal relevance. Does this the same issue: an event such as eating popcorn matter in practice? A high p(e2|e1) indicates ei- may raise the probability of the event crying, but ther: (1) a causal influence of e1 on e2, or (2) a it does so only through a shared association with a common cause e0 between them, meaning the re- movie watching context: the increase in probability lation between e1 and e2 is spurious. In the latter is not due to the eating popcorn itself. In other case, e0 acts as a confounder between e1 and e2. words, what is lacking is a direct causal effect be- Ge et al.(2016) acknowledges that the associ- tween these events, a quantity that can be formally ations picked up by correlational measures may defined using tools from the causal inference litera- often be spurious. Their solution relies on using ture (Hernan and Robins, 2019). trends of words in a temporal stream of newswire In this work we demonstrate how a measure data, and hence is fairly domain specific. based on causal effects can be derived, com- puted, and employed for the extraction of script 2.2 Defining Causal Relevance knowledge. Using crowdsourced human eval- Early works such as Schank and Abelson(1975) uations and a variant of the automatic cloze are vague with respect to the meaning of “causally evaluation, we show how this definition better chained.” Can one say that watching a movie has captures the notion of scripts as compared to causal influence on the subsequent event of eating prior standard measures, PMI and event sequence popcorn happening? Furthermore, can this defini- language models. Code and data available at tion be operationalized in practice? github.com/weberna/causalchains. We argue that both of these questions may be 2 Motivation elucidated by taking a manipulation-based view of causation. Roughly speaking, this view holds Does that fact that event e2 is often observed after that a causal relationship is one that is “potentially e1 in the data (i.e. p(e2|e1) is “high”) mean that e2 exploitable for the purposes of manipulation and prototypically follows e1, in the sense of being part control”– Woodward(2005). In other words, a of a script? In this section we argue that observed causal relationship between A and B means that (in associations are not sufficient for the purpose of some cases) manipulating the value of A should extracting script knowledge from text. We argue result in a change in the value of B. A primary from a conceptual standpoint that some notion of benefit of this view is that the meaning of a causal causal relevance is required. We then give exam- claim can be clarified by specifying what these ples showing the practical pitfalls that may arise ‘manipulations’ are exactly. We take this approach from ignoring this component. Finally, we propose below to clarify what exactly is meant by ‘causal our intervention based definition for script events, relevance’ between script events. and show how it both explicitly defines a notion of Imagine an agent reading a . After read- ‘causal relevance,’ while simultaneously fixing the ing a part of the discourse, the agent has some ex- aforementioned practical pitfalls. pectations for events that might happen next. Now imagine that, before the agents reads the next pas- 2.1 The Significance of Causal Relevance sage, we surreptitiously replace it with an alternate The original works defining scripts are unequivocal passage in which the event e1 happens. We then al- about the importance of causal linkage between low the agent to continue reading. If e1 is causally 2 script events, and other components of the origi- relevant to e2, then this replacement should, in nal script definition (e.g. what-ifs, preconditions, some contexts, raise the agents degree of belief in postconditions, etc.) are arguably causal in na- e2 happening next (contra the case where we didn’t ture. Early rule-based works on inducing scripts intervene to make e1 happen ). heavily used causal concepts in their schema rep- So, for example, if we replaced a passage such resentations (DeJong, 1983; Mooney and DeJong, that e1 = watching a movie was true, we could 2“...a script is not a simple list of events but rather a linked expect on average that the agent’s degree of belief causal chain”(Schank and Abelson, 1975) that e2 = eating popcorn happens next will be

7584 tween variables graphically in a manner similar to Bayesian networks; the key distinction being T T ... i-1 i-1 that the edges in a CBN posits a direction of causal influence between the variables 3. U e e i-1 i ... We will define our causal model from a top down, data generating perspective in a way that aligns Di-1 Di Di+1 with our conceptual story from the previous section. Below we describe the four types of variables in our model, as well as their causal dependencies. Figure 2: The diagram for our causal model up to time step i. Intervening on ei−1 acts to remove the dotted The World, U: The starting point for the gener- edges. See 3.1 for a description of the variables. ation of our data is the real world. This context is explicitly represented by the unmeasured variable U. This variable is unknowable and in general un- higher. In this way, we say these events are causally measurable: we don’t know how it is distributed, relevant, and are for our purposes, script events. nor even what ‘type’ of variable it is. This variable With this little ‘story,’ we have clarified the con- is represented by the hexagonal node in Figure2. ceptual notion of causal relevance in our problem. In the next section, we formalize this story and its The Text, T: The next type of variable represents notion of intervention into a causal model. the text of the document. For indexing purposes, we segment the text into chunks T1,...,TN , where 3 Method N is the number of realis events explicitly men- We would like to compute the effect of forcing an tioned in the text. The variable Ti is thus the text th event of a certain type to occur in the text. The chunk corresponding to the i event mentioned in event types that get the largest increase in probabil- text. These chunks may be overlapping, and may 4 ity due to this are held to be ‘script’ events. Com- skip over certain parts of the original text. The puting these quantities falls within the domain of causal relationship between various text chunks causal inference, and hence will require its tools be is thus ambiguous. We denote this by placing bi- used. There are three fundamental steps in causal directional arrows between the square text nodes inference we will need to work through to accom- in Figure2. The context of the world also causally plish this: (1) Define a Causal Model: Identify influences the content of the text, hence we include the variables of interest in the problem, and define an arrow from U to all text variables, Ti. causal assumptions regarding these variables, (2) Event Inferences, e: In our story in Section2, Establish Identifiability: With the given model, an agent reads a chunk of text and infers the type of determine whether the causal quantity can be com- event that was mentioned in the piece of text. This puted as a function of observed data. If it can, inference is represented (for the ith event in text) in derive this function and move to (3) Estimation: our model by the variable e ∈ E where E is the Estimate this function using observed data. We go i set of possible atomic event types (described at the through each step in the next three subsections. end of this section).5 To best contrast with prior work, we use the event representation of Chambers and Jurafsky 3See Pearl(2000); Bareinboim et al.(2012) for a compre- (2008) and others (Jans et al., 2012; Rudinger et al., hensive definition of CBNs and their properties. 4Keeping with prior work, we use the textual span of the 2015). A description of this representation is pro- event predicate syntactic dependents as the textual content vided in the Supplemental. of an event. The ordering of variables Ti corresponds to the positions of the event predicates in the text. 5 3.1 Step 1: Define a Causal Model For this study we use the output of information extraction tools as a proxy for the variable ei (see supplemental). As A causal model defines a set of causal assump- such, it is important to note that there will be bias in computa- tions due to measurement error. Fortunately, there do exists tions on the variables of interest in a problem. methods in the causal inference literature that can adjust for While there exists several formalisms that accom- this bias (Kuroki and Pearl, 2014; Miao et al., 2018). Wood- plish this, in this paper we make use of causal Doughty et al.(2018) derive equations in a case setting related to ours (i.e. with measurement bias on the variable being Bayesian networks (CBN) (Spirtes et al., 2000; intervened on). Dealing with this issue will be an important Pearl, 2000). CBNs model dependencies be- next step for future work.

7585 The textual content of Ti causally influences the on the distribution over the subsequent event ei. inferred type ei, hence directional connecting ar- Now that we have a causal model in the form of rows in Figure2. Fig.2, we can now define this effect. Using the notation of Pearl(2000), we write this as: Discourse Representation, D The variable ei represent a high level abstraction of part of the p(ei|do(ei−1 = k)) (1) semantic content found in Ti. Is this informa- The of do(ei−1 = k) are defined as tion about events used for later event inferences an ‘arrow breaking’ operation on Figure2 which by an agent reading the text? Prior results in deletes the incoming arrows to ei−1 (the dotted causal network/chain theories of discourse process- arrows in Figure2) and sets the variable to k. Be- ing (Black and Bower, 1980; Trabasso and Sperry, fore a causal query such as Eq.1 can be estimated 1985; Van den Broek, 1990) seem to strongly point we must first establish identifiability (Shpitser and to the affirmative. In brief, these theories hold that Pearl, 2008): can the causal query be written as a the identities of the events occurring in the text – function of (only) the observed data? and the causal relations among them – are a core Eq.1 is identified by noting that variables Ti−1 part of how a discourse is represented in human and Di−1 meet the ‘back-door criterion’ of Pearl memory while reading, and more-so, that this infor- (1995), allowing us to write Eq.1 as the following: mation significantly affects a reader’s event based h i inferences(Trabasso and Van Den Broek, 1985; ETi−1,Di−1 p(ei|ei−1 = k, Di−1,Ti−1) (2) Van den Broek and Lorch Jr, 1993). Thus we intro- Our next step is to estimate the above equa- duce a discourse representation variable, Di, itself I O tion. If one has an estimate for the conditional a combination of two sub-variables, Di and Di . I ∗ 6 p(ei|ei−1,Di−1,Ti−1), then one may ”plug it into” The variable Di ∈ E is a sequence of events that were explicitly stated in the text, up to step Eq.2 and use a Monte Carlo approximation of the i. After each step, the in-text event inferred at i expectation (using samples of (T,D)). This simple I plug in estimator is what we use here (the variable ei) is appended to Di+1. The causal I I It is important to be aware of the fact that parents of Di+1 are thus ei and Di (which is simply I This estimator, specifically when plugging in ma- copied over). We posit that the information in Di chine learning methods, is quite naive (e.g. Cher- provides information in the inference of ei, and thus draw an arrow from DI to e . nozhukov et al.(2018)), and will suffer from an i i 7 Unstated events not found in the text but inferred asymptotic (first order) bias. which prevents one by the reader also have an effect on event inferences from constructing meaningful confidence intervals (McKoon and Ratcliff, 1986, 1992; Graesser et al., or performing certain hypothesis tests. That said, 1994). We thus additionally take this into consid- in practice these machine learning based plug in es- eration in our causal model by including an out of timators can achieve quite reasonable performance O |E| (see for example, the results in Shalit et al.(2017)), text discourse representation variable, Di ∈ 2 . This variable is a bag of events that a reader may and since our current use case can be validated empirically, we save the usage of more sophisti- infer implicitly from the text chunk Ti using com- mon sense. Its causal parents are thus both the text cated estimators (and proper statistical inference) for future work8. chunk Ti, as well as the world context U; its causal children are ei. Obtaining this information is done 3.3 Step 3: Estimation via human annotation and discussed later. D is i Eq.2 depends on the conditional, p = thus equal to (DI ,DO), and inherits the incoming ei i i p(e |e ,D ,T ), which we estimate via stan- and outgoing arrows of both in Figure2. i i−1 i−1 i−1 7See Fisher and Kennedy(2018) for an introduction on 3.2 Step 2: Establishing Identifiability how this bias manifests. 8Semiparametric estimation of equations such as Eq.2 Our goal is to compute the effect that intervening involving high dimensional variables (like text) is an open and setting the preceding event ei−1 to k ∈ E has problem that we do not address here. See (D’Amour et al., 2020; Kennedy, 2016; Chernozhukov et al., 2018) for an analy- 6We don’t explicitly model the causal structure between sis of some of the problems that arise in both high dimensional events in Di, the importance of which is a key finding in the causal inference and semiparametric estimation (ie estimation above referenced literature. While this wouldn’t change the without full parametric assumptions). See Keith et al.(2020) structure of our causal model, it would impact the estimation for an overview of problems that arise particularly when deal- stage, and would be an interesting line of future work. ing with text.

7586 dard ML techniques with a dataset of samples for matrices A and B of dimension |E| × d. We drawn from p(ei, ei−1,Di−1,Ti−1). There are two then finetune this model on the 2000 annotated ex- O issues: (1) How do we deal with out-of-text events amples including Di−1. We add a new parameter in Di−1?, and (2) What form will pei take? matrix, C, to the previously trained model (allow- O ing it to take Di−1 as input) and model pei as: Dealing with Out-of-Text Events Recall that D is combination of the variables DI and DO. i i i pei ∝ Ave + Bvt + Cvo To learn a model for pei we require samples from the full joint. Out of the box however, we only have The input vo is the average of the embeddings for I O access to p(ei, ei−1,Di−1,Ti−1). If, for the sam- the events found in Di−1. The parameter matrix ples in our current dataset, we could draw samples C is thus the only set of parameters trained ‘from O I from pD = p(Di−1|ei, ei−1,Di−1,Ti−1), then we scratch,’ on the 2000 annotated examples. The rest would have access to a dataset with samples drawn of the parameters are initialized and finetuned from from the full joint. the previously trained model. See Appendix for In order to ‘draw’ samples from pD we employ further training details. human annotation. Annotators are presented with I 9 3.4 Extracting Script Knowledge a human readable form of (ei, ei−1,Di−1,Ti−1) and are asked to annotate for possible events be- Provided a model of the conditional pei we can O longing in Di−1. Rather than opt for noisy annota- approximate Eq.2 via Monte Carlo by taking our tions obtained via freeform elicitation, we instead annotated dataset of N = 2000 examples and com- provide users with a set of 6 candidate choices for puting the following average: O members of Di−1. The candidates are obtained N from various knowledge sources: ConceptNet 1 X Pˆ = p(e |e = k, D ,T ) (3) (Speer and Havasi, 2012), VerbOcean (Chklovski k N i i−1 j j and Pantel, 2004), and high PMI events from the j=1 NYT Gigaword corpus (Graff et al., 2003). The This gives us a length |E| vector Pˆ whose lth top two candidates are selected from each source. k component, Pˆ gives p(e = l|do(e = k)). We In a scheme similar to Zhang et al.(2017), we kl i i−1 compute this vector for all values of k. Note that ask users to rate candidates on an ordinal scale and this computation only needs to be done once. consider candidates rated at or above a 3 (out of 4) There are several ways one could extract script- to be considered within DO . We found annotator i−1 like knowledge using this information. In this pa- agreement to be quite high, with a Krippendorf’s per, we define a normalized score over intervened- α of 0.79. Under this scheme, we crowdsourced on events such that the script compatibility score a dataset of 2000 fully annotated examples on the between two concurrent events is defined as: Mechanical Turk platform. An image of our anno- tation interface is provided in the Appendix. Pˆ S(e = k, e = l) = kl (4) i−1 i PE ˆ The Conditional Model We use neural net- j=1 Pjl works to model pe . In order to deal with the small i We term this as the ‘Causal’ score in the eval below. amount of fully annotated data available, we em- ploy a finetuning paradigm. We first train a model 4 Experiments and Evaluation on a large dataset that does not include annota- O tions for Di−1. This model consists of a single Automatic evaluation of methods that extract script- layer, 300 dimensional GRU encoder which en- like knowledge is an open problem that we do not I d 10 codes [Di−1, ei−1] into a vector ve ∈ R and a attempt to tackle here, relying foremost on crowd- CNN-based encoder which encodes Ti−1 into a sourced human evaluations to validate our method. d vector vt ∈ R . The term pei is modeled as: However, as we aim to provide a contrast to prior script-induction approaches, we perform an experi-

pei ∝ Ave + Bvt ment looking at a variant of the popular automatic narrative cloze evaluation. 9In the final annotation experiment, we found it easier for 10 annotators to be only provided the text Ti−1, given that many See discussions by Rudinger et al.(2015) and Chambers I events in Di−1 are irrelevant. (2017).

7587 4.1 Dataset Method Average Score Average Rank (1-6) Causal 49.71 4.10 LM 35.95 3.39 For these experiments, we use the Toronto Books PMI 34.92 3.02 corpus (Zhu et al., 2015; Kiros et al., 2015), a col- lection of fiction novels spanning multiple genres. Table 1: Average Annotator Scores in Pairwise annota- The original corpus contains 11,040 books by un- tion experiment published authors. We remove duplicate books Causal LM PMI Target from the corpus (by exact file match), leaving a X tripped X came X featured X fell total of 7,101 books. The books are assigned ran- X lit X sat X laboured X inhaled domly to train, development, and test splits in 90%- X aimed X came X alarmed X fired 5%-5% proportions. Each book is then run through X poured X nodded X credited X refilled a pipeline of tokenization with CoreNLP 3.8 (Man- X radioed X made X fostered X ordered ning et al., 2014), parsing with CoreNLP’s univer- sal dependency parser (Nivre et al., 2016) and coref- Table 2: Examples from each system, each of erence resolution (Clark and Manning, 2016b), be- which outputs a previous event that maximizes the fore feeding the results into PredPatt (White et al., score/likelihood that the Targeted event follows in text. 2016). We additionally tag the events with factu- ality predictions from Rudinger et al.(2018b) (we 4.3 Eval I: Pairwise Event Associations only consider factual events). The end result is a large dataset of event chains centered around a Any system aimed at extracting script-like knowl- single protagonist entity, similar to (Chambers and edge should be able to answer the following ab- Jurafsky, 2008). We make this data public to facili- ductive question: given an event ei happened, what tate further work in this area. See the Appendix for previous event ei−1 best explains why ei is true? In a full detailed overview of our pipeline. other words, what ei−1, if it were true, would max- imize my belief that ei was true. We evaluate each method’s ability to do this via a human evaluation. On each task, annotators are presented with six event pairs (ei−1, ei), where ei is the same for all 4.2 Baselines pairs, but ei−1 is generated by one of the three sys- tems. Similar to the human evaluation in Pichotta and Mooney(2016), we filter out outputs in the In this paper, we compare against the two dominant top-20 most frequent events list for all systems. approaches for script induction (under a atomic For each system, we pick the top two events that event representation11): PMI (similar to Cham- maximize S(·, e ), PMI(·, e ), and p (·, e ), for bers and Jurafsky(2008, 2009)) and LMs over i i lm i the Causal, PMI, and LM systems respectively, and event sequences (Rudinger et al., 2015; Pichotta present them in random order. For each pair, users and Mooney, 2016). We defer definitions for these are asked to provide a scalar annotation (from 0%- models to the cited papers, below we provide the 100%, via a slider bar) on the chance that e is true relevant details for each baseline, with further train- i afterwards or happened as a result of e . The an- ing details provided in the Appendix. i−1 notation scheme is modeled after the one presented For computing PMI we follow many of the de- in Sakaguchi and Van Durme(2018), and shown to tails from (Jans et al., 2012). Due to the nature be effective for paraphrase evaluation in Hu et al. of the evaluations, we utilize their ‘ordered’ PMI- (2019). Example outputs for systems are provided variant. Also like Jans et al.(2012), we use skip- for several e1 choices for this task in Table2. 12 bigrams with a window of 2 to deal with count spar- The evaluation is done for 150 randomly cho- sity. Consistent with prior work we additionally sen instances of ei, each with 6 candidate ei−1. We employ the discount score of Pantel and Ravichan- 11There are also a related class of methods based on creating dran(2004). For the LM, we use a standard, 2 compositional event embeddings (Modi, 2016; Weber et al., layer, GRU-based neural network language model, 2018a). Since the event representation used here is atomic it makes little sense to use them here. with 512 dimensional hidden states, trained on a 12Note that we do manually filter out of the initial random log-likelihood objective. list events which we judge as difficult to understand

7588 Method Average Score Average Rank (1-3) to provide a scalar annotation (from 0%-100%, via Causal 60.12 2.19 a slider) on the chance that e4 is true afterwards LM 57.40 2.12 or happened as a result of the chain of context PMI 44.26 1.68 events. The evaluation is done for 150 tasks, with two annotators on each task. As before, we average Table 3: Average Annotator Scores in Chain annotation these annotations together for a gold annotation. experiment In Table3 we provide results of the experiment. Note the the rankings are now from 1 to 3 (higher is have two annotators provide annotations for each better). We find annotators usually rated the Causal task, and similar to Hu et al.(2019), average these system higher, though the LM model is much closer annotations together for a gold annotation. in this case. The differences (in both Score/Rank) In Table1 we provide the results of the experi- between the Causal and LM system outputs are ment, providing both the average annotation score not significant under a Wilcoxon signed-rank test, for the outputs of each system, as well as the av- though the differences between the Causal and PMI erage relative ranking (with a rank of 6 indicat- system is (p < 0.01). The fact that the pairwise ing the annotators ranked the output as the high- Causal model is still able to (at minimum) match est/best in the task, and a rank of 1 indicating the the full sequential model on a chain-wise evalua- opposite). We find that annotators consistently tion speaks to the robustness of the event associa- rated the Causal system higher. The differences tions mined from it, and further motivates work in (in both Score and Rank) between the Causal sys- extending the method to the sequential case. tem and the next best system are significant under 4.5 Diversity of System Outputs a Wilcoxon signed-rank test (p < 0.01). But what type of event associations are found from 4.4 Eval II: Event Chain Completion the Causal model? As noted both in Rudinger Of course, while pairwise knowledge between et al.(2015) and in Chambers(2017), PMI based events is a minimum prerequisite, we would also approaches can often extract intuitive event rela- like to generalize to handle chains of events con- tionships, but may sometimes overweight low fre- taining multiple events. In this section, we look at quency events or suffer problems from count spar- each system’s ability to provide an intuitive com- sity. LM based models, on the other hand, were pletion to an event chain. More specifically, the noted for their preference towards boring, uninfor- model is provided with a chain of three context mative, high frequency events (like ’sat’ or ’came’). events, (e1, e2, e3), and is tasked with providing a So where does the Causal model lay on this scale? suitable e4 that might follow given the first three We study this by looking at the percentage of events. We evaluate each method’s ability to do unique words used by each system in the previ- this via a human evaluation. ous evaluations, presented in Table5. Unsurpris- Since both PMI and the Causal model 13 work ingly, we find that PMI chooses a new word to only as pairwise models, we adopt the method of output often (77%-84% of the time), while the LM Chambers and Jurafsky(2008) for chains. For both model very rarely does (only 7%-13%). The Causal the PMI and Causal model, we pick the e4 that model, while not as adventurous as the PMI system, 1 P3 tends to produce very diverse output, generating a maximizes 3 i=1 Score(ei, e4), where Score is either PMI or Eq4. The LM model chooses an new output 60%-76% of the time. Both the PMI e4 that maximizes the joint over all events. and Causal system produce relatively less diverse Our annotation task is similar to the one in output on the chain task, which is expected due to 4.3, except the pairs provided consist of a context the averaging scheme used by each to select events. (e , e , e ) and a system generated e . Each system 1 2 3 4 4.6 Infrequent Narrative Cloze generates its top choice for e4, giving annotators 3 pairs14 to annotate for each task (i.e. each con- The narrative cloze task, or some variant of it, has text). On each task, human annotators are asked remained a popular automatic test for systems aim- ing to extract ‘script’ knowledge. The task is usu- 13 Generalizing the Causal model to multiple interventions, ally formulated as follows: given a chain of events though out of scope here, is a clear next step for future work. 14We found providing six pairs per task to be overwhelming e1, ...en−1 that occurs in the data, predict the held given the longer context out next event that occurs in the data, en. There

7589 Method Pairwise Chain We thus propose a Infrequent Cloze task. In this X awoke (2%) X collided (4%) task we create a variety of different cloze datasets Causal X parried (1%) pinched X (3%) (each with 2000 instances) from our test set. Each X came (30%) X made (23%) set is indexed by a value C, such that the indicated LM X sat (27%) X came (15%) dataset does not include instances from the top C X lurched (1%) bribed X (3%) most frequent events (C = 0 is the normal cloze PMI X patroled (1%) X swarmed (2%) setting). We compute a Recall@100 cloze task on 7 sets of various C and report results in Table6. Table 4: Two most used output events (and % of times At C = 0, as expected, the LM model is vastly it is used) for each system, for each human evaluation superior. The performance of the LM model dras- tically drops however, as soon as C increases, in- Method Pairwise Chain dicating an overreliance on prior probability. The Causal 76.0% 60.1% LM performance drops below 2% once C = 200, LM 7.30% 13.3% indicating almost no ability in predicting informa- PMI 84.0% 77.6% tive events such as drink or pay, both of which Table 5: % of times a system outputs a new event it occur in this set in our case. The PMI and Causal previously had not used before. model’s performance on the other hand, steadily improve while C increases, with the Causal model consistently outperforming PMI. This result, when exists various measures to calculate a models abil- combined with the results of the human evaluation, ity to perform in this task, but arguably the most give further evidence towards the relative robust- used one is the Recall@N measure introduced in ness of the Causal model in extracting informative Jans et al.(2012). Recall@N works as follows: core events. The precipitous drop in performance for a cloze instance, a system will return the top of the LM further underscores problems that a naive N guesses for en. Recall@N is the percentage of automatic cloze evaluation may cover up. times en is found anywhere in the top N list. The automatic version of the cloze task has 5 Related Work notable limitations. As noted in Rudinger et al. (2015), the cloze task is essentially a language mod- Our work looks at script like associations between eling task; it measures how well a model fits the events in a manner similar to Chambers and Ju- data. The question then becomes whether data fit rafsky(2008), and works along similar lines (Jans implies valid script knowledge was learned. The et al., 2012; Pichotta and Mooney, 2016). Related work of Chambers(2017) casts serious doubts on lines of work exist, such as work using generative this, with various experiments showing automatic models to induce probabilistic schemas(Chambers, cloze evaluations are biased to high frequency, un- 2013; Cheung et al., 2013; Ferraro and Van Durme, informative events, as opposed to informative, core, 2016), work showing how script knowledge may script events. They further posit human annotation be mined from user elicited event sequences (Reg- as a necessary requirement for evaluation. neri et al., 2010; Orr et al., 2014), and approaches In this experiment, we provide another datapoint take advantage of hand coded schematic knowledge for the inadequacy of the automatic cloze, while (Mooney and DeJong, 1985; Raskin et al., 2003). simultaneously showing the relative robustness of The literature is rich with work the knowledge extracted from our Causal system. studying the role of causal semantics in linguistic For the experiment, we make the following assump- constructions and argument structure (Talmy, 1988; tions: (1) Highly frequent events tend to appear in Croft, 1991, 2012), as well as the causal seman- many scenarios, and hence are less likely to be an tics of lexical items themselves (Wolff and Song, informative ‘core’ event for a script, and (2) Less 2003). Work in the NLP literature on extracting frequent events are more likely to appear only in causal relations has benefited from this line of work, specific scenarios, and are thus more likely to be utilizing the systematic way in which causation informative events. If these are true, then a system in expressed in language to mine relations (Girju that has extracted useful script knowledge should and Moldovan, 2002; Girju, 2003; Riaz and Girju, keep (or even improve) cloze performance when 2013; Blanco et al., 2008; Do et al., 2011; Bosse- the correct answer for en is a less frequent event. lut et al., 2019). This line work aims to extract

7590 Exclusion Threshold Method < 0 < 50 < 100 < 125 < 150 < 200 < 500 Causal 5.60 7.10 7.00 7.49 7.20 8.20 9.10 LM 65.3 28.1 9.70 6.30 3.60 1.70 0.25 PMI 1.80 3.30 3.36 4.10 4.00 4.90 7.00

Table 6: Recall@100 Narrative Cloze Results. < C indicates that instances whose cloze answer is one of the top C most frequent events are not evaluated on

causal relations between events that are in some Elias Bareinboim, Carlos Brito, and Judea Pearl. 2012. way explicitly expressed in the text (e.g. through Local characterizations of causal bayesian networks. the use of particular constructions).Taking advan- In Graph Structures for Knowledge Representation and Reasoning, pages 1–17. Springer. tage of how causation is expressed in language may benefit our causal model, and is a potential path for John B Black and Gordon H Bower. 1980. Story under- future work. standing as problem-solving. Poetics, 9(1-3):223– 250. 6 Conclusions and Future Work Eduardo Blanco, Nuria Castell, and Dan I Moldovan. 2008. Causal relation extraction. In Lrec. In this work we argued for a causal basis in script learning. We showed how this causal definition Antoine Bosselut, Hannah Rashkin, Maarten Sap, Chai- tanya Malaviya, Asli Celikyilmaz, and Yejin Choi. could be formalized and used in practice utiliz- 2019. Comet: Commonsense transformers for auto- ing the tools of causal inference, and verified our matic knowledge graph construction. ACL 2019. method with human and automatic evaluations. In Gordon H Bower, John B Black, and Terrence J Turner. the current work, we showed a method calculating 1979. Scripts in memory for text. Cognitive psychol- the ‘goodness’ of a script in the simplest case: be- ogy, 11(2):177–220. tween pairwise events, which we showed still to Paul Van den Broek. 1990. The causal inference maker: be quite useful. A causal definition is in no way Towards a process model of inference generation in limited to this pairwise case, and future work may text comprehension. Comprehension processes in generalize it to the sequential case or to event repre- reading, pages 423–445. sentations that are compositional. Having a causal Paul Van den Broek and Robert F Lorch Jr. 1993. Net- model shines a light on the assumptions made here, work representations of causal relations in memory and indeed, future work may further refine or over- for narrative texts: Evidence from primed recogni- haul them, a process which may further shine a tion. Discourse processes, 16(1-2):75–98. light on the nature of the knowledge we are after. Nathanael Chambers. 2013. Event schema induction with a probabilistic entity-driven model. In Proceed- Acknowledgements ings of EMNLP, volume 13. This work was supported by DARPA KAIROS and Nathanael Chambers. 2017. Behind the scenes of an the Allen Institute for AI. The U.S. Government is evolving event cloze test. In Proceedings of the 2nd Workshop on Linking Models of Lexical, Sentential authorized to reproduce and distribute reprints for and Discourse-level Semantics, pages 41–45. governmental purposes.The views and conclusions contained in this pub-lication are those of the au- Nathanael Chambers and Dan Jurafsky. 2008. Unsu- pervised learning of narrative event chains. In Pro- thors and should not be interpreted as representing ceedings of the Association for Computational Lin- official policies or endorsement. guistics (ACL), Hawaii, USA. Nathanael Chambers and Dan Jurafsky. 2009. Unsuper- References vised learning of narrative schemas and their partici- pants. In Proceedings of the Association for Compu- Valerie Abbott, John B Black, and Edward E Smith. tational Linguistics (ACL), Singapore. 1985. The representation of scripts in memory. Jour- nal of memory and language, 24(2):179–199. Danqi Chen and Christopher Manning. 2014. A fast and accurate dependency parser using neural net- Niranjan Balasubramanian, Stephen Soderland, works. In Proceedings of the 2014 Conference on Oren Etzioni Mausam, and Oren Etzioni. 2013. Empirical Methods in Natural Language Processing Generating coherent event schemas at scale. In (EMNLP), pages 740–750, Doha, Qatar. Association Proceedings of EMNLP. for Computational Linguistics.

7591 Victor Chernozhukov, Denis Chetverikov, Mert Tao Ge, Lei Cui, Heng Ji, Baobao Chang, and Zhifang Demirer, Esther Duflo, Christian Hansen, Whitney Sui. 2016. Discovering concept-level event associa- Newey, and James Robins. 2018. Double/debiased tions from a text stream. In Natural Language Un- machine learning for treatment and structural derstanding and Intelligent Applications, pages 413– parameters. 424. Springer.

Jackie Chi Kit Cheung, Hoifung Poon, and Lucy Van- Roxana Girju. 2003. Automatic detection of causal re- derwende. 2013. Probabilistic frame induction. In lations for question answering. In Proceedings of Proceedings of NAACL. the ACL 2003 workshop on Multilingual summariza- tion and question answering-Volume 12, pages 76– Timothy Chklovski and Patrick Pantel. 2004. Verbo- 83. Association for Computational Linguistics. cean: Mining the web for fine-grained semantic verb relations. In Proceedings of the 2004 Conference on Roxana Girju and Dan Moldovan. 2002. Mining an- Empirical Methods in Natural Language Processing, swers for causation questions. pages 33–40. Arthur C Graesser, Murray Singer, and Tom Trabasso. Kenneth Church and Patrick Hanks. 1990. Word as- 1994. Constructing inferences during narrative text sociation norms, mutual information, and lexicogra- comprehension. Psychological review, 101(3):371. phy. Computational Linguistics, 16(1). David Graff, Junbo Kong, Ke Chen, and Kazuaki Kevin Clark and Christopher D. Manning. 2016a. Maeda. 2003. English gigaword. Linguistic Data Deep reinforcement learning for mention-ranking Consortium, Philadelphia, 4(1):34. coreference models. In Proceedings of the 2016 Miguel A Hernan and James M Robins. 2019. Causal Conference on Empirical Methods in Natural Lan- inference: What if. CRC Boca Raton, FL;. guage Processing, pages 2256–2262, Austin, Texas. Association for Computational Linguistics. J Edward Hu, Rachel Rudinger, Matt Post, and Ben- jamin Van Durme. 2019. Parabank: Monolin- Kevin Clark and Christopher D. Manning. 2016b. Im- gual bitext generation and sentential paraphrasing proving coreference resolution by learning entity- via lexically-constrained neural machine translation. level distributed representations. In Proceedings arXiv preprint arXiv:1901.03644. of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Bram Jans, Steven Bethard, Ivan Vulic,´ and pages 643–653, Berlin, Germany. Association for Marie Francine Moens. 2012. Skip n-grams Computational Linguistics. and ranking functions for predicting script events. In Proceedings of the 13th Conference of the William Croft. 1991. Syntactic categories and gram- European Chapter of the Association for Computa- matical relations: The cognitive organization of in- tional Linguistics, pages 336–344. Association for formation. University of Chicago Press. Computational Linguistics.

William Croft. 2012. Verbs: Aspect and causal struc- Katherine A Keith, David Jensen, and Brendan ture. OUP Oxford. O’Connor. 2020. Text and causal inference: A review of using text to remove confounding from Gerald DeJong. 1983. Acquiring schemata through un- causal estimates. ACL 2020. derstanding and generalizing plans. In IJCAI. Edward H Kennedy. 2016. Semiparametric theory and Quang Xuan Do, Yee Seng Chan, and Dan Roth. empirical processes in causal inference. In Statisti- 2011. Minimally supervised event causality identi- cal causal inferences and their applications in pub- fication. In Proceedings of the Conference on Em- lic health research, pages 141–167. Springer. pirical Methods in Natural Language Processing, pages 294–303. Association for Computational Lin- Ryan Kiros, Yukun Zhu, Ruslan R Salakhutdinov, guistics. Richard Zemel, Raquel Urtasun, Antonio Torralba, and Sanja Fidler. 2015. Skip-thought vectors. In Alexander D’Amour, Peng Ding, Avi Feller, Lihua Lei, C. Cortes, N. D. Lawrence, D. D. Lee, M. Sugiyama, and Jasjeet Sekhon. 2020. Overlap in observational and R. Garnett, editors, Advances in Neural Informa- studies with high-dimensional covariates. Journal tion Processing Systems 28, pages 3294–3302. Cur- of Econometrics. ran Associates, Inc.

Francis Ferraro and Benjamin Van Durme. 2016. A Manabu Kuroki and Judea Pearl. 2014. Measure- unified bayesian model of scripts, frames and lan- ment bias and effect restoration in causal inference. guage. In Thirtieth AAAI Conference on Artificial Biometrika, 101(2):423–437. Intelligence. Christopher D. Manning, Mihai Surdeanu, John Bauer, Aaron Fisher and Edward H Kennedy. 2018. Visually Jenny Finkel, Steven J. Bethard, and David Mc- communicating and teaching intuition for influence Closky. 2014. The Stanford CoreNLP natural lan- functions. arXiv preprint arXiv:1810.03260. guage processing toolkit. In Proceedings of 52nd

7592 Annual Meeting of the Association for Computa- Haoruo Peng and Dan Roth. 2016. Two discourse tional Linguistics: System Demonstrations, pages driven language models for semantics. In Proceed- 55–60. ings of the 54th Annual Meeting of the Association for Computational Linguistics, ACL 2016, August 7- Gail McKoon and Roger Ratcliff. 1986. Inferences 12, 2016, Berlin, Germany, Volume 1: Long Papers. about predictable events. Journal of Experimen- tal Psychology: Learning, memory, and cognition, Karl Pichotta and Raymond J Mooney. 2016. Learn- 12(1):82. ing statistical scripts with lstm recurrent neural net- works. In AAAI, pages 2800–2806. Gail McKoon and Roger Ratcliff. 1992. Inference dur- ing reading. Psychological review, 99(3):440. Sameer Pradhan, Alessandro Moschitti, Nianwen Xue, Olga Uryupina, and Yuchen Zhang. 2012. Conll- Wang Miao, Zhi Geng, and Eric J Tchetgen Tchetgen. 2012 shared task: Modeling multilingual unre- 2018. Identifying causal effects with proxy vari- stricted coreference in ontonotes. In Joint Confer- ables of an unmeasured confounder. Biometrika, ence on EMNLP and CoNLL-Shared Task, pages 1– 105(4):987–993. 40. Association for Computational Linguistics.

Marvin Minsky. 1974. A framework for representing Victor Raskin, Sergei Nirenburg, Christian F. Hempel- knowledge. MIT Laboratory Memo 306. mann, Inna Nirenburg, and Katrina E. Triezenberg. 2003. The genesis of a script for bankruptcy in Ashutosh Modi. 2016. Event embeddings for seman- ontological semantics. In Proceedings of the HLT- tic script modeling. In Proceedings of The 20th NAACL 2003 Workshop on Text Meaning, pages 30– SIGNLL Conference on Computational Natural Lan- 37. guage Learning, pages 75–83. Michaela Regneri, Alexander Koller, and Manfred Ashutosh Modi and Ivan Titov. 2014. Inducing neu- Pinkal. 2010. Learning script knowledge with web ral models of script knowledge. In Proceedings of experiments. In Proceedings of the 48th Annual the Eighteenth Conference on Computational Nat- Meeting of the Association for Computational Lin- ural Language Learning, pages 49–57, Ann Arbor, guistics, pages 979–988. Association for Computa- Michigan. Association for Computational Linguis- tional Linguistics. tics. Mehwish Riaz and Roxana Girju. 2013. Toward a bet- Raymond Mooney and Gerald DeJong. 1985. Learning ter understanding of causality between verbal events: schemata for natural language processing. In Ninth Extraction and analysis of the causal power of verb- International Joint Conference on Artificial Intelli- verb associations. In Proceedings of the SIGDIAL 2013 Conference gence (IJCAI), pages 681–687. , pages 21–30. Rachel Rudinger, Pushpendre Rastogi, Francis Ferraro, Joakim Nivre, Marie-Catherine de Marneffe, Filip Gin- and Benjamin Van Durme. 2015. Script induction as ter, Yoav Goldberg, Jan Hajic, Christopher D. Man- language modeling. In Proceedings of the 2015 Con- ning, Ryan McDonald, Slav Petrov, Sampo Pyysalo, ference on Empirical Methods in Natural Language Natalia Silveira, Reut Tsarfaty, and Daniel Zeman. Processing, pages 1681–1686. 2016. Universal dependencies v1: A multilingual treebank collection. In Proceedings of the Tenth In- Rachel Rudinger, Adam Teichert, Ryan Culkin, Sheng ternational Conference on Language Resources and Zhang, and Benjamin Van Durme. 2018a. Neural- Evaluation (LREC 2016), Paris, France. European davidsonian semantic proto-role labeling. In Pro- Language Resources Association (ELRA). ceedings of the 2018 Conference on Empirical Meth- ods in Natural Language Processing, pages 944– John Walker Orr, Prasad Tadepalli, Janardhan Rao 955, Brussels, Belgium. Association for Computa- Doppa, Xiaoli Fern, and Thomas G Dietterich. 2014. tional Linguistics. Learning scripts as hidden markov models. In Twenty-Eighth AAAI Conference on Artificial Intel- Rachel Rudinger, Aaron Steven White, and Benjamin ligence. Van Durme. 2018b. Neural models of factuality. arXiv preprint arXiv:1804.02472. Patrick Pantel and Deepak Ravichandran. 2004. Auto- matically labeling semantic classes. In Proceedings Keisuke Sakaguchi and Benjamin Van Durme. 2018. of the Human Language Technology Conference of Efficient online scalar annotation with bounded sup- the North American Chapter of the Association for port. ACL. Computational Linguistics: HLT-NAACL 2004. Roger C Schank and Robert P Abelson. 1975. Scripts, Judea Pearl. 1995. Causal diagrams for empirical re- plans, and knowledge. IJCAI. search. Biometrika, 82(4):669–688. Roger C Schank and Robert P Abelson. 1977. Scripts, Judea Pearl. 2000. Causality: models, reasoning and plans, goals, and understanding: An inquiry into hu- inference, volume 29. Springer. man knowledge structures. Psychology Press.

7593 Uri Shalit, Fredrik D Johansson, and David Sontag. Zach Wood-Doughty, Ilya Shpitser, and Mark Dredze. 2017. Estimating individual treatment effect: gen- 2018. Challenges of using text classifiers for causal eralization bounds and algorithms. In Proceedings inference. In Proceedings of the Conference on Em- of the 34th International Conference on Machine pirical Methods in Natural Language Processing. Learning-Volume 70, pages 3076–3085. JMLR. org. Conference on Empirical Methods in Natural Lan- guage Processing, volume 2018, page 4586. NIH Ilya Shpitser and Judea Pearl. 2008. Complete identifi- Public Access. cation methods for the causal hierarchy. Journal of Machine Learning Research, 9(Sep):1941–1979. James Woodward. 2005. Making things happen: A the- ory of causal explanation. Oxford university press. Robyn Speer and Catherine Havasi. 2012. Represent- Sheng Zhang, Rachel Rudinger, Kevin Duh, and Ben- ing general relational knowledge in conceptnet 5. In jamin Van Durme. 2017. Ordinal common-sense in- LREC. ference. TACL.

Peter Spirtes, Clark N Glymour, Richard Scheines, Yukun Zhu, Ryan Kiros, Rich Zemel, Ruslan Salakhut- David Heckerman, Christopher Meek, Gregory dinov, Raquel Urtasun, Antonio Torralba, and Sanja Cooper, and Thomas Richardson. 2000. Causation, Fidler. 2015. Aligning books and movies: Towards prediction, and search. MIT press. story-like visual explanations by watching movies and reading books. In Proceedings of the IEEE inter- Leonard Talmy. 1988. Force dynamics in language and national conference on computer vision, pages 19– cognition. Cognitive science, 12(1):49–100. 27.

Kristina Toutanova, Dan Klein, Christopher D Man- A Appendix ning, and Yoram Singer. 2003. Feature-rich part-of- speech tagging with a cyclic dependency network. A.1 Event Representation In Proceedings of the 2003 Conference of the North American Chapter of the Association for Computa- To best contrast with prior work, we use the event tional Linguistics on Human Language Technology- representation of Chambers and Jurafsky(2008) Volume 1, pages 173–180. Association for computa- and others (Jans et al., 2012; Rudinger et al., 2015). tional Linguistics. Each event is a pair (p, d), where p is the event Tom Trabasso and Linda L Sperry. 1985. Causal relat- predicate (e.g. hit), and d is the dependency re- edness and importance of story events. Journal of lation (e.g. nsubj) between the predicate and the Memory and language, 24(5):595–611. protagonist entity. The protagonist is the entity that participates in every event in the considered event Tom Trabasso and Paul Van Den Broek. 1985. Causal chain, e.g., the ‘Bob’ in the chain ‘Bob sits, Bob thinking and the representation of narrative events. Journal of memory and language, 24(5):612–630. eats, Bob pays.’

Noah Weber, Niranjan Balasubramanian, and A.2 Data Pre-Processing Nathanael Chambers. 2018a. Event representations For these experiments, we use the Toronto Books with tensor-based compositions. In Thirty-Second AAAI Conference on Artificial Intelligence. corpus (Zhu et al., 2015; Kiros et al., 2015), a col- lection of fiction novels spanning multiple genres. Noah Weber, Leena Shekhar, Niranjan Balasubrama- The original corpus contains 11,040 books by un- nian, and Nate Chambers. 2018b. Hierarchical quan- published authors. We remove duplicate books tized representations for script generation. In Pro- from the corpus (by exact file match), leaving a ceedings of the 2018 Conference on Empirical Meth- ods in Natural Language Processing, pages 3783– total of 7,101 books; a distribution by genre is pro- 3792. vided in Table7. The books are assigned randomly to train, development, and test splits in 90%-5%- Aaron Steven White, D. Reisinger, Keisuke Sakaguchi, 5% proportions (6,405 books in train, and 348 in Tim Vieira, Sheng Zhang, Rachel Rudinger, Kyle development and test splits each). Each book is Rawlins, and Benjamin Van Durme. 2016. Univer- sal decompositional semantics on universal depen- then sentence-split and tokenized with CoreNLP dencies. In Proceedings of the 2016 Conference on 3.8 (Manning et al., 2014); these sentence and to- Empirical Methods in Natural Language Process- ken boundaries are observed in all downstream ing, pages 1713–1723, Austin, Texas. Association processing. for Computational Linguistics. A.2.1 Narrative Chain Extraction Pipeline Phillip Wolff and Grace Song. 2003. Models of cau- sation and the semantics of causal verbs. Cognitive In order to extract the narrative chains from the Psychology, 47(3):276–332. Toronto Books data, we implement the following

7594 Adventure 390 Other 284 take the dependency relation type (e.g., nsubj) be- Fantasy 1,440 Romance 1,437 tween the predicate head and argument head in- Historical 161 Science Fiction 425 dices (as determined by the UD parse); if a direct Horror 347 Teen 281 arc relation does not exist, we instead take the uni- Humor 237 Themes 32 directional dependency path from predicate to ar- Literature 289 Thriller 316 gument; if a unidirectional path does not exist, we Mystery 512 Vampires 131 use a generic “arg” relation. New Adult 702 Young Adult 117 To extract a factuality feature for each narrative event (i.e. whether the event happened or not, ac- Table 7: Distribution of books within each genre of the cording to the meaning of the text), we use the deduplicated Toronto Books corpus. neural model of Rudinger et al.(2018a).As input to this model, we provide the full sentence in which pipeline. First, we note that coreference resolution the event appears, as well as the index of the event systems are trained on documents much smaller predicate’s head token. The model returns a fac- than full novels (Pradhan et al., 2012); to accom- tuality score on a [−3, 3] scale, which is then dis- modate this limitation, we partition each novel into cretized using the following intervals: [1, 3] is “pos- non-overlapping windows that are 100 sentences in itive” (+), (−1, 1) is “uncertain,” and [−3, −1] is length, yielding approximately 400,000 windows “negative” (−). in total. We then run CoreNLP’s universal depen- From this extraction pipeline, we yield one se- dency parser (Nivre et al., 2016; Chen and Man- quence of narrative events (i.e., narrative chain) per ning, 2014), part of speech tagger (Toutanova et al., text window. 2003), and neural coreference resolution system A.3 Training and Model Details - Causal (Clark and Manning, 2016a,b) over each window Model of text. For each window, we select the longest coreference chain and call the entity in that chain A.3.1 RNN Encoder the “protagonist,” following Chambers and Juraf- We use a single layer GRU based RNN encoder sky(2008). with a 300 dimensional hidden state and 300 di- We feed the resulting universal dependency (UD) mensional input event embeddings to encode the parses into PredPatt (White et al., 2016), a rule- previous events into a single 300 dimensional vec- based predicate-argument extraction system that tor. runs over universal dependency parses. From Pred- A.3.2 CNN Encoder Patt output, we extract predicate-argument edges, i.e., a pair of token indices in a given sentence We use a CNN to encode the text into a 300 dimen- where the first index is the head of a predicate, and sional output vector. The CNN uses 4 filters with the second index is the head of an argument to that ngram windows of (2, 3, 4, 5) and max pooling. predicate. Edges with non-verbal predicates are A.3.3 Training Details - Pretraining discarded. The conditional for the Causal model is trained At this stage in the pipeline, we merge infor- using Adam with a learning rate of 0.001, gradient mation from the coreference chain and predicate- clipping at 10, and a batch size of 512. The model argument edges to determine which events the pro- is trained to minimize cross entropy loss. We train tagonist is participating in. For each predicate- the model until loss on the validation set does not argument edge in every sentence, we discard it go down after three epochs, afterwhich we keep the if the argument index does not match the head model with the best validation performance, which of a protagonist mention. Each of the remaining in our case was epoch 4 predicate-argument edges therefore represents an event that the protagonist participated in. A.3.4 Training Details - Finetuning With a list of PredPatt-determined predicate- The model is then finetuned on our dataset of 2000 argument edges (and their corresponding sen- annotated examples. We use the same objective as tences), we are now able to extract the narrative above, training using Adam with a learning rate of event representations, (p, d) For p, we take the 0.00001, gradient clipping at 10, and a batch size lemma of the (verbal) predicate head. For d, we of 512. We split our 2000 samples into a train set of

7595 Figure 3: The annotation interface for the out-of-text events annotation.

1800 examples and a dev set of 200 examples. We train the model in a way similar to above, keeping the best validation model (at epoch 28).

A.4 Training and Model Details - LM Figure 4: The annotation interface for the pairwise hu- Baseline man evaluation annotation experiment. We use a 2 layer GRU based RNN encoder with a 512 dimensional hidden state and 300 dimensional input event embeddings as our baseline event se- quence LM model. A.4.1 Training Details The LM model is trained using Adam with a learn- ing rate of 0.001, gradient clipping at 10, and a batch size of 64. We found using dropout at the embedding layer and the output layers to be helpful (with dropout probability of 0.1). The model is trained to minimize cross entropy loss. We train the model until loss on the validation set does not go down after three epochs, afterwhich we keep the model with the best validation performance, which in our case was epoch 5.

A.5 Annotation Interfaces To get an idea for about the annotation set ups used here, we also provide screen shots of the annotation suites for all three annotation experiments. The out-of-text annotation experiment of Section 3.3 is shown in Figure3. The pairwise annotation evaluation of Section 4.3 is shown in Figure4. The chain completion annotation evaluation of Section 4.4 is shown in Figure5. Figure 5: The annotation interface for the chain com- pletion human evaluation annotation experiment.

7596