Event Narrative Module, version 3 Deliverable D5.1.3
Transcription
Event Narrative Module, version 3 Deliverable D5.1.3
Event Narrative Module, version 3 Deliverable D5.1.3 Version FINAL Authors: Piek Vossen1 , Tommaso Caselli1 , Agata Cybulska1 , Antske Fokkens1 , Filip Ilievski1 , Anne-Lyse Minard2 , Paramita Mirza2 , Itziar Aldabe3 , Egoitz Laparra3 , German Rigau3 Affiliation: (1) VUA, (2) FBK, (3) UPV/EHU Building structured event indexes of large volumes of financial and economic data for decision making ICT 316404 Event Narrative Module, version 3 2/148 Grant Agreement No. Project Acronym Project Full Title 316404 NEWSREADER Building structured event indexes of large volumes of financial and economic data for decision making. Funding Scheme FP7-ICT-2011-8 Project Website http://www.newsreader-project.eu/ Prof. dr. Piek T.J.M. Vossen VU University Amsterdam Tel. + 31 (0) 20 5986466 Project Coordinator Fax. + 31 (0) 20 5986500 Email: piek.vossen@vu.nl Document Number Deliverable D5.1.3 Status & Version FINAL Contractual Date of Delivery October 2015 Actual Date of Delivery January 2016 Type Report Security (distribution level) Public Number of Pages 148 WP Contributing to the Deliverable WP05 WP Responsible VUA EC Project Officer Susan Fraser 1 1 Authors: Piek Vossen , Tommaso Caselli , Agata Cybulska1 , Antske Fokkens1 , Filip Ilievski1 , Anne-Lyse Minard2 , Paramita Mirza2 , Itziar Aldabe3 , Egoitz Laparra3 , German Rigau3 Keywords: Event detection, event-coreference, event components, NAF, RDF, SEM, GAF, event relations, timelines, storylines, attribution and perspective, crosslingual event extraction Abstract: This deliverable describes the final version of the modules that convert the Natural Language Processing output in NAF to the Semantic Web interpretation in SEM-RDF. We describe the way we represent instances in SEM-RDF, relations between these instances and the GAF pointers to the mentions in the text. Since instances can have many different mentions in different sources, we define identity criteria for mentions. The resulting SEM-RDF representations are loaded into the KnowledgeStore as triples that can be queried through SPARQL. In addition to the data on individual events, we also extract causal and temporal relations between events. Time anchoring of events and the relations are used to create timelines of events. Timelines are then used to derive storylines. In addition to the event data, we also derive the perspective of the sources of information with respect to the data. This is represented in a separate RDF structure that models provenance and attribution. Finally, we describe how the system extract semantic data across different languages. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 3/148 Table of Revisions Version Date Description and reason By 1.1 1 October 2015 First structure 1.2 5 October 2015 Streaming architecture 1.3 11 October 2015 Event relations 1.4 13 October 2015 Perspectives 1.5 13 October 2015 Storylines 1.6 21 October 2015 RDF evaluation and clean up 1.7 21 October 2015 Review 1.8 5 October 2015 Streaming architecture revision 1.9 21 October 2015 Revision all 2.0 January 2015 Event coreference 2.1 January 2015 Event coreference 2.2 January 2016 Cross-lingual extraction 2.2 January 2016 Reviewed by 2.3 January 2016 Revised after review 2.4 29 January 2016 Check by coordinator Piek Vossen, VUA Filip Ilievski, VUA Anne-Lyse Minard, FBK Antske Fokkens, VUA Tommaso Caselli, VUA Piek Vossen, VUA Marieke van Erp, VUA Filip Ilievski, VUA Piek Vossen, VUA Piek Vossen, VUA Agata Cybulska, VUA Piek Vossen, VUA Egoitz Laparra, EHU Piek Vossen, VUA VUA NewsReader: ICT-316404 Affected tions all 2 4 6 5 2 all 2 all 3 3 7 all all - February 1, 2016 sec- Event Narrative Module, version 3 NewsReader: ICT-316404 4/148 February 1, 2016 Event Narrative Module, version 3 5/148 Executive Summary This deliverable describes the final version of the modules that convert the Natural Language Processing output in NAF to the Semantic Web interpretation in SEM-RDF. We describe the way we represent instances in SEM-RDF, relations between these instances and the GAF pointers to the mentions in the text. Since instances can have many different mentions in different sources, we define identity criteria for mentions. The resulting SEMRDF representations are loaded into the KnowledgeStore as triples that can be queried through SPARQL. Two different approaches have been defined for processing NAF files: 1) assuming an empty KnowledgeStore and a batch of NAF files, all NAF files are processed and compared after which the KnowledgeStore is populated with the RDF and 2) assuming a streaming set-up in which each NAF file is processed one-by-one and the result is compared with the data that is already in the KnowledgeStore. The data is event-centric, which means we provide only the data relevant for the events detected. Every event is assumed to be anchored to a date or period in time and have at least one participant. Next to the data on individual events, we also extract causal and temporal relations between events. Time anchoring of events and the relations are used to create timelines of events. Timelines are then used to derive storylines. In addition to the event data, we also derive the perspective of the sources of information with respect to the data. This is represented in a separate RDF structure that models provenance and attribution. Finally, since the RDF representation is agnostic for the expression in language and our NLP modules for English, Spanish, Dutch and Italian are interoperable through NAF, the NAF2SEM module can extract the same RDF representation across different languages. This provides a unique opportunity for comparing the semantic processing of text across languages. We describe the cross-lingual processing and the comparison across the language-system output. In comparison with the previous deliverable D5.1.2 almost all sections changed. Minor changes were made for sections 4, 5 and 6. Other sections were changed drastically. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 NewsReader: ICT-316404 6/148 February 1, 2016 Event Narrative Module, version 3 7/148 Contents Table of Revisions 3 1 Introduction 13 2 Interpreting NAF-XML as SEM-RDF 2.1 Extracting instances from NAF layers . . . . 2.1.1 Entities and non-entities . . . . . . . 2.1.2 Events . . . . . . . . . . . . . . . . . 2.1.3 Participant and event relations . . . 2.1.4 Temporal anchoring . . . . . . . . . . 2.2 Identity across events . . . . . . . . . . . . . 2.2.1 Event comparison in batch mode . . 2.2.2 Event comparison in streaming mode 2.3 Evaluation . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16 19 19 23 25 28 32 35 39 44 . . . . . . . . . . . . . . . 47 47 47 48 49 51 51 52 52 54 54 57 58 61 67 77 . . . . . . . . 79 79 79 80 83 83 84 87 87 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3 Event Coreference 3.1 Bag of Events Approach . . . . . . . . . . . . . . . . . . . . . . . . 3.1.1 The Overall Approach . . . . . . . . . . . . . . . . . . . . . 3.1.2 Two-step Bag of Events Approach . . . . . . . . . . . . . . . 3.1.3 Step 1: Clustering Documents Using Bag of Events Features 3.1.4 Step 2: Clustering Sentence Templates . . . . . . . . . . . . 3.1.5 One-step Bag of Events Approach . . . . . . . . . . . . . . . 3.1.6 Corpus . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.1.7 Experimental Set Up . . . . . . . . . . . . . . . . . . . . . . 3.1.8 Baseline . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.1.9 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.1.10 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.2 Evaluation of the NewsReader pipeline . . . . . . . . . . . . . . . . 3.2.1 NewsReader output . . . . . . . . . . . . . . . . . . . . . . . 3.2.2 Maximizing the event detection . . . . . . . . . . . . . . . . 3.2.3 Conclusion NewsReader cross-document event coreference . 4 Event Relations 4.1 Temporal Relations . . . . . . . . . . 4.1.1 Annotation Schema . . . . . . 4.1.2 Temporal Relation Extraction 4.2 Causal Relation . . . . . . . . . . . . 4.2.1 Annotation Scheme . . . . . . 4.2.2 Causal Relation Extraction . 4.3 Predicate Time Anchors . . . . . . . 4.3.1 Annotation Scheme . . . . . . NewsReader: ICT-316404 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . February 1, 2016 Event Narrative Module, version 3 4.3.2 8/148 Predicate Time Anchor Relation Extraction . . . . . . . . . . . . . 5 From TimeLines to StoryLines 5.1 TimeLine extraction . . . . . . . . . . . . . . . . . . . . . . . 5.1.1 TimeLines: task description . . . . . . . . . . . . . . . 5.1.2 System Description and Evaluation . . . . . . . . . . . 5.1.3 Document level time-anchoring for TimeLine extraction 5.2 Storylines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.2.1 StoryLines aggregated from entity-centered TimeLines 5.2.2 Storylines aggregated from climax events . . . . . . . . 5.2.3 Workshop on Computing News Storylines . . . . . . . 6 Perspectives 6.1 Basic Perspective Module . . 6.2 Factuality module . . . . . . . 6.2.1 Event factuality . . . . 6.2.2 Identifying factualities 6.2.3 Factuality module . . . 6.2.4 Future work . . . . . . 6.3 A perspective model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88 . . . . . . . . 89 . 90 . 90 . 92 . 93 . 97 . 98 . 101 . 107 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 108 108 113 113 114 116 117 118 7 Cross-lingual extraction 7.1 Crosslingual extraction of entities . 7.2 Crosslingual extraction of events . . 7.3 Crosslingual extraction of relations 7.4 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121 127 130 134 137 . . . . . . . . . . . . . . 8 Conclusions 138 9 Appendix 139 NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 9/148 List of Tables 1 Cross-document event coreference arguments for stream processing . . . . . 41 2 Quality triple evaluation of SEM-RDF extracted from Wikinews. . . . . . . 45 3 Detailed quality triple evaluation of SEM-RDF extracted from Wikinews with and without taking even-coreference into account. . . . . . . . . . . . 46 4 Sentence template ECB topic 1, text 7, sentence 1 . . . . . . . . . . . . . . 49 5 Sentence template ECB topic 1, text 7, sentence 2 . . . . . . . . . . . . . . 49 6 Document template ECB topic 1, text 7, sentences 1-2 . . . . . . . . . . . 49 7 ECB+ statistics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50 8 Features grouped into four categories: L-Lemma based, A-Action similarity, D-location within Discourse, E-Entity coreference and S-Synset based. . . . 53 9 Bag of events approach to event coreference resolution, evaluated on the ECB+ in MUC, B3, mention-based CEAF, BLANC and CoNLL F measures. 55 10 Baseline results on the ECB+: singleton baseline and lemma match of event triggers evaluated in MUC, B3, mention-based CEAF, BLANC and CoNLL F. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55 11 Best scoring two-step bag of events approach, evaluated in MUC, B3, entitybased CEAF, BLANC and CoNLL F in comparison with related studies. Note that the BOE approach uses gold while related studies system mentions. 56 12 BLANC refererence results macro averaged over ECB+ topics in terms of recall (R), precision (P) and F1 (F) for NewsReader output with different proportions of WordNet synsets to match: S=only synset matches, SL= Synsets matches if synsets and lemma matches if no synsets associated, L=lemmas only. Different columns represent proportions in steps of 10% from 1% to 100%. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63 BLANC refererence results macro averaged over ECB+ topics in terms of recall (R-SL30), precision (P-SL30) and F1 (F-SL30). AR is stable across the results, meaning that a single participant in any role needs to match. We varied the hypernyms (H) and lowest-common-subsumer (L) for action matches and the time constraints: no time constraint (NT), year (Y), month (M) and day (D) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66 BLANC refererence results macro averaged over ECB+ topics in terms of recall (R-SL30), precision (P-SL30) and F1 (F-SL30). The hypernyms (H), lowest-common-subsumer (L) and time constraint month (M) are kept stable. We varied the role-participant constraints: NR=no constraint, A0 role participant should match, A1 should match, A2 should match, A0 and A1 should match, A0 and A2 should match, A1 and A2 should match . . . . . 66 13 14 NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 15 16 17 18 19 20 21 22 23 24 10/148 Macro averaged Mention identification for ECB+ topics. NWR=NewsReader pipeline v3.0 without adaptation, EDg(old)=NWR augmented with EventDetection trained with gold data, EDg(old)EC= same as EDg(old) but skipping predicates with an Event class, EDs(ilver)= NWR augmented with EventDetection trained with silver data, EDs(ilver)EC= same as EDs but skipping predicates with an Event class. . . . . . . . . . . . . . . . . . . . Predicates missed more than once by NewsReader extended with EventDetection (silver) and Event class filter as events in ECB+ . . . . . . . . . . Predicates missed once by NewsReader extended with EventDetection (silver) and Event class filter as events in ECB+ . . . . . . . . . . . . . . . . Predicates invented and occurring more than once by NewsReader extended with EventDetection (silver) and Event class filter as events in ECB+ . . . Predicates invented and occurring only once by NewsReader extended with EventDetection (silver) and Event class filter as events in ECB+ . . . . . . Reference results macro averaged over ECB+ topics with different options for event detection. kNWR=NewsReader event detection without invented mentions, maximizing precision, NWR=NewsReader pipeline v3.0 without adaptation, EDg=NWR augmented with EventDetection trained with gold data, EDgEC= same as EDg but skipping predicates with an Event class, EDs= NWR augmented with EventDetection trained with silver data, EDsEC= same as EDs but skipping predicates with an Event class. ARM= standard setting one participant in any role (AR), time month match and action concept and phrase match 30%, mR= maximizing recall by no constraints on participant match and time, action concept and phrase match 1%, mP=maximizing precision by participant roles A0A1, time day match and action concept and phrase match is set to 100%. . . . . . . . . . . . . Reference results macro averaged over ECB+ corpus with different options for event detection. kNWR=NewsReader event detection without invented mentions, maximizing precision, NWR=NewsReader pipeline v3.0 without adaptation, EDg=NWR augmented with EventDetection trained with gold data, EDgEC= same as EDg but skipping predicates with an Event class, EDs= NWR augmented with EventDetection trained with silver data, EDsEC= same as EDs but skipping predicates with an Event class. ARM= standard setting one participant in any role (AR), time month match and action concept and phrase match 30%, mR= maximizing recall by no constraints on participant match and time, action concept and phrase match 1%, mP=maximizing precision by participant roles A0A1, time day match and action concept and phrase match is set to 100%. . . . . . . . . . . . . Reference results macro averaged over ECB+ corpus as reported by Yang et al. (2015) for state-of-the-art machine learning systems . . . . . . . . . . Distribution of tell, kill and election over all text and annotated text per mention, document and topic in ECB+ . . . . . . . . . . . . . . . . . . . . Temporal relations in TimeML annotation . . . . . . . . . . . . . . . . . . NewsReader: ICT-316404 67 69 70 71 72 73 75 75 76 79 February 1, 2016 Event Narrative Module, version 3 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 11/148 Tempeval-3 evaluation on temporal relation classification . . . . . . . . . . 82 CLINK extraction system’s performance. . . . . . . . . . . . . . . . . . . . 87 System Results (micro F1 score) for the SemEval 2015 Task 4 Task A - Main 92 System Results (micro F1 score) for the SemEval 2015 Task 4 Task A Subtask . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92 Results on the SemEval-2015 task . . . . . . . . . . . . . . . . . . . . . . . 97 Figures of the StoryLine gold dataset. . . . . . . . . . . . . . . . . . . . . . 99 Results of the StoryLine extraction process. . . . . . . . . . . . . . . . . . 101 Certainty, polarity and tense values . . . . . . . . . . . . . . . . . . . . . . 114 DBpedia entities extracted for English, Spanish, Italian and Dutch Wikinews with proportion of coverage, measured as macro and micro coverage. I=instances, M=mentions, O=overlap, maC=macro-average over all document results, miC=microAverage over all mentions . . . . . . . . . . . . . . . . . . . . . 130 DBpedia entities in the Wikinews Airbus corpus most frequent in English with Spanish, Italian and Dutch frequencies . . . . . . . . . . . . . . . . . 132 DBpedia entities in the Wikinews Apple corpus most frequent in English with Spanish, Italian and Dutch frequencies . . . . . . . . . . . . . . . . . 132 DBpedia entities in the Wikinews GM, Chrysler, Ford corpus most frequent in English with Spanish, Italian and Dutch frequencies . . . . . . . . . . . 133 DBpedia entities in the Wikinews stock market corpus most frequent in English with Spanish, Italian and Dutch frequencies . . . . . . . . . . . . . 133 ILI-based events extracted for English, Spanish, Italian and Dutch Wikinews with proportion of coverage, measured as macro and micro coverage. I=instances, M=mentions, O=overlap, maC=macro-average over all document results, miC=microAverage over all mentions . . . . . . . . . . . . . . . . . . . . . 134 ILI-based events in the Wikinews Airbus corpus most frequent in English with Spanish, Italian and Dutch frequencies . . . . . . . . . . . . . . . . . 134 ILI-based events in the Wikinews Apple corpus most frequent in English with Spanish, Italian and Dutch frequencies . . . . . . . . . . . . . . . . . 135 ILI-based events in the Wikinews GM, Chrysler, Ford corpus most frequent in English with Spanish, Italian and Dutch frequencies . . . . . . . . . . . 135 ILI-based events in the Wikinews stock market corpus most frequent in English with Spanish and Dutch frequencies . . . . . . . . . . . . . . . . . 135 Triple predicates that are most frequent in the English Wikinews corpus with coverage in Spanish, Italian and Dutch . . . . . . . . . . . . . . . . . 136 ILI-based Triples extracted for English, Spanish, Italian and Dutch Wikinews with proportion of coverage, measured as macro and micro coverage. I=instances, M=mentions, O=overlap, maC=macro-average over all document results, miC=microAverage over all mentions . . . . . . . . . . . . . . . . . . . . . 137 FrameNet frames for contextualEvents . . . . . . . . . . . . . . . . . . . . 139 FrameNet from for sourceEvents . . . . . . . . . . . . . . . . . . . . . . . . 140 FrameNet frames for grammaticalEvents . . . . . . . . . . . . . . . . . . . 140 NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 NewsReader: ICT-316404 12/148 February 1, 2016 Event Narrative Module, version 3 1 13/148 Introduction The goal of the NewsReader project1 is to automatically process massive streams of daily news in 4 different languages to reconstruct longer term story lines of events. For this purpose, we extract events mentioned in news articles, the place and date of their occurrence and who is involved, using a pipeline of Natural Language Processing modules (for the details see Agerri et al. (2015)) that process each news article and store the interpretation in the form of the Natural Language Processing Annotation Format (NAF (Fokkens et al., 2014)). Representations in NAF are mention-based. Mentions of an event, entity, place or time are not unique. The same instance (event, entity, place or date) can be mentioned several times in a single text or in different texts and, likewise NAF represents each mention separately. Consider the following short fragments of two news articles published on the same day: http://www.telegraph.co.uk/finance/newsbysector/industry/engineering/10125280/Porsche-family-buys-back10pc-stake-from-Qatar.html 17 Jun 2013 Porsche family buys back 10pc stake from Qatar Descendants of the German car pioneer Ferdinand Porsche have bought back a 10pc stake in the company that bears the family name from Qatar Holding, the investment arm of the Gulf State’s sovereign wealth fund. ————————————————————————————– http://english.alarabiya.net/en/business/banking-and-finance/2013/06/17/Qatar-Holding-sells-10-stake-in-Porscheto-family-shareholders.html Monday, 17 June 2013 Qatar Holding sells 10% stake in Porsche to founding families Qatar Holding, the investment arm of the Gulf state’s sovereign wealth fund, has sold its 10 percent stake in Porsche SE to the luxury carmaker’s family shareholders, four years after it first invested in the firm. Both fragments describe the same event but do this very differently. The first fragment talks about a buy event of 10pc stake in the Porsche company from Qatar holding by the Porsche family. The second fragments frames this as a sell event of the same stake by Qatar Holding to Porsche. Both articles make reference to the Porsche company, the Porsche family and Qatar Holding in various ways e.g. Descendants of the German car pioneer Ferdinand Porsche and the investment arm of the Gulf state’s sovereign wealth fund. However, there is no difference in the content across the texts. The NewsReader pipeline for processing text deals with this by detecting DBpedia URIs for each entity and resolving coreference relations in the text to connect other expressions to these entities. The same is done for events mentioned in the text: buys back and bought back are represented through a unique URI as a single event, and so are sells and solds in the second fragment. Obviously, these events are not in DBpedia and therefore unique blank URIs are created. In NewsReader, we ultimately generate representations for unique instances represented from these URIs in RDF across all these mentions and the relations between them as RDFtriples according to the Simple Event Model (SEM, van Hage et al. (2011)). Within SEM, 1 FP7-ICT-316404 Building structured event indexes of large volumes of financial and economic data for decision making, www.newsreader-project.eu/ NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 : e v e n t #23 a rdfs : label f n : Buyer fn : S e l l e r f n : Goods sem : hasAtTime 14/148 sem : Event , f n : C o m m e r c e s e l l , f n : Commerce buy ; ” buy ” , ” s e l l ” ; dbp : r e s o u r c e / P o r s c h e ; dbp : r e s o u r c e / Q a t a r I n v e s t m e n t A u t h o r i t y ; : non− e n t i t i e s /10 p c s t a k e ; :20150120. Figure 1: SEM event instance 1 2 3 4 5 6 7 8 9 : e v e n t #23 g a f : denotedBy <h t t p : / /www. t e l e g r a p h . c o . uk#c h a r =15,19> , <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#c h a r =13 ,19 >. dbp : r e s o u r c e / P o r s c h e g a f : denotedBy <h t t p : / /www. t e l e g r a p h . c o . uk#c h a r =0 ,7 >. : non− e n t i t i e s /10 p c s t a k e g a f : denotedBy <h t t p : / /www. t e l e g r a p h . c o . uk#c h a r =25 ,35 >. dbp : r e s o u r c e / Q a t a r I n v e s t m e n t A u t h o r i t y g a f : denotedBy <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#c h a r =0 ,13 >. Figure 2: SEM event instance with mentions data objects are defined for events, actors, places and time with relations between them. Sources that contain the same events, entities and relations thus should result in the same structure in SEM. A SEM representation for the above two fragments would look as in Figure 1. For the event, we created the so-called blank URI event#23 which is an instance of the ontological types sem:Event, fn:Commerce sell, fn:Commerce buy. The two latter types come from FrameNet Baker et al. (1998). Furthermore, we see the labels buy and sell that are aggregated from the two sources. Furthermore, there are triples that relate the event to the entities in the text through FrameNet relations, while the event is anchored to a data through a sem:hasAtTime relation. From both text, only a single event representation is derived with the same set of triples. Since we do not want to lose the connection to the textual mentions of the information, we developed the Grounded Annotation Framework (GAF, Fokkens et al. (2013)), which formally distinguishes between mentions of events and entities in NAF and instances of events and entities in SEM, connected through denotedBy links between representations. For each object in SEM, we therefore provide the pointers to the character offsets in the text where the information is mentioned as additional triples, as is shown in Figure 2. In this deliverable, we describe the modules that read NLP interpretations from NAF files and create the RDF representations according to SEM and GAF. The RDF representations are eventually stored into the KnowledgeStore (Corcoglioniti et al. (2013)). The following Figure 3 shows the position of modules in the overall process. In this deliverable, we describe the final implementation of the modules that interpret NAF data as SEM/GAF. The document is structured as follows. In section 2, we describe the implementation of the NAF2SEM module that carries out the conversion. There are two implementations: one for batch processing and one for stream processing. The latter implementation is part of the streaming end-to-end architecture through which a text file is completely processed through the NLP pipeline, interpreted as SEM-RDF and integrated into the KnowledgeStore without intermediate storage. In section 3, we report NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 15/148 Figure 3: Input-output schema for Work Packages in NewsReader on the progress made on event coreference, which is the basis for establishing instance representation, detailing the current implementation of the NAF2SEM module. In section 4, we describe two modules for detecting temporal and causal relations. These relations are used in section 5.2, where we report on the modules that extract timelines and storylines. Section 6 discusses the implementation of the perspective and attribution module. Events are divided into contextual events and source events. The former describe the statements about the changes in the world, while the latter events indicate the relations between sources and these statements. We developed a separate module that derives the perspective and attribution values from these two event types, incorporating the output of the opinion layer and the attribution layer in NAF. Finally in section 7, we provide our results on crosslingual event extraction (i.e. combining the event information extracted from documents in different languages). The English Wikinews corpus was translated to Dutch, Italian and Spanish. By processing the translations with the Spanish, Italian and Dutch pipelines, we were able to create SEM representation from each translated data set. These SEM representations should entail the same instance information. We report on the results of this comparison. Conclusions are given in section 8. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 2 16/148 Interpreting NAF-XML as SEM-RDF Where NAF contains the semantic interpretations of mentions in text, SEM represents instances of entities, events and relations linked to these mentions through GAF. Following GAF, we create SEM instances from NAF layers by combining different mentions of the same event or entity into a unique URI representation with gaf:denotedBy links to their mentions. Identity across mentions is based on various measures, among which identical DBPedia URIs, overlap with spans of tokens or terms in coreference sets, normalisation of dates, and similarity in WordNet. Next we show the output for processing just the titles in the two Qatar-Porsche examples as separate NAF sources. The RDF-TRiG has 3 parts: nwr:instances a named graph of all the instances detected: events, participants and time instances or intervals. See Figure 4. event relations a set of named-graphs, each representing a relation with an event instance. See Figure 5. nwr:provenance a named graph with gaf:denotedBy relations between the named graphs of the event relations and the offset positions in the sources . See Figure 6. The TRiG example shows a graph that includes all the instances. We see here 4 types of instances: events, entities, non-entities and time descriptions. Instances are based on coreference of mentions. Entity mention coreference is established by a the nominal coreference module but also indirectly by the URI assigned to the entities. Since URIs are unique and most URIs are based on DBpedia, mentions for which we create the same URI are automatically merged and all statements will apply to this unique representation across different NAF files. In this example we see that the entity <dbp:Porsche> has two different mentions originated from the two different sources. In other cases such as Qatar and Qatar Holding, different URIs have been chosen by the Named Entity Disambiguation (NED) module and consequently the entities are represented as distinct instances. There are also event components that are not considered as entities, such as: <nwr:data/porsche/non-entities/10pc+stake> To identify the concept, we create a URI based on the phrase that is only unique across data sets. So all reference to 10p stake within the database will become corerefential and point to the same instance representation. Obviously, this is not a proper representation since each 10 percent can be a different one. We also see that the events sell and buy are kept separately here by the software. This is because the similarity between these words is not high enough to merge them. Finally, the two time expressions in the two documents represent the day on which the document was processed, which resolves to the same time:Instant instance. This is because the meta data was lacking and there is no other time description in the title. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 @prefix @prefix @prefix @prefix @prefix @prefix @prefix @prefix @prefix @prefix @prefix @prefix @prefix @prefix @prefix @prefix 17/148 prov : <h t t p : / /www. w3 . o r g / n s / p r o v#> . gaf : <h t t p : / / g r o u n d e d a n n o t a t i o n f r a m e w o r k . o r g / g a f#> . wn : <h t t p : / /www. n e w s r e a d e r −p r o j e c t . eu / o n t o l o g i e s /pwn3.0/ > . n w r o n t o l o g y : <h t t p : / /www. n e w s r e a d e r −p r o j e c t . eu / o n t o l o g i e s /> . ili : <h t t p : / / g l o b a l w o r d n e t . o r g / i l i /> . rdfs : <h t t p : / /www. w3 . o r g / 2 0 0 0 / 0 1 / r d f −schema#> . time : <h t t p : / /www. w3 . o r g /TR/ owl−t i m e#> . eso : <h t t p : / /www. n e w s r e a d e r −p r o j e c t . eu / domain−o n t o l o g y#> . pb : <h t t p : / /www. n e w s r e a d e r −p r o j e c t . eu / o n t o l o g i e s / propbank/> . owl : <h t t p : / /www. w3 . o r g / 2 0 0 2 / 0 7 / owl#> . nwr : <h t t p : / /www. n e w s r e a d e r −p r o j e c t . eu/> . rdf : <h t t p : / /www. w3 . o r g /1999/02/22 − r d f −s y n t a x−n s#> . sem : <h t t p : / / s e m a n t i c w e b . c s . vu . n l / 2 0 0 9 / 1 1 / sem/> . fn : <h t t p : / /www. n e w s r e a d e r −p r o j e c t . eu / o n t o l o g i e s / f r a m e n e t/> . skos : <h t t p : / /www. w3 . o r g / 2 0 0 4 / 0 2 / s k o s / c o r e#> . nwrdata : <h t t p : / /www. n e w s r e a d e r −p r o j e c t . eu / d a t a/> . nwr : i n s t a n c e s { #e n t i t i e s <dbp : P o r s c h e> rdfs : label g a f : denotedBy <dbp : Qatar> rdfs : label g a f : denotedBy ” Porsche ” ; <h t t p : / /www. t e l e g r a p h . c o . uk#c h a r =0,7> , <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#c h a r =33,40> . ” Qatar ” ; <h t t p : / /www. t e l e g r a p h . c o . uk#c h a r =41,46> . <dbp : Q a t a r I n v e s t m e n t A u t h o r i t y > rdfs : label ” Qatar H o l d i n g ” ; g a f : denotedBy <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#c h a r =0,13> . <nwr : d a t a / p o r s c h e / non− e n t i t i e s /10 pc+s t a k e > rdfs : label ”10 pc s t a k e ” ; g a f : denotedBy <h t t p : / /www. t e l e g r a p h . c o . uk#c h a r =25,35> . #non− e n t i t i e s <nwr : d a t a / p o r s c h e / non− e n t i t i e s /10+\%25+ s t a k e+i n+p o r s c h e > rdfs : label ”10 \% s t a k e i n P o r s c h e ” ; g a f : denotedBy <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#c h a r =20,40> . <nwr : d a t a / p o r s c h e / non− e n t i t i e s / t o+f o u n d i n g+f a m i l i e s > rdfs : label ” to founding family ” ; g a f : denotedBy <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#c h a r =41,61> . #e v e n t s <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#ev1> a sem : Event , n w r o n t o l o g y : c o n t e x t u a l E v e n t , fn : Commerce sell , eso : S e l l i n g , i l i : i32963 , i l i : i32953 rdfs : label ”sell” ; g a f : denotedBy <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#c h a r =14,19> . <h t t p : / /www. t e l e g r a p h . c o . uk#ev2> a sem : Event , n w r o n t o l o g y : c o n t e x t u a l E v e n t , f n : Commerce buy , e s o : Buying , i l i : i 3 2 7 8 8 , i l i : i 3 4 9 0 1 rdfs : label ” buy ” ; g a f : denotedBy <h t t p : / /www. t e l e g r a p h . c o . uk#c h a r =15,19> . #t i m e <nwr : t i m e /20150324 > a t i m e : day t i m e : month time : unitType time : year ; ; time : DateTimeDescription ; ”−−−24”ˆˆ< h t t p : / /www. w3 . o r g / 2 0 0 1 /XMLSchema#gDay> ; ”−−03”ˆˆ< h t t p : / /www. w3 . o r g / 2 0 0 1 /XMLSchema#gMonth> ; t i m e : unitDay ; ”2015”ˆˆ < h t t p : / /www. w3 . o r g / 2 0 0 1 /XMLSchema#gYear> . <h t t p : / /www. t e l e g r a p h . c o . uk#tmx0> a time : I n s t a n t ; t i m e : inDateTime <nwr : t i m e /20150324 > . <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#tmx0> a time : I n s t a n t ; t i m e : inDateTime <nwr : t i m e /20150324 > . } Figure 4: Instances of events, entities, non-entities and time NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 18/148 <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#pr1 , r l 1 > { <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#ev1> sem : h a s A c t o r <dbp : Q a t a r I n v e s t m e n t A u t h o r i t y > ; e s o : p o s s e s s i o n −o w n e r 1 <dbp : Q a t a r I n v e s t m e n t A u t h o r i t y > ; <nwr : o n t o l o g i e s / f r a m e n e t / C o m m e r c e s e l l @ S e l l e r > <dbp : Q a t a r I n v e s t m e n t A u t h o r i t y > ; pb : A0 <dbp : Q a t a r I n v e s t m e n t A u t h o r i t y > . } <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#pr1 , r l 2 > { <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#ev1> sem : h a s A c t o r <nwr : d a t a / p o r s c h e / non− e n t i t i e s /10+\%25+ s t a k e+i n+p o r s c h e > ; <nwr : o n t o l o g i e s / f r a m e n e t / Commerce sell@Goods> <nwr : d a t a / p o r s c h e / non− e n t i t i e s /10+\%25+ s t a k e+i n+p o r s c h e > ; pb : A1 <nwr : d a t a / p o r s c h e / non− e n t i t i e s /10+\%25+ s t a k e+i n+p o r s c h e > . } <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#pr1 , r l 3 > { <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#ev1> sem : h a s A c t o r <nwr : d a t a / p o r s c h e / non− e n t i t i e s / t o+f o u n d i n g+f a m i l i e s > ; e s o : p o s s e s s i o n −o w n e r 2 <nwr : d a t a / p o r s c h e / non− e n t i t i e s / t o+f o u n d i n g+f a m i l i e s > ; <nwr : o n t o l o g i e s / f r a m e n e t / Commerce sell@Buyer> <nwr : d a t a / p o r s c h e / non− e n t i t i e s / t o+f o u n d i n g+f a m i l i e s > ; pb : A2 <nwr : d a t a / p o r s c h e / non− e n t i t i e s / t o+f o u n d i n g+f a m i l i e s > . } <h t t p : / /www. t e l e g r a p h . c o . uk#pr2 , r l 2 > { <h t t p : / /www. t e l e g r a p h . c o . uk#ev2> sem : h a s A c t o r <dbp : P o r s c h e> ; e s o : p o s s e s s i o n −o w n e r 2 <dbp : P o r s c h e> ; <nwr : o n t o l o g i e s / f r a m e n e t / Commerce buy@Buyer> <dbp : P o r s c h e> ; pb : A0 <dbp : P o r s c h e> . } <h t t p : / /www. t e l e g r a p h . c o . uk#pr2 , r l 4 > { <h t t p : / /www. t e l e g r a p h . c o . uk#ev2> sem : h a s A c t o r <nwr : d a t a / p o r s c h e / non− e n t i t i e s /10 pc+s t a k e > ; <nwr : o n t o l o g i e s / f r a m e n e t / Commerce buy@Goods> <nwr : d a t a / p o r s c h e / non− e n t i t i e s /10 pc+s t a k e > ; pb : A1 <nwr : d a t a / p o r s c h e / non− e n t i t i e s /10 pc+s t a k e > . } <h t t p : / /www. t e l e g r a p h . c o . uk#pr2 , r l 5 > { <h t t p : / /www. t e l e g r a p h . c o . uk#ev2> sem : h a s A c t o r <dbp : Qatar> ; <nwr : o n t o l o g i e s / f r a m e n e t / Commerce buy@Means> <dbp : Qatar> ; pb : A2 <dbp : Qatar> . } <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#t r 1 > { <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#ev1> sem : hasAtTime <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#tmx0> ; sem : hasTime <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#tmx0> . } <h t t p : / /www. t e l e g r a p h . c o . uk#t r 2 > { <h t t p : / /www. t e l e g r a p h . c o . uk#ev2> sem : hasAtTime <h t t p : / /www. t e l e g r a p h . c o . uk#tmx0> ; sem : hasTime <h t t p : / /www. t e l e g r a p h . c o . uk#tmx0> . } Figure 5: Sem triples embedded in named-graphs NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 19/148 nwr : p r o v e n a n c e { <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#pr1 , r l 1 > g a f : denotedBy <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#c h a r =0,19> . <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#pr1 , r l 2 > g a f : denotedBy <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#c h a r =14,40> . <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#pr1 , r l 3 > g a f : denotedBy <h t t p : / / e n g l i s h . a l a r a b i y a . n e t#c h a r =14,61> . <h t t p : / /www. t e l e g r a p h . c o . uk#pr2 , r l 2 > g a f : denotedBy <h t t p : / /www. t e l e g r a p h . c o . uk#c h a r =0,19> . <h t t p : / /www. t e l e g r a p h . c o . uk#pr2 , r l 4 > g a f : denotedBy <h t t p : / /www. t e l e g r a p h . c o . uk#c h a r =15,35> . <h t t p : / /www. t e l e g r a p h . c o . uk#pr2 , r l 5 > g a f : denotedBy <h t t p : / /www. t e l e g r a p h . c o . uk#c h a r =15,46> . } Figure 6: Provenance information indicating the mentions of relations These examples show that the interpretation of mentions as instances, is a delicate process that is on the one hand the result of creating identical URIs for the interpretation of a mention (e.g. DBpedia references or normalised dates) and on the other hand the result of specific strategies to match the semantics. The module NAF2SEM2 defines this interpretation. The core function to interpret the mentions in NAF is the class GetSemFromNaf.java. It creates the SEM objects and relations in memory. At that point there are two API options: either the information is stored on disk as a binary object file or the SEM objects are serialized to an RDF file or stream. The former is used for batch processing described in Section 2.2.1, whereas the latter is used for the streaming architecture described in Section 2.2.2. Below we first describe the representation of the instances and their relations in more detail and next the general way in which we establish identity across events. In the next subsections, we describe in detail the interpretation as instances and relations and how identity is established. 2.1 Extracting instances from NAF layers The NAF2SEM module combines information from various NAF layers to define the SEM Objects (internal data structures defined in Java) for entities, events, time and event relations. We discuss each type of object in more details below. 2.1.1 Entities and non-entities Genuine entities are represented in the entity layer in NAF and have a DBpedia URI that identifies them and are a participant in an event that is extracted. However, there are many important objects that do not fit all 3 of these requirements. There are entities without a reference to DBPedia, there are entities that do not play a role in an event that is represented and sometimes important participants in events are not detected as entities. As a result, we can find the different entities and entity-like objects in the RDF representations. Regular entities that have been detected by the Named Entity Recognizer (NERC) and got an external reference to an URI by the spotlight program are represented as discussed 2 https://github.com/cltl/EventCoreference NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 <e n t i t y i d =”e 1 ” t y p e=”ORGANIZATION”> <r e f e r e n c e s > <!−−P o r s c h e−−> <span> <t a r g e t i d =”t 1 ” /> </span> </ r e f e r e n c e s > <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =” s p o t l i g h t r e f t y p e =”en ” /> <e x t e r n a l R e f r e s o u r c e =” s p o t l i g h t r e f t y p e =”en ” /> <e x t e r n a l R e f r e s o u r c e =” s p o t l i g h t r e f t y p e =”en ” /> <e x t e r n a l R e f r e s o u r c e =” s p o t l i g h t r e f t y p e =”en ” /> <e x t e r n a l R e f r e s o u r c e =” s p o t l i g h t r e f t y p e =”en ” /> <e x t e r n a l R e f r e s o u r c e =” s p o t l i g h t r e f t y p e =”en ” /> <e x t e r n a l R e f r e s o u r c e =” s p o t l i g h t r e f t y p e =”en ” /> <e x t e r n a l R e f r e s o u r c e =” s p o t l i g h t r e f t y p e =”en ” /> <e x t e r n a l R e f r e s o u r c e =” s p o t l i g h t r e f t y p e =”en ” /> <e x t e r n a l R e f r e s o u r c e =” s p o t l i g h t r e f t y p e =”en ” /> </ e x t e r n a l R e f e r e n c e s > </ e n t i t y > 20/148 v 1 ” r e f e r e n c e =”dbp : P o r s c h e ” c o n f i d e n c e = ” 0 . 9 9 9 9 3 2 1 7 ” v 1 ” r e f e r e n c e =”dbp : P o r s c h e D e s i g n G r o u p ” c o n f i d e n c e = ” 3 . 9 7 6 1 5 2E−5” v 1 ” r e f e r e n c e =”dbp : P o r s c h e 9 1 1 ” c o n f i d e n c e = ” 1 . 5 3 5 9 3 9 7E−5” v 1 ” r e f e r e n c e =”dbp : P o r s c h e 9 1 4 ” c o n f i d e n c e = ” 7 . 1 7 1 3 5 8E−6” v 1 ” r e f e r e n c e =”dbp : F e r d i n a n d P o r s c h e ” c o n f i d e n c e = ” 5 . 5 2 2 5 6 3 7E−6” v 1 ” r e f e r e n c e =”dbp : P o r s c h e f a m i l y ” c o n f i d e n c e = ” 2 . 4 9 3 9 7 5 6E−10” v 1 ” r e f e r e n c e =”dbp : P o r s c h e i n m o t o r s p o r t ” c o n f i d e n c e = ” 2 . 9 5 7 0 8 0 8E−14” v 1 ” r e f e r e n c e =”dbp : P o r s c h e 5 5 0 ” c o n f i d e n c e = ” 8 . 5 1 7 1 6 5E−17” v 1 ” r e f e r e n c e =”dbp : P o r s c h e 3 5 1 2 ” c o n f i d e n c e = ” 2 . 5 0 4 6 7 4 6E−17” v 1 ” r e f e r e n c e =”dbp : P o r s c h e R S S p y d e r ” c o n f i d e n c e = ” 1 . 9 3 9 9 4 5 2E−17” Figure 7: Organisation Entity in NAF with URIs 1 2 3 4 5 6 7 8 <e n t i t y i d =”e 4 5 ” t y p e=”LOCATION”> <r e f e r e n c e s > <!−−N o r t h e a s t China ’ s−−> <span><t a r g e t i d =”t 6 9 7”/>< t a r g e t </span> </ r e f e r e n c e s > </ e n t i t y > <e n t i t y i d =”e 4 6 ” t y p e=”LOCATION”> i d =”t 6 9 8”/>< t a r g e t i d =”t 6 9 9 ”/> Figure 8: Location Entity in NAF without URI before through their URI. The NAF representation for the Porsche instance shown in TRiG example above is shown in Figure 7. There are however many phrases detected as entities by the NERC that did not receive an external reference to DBPedia by spotlight as is shown in Figure 8. For these so-called dark entities we create a blank URI within the set of entities in a project. The entity type assigned by the NERC is used to create a subclass relation. This is shown in Figure 9. When the same URI was recovered for different mentions of an entity, this results in a single representation of that entity with gaf:denotedBy links to each mention. The spans of these mentions can however overlap with the span in the entity coreference set in a NAF layer. In these cases, we can extend the mentions of an entity with the mentions of the coreference set but in some cases, distinct entities also get tight to the same instance 1 2 3 4 <nwr : d a t a / c a r s / e n t i t i e s / N o r t h e a s t C h i n a s > a n w r o n t o l o g y : LOCATION ; rdfs : label ” N o r t h e a s t China ’ s ” ; g a f : denotedBy <nwr : d a t a / c a r s /57DD−HR81−JB4B−V3T6 . xml#c h a r =3582 ,3599 > . Figure 9: Entity in SEM with a blank URI NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 21/148 <e n t i t y i d =”e 5 ” t y p e=”PERSON”> <r e f e r e n c e s > <!−− D i d i e r Drogba−−> <span> <t a r g e t i d =”t 6 8 ”/> <t a r g e t i d =”t 6 9 ”/> </span> </ r e f e r e n c e s > <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =” s p o t l i g h t v 1 ” r e f e r e n c e =”dbp : D i d i e r D r o g b a ” c o n f i d e n c e =”1.0” r e f t y p e =”en ” s o u r c e =”en”/> <e x t e r n a l R e f r e s o u r c e =”dbp ” r e f e r e n c e =”dbp : D i d i e r D r o g b a ” c o n f i d e n c e =”1.0” s o u r c e =”POCUS”/> </ e x t e r n a l R e f e r e n c e s > </ e n t i t y > <e n t i t y i d =”e 2 ” t y p e=”PERSON”> <r e f e r e n c e s > <!−− D i d i e r Yves Drogba T e b i l y −−> <span> <t a r g e t i d =”t 2 ”/> <t a r g e t i d =”t 3 ”/> <t a r g e t i d =”t 4 ”/> <t a r g e t i d =”t 5 ”/> </span> </ r e f e r e n c e s > </ e n t i t y > Figure 10: Drogba as an entity in NAF through the coreference set. Consider the example of Didier Drogba and Didier Yves Drogba Tébily in Figure 10 that was in both cases detected as an entity but only the former is mapped to DBpedia. Within the same NAF we also find the coreference set shown in Figure 11. The spans in the coreference set overlap with the spans of the two entities. As a result of that, we can merge all these representation into a single entity and list all of the mentions of the entity layer and the coreference set. Consequently, the rdfs:label predicate is also extended with all the different labels used to make reference. This is shown in the SEM representation in Figure 12. Although the NAF2SEM module reads all entities from a NAF file, we currently only output entities that play a role in events. To determine whether an entity plays a role in an event, we need to match the mentions of an entity with the span of the roles of events. We will discuss this later below. In addition, there are phrases that play a role but are not detected as entities. We have seen two examples in the TRiG example shown in the beginning of this section. To limit the amount of entities and triples, we only consider roles that have a FrameNet role element assigned. We consider these roles essential for understanding what the event is about. In all cases that these roles cannot be assigned to an entity, we represent the concept as a so-called non-entity. To find these non-entities, we match the span of the roles with all the spans of the entities extracted before. We assign the type NONENTITY to these instances as showin in Figure 13. Spotlight is not only applied to entities in the text but also to all other content phrases. This is represented in the NAF layer with the markables, see Figure 14. We use these markables to find potential DBpedia references that somehow relate to the non-entity. Again the overlap in the span elements is used to determine relevance. Since we do not know the precise semantic relation, we use the skos:relatedMatch predicate to relate the non-entity to the DBpedia entry. Through the skos:relatedMatch it is possible to query the NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 < c o r e f i d =”c o 2”> <!−−His−−> <span> <t a r g e t i d =”t 1 2 9 ”/> </span> <!−− D i d i e r Drogba−−> <span> <t a r g e t i d =”t 5 3 1 ”/> <t a r g e t i d =”t 5 3 2 ”/> </span> <!−−Drogba−−> <span> <t a r g e t i d =”t 4 2 9 ”/> </span> <!−−he−−> <span> <t a r g e t i d =”t 7 4 ”/> </span> <!−−h i s −−> <span> <t a r g e t i d =”t 9 ”/> </span> <!−−Drogba−−> <span> <t a r g e t i d =”t 3 3 3 ”/> </span> <!−−Drogba−−> <span> <t a r g e t i d =”t 3 5 3 ”/> </span> <!−−Drogba−−> <span> <t a r g e t i d =”t 2 0 2 ”/> </span> <!−− D i d i e r Drogba−−> <span> <t a r g e t i d =”t 6 2 6 ”/> <t a r g e t i d =”t 6 2 7 ”/> </span> <!−−Drogba−−> <span> <t a r g e t i d =”t 7 6 6 ”/> </span> <!−− D i d i e r Drogba−−> <span> <t a r g e t i d =”t 6 8 ”/> <t a r g e t i d =”t 6 9 ”/></span> <!−−His−−> <span><t a r g e t i d =”t 1 0 8 ”/></span> <!−−h i s −−> <span><t a r g e t i d =”t 8 4 ”/></span> <!−− D i d i e r Yves Drogba T e b i l y −−> <span><t a r g e t i d =”t 2”/>< t a r g e t i d =”t 3”/>< t a r g e t i d =”t 4”/>< t a r g e t </ c o r e f > 22/148 i d =”t 5 ”/></span> Figure 11: Coreference set with Drogba in NAF 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 dbp : D i d i e r D r o g b a rdfs : label ” Drogba ” , ” D i d i e r Yves Drogba T e b i l y ” , ” h i s ” ” D i d i e r Drogba ” , ” D i d i e r Drogba ’ s ” ; g a f : denotedBy <nwr : d a t a / c a r s /57R0−J5K1−JC86−C1N7 . xml#c h a r =11,36> , <nwr : d a t a / c a r s /57R0−J5K1−JC86−C1N7 . xml#c h a r =694 ,697 > , <nwr : d a t a / c a r s /57R0−J5K1−JC86−C1N7 . xml#c h a r =2795 ,2808 > , <nwr : d a t a / c a r s /57R0−J5K1−JC86−C1N7 . xml#c h a r =2274 ,2280 > , <nwr : d a t a / c a r s /57R0−J5K1−JC86−C1N7 . xml#c h a r =395 ,397 > , <nwr : d a t a / c a r s /57R0−J5K1−JC86−C1N7 . xml#c h a r =53,56> , <nwr : d a t a / c a r s /57R0−J5K1−JC86−C1N7 . xml#c h a r =1752 ,1758 > , <nwr : d a t a / c a r s /57R0−J5K1−JC86−C1N7 . xml#c h a r =1858 ,1864 > , <nwr : d a t a / c a r s /57R0−J5K1−JC86−C1N7 . xml#c h a r =1052 ,1058 > , <nwr : d a t a / c a r s /57R0−J5K1−JC86−C1N7 . xml#c h a r =3333 ,3346 > , <nwr : d a t a / c a r s /57R0−J5K1−JC86−C1N7 . xml#c h a r =4120 ,4126 > , <nwr : d a t a / c a r s /57R0−J5K1−JC86−C1N7 . xml#c h a r =357 ,370 > , <nwr : d a t a / c a r s /57R0−J5K1−JC86−C1N7 . xml#c h a r =580 ,583 > , <nwr : d a t a / c a r s /57R0−J5K1−JC86−C1N7 . xml#c h a r =443 ,446 > , <nwr : d a t a / c a r s /57R0−J5K1−JC86−C1N7 . xml#c h a r =4344 ,4350 > , <nwr : d a t a / c a r s /59 JJ −0761−DY2M−33MF. xml#c h a r =4277 ,4292 > , <nwr : d a t a / c a r s /59 JJ −0761−DY2M−33MF. xml#c h a r =4693 ,4699 > . , Figure 12: Drogba as an instance in SEM 1 2 3 4 5 6 7 8 9 10 11 <nwr : d a t a / c a r s / non− e n t i t i e s / from+a+n e w s p a p e r+o r+magazine> a n w r o n t o l o g y :NONENTITY ; rdfs : label ” from a n e w s p a p e r o r m a g a z i n e ” ; g a f : denotedBy <nwr : d a t a / 2 0 0 4 / 0 3 / 2 6 / 4 C16−2HY0−01JV−13BX. xml#c h a r =4099 ,4127 > ; skos : relatedMatch dbp : Magazine <nwr : d a t a / c a r s / non− e n t i t i e s / a+l a w s u i t > a n w r o n t o l o g y :NONENTITY ; rdfs : label ”a l a w s u i t ” ; g a f : denotedBy <nwr : d a t a / c a r s /5629−K3P1−F190−V098 . xml#c h a r =3239 ,3248 > ; skos : relatedMatch dbp : L a w s u i t . Figure 13: Non-entity instance in SEM NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 23/148 <!−−magazine−−> <mark i d =”m4” s o u r c e =”DBpedia ” lemma=”m a g a z i n e”> <span> <t a r g e t i d =”w23”/> </span> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =” s p o t l i g h t ” r e f e r e n c e =”dbp : Magazine ” c o n f i d e n c e =”1.0” r e f t y p e =”en ” s o u r c e =”en”/> </ e x t e r n a l R e f e r e n c e s > </mark> <!−− l a w s u i t −−> <mark i d =”m138” s o u r c e =”DBpedia ” lemma=” l a w s u i t ”> <span> <t a r g e t i d =”w601”/> </span> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =” s p o t l i g h t ” r e f e r e n c e =”dbp : L a w s u i t ” c o n f i d e n c e =”1.0” r e f t y p e =”en ” s o u r c e =”en”/> </ e x t e r n a l R e f e r e n c e s > </mark> Figure 14: Markables in NAF non-entities through the DBpedia entries and classes rather than just on the string values. 2.1.2 Events We cannot recover a URI from DBpedia to represent an event instance. We therefore create a blank URI using the meta data of the document and an event counter. Furthermore, we give the type information, the labels, the mentions for each event object. The event coreference sets in NAF are the basis for defining the event instances in a single document. In Figure 15, we show several examples of event coreference sets. Event coreference sets are derived from the predicates in the Semantic Role Layer (SRL). Every predicate is represented in a coreference set. Predicates with the same lemma are represented in the same set, as well as predicates that are semantically similar. As a result we get both singleton sets (noreference) and multiform sets (coreference). For each mention, we determine the highest scoring wordnet synsets and across different mentions we select the most dominant senses from the highest scoring senses. Likewise, each coreference set gets external references to the most dominant senses with the averaged word-sense-disambiguation score. Whenever the similarity across these dominant senses is above the threshold, we merge coreference sets and add the lowest common subsumer to the merged set. This is shown for the last example in Figure 15, were shot and injured are merged into a single set with the lowest common subsumer synset eng-30-00069879-v, injure:1, wound:1 and a similarity score of 2.6390574 using the method described by Leacock and Chodorow (1998). Each coreference set becomes a potential event instance, where we use the mentions for the labels and the references to the WordNet synset as a subclass relation.3 Furthermore, we collect a subset of the ontological labels assigned to each predicate in the SRL. For example for chased, we find the following typing in the SRL layer shown in Figure 16, from which we copy the FrameNet and ESO classes into the instance representation. 3 We convert all reference to WordNet synset to InterLingualIndex concepts to allow cross-lingual match- ing NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 < c o r e f i d =” c o e v e n t 7 2 ” t y p e=” e v e n t ”> <!−−c h a s e d−−> <span> <t a r g e t i d =”t 3 7 1 ”/> </span> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =”WordNet −3.0” s o u r c e =” d o m i n a n t s e n s e ”/> </ e x t e r n a l R e f e r e n c e s > </ c o r e f > < c o r e f i d =” c o e v e n t 4 9 ” t y p e=” e v e n t ”> <!−− g i v i n g −−> <span> <t a r g e t i d =”t 1 5 0 ”/> </span> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =”WordNet −3.0” s o u r c e =” d o m i n a n t s e n s e ”/> <e x t e r n a l R e f r e s o u r c e =”WordNet −3.0” s o u r c e =” d o m i n a n t s e n s e ”/> <e x t e r n a l R e f r e s o u r c e =”WordNet −3.0” s o u r c e =” d o m i n a n t s e n s e ”/> <e x t e r n a l R e f r e s o u r c e =”WordNet −3.0” s o u r c e =” d o m i n a n t s e n s e ”/> <e x t e r n a l R e f r e s o u r c e =”WordNet −3.0” s o u r c e =” d o m i n a n t s e n s e ”/> <e x t e r n a l R e f r e s o u r c e =”WordNet −3.0” s o u r c e =” d o m i n a n t s e n s e ”/> </ e x t e r n a l R e f e r e n c e s > </ c o r e f < c o r e f i d =” c o e v e n t 5 7 ” t y p e=” e v e n t ”> <!−−w i n n i n g−−> <span><t a r g e t i d =”t 2 8 4 ”/></span> <!−−won−−> <span><t a r g e t i d =”t 2 9 1 ”/></span> <!−−won−−> <span><t a r g e t i d =”t 4 0 5 ”/></span> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =”WordNet −3.0” s o u r c e =” d o m i n a n t s e n s e ”/> <e x t e r n a l R e f r e s o u r c e =”WordNet −3.0” s o u r c e =” d o m i n a n t s e n s e ”/> </ e x t e r n a l R e f e r e n c e s > </ c o r e f > 24/148 r e f e r e n c e =” i l i −30−02001858−v ” c o n f i d e n c e =”1.0” r e f e r e n c e =” i l i −30−02200686−v ” c o n f i d e n c e = ” 0 . 8 6 5 3 8 1 9 6 ” r e f e r e n c e =” i l i −30−02339171−v ” c o n f i d e n c e =”1.0” r e f e r e n c e =” i l i −30−02199590−v ” c o n f i d e n c e = ” 0 . 9 1 3 9 8 1 ” r e f e r e n c e =” i l i −30−01629403−v ” c o n f i d e n c e = ” 0 . 9 7 3 0 2 9 3 ” r e f e r e n c e =” i l i −30−02316868−v ” c o n f i d e n c e = ” 0 . 8 5 7 6 5 4 9 ” r e f e r e n c e =” i l i −30−02235842−v ” c o n f i d e n c e = ” 0 . 8 0 7 2 3 5 1 ” r e f e r e n c e =” i l i −30−02288295−v ” c o n f i d e n c e = ” 0 . 9 7 5 4 5 7 1 3 ” r e f e r e n c e =” i l i −30−01100145−v ” c o n f i d e n c e =”1.0” < c o r e f i d =” c o e v e n t 6 1 ” t y p e=” e v e n t ”> <!−−s h o t −−> <span><t a r g e t i d =”t 3 4 5 ”/></span> <!−−s h o t −−> <span><t a r g e t i d =”t 4 1 4 ”/></span> <!−− i n j u r e d −−> <span> <t a r g e t i d =”t 6 8 4 ”/></span> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =” P r i n c e t o n WordNet 3 . 0 ” r e f e r e n c e =”eng −30−00069879−v ” c o n f i d e n c e = ” 2 . 6 3 9 0 5 7 4 ” s o u r c e =”l o w e s t c o m m o n s u b s u m e r ”/> <e x t e r n a l R e f r e s o u r c e =”WordNet −3.0” r e f e r e n c e =” i l i −30−02055267−v ” c o n f i d e n c e = ” 0 . 8 1 2 2 6 4 8 6 ” s o u r c e =” d o m i n a n t s e n s e ”/> <e x t e r n a l R e f r e s o u r c e =”WordNet −3.0” r e f e r e n c e =” i l i −30−01134781−v ” c o n f i d e n c e = ” 0 . 9 5 1 7 7 9 2 ” s o u r c e =” d o m i n a n t s e n s e ”/> <e x t e r n a l R e f r e s o u r c e =”WordNet −3.0” r e f e r e n c e =” i l i −30−01137138−v ” c o n f i d e n c e = ” 0 . 9 2 4 6 8 0 7 ” s o u r c e =” d o m i n a n t s e n s e ”/> <e x t e r n a l R e f r e s o u r c e =”WordNet −3.0” r e f e r e n c e =” i l i −30−02484570−v ” c o n f i d e n c e = ” 0 . 9 3 0 1 4 6 4 6 ” s o u r c e =” d o m i n a n t s e n s e ”/> </ e x t e r n a l R e f e r e n c e s > </ c o r e f > Figure 15: Event coreference set in NAF NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 25/148 <!−−t 3 7 1 c h a s e d : A1 [ t 3 6 7 England ] AM−DIR [ t 3 7 2 down]−−> <p r e d i c a t e i d =”p r 7 1”> <!−−c h a s e d−−> <span> <t a r g e t i d =”t 3 7 1 ”/> </span> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =”PropBank ” r e f e r e n c e =” c h a s e .01”/ > <e x t e r n a l R e f r e s o u r c e =”VerbNet ” r e f e r e n c e =”c h a s e −51.6”/ > <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =”Cotheme”/> <e x t e r n a l R e f r e s o u r c e =”PropBank ” r e f e r e n c e =” c h a s e .01”/ > <e x t e r n a l R e f r e s o u r c e =”ESO” r e f e r e n c e =” T r a n s l o c a t i o n ”/> <e x t e r n a l R e f r e s o u r c e =”EventType ” r e f e r e n c e =” c o n t e x t u a l ”/> <e x t e r n a l R e f r e s o u r c e =”WordNet” r e f e r e n c e =” i l i −30−02001858−v”/> <e x t e r n a l R e f r e s o u r c e =”WordNet” r e f e r e n c e =” i l i −30−02535093−v”/> </ e x t e r n a l R e f e r e n c e s > <nwr : d a t a / c a r s /55XK−XGX1−JBKJ−C3CF . xml#ev72> a sem : Event , n w r o n t o l o g y : c o n t e x t u a l E v e n t , e s o : T r a n s l o c a t i o n , f n : Cotheme , i l i : i 3 1 7 4 7 ; rdfs : label ” chase ” ; g a f : denotedBy <nwr : d a t a / c a r s /55XK−XGX1−JBKJ−C3CF . xml#c h a r =2074 ,2080 > . Figure 16: SRL in NAF with event types for the predicate chased In the case of the coreference set of injured and shot, we derive no classes for injured but various FrameNet classes from the predicate shot, seen in Figure 17. In addition to the specific semantic classes, every event is of the type sem:Event and one of the 3 main event classes in NewsReader: sourceEvent events that introduce a source of information as the semantic subject, such as speech-acts (say, claim, declare) and cognitive-verbs (think, believe, hope, fear ). grammaticalEvent auxiliary verbs (be, have, should, will ), aspectual verbs (stop, begin, take place, happen) that do not introduce other participants in an event, do not define a different time interval for events and express properties of other events. contextualEvent all other events than the above that take place in some world and do not introduce a source of events or express a property of an event. On the basis of an analysis of the car data set, we create lists of FrameNet frames classified as contextual, source or grammatical. In Appendix 9, we give a complete list of FrameNet frames that distinguish between these three main event types. 2.1.3 Participant and event relations Once the instances for entities, non-entities and events have been established, we determine the relations between them. We first extract from the SRL layer all the roles with a valid Propbank role. The following role labels are considered valid: PRIMEPARTICIPANT a0, arg0, a-0, arg-0 NONPRIMEPARTICIPANT a1, a2, a3, a4, arg1, arg2, arg3, a-1, a-2, a-3, a-4, arg-1, arg-2, arg-3, arg-4, am-dir, argm-dir LOCATION am-loc, argm-loc, am-dir NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 26/148 <!−−t 5 7 9 i n j u r e d : A0 [ t 5 7 7 he ] AM−MNR[ t 5 8 0 s o ] A1 [ t 5 8 1 t h a t]−−> <p r e d i c a t e i d =”p r 1 0 7”> <!−− i n j u r e d −−> <span> <t a r g e t i d =”t 5 7 9 ”/> </span> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =”PropBank ” r e f e r e n c e =” i n j u r e .01”/ > </ e x t e r n a l R e f e r e n c e s > <p r e d i c a t e i d =”p r 8 3”> <!−−s h o t −−> <span> <t a r g e t i d =”t 5 0 0 ”/> </span> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =”PropBank ” r e f e r e n c e =” s h o o t .02”/ > <e x t e r n a l R e f r e s o u r c e =”VerbNet ” r e f e r e n c e =”p o i s o n −42.2”/ > <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =” H i t t a r g e t ”/> <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =” S h o o t p r o j e c t i l e s ”/> <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =” U s e f i r e a r m ”/> <e x t e r n a l R e f r e s o u r c e =”PropBank ” r e f e r e n c e =” s h o o t .02”/ > <e x t e r n a l R e f r e s o u r c e =”EventType ” r e f e r e n c e =” c o n t e x t u a l ”/> <e x t e r n a l R e f r e s o u r c e =”WordNet” r e f e r e n c e =” i l i −30−02484570−v”/> </ e x t e r n a l R e f e r e n c e s > <nwr : d a t a / c a r s /59JB−GV01−JBSN−30SP . xml#ev84> a sem : Event , n w r o n t o l o g y : c o n t e x t u a l E v e n t , f n : H i t t a r g e t , i l i : i27293 , i l i : i27278 , i l i : i34141 , fn : S h o o t p r o j e c t i l e s i l i : i22125 , fn : Use firearm ; rdfs : label ” shoot ” , ” i n j u r e ” ; g a f : denotedBy <nwr : d a t a / c a r s /59JB−GV01−JBSN−30SP . xml#c h a r =2215 ,2219 > , <nwr : d a t a / c a r s /59JB−GV01−JBSN−30SP . xml#c h a r =2588 ,2595 > . , Figure 17: SRL in NAF for predicate injured Next, we intersect the span of the role with the span of any mention of the entities and nonentities that were extracted before. Since the spans are established in very different ways across the NLP modules that create the NAF layers, we implemented a loose matching principle that the number of matching content words across the role and the entity needs exceed 75% of each. This prevents excessively long spans to match with short spans. Content words are terms with part-of-speech that start with R (adverb), N (noun), V (verb), A (adjective) or G (adjective). Note that non-entities always exactly match at least one role since they are derived from the roles. If there is a match between an entity and a role, we create a SEM relation between the event instance and the entity instance and copy the semRole value and the external references from the role as predicates for the relation. In Figure 18 we show the SRL structure for the chased predicate given before, followed by the SEM relations that are extracted from it. The SEM relations are combined in named-graphs for which we created blank URIs based on the identifiers for the predicate and the role, e.g. pr71,rl175. Since the role is considered to be an actor role, the sem:hasActor predicate is added. Furthermore, only the PropBank, ESO and FrameNet roles are kept. Note that the NAF role down the ball is not represented in RDF. This is because there is no matching entity for this role and there is no FrameNet role to promote it as a non-entity. We thus constraint the representation of the events to components that are somehow modeled and crystallized. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 27/148 <!−−t 3 7 1 c h a s e d : A1 [ t 3 6 7 England ] AM−DIR [ t 3 7 2 down]−−> <p r e d i c a t e i d =”p r 7 1”> <!−−c h a s e d−−> <span> <t a r g e t i d =”t 3 7 1 ”/> </span> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =”PropBank ” r e f e r e n c e =” c h a s e .01”/ > <e x t e r n a l R e f r e s o u r c e =”VerbNet ” r e f e r e n c e =”c h a s e −51.6”/ > <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =”Cotheme”/> <e x t e r n a l R e f r e s o u r c e =”PropBank ” r e f e r e n c e =” c h a s e .01”/ > <e x t e r n a l R e f r e s o u r c e =”ESO” r e f e r e n c e =” T r a n s l o c a t i o n ”/> <e x t e r n a l R e f r e s o u r c e =”EventType ” r e f e r e n c e =” c o n t e x t u a l ”/> <e x t e r n a l R e f r e s o u r c e =”WordNet” r e f e r e n c e =” i l i −30−02001858−v”/> <e x t e r n a l R e f r e s o u r c e =”WordNet” r e f e r e n c e =” i l i −30−02535093−v”/> </ e x t e r n a l R e f e r e n c e s > < r o l e i d =” r l 1 7 5 ” semRole=”A1”> <!−−England d e f e n d e r John Terry−−> <span> <t a r g e t i d =”t 3 6 7 ”/> <t a r g e t i d =”t 3 6 8 ” head=” y e s ”/> <t a r g e t i d =”t 3 6 9 ”/> <t a r g e t i d =”t 3 7 0 ”/> </span> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =”VerbNet ” r e f e r e n c e =”c h a s e −51.6@Theme”/> <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =”Cotheme@Cotheme”/> <e x t e r n a l R e f r e s o u r c e =”PropBank ” r e f e r e n c e =” c h a s e . 0 1 @1”/> <e x t e r n a l R e f r e s o u r c e =”ESO” r e f e r e n c e =” T r a n s l o c a t i o n @ t r a n s l o c a t i o n −theme”/> </ e x t e r n a l R e f e r e n c e s > </ r o l e > < r o l e i d =” r l 1 7 6 ” semRole=”AM−DIR”> <!−−down t h e b a l l −−> <span> <t a r g e t i d =”t 3 7 2 ” head=” y e s ”/> <t a r g e t i d =”t 3 7 3 ”/> <t a r g e t i d =”t 3 7 4 ”/> </span> </ r o l e > </ p r e d i c a t e > <nwr : d a t a / c a r s /55XK−XGX1−JBKJ−C3CF . xml#pr71 , r l 1 7 5 > { <nwr : d a t a / c a r s /55XK−XGX1−JBKJ−C3CF . xml#ev72> sem : h a s A c t o r dbp : J o h n T e r r y ; f n : Cotheme@Cotheme dbp : J o h n T e r r y ; e s o : t r a n s l o c a t i o n −theme dbp : J o h n T e r r y ; pb : A1 dbp : J o h n T e r r y . } Figure 18: SRL in NAF with roles and corresponding SEM triples for actor relations NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 2.1.4 28/148 Temporal anchoring Time objects and temporal relations play an important role in NewsReader. Without a proper time anchoring, we cannot compare one event to another. The same type of event with the same participants on a different day is by definition not the same event: compare telling the same story today and next week. As disjoint spatial boundaries define distinct objects, so do disjoint temporal boundaries for events. Time objects for delineating events are derived from the timex3 layer in NAF. There are two types of time expressions: DATE and DURATION. Both have span elements to the words expressing it and can have value attributes. DATE expressions have a value attribute that usually points to a specific normalised ISO date. DURATION expressions usually have ISO periods as values but can have optional attributes for beginPoint and endPoint whose values are DATE expressions represented elsewhere. A special timex3 element is the document creation time (if known). This is a DATE expression with the attribute functionInDocument=”CREATION TIME” and usually the identifier tm0. The document creation time is derived from the meta data of a document and has no span element pointing to its expression in the document. Some examples for each type taken from different NAF files are given in Figure 19. From the timex3 elements, we derive two types of instances: time:Instant and time:Interval. We can only do this if we can obtain at least the year from the normalised value (the month and day can remain underspecified). If the values are relative and the year is not explicit, we cannot interpret the expression as a time object and we have to ignore it. The next examples in Figure 20show the representation of time objects in RDF-TRiG derived from different time expressions such as the ones shown above. Time expressions as well as the document creation time are represented as instances with unique URIs. Instances of the type time:Instant have a time:inDateTime relation to a date object, whereas instances of time:Interval have a time:hasBeginning and/or time:hasEnd relation to a date. Dates are represented as separate instances of the type time:DateTimeDescription with values for the year, month and/or day according to owl-time4 . Each event in SEM-RDF needs to be anchored to at least one time expression that resolves to a time instance. Events without time anchoring are ignored in the output. Anchoring relations are expressed using a sem:hasTime relation between an event URI and a time expression URI. The triples are embedded inside a named graph just as the participant-event relations we have seen before: <http://www.telegraph.co.uk#tr2> { <http://www.telegraph.co.uk#ev2> sem:hasTime <http://www.telegraph.co.uk#tmx0> . } We use the following heuristic to detect relations between events and time instances in a NAF file: 1. The NAF layer temporalRelations provides explicit anchor relations between predicates and time expressions; 4 http://www.w3.org/TR/owl-time NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 29/148 <t i m e x 3 i d =”tmx0” t y p e=”DATE” f u n c t i o n I n D o c u m e n t=”CREATION\ TIME” v a l u e =”2007−01−10T00 : 0 0 : 0 0 ” / > <t i m e x 3 i d =”tmx1” t y p e=”DATE” v a l u e =”2007−01−10”> <!−−J a n u a r y 10 , 2007−−> <span><t a r g e t i d =”w7”/>< t a r g e t i d =”w8”/>< t a r g e t </timex3> i d =”w9”/>< t a r g e t i d =”w10”/></span> <t i m e x 3 i d =”tmx1” t y p e=”DURATION” b e g i n P o i n t =”tmx19 ” e n d P o i n t=”tmx0 ” v a l u e =”P10Y”> <!−−10 y e a r s −−> <span> <t a r g e t i d =”w70”/>< t a r g e t i d =”w71”/></span> </timex3> <t i m e x 3 i d =”tmx2” t y p e=”DATE” v a l u e =”PRESENT REF”> <!−−now−−> <span><t a r g e t i d =”w73”/></span> </timex3> <t i m e x 3 i d =”tmx2” t y p e=”DATE” v a l u e =”2008−09”> <!−−September 2008−−> <span><t a r g e t i d =”w76”/>< t a r g e t i d =”w77”/></span> </timex3> <t i m e x 3 i d =”tmx3” t y p e=”DATE” v a l u e =”2015−03”> <!−−March−−> <span><t a r g e t i d =”w141”/></span> </timex3> <t i m e x 3 i d =”tmx4” t y p e=”DATE” v a l u e =”2015−06”> <!−−June−−> <span><t a r g e t i d =”w168”/></span> </timex3> <t i m e x 3 i d =”tmx5” t y p e=”DURATION” e n d P o i n t=”tmx0 ” v a l u e =”PXM”> <!−− r e c e n t months−−> <span><t a r g e t i d =”w202”/>< t a r g e t i d =”w203”/></span> </timex3> <t i m e x 3 i d =”tmx4” t y p e=”DURATION” b e g i n P o i n t =”tmx7 ” e n d P o i n t=”tmx0” v a l u e =”P8Y”> <!−− e i g h t y e a r s −−> <span><t a r g e t i d =”w90”/>< t a r g e t i d =”w91”/></span> </timex3> <t i m e x 3 i d =”tmx5” t y p e=”DURATION” v a l u e =”PXY”> <!−−t h e y e a r s −−> <span><t a r g e t i d =”w176”/>< t a r g e t i d =”w177”/></span> </timex3> <t i m e x 3 i d =”tmx6” t y p e=”DURATION” b e g i n P o i n t =”tmx8 ” e n d P o i n t=”tmx0” v a l u e =”P21Y”> <!−−21 y e a r s −−> <span><t a r g e t i d =”w198”/>< t a r g e t i d =”w199”/></span> </timex3> <t i m e x 3 i d =”tmx7” t y p e=”DATE” v a l u e =”2003−10−20”/> <t i m e x 3 i d =”tmx8” t y p e=”DATE” v a l u e =”1990−10−20”/> Figure 19: Timex3 elements in NAF NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 # document c r e a t i o n time , which h a s no m e n t i o n s <nwr : w s j 1 0 1 3 . xml#tmx0> a time : I n s t a n t ; rdfs : label ” nwr : t i m e / 1 9 8 9 1 0 2 6 ” t i m e : inDateTime nwr : t i m e / 1 9 8 9 1 0 2 6 . 30/148 ; # DURATION w i t h b e g i n and end p o i n t <nwr : 4 M1J−3MC0−TWKJ−V1W8 . xml#tmx2> a time : I n t e r v a l ; rdfs : label ” week ” ; g a f : denotedBy nwr : 4 M1J−3MC0−TWKJ−V1W8 . xml#c h a r =822 ,825 , nwr : 4 M1J−3MC0−TWKJ−V1W8 . xml#c h a r =826 ,830 nwr : 4 M1J−3MC0−TWKJ−V1W8 . xml#c h a r =831 ,833 nwr : 4 M1J−3MC0−TWKJ−V1W8 . xml#c h a r =834 ,839 t i m e : h a s B e g i n n i n g nwr : t i m e / 2 0 0 5 1 0 0 3 ; t i m e : hasEnd nwr : 2 0 0 6 1 0 0 2 . , , ; # Q u a r t e r i n t e r p r e t e d a s I n t e r v a l w i t h b e g i n and end p o i n t <nwr : w s j 1 0 1 3 . xml#tmx3> a time : I n t e r v a l ; rdfs : label ” quarter ” ; g a f : denotedBy nwr : w s j 1 0 1 3 . xml#c h a r =363 ,366 , nwr : w s j 1 0 1 3 . xml#c h a r =367 ,373 nwr : w s j 1 0 1 3 . xml#c h a r =374 ,381 ; t i m e : h a s B e g i n n i n g nwr : t i m e / 1 9 8 9 0 7 0 1 ; t i m e : hasEnd nwr : t i m e / 1 9 8 9 0 9 3 0 . # DATE i n t e r p r e t e d a s a I n s t a n t <nwr : w s j 1 0 1 3 . xml#tmx4> a time : I n s t a n t ; rdfs : label ”earlier” ; g a f : denotedBy nwr : w s j 1 0 1 3 . xml#c h a r =401 ,402 , nwr : w s j 1 0 1 3 . xml#c h a r =403 ,407 nwr : w s j 1 0 1 3 . xml#c h a r =408 ,415 ; t i m e : inDateTime nwr : t i m e /1988> . <nwr : t i m e /19890701 > a t i m e : day t i m e : month time : unitType time : year time : DateTimeDescription ; ”−−−01”ˆˆ< h t t p : / /www. w3 . o r g / 2 0 0 1 /XMLSchema#gDay> ; ”−−07”ˆˆ< h t t p : / /www. w3 . o r g / 2 0 0 1 /XMLSchema#gMonth> ; t i m e : unitDay ; ”1989”ˆˆ < h t t p : / /www. w3 . o r g / 2 0 0 1 /XMLSchema#gYear> . <nwr : t i m e /19890930 > a t i m e : day t i m e : month time : unitType time : year time : DateTimeDescription ; ”−−−30”ˆˆ< h t t p : / /www. w3 . o r g / 2 0 0 1 /XMLSchema#gDay> ; ”−−09”ˆˆ< h t t p : / /www. w3 . o r g / 2 0 0 1 /XMLSchema#gMonth> ; t i m e : unitDay ; ”1989”ˆˆ < h t t p : / /www. w3 . o r g / 2 0 0 1 /XMLSchema#gYear> . <nwr : t i m e /1988> a time : unitType time : year time : DateTimeDescription ; t i m e : unitDay ; ”1988”ˆˆ < h t t p : / /www. w3 . o r g / 2 0 0 1 /XMLSchema#gYear> . , , Figure 20: SEM representations for time instants and time intervals NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 <p r e d i c a t e A n c h o r <p r e d i c a t e A n c h o r <p r e d i c a t e A n c h o r <p r e d i c a t e A n c h o r i d =”an1 ” i d =”an5 ” i d =”an8 ” i d =”an8 ” 31/148 anchorTime=”tmx0”><span><t a r g e t i d =”p r 1”/></span></p r e d i c a t e A n c h o r > anchorTime=”tmx2”><span><t a r g e t i d =”p r 6”/></span></p r e d i c a t e A n c h o r > b e g i n P o i n t =”tmx2”><span><t a r g e t i d =”p r 9”/></span> </ p r e d i c a t e A n c h o r > e n d P o i n t=”tmx0”><span><t a r g e t i d =”p r 4 3”/></span></p r e d i c a t e A n c h o r > Figure 21: Time anchoring of predicates in NAF 2. If there is no anchor relation through the temporalRelations, check the sentence in which the event is mentioned for time expressions; else the preceding and following sentence, and finally two sentences before the event mention; 3. If there is still no anchor relation, then attach the event to the document creation time; 4. If there is also no document creation time, then anchor the event to the year zero: 0000-12-25. The next example in Figure 21 shows how predicates are anchored to time expressions in the temporalRelations layer in NAF, where the predicate identifier is given as the span and the time expression as an attribute value for either the anchorTime, beginPoint or endPoint. To deal with the complexity of the temporal relations, we had to extend SEM with more specific time relations. In addition to the generic sem:hasTime, events are linked through any of the following relations: sem:hasAtTime: we assume the event took place at this time Instant or during this Interval; sem:hasFutureTime: we assume the event took place in the future relative to this time Instant object; sem:hasEarliestBeginTime: we assume the event began at this time Instant object; sem:hasEarliestEndTime: we assume the event ended at this time Instant object; The procedure is as follows: we first check if the event is explicitly anchored in NAF to a time expression, where: 1. an anchorTime attribute results in a sem:hasAtTime relation 2. a value for beginPoint results in a sem:hasEarliestBeginTime relation 3. an endPoint value in sem:hasEarliestEndTime. If there is no such relation, we check if the factuality module located the event in the future. If that is the case, we create a sem:hasFutureTime relation relative to the document NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 32/148 : e v e n t#Worked sem : hasTime : tmxWeek . : e v e n t#S l e p t sem : h a s E a r l i e s t B e g i n T i m e sem : h a s E a r l i e s t B e g i n T i m e : tmxMonday ; : tmxFriday . : tmxWeek a time : I n t e r v a l rdfs : label ” week ” ; t i m e : h a s B e g i n n i n g nwr : t i m e / 1 9 8 9 0 7 0 1 ; t i m e : hasEnd nwr : t i m e / 1 9 8 9 0 7 0 7 . ; : tmxMonday a rdfs : label t i m e : inDateTime time : I n s t a n t ”Monday” ; nwr : t i m e / 1 9 8 9 0 7 0 1 . ; a rdfs : label t i m e : inDateTime time : I n s t a n t ” Friday ” ; nwr : t i m e / 1 9 8 9 0 7 0 5 . ; : tmxFriday Figure 22: SEM relations between events and time expressions , creation time.5 For all other cases: the event is related to a time expression in the same or a close sentence or it is simply related to the document creation time, we relate the event through the sem:hasAtTime relation. It is important to realise that intervals are represented in two different ways. An event can have a sem:hasAtTime relation to an interval object or it can have a: sem:hasEarliestBeginTime and/or sem:hasEarliestEndTime relation to two different time instance objects. The former applies to cases where the interval itself is explicitly referred to in the text, e.g. I worked for a week, whereas the latter applies to cases where the interval is not mentioned directly in the text but the boundaries of the event are mentioned, e.g. I slept from Monday till Friday. We represent these cases as shown in Figure 22, assuming that the expressions have been normalised to ISO dates. The begin and end points of intervals are therefore either found at the instance level or at the SEM relation level. Although this complicates the querying of the data for events at points in time, we believe it better expresses the way time intervals are represented in language. To ease the retrieval of the events, we always provide a sem:hasTime relation in addition to any of these specific relations. All events related to a specific point in time can be retrieved through this relation, regardless of the specific time relation. 2.2 Identity across events So far, we discussed the way data in a single NAF file is interpreted as SEM instances and represented in RDF-TRiG. However, the GAF framework is designed to capture identity of mentions across sources. This means that we need to compare SEM objects across different 5 Note that events can be located in the future in two ways: we either created an explicit sem:hasFutureTime relation with respect to the document creation time as described here or there was an explicit anchoring to a time that is in the future with respect to the document creation time. In the latter case, the time is known whereas in the former case we only know it may happen after the document creation time NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 33/148 sources to establish identity. For entities, non-entities and normalised time objects this comes naturally through the way the URIs are defined. For these instances, we assume that the URI is defining the identity and no further measures are taken. In the case of events, this is more complicated. The blank URIs are meaningless across documents and we need to define identity on terms of the triples defined for each event. We follow Quine (1985) here who assumes that time and place are defining criteria. Without time and place, actions are just denotations of abstract classes of concepts. They need to be anchored in time and space to become instantiated. Furthermore, we assume that the type of action and the participants play a role in establishing identity (Cybulska and Vossen, 2013a). We therefore defined so-called Composite Event structure combining all the information relevant for an event. It contains the following SEM components: 1. the event instance 2. the entity instances that participate in the event (both actors and locations) 3. time instants and intervals that apply to the event 4. the relations between the event and the other elements Since the instance representations generalise over the mentions, relations between event components are aggregated from different mentions. For example, the time of an event can be mentioned in one sentence whereas a participant can be mentioned in another sentence that is coreferential. Likewise, the instance representation is the aggregation of all the information that directly relates to the events expressed over the complete document. The Composite Event structures are compared across different NAF files to establish cross-document coreference. For this comparison, again all the information related to the instances is used. In case of identify, Composite Event structures are merged, if not, they are kept separate. Figure 23 shows an overview of the two-step approach. The NAF files represent mentions of events (labeled as e), entities that participate in these events (labeled as p), time expressions linked to the events (labelled as t) and locations labeled as l. As far as mentions have identity relations they are coreferential across the mentions in a single NAF file. NAF2SEM then first creates SEM instances from these mentions in a single NAF file and next compares the Composite Events based on the SEM representation across NAF files to establish cross-document identity of events on the basis of the similarity of the action, the participants and the time. The details of this approach are described in more detail below. We developed two different implementations of the above general approach. One for batch processing in which an empty KnowledgeStore is filled with the result of processing a fixed set of NAF documents. Another for a continuous stream of NAF documents, in which the interpretation of incoming NAF files is compared to data already stored in the KnowledgeStore. In the former case, a file structure is used to define which Composite Events need to be compared. In the latter case, each Composite Event extracted from a NAF file is compared to the data in the KnowledgeStore and the result is directly stored. We describe both processes in more detail in next subsection. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 34/148 Figure 23: Identity and event relations between NAF mentions and SEM instances using Composite Event structures NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 2.2.1 35/148 Event comparison in batch mode The main steps for batch processing are: 1. ClusterEventObjects (a) Read each NAF file one by one; (b) Create Composite Event objects from single NAF files; (c) Store the Composite Events into temporal buckets as NAF objects; 2. MatchEventObjects (a) Read all Composite Events from the NAF objects in the same temporal bucket; (b) Compare the events to decide on cross-document event coreference and merge data if necessary; (c) Output the Composite event data to a SEM.trig file for each temporal bucket; 3. Populate the KnowledgeStore with the SEM.trig files; In Figure 24, we show an overview of the batch processing architecture. From single NAF files, we extract Composite Event structures which are stored in so-called temporal buckets for the different main types of events: contextualEvent, source and grammatical events (see below for more details). After processing all NAF files, the binary object files in each bucket are loaded in memory and compared for identity. Identical events are merged. The result is stored in SEM RDF-TRiG files. When the process has finished, the SEM RDF-TRiG files are used to populate the KnowledgeStore. After creating the time-event relations, we use the time relation to create temporal buckets, where we apply the following rules: 1. Events without a temporal relation or with too many temporal relations are ignored. The threshold for the maximum number of temporal relations is now set to 5. If more than 5 time expressions are associated with the event, we assume the temporal relation is too complex to interpret. 2. If there is a single time relation to a time:Instant Object, we use the date to create a folder, e.g. e-1989-10-26 to store the event data. If the time Object is of the type time:Interval, we create a period using the begin and end points, e.g. e-2005-10-032006-10-02. 3. If there is more than one time Object related to an event we check if the relations are sem:hasEarliestBeginTime and sem:hasEarliestEndTime. These relations implicitly define a period without the use of an interval expression, i.e. there is no linguistic expression of the interval but only of the begin and end of the event. If so, we create a period bucket by first taking the begin points and then taking the end points, e.g. e-1989-10-12-1989-10-26. Note that events that have only a begin or end point, still get a simple event bucket similar to events with a single event relation. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 36/148 Figure 24: NAF2SEM batch processing overview 4. If the multiple time Objects are not the begin or end points, we create different time buckets for each time object and the event is duplicated to each bucket. This means that we assume that the event can be matched with different events at different points of time with other events.6 The above time buckets are created within the basic event type distinction between contextual, source and grammatical events except for those events that are positioned in the future (see below). For each event, we check the associated frames against these lists to decide on the main event type. If no frame is associated, the event is considered as contextual. Events with the sem:hasFutureEvent relation are stored in a separate folder without distinguishing the basic event types. Figure 25 shows the structure that is created for storing Composite Event data in batch processing mode. The events folder is subdivided into 4 subfolders: contextualEvent, sourceEvent, grammaticalEvent and futureEvent. Within each subfolder, binary object files (.obj) are stored in temporal buckets for events anchored to that time from each NAF file with all the data relevant for that event. 6 There is a positive and negative side effect to this strategy. If the events get different instance URIs they get automatically split. If the URIs remain the same they get merged back into a single instance eventually. Whether this happens depends on the order of the comparison since the event instance URI is based on the first URI used in the comparison. A future fix of this arbitrary effect is that merging events forces the creation of a new unique URI. A side-effect of this fix is that split events will never get merged and are treated as distinct events NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 37/148 Figure 25: Example of an event type and temporal bucket structure created by the NAF2SEM module. First the events are divided into contextual, source and grammatical events, and within these into temporal buckets. In each temporal bucket, an object file is created for the events from each NAF file that are associated with the time. Across the different NAF object files, a single sem.trig file is created with all the event data. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 38/148 The purpose of the division into event type and temporal buckets is to compare events for cross-document event coreference, after which the SEM RDF-TRiG files are created. We create a sem.trig file for each bucket. We apply different strategies for the different event types and for future events. Events are considered identical: IFF events have the same time anchoring; IFF events share the same dominant senses or the same lemmas; IFF events shared sufficient participants, which varies for the type of event and whether events are labeled as future or not; The temporal anchoring follows from the comparison of events within the same temporal buckets. The events with the same time anchor are first compared for the type of action expressed. Events are derived from the coreference sets in the NAF file and each Composite Event can be based on more than one mention of the event. These mentions can have the same lemma and share wordnet synsets. Across the Composite Events, we first check if they have sufficient overlap in the wordnet synsets (defined by a threshold that can be adapted, default is set to 50%). This prevents different meanings of the same word to be matched for the wrong reasons. If there are no wordnet synsets associated with either of the Composite Events, we check if the lemmas sufficiently overlap across the coreference set (this can also be set through a proportional threshold, default 50%).7 Finally, events need to share participants. What participants and how many participants need to be matched can be specified through the API by specifying the role-labels that need to be matched. For the different types of events, we follow the following principles: 1. source events must share the PropBank A0 participant to be coreferential. This correlates to the source of the event. The target of the source event is more difficult to compare because the descriptions are usually longer and can vary across mentions.8 2. grammatical events typically have PropBank A1 and A2 arguments that need to match since these arguments usually refer to the main event, e.g. They stopped the meeting on Monday, The meeting was stopped on Monday. We also assume that the lemmas need to match instead of synsets since the meanings of these expressions are abstract. 3. contextual events typically can vary with respect to the participants expressed. We typically assume that the A1 role is most informative and needs to match. Depending on the domain other roles can be choosen. 7 We also may have the synset of Lowest-Common-Subsumers (LCS) from NAF coreference sets that have different lemmas. The LCS represents the hypernym synset that mention pairs share. Currently, this is not used to match events. 8 A future extension could do a word vector comparison of the target roles of these events to constraint coreference more NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 39/148 j a v a −Xmx2000m −cp . . / l i b / E v e n t C o r e f e r e n c e −1.0−SNAPSHOT−j a r −with−d e p e n d e n c i e s . j a r eu . n e w s r e a d e r . e v e n t c o r e f e r e n c e . n a f . C l u s t e r E v e n t O b j e c t s −−n a f−f o l d e r ” . . / t e s t ” −−e v e n t −f o l d e r ” . . / t e s t ” −−e x t e n s i o n ” . xml ” −−p r o j e c t c a r s −−s o u r c e −f r a m e s ” . . / r e s o u r c e s / s o u r c e . t x t ” −−g r a m m a t i c a l −f r a m e s ” . . / r e s o u r c e s / g r a m m a t i c a l . t x t ” −−c o n t e x t u a l −f r a m e s ” . . / r e s o u r c e s / c o n t e x t u a l . t x t ” −−non− e n t i t i e s j a v a −Xmx2000m −cp . . / l i b / E v e n t C o r e f e r e n c e −1.0−SNAPSHOT−j a r −with−d e p e n d e n c i e s . j a r eu . n e w s r e a d e r . e v e n t c o r e f e r e n c e . n a f . M a t c h E v e n t O b j e c t s −−e v e n t −f o l d e r ” . . / t e s t / e v e n t s / c o n t e x t u a l E v e n t ” −−match−t y p e i l i l e m m a −− r o l e s ” a n y r o l e ” −−c o n c e p t −match 30 −−p h r a s e −match 10 −− i l i . . / r e s o u r c e s / i l i . t t l −−p e r s p e c t i v e j a v a −Xmx2000m −cp . . / l i b / E v e n t C o r e f e r e n c e −1.0−SNAPSHOT−j a r −with−d e p e n d e n c i e s . j a r eu . n e w s r e a d e r . e v e n t c o r e f e r e n c e . n a f . M a t c h E v e n t O b j e c t s −−e v e n t −f o l d e r ” . . / t e s t / e v e n t s / s o u r c e E v e n t ” −−match−t y p e i l i l e m m a −− r o l e s ” a0 , a1 ” −−c o n c e p t −match 50 −−p h r a s e −match 50 −− i l i . . / r e s o u r c e s / i l i . t t l −−p e r s p e c t i v e j a v a −Xmx2000m −cp . . / l i b / E v e n t C o r e f e r e n c e −1.0−SNAPSHOT−j a r −with−d e p e n d e n c i e s . j a r eu . n e w s r e a d e r . e v e n t c o r e f e r e n c e . n a f . M a t c h E v e n t O b j e c t s −−e v e n t −f o l d e r ” . . / t e s t / e v e n t s / g r a m m a t i c a l E v e n t ” −−match−t y p e lemma −− r o l e s ” a1 , a2 ” −−p h r a s e −match 50 −− i l i . . / r e s o u r c e s / i l i . t t l −−p e r s p e c t i v e j a v a −Xmx2000m −cp . . / l i b / E v e n t C o r e f e r e n c e −1.0−SNAPSHOT−j a r −with−d e p e n d e n c i e s . j a r eu . n e w s r e a d e r . e v e n t c o r e f e r e n c e . n a f . M a t c h E v e n t O b j e c t s −−e v e n t −f o l d e r ” . . / t e s t / e v e n t s / f u t u r e E v e n t ” −−match−t y p e lemma −− r o l e s ” a l l ” −−p h r a s e −match 80 −− i l i . . / r e s o u r c e s / i l i . t t l −−p e r s p e c t i v e Figure 26: NAF2SEM calls for clustering and matching with different paramters 1 2 3 4 5 6 7 8 9 10 11 j a v a −Xmx2000m −cp . . / l i b / E v e n t C o r e f e r e n c e −1.0−SNAPSHOT−j a r −with−d e p e n d e n c i e s . j a r eu . n e w s r e a d e r . e v e n t c o r e f e r e n c e . n a f . N o C l u s t e r E v e n t O b j e c t s −−n a f−f o l d e r ” . . / t e s t ” −−e v e n t −f o l d e r ” . . / t e s t ” −−e x t e n s i o n ” . xml ” −−p r o j e c t c a r s −−s o u r c e −f r a m e s ” . . / r e s o u r c e s / s o u r c e . t x t ” −−g r a m m a t i c a l −f r a m e s ” . . / r e s o u r c e s / g r a m m a t i c a l . t x t ” −−c o n t e x t u a l −f r a m e s ” . . / r e s o u r c e s / c o n t e x t u a l . t x t ” −−non− e n t i t i e s −−t o p i c s j a v a −Xmx2000m −cp . . / l i b / E v e n t C o r e f e r e n c e −1.0−SNAPSHOT−j a r −with−d e p e n d e n c i e s . j a r eu . n e w s r e a d e r . e v e n t c o r e f e r e n c e . n a f . M a t c h E v e n t O b j e c t s −−e v e n t −f o l d e r ” . . / t e s t / e v e n t s / a l l ” −−match−t y p e i l i l e m m a −− r o l e s ” a n y r o l e ” −−t i m e ”month” −−c o n c e p t −match 30 −−p h r a s e −match 10 −− i l i . . / r e s o u r c e s / i l i . t t l −−p e r s p e c t i v e Figure 27: NAF2SEM calls without clustering and matching all events 4. future events need to have identical participants. Since there is no specific time anchoring, we can only assume identity if all other information matches. In Figure 26 we show the current settings that are used to call the function for creating the binary object files for NAF files (ClusterEventObjects) and the functions for processing the temporal buckets in each type of event (MatchEventObjects): Alternatively, one can choose also not to cluster events using the class: eu.newsreader.eventcoreference.naf.NoClusterEventObjects and in addition use the temporal anchoring of the events as a matching constraint, as shown in Figure 27. In the latter case, you can set the granularity of the temporal dimension to year, month or day. Optionally, you can use the –topic parameter to subdivide into topics within the main folder. Note that the –topic option also works for the above event-type and temporal clustering option. 2.2.2 Event comparison in streaming mode The NAF2SEM stream architecture has been introduced to enable real-time, incremental processing of NAF documents. The batch processing mode works quite solid for transformNewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 40/148 nwr : i n s t a n c e s { <nwr : d a t a / 2 0 1 4 / 0 6 / 1 1 / 5CDF−W191−F15C−G01S . xml#ev13> a sem : Event , n w r o n t o l o g y : c o n t e x t u a l E v e n t , wn : eng −14422488−n , wn : eng −13456567−n , wn : eng −13457378−n , f n : C h a n g e p o s i t i o n o n a s c a l e , e s o : QuantityChange , wn : eng −00203866−v ; rdfs : label ” decline ” ; g a f : denotedBy <nwr : d a t a / 2 0 1 4 / 0 6 / 1 1 / 5CDF−W191−F15C−G01S . xml#c h a r =33,40> . <nwr : d a t a / c a r s / non− e n t i t i e s /30+%25> rdfs : label ”30 %” ; g a f : denotedBy <nwr : d a t a / 2 0 1 4 / 0 6 / 1 1 / 5CDF−W191−F15C−G01S . xml#c h a r =29,32> . <nwr : t i m e /20140611 > a t i m e : day t i m e : month time : unitType time : year time : DateTimeDescription ; ”−−−11”ˆˆ< h t t p : / /www. w3 . o r g / 2 0 0 1 /XMLSchema#gDay> ; ”−−06”ˆˆ< h t t p : / /www. w3 . o r g / 2 0 0 1 /XMLSchema#gMonth> ; t i m e : unitDay ; ”2014”ˆˆ < h t t p : / /www. w3 . o r g / 2 0 0 1 /XMLSchema#gYear> . <nwr : d a t a / 2 0 1 4 / 0 6 / 1 1 / 5CDF−W191−F15C−G01S . xml#tmx0> a time : I n s t a n t ; t i m e : inDateTime <nwr : t i m e /20140611 > . ... } <nwr : d a t a / 2 0 1 4 / 0 6 / 1 1 / 5CDF−W191−F15C−G01S . xml#t r 3 > { <nwr : d a t a / 2 0 1 4 / 0 6 / 1 1 / 5CDF−W191−F15C−G01S . xml#ev13> sem : hasTime <nwr : d a t a / 2 0 1 4 / 0 6 / 1 1 / 5CDF−W191−F15C−G01S . xml#tmx0> . } nwr : p r o v e n a n c e { <nwr : d a t a / 2 0 1 4 / 0 6 / 1 1 / 5CDF−W191−F15C−G01S . xml#pr3 , r l 4 > g a f : denotedBy <nwr : d a t a / 2 0 1 4 / 0 6 / 1 1 / 5CDF−W191−F15C−G01S . xml#c h a r =29,40> . <nwr : d a t a / 2 0 1 4 / 0 6 / 1 1 / 5CDF−W191−F15C−G01S . xml#t r 1 0 > g a f : denotedBy <nwr : d a t a / 2 0 1 4 / 0 6 / 1 1 / 5CDF−W191−F15C−G01S . xml#> . <nwr : d a t a / 2 0 1 4 / 0 6 / 1 1 / 5CDF−W191−F15C−G01S . xml#pr15 , r l 3 4 > g a f : denotedBy <nwr : d a t a / 2 0 1 4 / 0 6 / 1 1 / 5CDF−W191−F15C−G01S . xml#c h a r =296 ,330 > . ... } Figure 28: Example snippet of the TRiG stream generated by NAF2SEM ing very big sets of news articles from NAF to SEM. However, in realistic scenarios the news articles come incrementally, minute by minute, day by day. The batch architecture is unable to handle this situation, as it requires the batch to be reprocessed each time new NAFs need to be processed. Clearly, the streaming architecture is a promising solution for this challenge. Once the initial set of NAF files would be processed and inserted using the batch mode, the follow-up NAF files are being processed through a streaming architecture. Unlike the batch architecture which mainly pays off once we have a significant amount of NAF files, the stream architecture is meant to work with a small set of NAF files at a time (maximum of 1000 at a time). Figure 31 presents an overview of the streaming architecture. Once a new (set of) NAF(s) has arrived, the NAF2SEM module is first fired to convert each NAF file into an RDF-TRiG representation. In the streaming architecture, NAF2SEM simply creates a TRiG set of events by looking at the current file. Similarly as in the batch mode, the events are extracted together with their attached actors, relations and time expressions. What is different than in the batch mode, is that the extracted events are not matched with events coming from other news articles. An example snippet of a TRiG created by NAF2SEM is given in Figure 28. The output from the NAF2SEM module is then used as an input for the streaming NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 41/148 Table 1: Cross-document event coreference arguments for stream processing Argument –concept-match INT –phrase-match INT –contextual-match-type PATH –contextual-lcs –contextual-roles PATH –source-match-type PATH –source-lcs –source-roles PATH –grammatical-match-type PATH –grammatical-lcs –grammatical-roles PATH –future-match-type PATH –future-lcs –recent-span INT Comment Threshold for conceptual matches of events, default is 50. Threshold for phrase matches of events, default is 50. Indicates what is used to match events across resources. Default value is ILILEMMA. Use lowest-common-subsumers. Default value is ON. String with roles for which there must be a match, e.g. ”pb:A1, sem:hasActor” Indicates what is used to match events across resources. Default value is ILILEMMA. Use lowest-common-subsumers. Default value is OFF. String with roles for which there must be a match, e.g. ”pb:A1, sem:hasActor” Indicates what is used to match events across resources. Default value is LEMMA. Use lowest-common-subsumers. Default value is OFF. String with roles for which there must be a match, e.g. ”pb:A1, sem:hasActor” Indicates what is used to match events across resources. Default value is ”LEMMA” Use lowest-common-subsumers. Default value is OFF. Amount of past days which are still considered recent and are treated differently. cross-document event coreference module. Apart from the SEM-TRiG input, the user is also allowed to specify a rich set of matching arguments, such as match type and a list of needed roles for each event type. The full list of input arguments for event coreference is given in Table 1. The cross-document event coreference module then processes the piped TRiG stream based on the argument configuration specified by the user. This module essentially builds a SPARQL SELECT query for each of the events from the TRiG, based on its components: role participants, relations and time expressions (see Figure 32 for an example query). The SPARQL query incorporates all criteria for matching of the current event with respect to the KnowledgeStore. Once it has been composed, the query is sent to the KnowledgeStore. Every result returned from the KnowledgeStore (if any) is a coreferential event to the event in question. For each of these events, we create an owl:sameAs relation to the current event, and append all the sameAs identity relations to the input TRiG in a separate graph (nwr:identities). The nwr:identity graph is shown in Figure 29. Now the TRiG contains all the events from the current news article and their identity relations to already processed events in the KnowledgeStore. It can be the case that the current event has no coreferential event yet in the KnowledgeStore: in this case, the event has no entry in the graph with identity relations. Nevertheless, at this point the TRiG is ready to be inserted into the KnowledgeStore. The module functionality hence finishes with an INSERT request to the KnowledgeStore to store the resulting TRiG. Once this is done, the new events still need to be merged to their coreferential events (if any) in the KnowledgeStore. The remainder of the events (with no coreferential correspondent) is directly inserted in KS. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 42/148 nwr : i d e n t i t y { <nwr : d a t a / 2 0 1 4 / 0 6 / 1 1 / 5CDF−W191−F15C−G01S . xml#ev40> owl : sameAs <nwr : d a t a / c a r s /52 F6−KBJ1−DYWJ−P42R . xml#ev113> . <nwr : d a t a / 2 0 1 4 / 0 6 / 1 1 / 5CDF−W191−F15C−G01S . xml#ev48> owl : sameAs <nwr : d a t a / 2 0 1 4 / 0 6 / 1 1 / 5CDF−W191−F15C−G01S . xml#ev47> . <nwr : d a t a / 2 0 1 4 / 0 6 / 1 1 / 5CDF−W191−F15C−G01S . xml#ev50> owl : sameAs <nwr : d a t a / 2 0 1 4 / 0 6 / 1 1 / 5CDF−W191−F15C−G01S . xml#ev45> . <nwr : d a t a / 2 0 1 4 / 0 6 / 1 1 / 5CDF−W191−F15C−G01S . xml#ev51> owl : sameAs <nwr : d a t a / 2 0 1 4 / 0 6 / 1 1 / 5CDF−W191−F15C−G01S . xml#ev50> . <nwr : d a t a / 2 0 1 4 / 0 6 / 1 1 / 5CDF−W191−F15C−G01S . xml#ev50 tmx0> owl : sameAs <nwr : d a t a / 2 0 1 4 / 0 6 / 1 1 / 5CDF−W191−F15C−G01S . xml#ev45> . } Figure 29: Examples of owl:sameAs relations generated in streaming mode 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 #!/ b i n / bash FILES = ” . . / n a f s t o p r o c e s s / ∗ . xml ” f o r f i n $FILES ; do c a t $ f | j a v a −Xmx2000m −cp . . / t a r g e t / E v e n t C o r e f e r e n c e −1.0−SNAPSHOT−j a r −with−d e p e n d e n c i e s . j a r eu . n e w s r e a d e r . e v e n t c o r e f e r e n c e . n a f . GetSemFromNafStream −−p r o j e c t c a r s −−s o u r c e −f r a m e s ” . . / r e s o u r c e s / s o u r c e . t x t ” −−g r a m m a t i c a l −f r a m e s ” . . / r e s o u r c e s / g r a m m a t i c a l . t x t ” −−c o n t e x t u a l −f r a m e s ” . . / r e s o u r c e s / c o n t e x t u a l . t x t ” −−non− e n t i t i e s −−timex−max 5 | j a v a −Xmx2000m −cp . . / t a r g e t / E v e n t C o r e f e r e n c e −1.0−SNAPSHOT−j a r −with−d e p e n d e n c i e s . j a r eu . n e w s r e a d e r . e v e n t c o r e f e r e n c e . n a f . P r o c e s s E v e n t O b j e c t s S t r e a m −−c o n t e x t u a l −match−t y p e ”ILILEMMA” −−c o n t e x t u a l −l c s −−s o u r c e −match−t y p e ”ILILEMMA” −−s o u r c e −r o l e s ”pb : A0” −−g r a m m a t i c a l −match−t y p e ”LEMMA” −−g r a m m a t i c a l −r o l e s ”pb : A1” −−c o n c e p t −match 25 −−p h r a s e −match 25 > . . / t r i g s / $ f . t r i g done Figure 30: Script for NAF2SEM in streaming architecture An example call to the streaming architecture is shown in Figure 30. The streaming architecture is an end-to-end architecture spreading from NAF to the KnowledgeStore. It turns NewsReader into a real-time incremental system, ranging from a news article to KnowledgeStore, in the following manner: 1. A fresh news article appears 2. This file is processed by the NewsReader pipeline. 3. The pipeline creates a NAF file, which is fed to NAF2SEM. 4. NAF2SEM extracts RDF-TRiG data from NAF and pipes this to the cross-document Event coreference module. 5. This module communicates with the KnowledgeStore and decides on the final set of events and their coreferential relations to the KS events. This is represented in RDF-TRiG format. 6. KS finally “digests” the new RDF, by merging the coreferential events and by simply inserting the events which do not have coreferential event in KS. From a scientific perspective, this architecture introduces the flexibility to adjust the manner of event matching and facilitates experimentation with different parameters. Through NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 43/148 Figure 31: Overview of the NAF2SEM and event coreference stream architecture. Figure 32: Example SPARQL query used to match events from code in the KnowledgeStore. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 44/148 the list of configuration arguments, we are offered an easy option to experiment with various matching options concerning: roles, matching thresholds, matching types, etc. This list of arguments is furthermore not finite and may easily expand/shrink as we go along. We are planning to experiment with using lowest-common-subsumers (LCS) in the matching of event relations. We also plan to work towards using WordNet relations for coreference. Additionally, time is an important factor to be further researched. We may want to introduce a recency relevance, which would first try to match an event to the most recent set of events (from the last few days). This idea comes from the fact that the news articles about an event often spread in a rather small interval of time. Another time aspect which we may look into is the matching of future events. As mentioned above, this architectural solution requires certain caution with respect to the amount of files being fed to the system. Components from this architecture may easily become bottlenecks, for instance, the KnowledgeStore may become slow if the frequency and number of event merges is too high. 2.3 Evaluation In this section, we present an evaluation of the quality of the SEM-RDF built with our approach (forthcoming: Rospocher et al. (2016)). Due to the lack of a proper gold standard to compare with, we relied on human judgment for the triples describing some randomly sampled events of the graph. A similar approach was applied to evaluate YAGO2 Hoffart et al. (2013), a large knowledge graph automatically extracted from Wikipedia pages. We conducted an evaluation of the SEM-RDF triples extracted from the NewsReader Wikinews corpus consisting of 120 news documents. We sampled 100 events from the resulting RDF, splitting them in 4 groups of 25 events each. For each event, we retrieved all its triples, obtaining 4 subgraphs (labeled: S1 , S2 , S3 , S4 ) of approx. 260 triples each. Each subgraph was submitted to a pair of human raters, which independently evaluated each triple of their subgraph. The triples of each subgraph were presented to the raters grouped by event, and for each event the link to its corresponding mentions in the text were provided, so that raters were able to inspect the original text to assess the correctness of the extracted triples. In total, 8 human raters evaluated a total of 1,043 triples of the RDF-SEM data, with each triple independently evaluated by two raters. Raters were given precise criteria to follow for evaluating their subgraph. For instance, in case of an event extracted from many event mentions, raters were instructed to first assess if all its mentions actually refer to the same event: if at least one of these mentions is referring to an event different than the other ones, all triples of the resulting instance have to be considered incorrect.9 This is a quite strict and potentially highly-penalizing criterion, if considered in absolute terms from a triple perspective: one “wrong” mention out of many coreferring mentions, potentially contributing with few triples to the event, may hijack all the triples of the corresponding event. There were for example several instances in which 4 mentions were identified by the pipeline as referring to the same 9 A similar criterion was adopted for cases where something was wrongly identified as an event. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 45/148 Table 2: Quality triple evaluation of SEM-RDF extracted from Wikinews. S1 S2 S3 S4 All Triples 267 256 261 259 1043 Accuracy 0.607 0.525 0.552 0.548 0.551 κ 0.623 0.570 0.690 0.751 event instance, of which 3 were indeed referring to the same instance. Due to our strict evaluation method, all four mentions were considered incorrect. Performing a pairwise evaluation would have been less strict, but as our goal is to accurately extract knowledge graphs from text, and in particular to obtain correctly structured description of events, we believe this criterion goes in this direction. Table 2.3 presents the resulting triple accuracy on the whole evaluation dataset, as well as the accuracy on each subgraph composing it, obtained as average of the assessment of the each rater pair. For each subgraph, the agreement between the rater pair is also reported, computed according to the Cohen’s kappa coefficient (κ). The results show an overall accuracy of 0.551, varying between 0.525 and 0.607 on each subgraph. The Cohen’s kappa values, ranging from 0.570 and 0.751, show a substantial agreement between the raters of each pair. Drilling down these numbers on the type of triples considered — typing triples (rdf:type), annotation triples (rdfs:label), participation triples (properties modelling event roles according to PropBank, FrameNet, and ESO), the accuracy on annotation triples is higher (0.772 on a total of 101 triples), while it is slightly lower for typing (0.522 on 496 triples) and participation triples (0.534 on 446 triples). Indeed, further drilling down on participation triples, the accuracy is higher for PropBank roles (0.559) while it is lower on FrameNet (0.438) and ESO roles (0.407), which reflects the fact that the SRL tool used is trained on PropBank, while FrameNet and ESO triples are obtained via mapping. Looking at the event candidates in the evaluation dataset, 69 of them (out of 100) were confirmed as proper events by both raters. Of the 17 candidate coreferring events (i.e. those having multiple mentions), only 4 of them were marked as correct by both raters (i.e. both raters stated that all mentions were actually referring to the same event) while in a couple of cases an event was marked as incorrect because of one wrong mention out of 4, thus causing all the triples of the event to be marked as incorrect. To remark the aforementioned strict evaluation criterion adopted, we note that ignoring all coreferring events (and their corresponding triples) in the evaluation dataset, the triple accuracy rises to 0.697 on a total of 782 triples. Table 3 shows the details for both the full evaluation and when ignoring the even-coreference effects. Finally, these numbers need to be interpreted in relation to the task-specific evaluation of event detection and event coreference on benchmark corpora. In Wikinews recall of event mention detection is about 77% against the gold annotation of events, while within document event-coreference has a precision of 40.7%. This is a different type of evaluation NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 46/148 Table 3: Detailed quality triple evaluation of SEM-RDF extracted from Wikinews with and without taking even-coreference into account. Group 1 Trp Avg κ ALL 267 0.607 0.623 TYPES 122 0.594 0.649 LABELS 28 0.768 0.7 ROLES 117 0.581 0.591 PROPBANK 39 0.628 0.629 FRAMENET 77 0.461 0.559 ESO 26 0.5 0.692 Without Event Coreference: ALL 207 0.732 0.447 TYPES 91 0.731 0.483 LABELS 23 0.87 0.617 ROLES 93 0.699 0.419 PROPBANK 32 0.734 0.472 FRAMENET 56 0.58 0.387 ESO 17 0.706 0.433 Trp 256 115 26 115 39 73 25 188 91 20 77 25 58 20 Group 2 Avg κ 0.525 0.57 0.539 0.585 0.788 0.661 0.452 0.509 0.423 0.633 0.438 0.445 0.4 0.5 0.601 0.566 0.925 0.558 0.56 0.457 0.375 0.491 0.578 0 0.37 0.516 0.41 0.468 Trp 261 137 24 100 30 90 38 184 96 17 71 20 70 21 Group 3 Avg κ 0.552 0.69 0.504 0.65 0.729 0.684 0.575 0.735 0.583 0.796 0.489 0.645 0.368 0.435 0.731 0.672 0.912 0.768 0.825 0.607 0.571 0.683 0.649 0.638 0.724 0.828 0.615 0.438 Trp 259 122 23 114 28 105 29 203 93 19 91 23 79 21 Group 4 Avg κ 0.548 0.751 0.578 0.748 0.804 0.862 0.465 0.718 0.554 0.928 0.424 0.63 0.345 0.545 0.7 0.758 0.974 0.582 0.674 0.563 0.476 0.625 0.561 0 0.639 0.901 0.512 0.432 Overall Trp Avg 1043 0.551 496 0.522 101 0.772 446 0.534 136 0.559 345 0.438 118 0.407 782 371 79 332 100 263 79 0.697 0.663 0.937 0.678 0.72 0.548 0.557 since the annotators only labelled the important events and systems tend to recognise all possible events in text, which down-ranks the precision. We cannot expect coreference results within a document and cross-document to exceed the recall of event mentions. In the next section, we describe the evaluation for cross-document event coreference. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 3 47/148 Event Coreference Event coreference resolution is the task of determining whether two event mentions refer to the same event instance. In this section, we describe a new, robust “bag of events” approach to cross-textual event coreference resolution on news articles. We discuss two variations of the approach, a one- and a two-step implementation. We first delineate the approach and then describe some experiments with the new method on the ECB+ data set. 10 In the following section, we describe the current implementation within the NewsReader pipeline. 3.1 3.1.1 Bag of Events Approach The Overall Approach It is pretty much common practice to use information coming from event arguments for event coreference resolution (Humphreys et al. (1997), Chen and Ji (2009a), Chen and Ji (2009b), Chen et al. (2011), Bejan and Harabagiu (2010a), Lee et al. (2012), Cybulska and Vossen (2013b), Liu et al. (2014) amongst others). The research community seems to agree that event context information regarding time and place of an event as well as information about other participants play an important role in resolution of coreference between event mentions. Using entities for event coreference resolution is complicated by the fact that event descriptions within a sentence often lack pieces of information. As pointed out by Humphreys et al. (1997) it could be the case however that a lacking chunk of information might be available elsewhere within discourse borders. News articles, which are the focus of the NewsReader project, can be seen as a form of public discourse (van Dijk (1988)). As such the news follows the Gricean Maxim of quantity (Grice (1975)). Authors do not make their contribution more informative than necessary. This means that information previously communicated within a unit of discourse, unless required, will not be mentioned again. This is a challenge for models comparing separate event mentions with one another on the sentence level. To be able to fully make use of information coming from event arguments, instead of looking at event information available within the same sentence, we propose to take a broader look at event descriptions surrounding the event mention in question within a unit of discourse. For the purpose of this study, we consider a document (a news article) to be our unit of discourse. We experimented with an “event template” approach which employs the structure of event descriptions for event coreference resolution. In the proposed heuristic event mentions are looked at through the perspective of five slots, as annotated in the ECB+ dataset created within the NewsReader project (Cybulska and Vossen (2014b)). The five slots correspond to different elements of event information such as the action slot (or event trigger following the ACE (LDC (2005)) terminology) and four kinds of event arguments: time, location, human and non-human participant slots (see Cybulska and Vossen (2014)). 10 The experiments with the two-step bag of events approach that are reported in this section are described in Cybulska and Vossen (2015). NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 48/148 The ECB+ corpus is used in the experiments described here. Our approach determines coreference between descriptions of events through compatibility of slots of the five slot template. The next quote shows an excerpt from topic one, text number seven of the ECB corpus (Bejan and Harabagiu (2010b)). The “American Pie” actress has entered Promises for undisclosed reasons. The actress, 33, reportedly headed to a Malibu treatment facility on Tuesday. Consider two event templates presenting the distribution of event information over the five event slots in the two example sentences (tables 4 and 5). An event template can be created on different levels of information, such as a sentence, a paragraph or an entire document. We propose a novel “bag of events” approach to event coreference that explicitly employs event- and discourse structure to account for implications of Gricean Maxim of quantity. The approach fills in two event templates: a sentence and a document template. A “sentence template” collects event information from the sentence of an active action mention (tables 4 and 5). By filling in a “document template”, one creates a “bag of events” for a document, that could be seen as a kind of document “summary” (table 6). The bag of events heuristic employs clues coming from discourse structure and namely those implied by discourse borders. Descriptions of different event mentions occurring within a discourse unit, whether coreferent or related in some other way, unless stated otherwise, tend to share elements of their context. In our example text fragment the first sentence reveals that an actress has entered a rehab facility. From the second sentence the reader finds out where the facility is located (Malibu) and when the “American Pie” actress headed to the treatment center. It is clear to the reader of the example text fragment from the quotation that both events described in sentence one and two, happened on Tuesday. Also both sentences mention the same rehab center in Malibu. These observations are crucial for the “bag of events” approach proposed here. The bag of events method can be implemented as a one- or two-step classification. In a two-step approach bag of events (document) features are used for preliminary document clustering. Then per document cluster coreference is solved between action mentions in a pairwise model, based on information available in the sentence. In a one-step implementation bag of events features are added to sentence-based feature vectors generated per action mention. Coreference is solved by a classifier in a pairwise model. 3.1.2 Two-step Bag of Events Approach As the first step of the approach a document template is filled, accumulating instances of the five event slot mentions from a document, as exemplified in table 6. Pairs of document templates are clustered by means of supervised classification. In the second step of the approach coreference is solved between event mentions within document clusters created in step 1. For this task again an event template is filled but this time, it is a “sentence template” which per event mention gathers information from the sentence. A supervised classifier solves coreference between pairs of event mentions and finally pairs sharing comNewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 49/148 mon mentions are chained into coreference clusters. Figure 33 depicts the implications of the approach for the training data. Figure 34 presents how the test set is processed. Table 4: Sentence template ECB topic 1, text 7, sentence 1 Action Time Location Human Participant Non-Human Participant entered N/A Promises actress N/A Table 5: Sentence template ECB topic 1, text 7, sentence 2 Action Time Location Human Participant Non-Human Participant headed on Tuesday to a Malibu treatment facility actress N/A Table 6: Document template ECB topic 1, text 7, sentences 1-2 Action Time Location Human Participant Non-Human Participant 3.1.3 entered, headed on Tuesday Promises, to a Malibu treatment facility actress N/A Step 1: Clustering Documents Using Bag of Events Features The first step in this approach is filling in an event template per document. We create a document template by collecting mentions of the five event slots: actions, locations, times, human and non-human participants from a single document. In a document template there is no distinction made between pieces of event information coming from different sentences of a document and no information is kept about elements being part of different mentions. A document template can be seen as a bag of events and event arguments. The template stores unique lemmas, to be precise a set of unique lemmas per event template slot. On the training set of the data, we train a pairwise binary classifier determining whether two document templates share corefering event mentions. This is a supervised learning task in which we determine “compatibility” of two document templates if any two mentions from those templates were annotated in the corpus as coreferent. Let m be an event mention, and doc a collection of mentions from a single document template such that {mı : 1 ≤ ı ≤ doc} where ı is the index of a mention and NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 50/148 Table 7: ECB+ statistics ECB+ Topics Texts Action mentions Location mentions Time mentions Human participant mentions Non-human participant mentions Coreference chains # 43 982 6833 1173 1093 4615 1408 1958 Figure 33: Bag of events approach - training set processing Figure 34: Bag of events approach - test set processing NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 51/148 indexes document templates; doc : 1 ≤ ≤ DOC where DOC are all document templates from the corpus. Let ma and mb be mentions from different document templates. “Compatibility” of a pair of document templates (doc, doc+1 ) is determined based on coreference of any mentions (maı, mbı) from a pair of document templates such that: coref erence(∃maı ∈ doc, ∃mbı ∈ doc+1 ) =⇒ compatibility (doc, doc+1 ). On the training set we train a binary decision tree classifier (hereafter DT ) to find pairs of document templates containing corefering event mentions. After all unique pairs of document templates from the test set have been classified by means of the DT document template classifier, “compatible” pairs are merged into document clusters based on pair overlap. 3.1.4 Step 2: Clustering Sentence Templates The aim of the second step is to solve coreference between event mentions from document clusters which are the output of the classification task from step 1. We experiment with a supervised decision tree sentence template classifier but this time in the classification task pairs of sentence templates are considered. A sentence template is created for every action mention annotated in the data set (see examples of sentence templates in table 4 and 5). All possible unique pairs of event mentions (and their sentence templates) are generated within clusters of document templates sharing corefering event mentions in the training set. Pairs of sentence templates that translate into features indicating compatibility across five template slots are used to train a DT sentence template classifier. On the test set; after output clusters of the DT document template classifier from step 1 are turned to mention pairs (all unique pairs within a document cluster), pairs of sentence templates are classified by means of the DT sentence template classifier. To identify the final equivalence classes of corefering event mentions, within each document cluster, event mentions are grouped based on corefering pair overlap. 3.1.5 One-step Bag of Events Approach In the one-step implementation of the approach all possible unique pairs of action mentions from the corpus are used as the starting point for classification. No initial document clustering is performed. For every action mention a sentence template is filled (see examples in table 4 and 5). Also, for every corpus document a document template is filled. Five bag of events features indicating the degree of overlap between documents, from which two active mentions come from, are used for classification. In the one-step approach document features are used by a classifier together with sentence-based features; all in one go. One DT classifier is trained to determine event coreference. Pairs of mentions are classified based on a mix of information from a sentence and from a document. Corefering pairs with overlap are merged into equivalence classes. The one-step classification is implementationwise simpler but it is computationally much more expensive. Ultimately every action mention has to be compared with every other action mention. This is a drawback of the one-step method. On the other hand, it could be of advantage to have different types of NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 52/148 information (sentence- and document-based) available simultaneously to determine event mention coreference. 3.1.6 Corpus For the experiments we used true mentions from the ECB+ corpus (Cybulska and Vossen (2014b)) which is an extended and re-annotated version of the ECB corpus (Bejan and Harabagiu (2010b)). ECB+ is particularly interesting for this experiment because we extended the ECB topics with texts about different event instances but from the same event type (see Cybulska and Vossen (2014)). For example in addition to the earlier mentioned topic of a celebrity checking into a rehab, we added descriptions of another event involving a different celebrity checking into another rehab facility. Likewise, we increased the referential ambiguity for the event mentions. Since the events are similar, we expect that the only way to solve this is through analysis of the event slots. Figure 35 shows some examples of the seminal events represented in ECB+ with different event instances. Figure 35: Overview of seminal events in ECB and ECB+, topics 1-10 For the experiments on event coreference we used a subset of ECB+ annotations (based on a list of 1840 selected sentences), that were additionally reviewed with focus on coreference relations. Table 7 presents information about the data set used for the experiments. We divided the corpus into a training set (topics 1-35) and test set (topics 36-45). 3.1.7 Experimental Set Up The ECB+ texts are available in the XML format. The texts are tokenized, hence no sentence segmentation nor tokenization needed to be done. We POS-tagged (for the purpose of proper verb lemmatization) and lemmatized the corpus sentences. For the experiments we used tools from the Natural Language Toolkit (Bird et al. (2009), NLTK version 2.0.4): NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 53/148 the NLTK’s default POS tagger, Word-Net lemmatizer11 as well as WordNet synset assignment by the NLTK12 . For machine learning experiments we used scikit-learn (Pedregosa et al. (2011)). Table 8: Features grouped into four categories: L-Lemma based, A-Action similarity, Dlocation within Discourse, E-Entity coreference and S-Synset based. Event Slot Action Location Mentions Active mentions Sent. or doc. mentions Sent. or doc mentions Time Sent. or doc mentions Human Participant Sent. or doc mentions NonHuman Participant Sent. or doc mentions Feature Kind Lemma overlap (L) Synset overlap (S) Action similarity (A) Discourse location (D) - document - sentence Lemma overlap (L) Synset overlap (S) Lemma overlap (L) Entity coreference (E) Synset overlap (S) Lemma overlap (L) Entity coreference (E) Synset overlap (S) Lemma overlap (L) Entity coreference (E) Synset overlap (S) Lemma overlap (L) Entity coreference (E) Synset overlap (S) Explanation Numeric feature: overlap %. Numeric: overlap %. Numeric: Leacock and Chodorow. Binary: - the same document or not. - the same sentence or not. Numeric: overlap %. Numeric: overlap %. Numeric: overlap %. Numeric: cosine similarity. Numeric: overlap %. Numeric: overlap %. Numeric: cosine similarity. Numeric: overlap %. Numeric: overlap %. Numeric: cosine similarity. Numeric: overlap %. Numeric: overlap %. Numeric: cosine similarity. Numeric: overlap %. In the experiments different features were assigned values per event slot (see Table 8). The lemma overlap feature (L) expresses a percentage of overlapping lemmas between two instances of an event slot, if instantiated in the sentence or in a document (with the exclusion of stop words). Frequently, one ends up with multiple entity mentions from the same sentence for an action mention (the relation between an action and involved entities is not annotated in ECB+). All entity mentions from the sentence (or a document in case of bag of events features) are considered. There are two features indicating action mentions’ location within discourse (D), specifying if two active mentions come from the same sentence and the same document. Action similarity (A) was calculated for a pair of 11 12 www.nltk.org/modules/nltk/stem/wordnet.html http://nltk.org/ modules/nltk/corpus/reader/wordnet.html NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 54/148 active action mentions using the Leacock and Chodorow measure Leacock and Chodorow (1998). Per entity slot (location, time, human and non-human participant) we checked if there is coreference between entity mentions from the sentence of the two compared actions; we used cosine similarity to express this feature (E). For all five slots a percentage of synset overlap is calculated (S). In case of document templates features referring to active action mentions were disregarded, instead action mentions from a document were considered. All feature values were rounded to the first decimal point. We experimented with a few feature sets, considering per event slot lemma features only (L), or combining them with other features described in Table 8. Before fed to a classifier, missing values were imputed (no normalization was needed for the scikit-learn DT algorithm). All classifiers were trained on an unbalanced number of pairs of examples from the training set. We used grid search with ten fold cross-validation to optimize the hyper-parameters (maximum depth, criterion, minimum samples leafs and split) of the decision-tree algorithm. 3.1.8 Baseline We will look at two baselines: a singleton baseline and a rule-based lemma match baseline. The singleton baseline considers event coreference evaluation scores generated taking into account all action mentions as singletons. In the singleton baseline response there are no “coreference chains” of more than one element. The rule-based lemma baseline generates event coreference clusters based on full overlap between lemma or lemmas of compared event triggers (action slot) from the test set. Table 10 presents baselines’ results in terms of recall (R), precision (P) and F-score (F) by employing the coreference resolution evaluation metrics: MUC (Vilain et al. (1995)), B3 (Bagga and Baldwin (1998)), CEAF (Luo (2005)), BLANC (Recasens and Hovy (2011)), and CoNLL F1 (Pradhan et al. (2011)). When discussing event coreference scores must be noted that some of the commonly used metrics depend on the evaluation data set, with scores going up or down with the number of singleton items in the data Recasens and Hovy (2011). Our singleton baseline gives us zero scores in MUC, which is understandable due to the fact that the MUC measure promotes longer chains. B3 on the other hand seems to give additional points to responses with more singletons, hence the remarkably high scores achieved by the baseline in B3. CEAF and BLANC as well as the CoNLL measures (the latter being an average of MUC, B3 and entity CEAF) give more realistic results. The lemma baseline reaches 62% CoNLL F1. A baseline only considering event triggers, will allow for an interesting comparison with our event template approach, employing event argument features. 3.1.9 Results Table 9 evaluates the final clusters of corefering event mentions produced in the experiments by means of the DT algorithm when employing different features. When considering bag of events classifiers using exclusively lemma features L (row two and three), the two-step approach reached a 1% higher CoNLL F-score than the one-step NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 55/148 approach with document-based lemma features (docL). The one-step method achieved in BLANC a 2% better precision but a 2% lower recall. This is understandable. In a two-step implementation when document clusters are created some precision is lost. In a one-step classification specific sentence information is always available for the classifier hence we see slightly higher precision scores (also in other metrics). The best coreference evaluation scores with the highest CoNLL F-score of 73% and BLANC F of 72% were reached by the two-step bag of events approach with a combination of the DT document classifier using feature set L (document-based hence docL) across five event slots and the DT sentence classifier when employing features LDES (see Table 8 for a description of features). Adding action similarity (A) on top of LDES features in step two, does not make any difference on decision tree classifiers with a maximum depth of 5 using five slot templates. Our best CoNLL F-score of 73% is an 11% improvement over the strong rule based event trigger lemma baseline, and a 34% increase over the singleton baseline. Table 9: Bag of events approach to event coreference resolution, evaluated on the ECB+ in MUC, B3, mention-based CEAF, BLANC and CoNLL F measures. Alg DT DT DT DT Step1 Slot Nr 5 5 2 5 Features docL docL docL docL Alg DT DT DT DT DT DT Step2 Slot FeaNr tures 5 L 5 L+docL 5 L 5 LDES 2 LDES 5 LADES R 61 65 71 71 76 71 MUC P F 76 80 75 75 70 75 68 71 73 73 73 73 R B3 P F CEAF F R 66 68 71 71 74 71 79 83 77 78 68 78 72 75 74 74 71 74 61 64 64 64 61 64 67 69 71 72 74 72 BLANC P F 69 73 71 71 68 71 68 71 71 72 70 72 CoNLL F 70 72 73 73 70 73 Table 10: Baseline results on the ECB+: singleton baseline and lemma match of event triggers evaluated in MUC, B3, mention-based CEAF, BLANC and CoNLL F. Baseline Singleton Baseline Action Lemma Baseline R 0 71 MUC P F 0 0 60 65 R 45 68 B3 P 100 58 F 62 63 CEAF R/P/F 45 51 R 50 65 BLANC P F 50 50 62 63 CoNLL F 39 62 To quantify the contribution of document features, we contrast the results of classifiers using bag of events features with scores achieved when disregarding document features. The results reached with sentence template classification only (without any document features, row one in table 9), give us some insights into the impact of the document features on our experiment. Note that one-step classification without preliminary document template clustering is computationally much more expensive than a two-step approach, which ultimately takes into account much less item pairs thanks to the initial document template clustering. The DT sentence template classifier trained on an unbalanced training NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 56/148 set reaches 70% CoNLL F. This is 8% better than the strong baseline disregarding event arguments, but only 3% less than the two-step bag of events approach and 2% less than the one-step classification with document features. The reason for the relatively small contribution by document features could owe to the fact that in the ECB+ corpus not that many sentences are annotated per text. 1840 sentences are annotated in 982 corpus texts, i.e. 1.87 sentence per text. We expect that the impact of document features would be bigger, if more event descriptions from a discourse unit were taken into account than only the ground truth mentions. We run an additional experiment with the two-step approach in which four entity types were bundled into one entity slot. Locations, times, human and non-human participants were combined into a cumulative entity slot resulting in a simplified two-slot template. When using two-slot templates for both, document and sentence classification on the ECB+ 70% CoNLL F score was reached. This is 3% less than with five-slot templates. Table 11: Best scoring two-step bag of events approach, evaluated in MUC, B3, entitybased CEAF, BLANC and CoNLL F in comparison with related studies. Note that the BOE approach uses gold while related studies system mentions. Approach Data B&H LEE BOE-2 BOE-5 BOE-2 BOE-5 ECB B&H 2010 ECB Lee et al. 2012 ECB annot. ECB+ ECB annot. ECB+ ECB+ ECB+ Model HDp LR DT+DT DT+DT DT+DT DT+DT MUC R 52 63 65 64 76 71 P 90 63 59 52 70 75 B3 F 66 63 62 57 73 73 R 69 63 77 76 74 71 P 96 74 75 68 68 78 CEAF F 80 68 76 72 71 74 F 71 34 72 68 67 71 BLANC R NA 68 66 65 74 72 P NA 79 70 66 68 71 CoNLL F NA 72 67 65 70 72 F NA 55 70 66 70 73 To the best of our knowledge, the only related study using clues coming from discourse structure for event coreference resolution was done by Humphreys et al. (1997) who perform coreference merging between event template structures. Both approaches determine event compatibility within a discourse representation but we achieve that in a different way, with a much more restricted template (five slots only) which in our two-step approach facilitates merging of all event and entity mentions from a text as the starting point. Humphreys et al. consider discourse events and entities for event coreference resolution while operating on the level of mentions, more similar to our one-step approach. They did not report any event coreference evaluation scores. Some of the metrics used to score event coreference resolution are dependent on the number of singleton events in the evaluation data set (Recasens and Hovy, 2011). Hence for the sake of a meaningful comparison it is important to consider similar data sets. The ECB and ECB+ are the only available resources annotated with both: within- and crossdocument event coreference. To the best of our knowledge no baseline has been set yet for event coreference resolution on the ECB+ corpus. So in Table 11 we will also look at results achieved on the ECB corpus which is a subset of ECB+, and so the closest to the data set used in our experiments but capturing less ambiguity of the annotated event types NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 57/148 (Cybulska and Vossen, 2014b). We will focus on the CoNLL F measure that was used for comparison of competing coreference resolution systems in the CoNLL 2011 shared task. The best results of 73% CoNLL F were achieved on the ECB+ by the two-step bag of events approach using five slot event templates (BOE-5 in Table 11). When using twoslot templates we get 3% less CoNLL F on ECB+. For the sake of comparison, we run an additional experiment on the ECB part of the corpus (annotation by Cybulska and Vossen (2014b)). The ECB was used in related work although with different versions of annotation so not entirely comparable. We run two tests, one with the simplified templates considering two slots only: action and entity slot (as annotated in the ECB by Lee et al. (2012)) and one with five-slot templates. The two slot bag of events (BOE-2 ) on the ECB part of the corpus reached comparable results to related works: 70% CoNLL F, while the five-slot template experiment (BOE-5 ) results in 66% CoNLL F. The approach of Lee et al. (2012) (LEE ) using linear regression (LR) reached 55% CoNLL F although on a much more difficult task entailing event extraction as well. The component similarity method of Cybulska and Vossen (2013b) resulted in 70% CoNLL F but on a simpler within topic task (not considered in Table 11). B&H in the table refers to the approach of Bejan and Harabagiu (2010) using hierarchical Dirichlet process (in the table referred to by HDp); for this study no CoNLL F was reported. In the BOE experiments reported in Table 11 we used the two-step approach. During step 1 only (document based) lemma features (docL) were used and for sentence template classification (step 2) LDES features were employed. In the tests with the bag of events approach, ground truth mentions were used. 3.1.10 Conclusion In this section we experimented with two variations of a new bag of events approach to event coreference resolution: a one-step method and a higher scoring two-step heuristic. Instead of performing topic classification before solving coreference between event mentions, as is done in most studies, the two-step bag of events approach first compares document templates created per discourse unit and only after that, does it compare single event mentions and their arguments. In contrast to a heuristic using a topic classifier, that might have problems distinguishing between different instances of the same event type, the bag of events approach facilitates context disambiguation between event mentions from different discourse units. Grouping events depending on compatibility of event context (time, place and participants) on the discourse level, allows one to take advantage of event context information, which is mentioned only once per unit of discourse and consequently is not always available on the sentence level. From the perspective of performance, the robust two-step bag of events approach using a very small feature set, also significantly restricts the number of compared items. Therefore, it has much lower memory requirements than a pairwise approach operating on the mention level. Given that this approach does not consider any syntactic features and that the evaluation data set is only annotated with 1.8 sentences per text, the evaluation results are highly encouraging. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 3.2 58/148 Evaluation of the NewsReader pipeline The ’bag-of-events’ approach described in the previous sections uses the annotations of the event components. By taking the gold annotation, we can more purely evaluate the impact of the approach without any impact of the other processing that may introduce errors. In this section, we describe the performance of the NewsReader pipeline as is on the same ECB+ data set starting from the text. The same tokenized text was processed with the NewsReader pipeline version 3.0 (Agerri et al. (2015)). There were 4 files out of 982 for which the pipeline gave no output. This resulted in 20 events that were not recovered. The starting point for the cross-document coreference is the intra-document coreference layer in NAF. Event coreference sets are generated using the EventCoreference module that makes a distinction between the event types contextualEvent, sourceEvent and grammaticalEvent. For the latter two, no coreference relations are generated within the same document. For contextualEvents, we first group all lemmas into a candidate coreference set, next decide on the dominant sense of the lemma and finally measure the similarity across candidate coreference sets. If the dominant senses of lemmas are sufficiently similar, the candidate sets are merged. Dominant senses are derived from all occurrences of a lemma in a document by cumulating the WSD score of each occurrence. We take those senses with the 80% highest cumulated WSD scores as the dominant senses (program setting: – wsd 0.8). When we compare different lemma-based coreference sets we use these dominant senses to measure similarity in WordNet according to the Leacock-Chodorow method (Leacock and Chodorow (1998)). We use the hypernym relations and the cross-part-of-speech event relations from WordNet to establish similarity (program setting: –wn-lmf wneng30.lmf.xml.xpos). The threshold for similarity was set to 2.0 (program setting –sim 2.0). If different lemmas are considered coreferential, we store the lowest-common-subsumer synset that established the similarity as an external reference in the coreference set. Once all the coreference sets are established (both singletons and multiforms), we add all the hypernym synsets for the dominant senses as external references. For the cross-document evaluation, we merged all the ECB+ CoNLL files from each topic into a single key file, effectively mixing ECB and ECB+ files into a single task. Since ECB+ reflects a systematic referential ambiguity for seminal topic events, we thus create a task that reflects this ambiguity as well. Since ECB has 43 different topics, 43 unique key files were created. Below, we show a fragment from such a key file for topic 1 in which Tara Reid is checking into rehab in the ECB file number 10, whereas Lindsay Lohan is checking into rehab in the ECB+ file number 10. The identifiers in the key file are created using the CROMER tool across the different documents, each annotated in NAF, as explained in Cybulska and Vossen (2014b).13 #begin document (1); 1_10ecb 0 1 Perennial 1_10ecb 0 2 party 1_10ecb 0 3 girl 1_10ecb 0 4 Tara - 13 Note that we reduced multiword phrases such as checked into to the first token only, since the NewsReader system does not mark multiwords as events. Hence in the example below, we removed the idenitifer (132016236402809085484) from token 1 10ecb 0 8 into. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 0 0 0 0 0 0 0 0 0 0 0 0 0 5 Reid 6 checked (132016236402809085484) 7 herself 8 into 9 Promises 10 Treatment 11 Center 12 ,13 her 14 rep 15 told (132016235311629112331) 16 People 17 . - 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 59/148 A friend of the actress told (110088372) People she went (132016236402809085484) to Promises on Tuesday and that her friends and family supported (110088386) her decision (132016236402809085484) . - 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1 1 1 1 1 1 1 30 31 32 33 34 35 36 Lindsay Lohan checks (132015738091639707092) into Betty Ford Center - 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 After skipping out on entering (132015832182464413376) a Newport Beach rehabilitation facility and facing (132015992713150306962) the prospect of arrest for violating (132015992916565253818) her probation (132015993252785693471) ,Lindsay Lohan has checked (132015738091639707092) into the Betty Ford Center to begin (132015992988761097172) a 90 -day court -mandated stay (132015736700251185985) in her - NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 3 3 3 3 83 84 85 86 60/148 reckless driving (132015732766174435569) conviction (132015993650409892802) . - The NAF2SEM module from NewsReader was used to generate SEM-RDF files for each topic by processing all the NAF files within that topic. From the SEM-RDF files, we extract unique numerical identifiers from the event instances and insert them in a CoNLL response file for the tokens that form the mentions of the event. Below we show the same fragment for topic 1 with the output from the NewsReader system added to the tokens: #begin document (1); 1_10ecb 0 1 Perennial 1_10ecb 0 2 party 1_10ecb 0 3 girl 1_10ecb 0 4 Tara 1_10ecb 0 5 Reid 1_10ecb 0 6 checked (139) 1_10ecb 0 7 herself 1_10ecb 0 8 into 1_10ecb 0 9 Promises 1_10ecb 0 10 Treatment 1_10ecb 0 11 Center 1_10ecb 0 12 ,1_10ecb 0 13 her 1_10ecb 0 14 rep 1_10ecb 0 15 told (139) 1_10ecb 0 16 People 1_10ecb 0 17 . 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 1_10ecb 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 A friend of the actress told (139) People she went to Promises (445) on Tuesday and that her friends and family supported (239) her decision (18) . - 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1 1 1 1 1 1 1 30 31 32 33 34 35 36 Lindsay Lohan checks (153) into Betty Ford Center - 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 After skipping (399) out on entering (173) a Newport Beach rehabilitation facility and facing (499) the prospect (265) of arrest (42) for violating (138) her - NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 1_10ecbplus 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 61/148 probation (367) ,Lindsay Lohan has checked (153) into the Betty Ford Center to begin (464) a 90 -day court -mandated stay (407) in her reckless driving conviction (10) . - We used the latest version (v8.01) of the official CoNLL scorer (Luo et al., 2014) to compare the response output with the key data. We generate the BLANC score for each topic and then macro average the results across all topics. In the next subsection, we describe the results for the NewsReader pipeline without any adaptation. We vary the parameters in the NAF2SEM module to see the impact on the performance. In subsection 3.2.2, we measure the result of augmenting the event detection in the NewsReader pipeline output. 3.2.1 NewsReader output We first applied the NAF2SEM system on the NewsReader pipeline output as is. Since ECB+ is an experimental data set with systematic but limited ambiguity (each seminal event mention within a topic can have two referential values), we ran the NAF2SEM system by comparing all CompositeEvents within each topic, using temporal anchoring as an additional parameter. The CompositeEvent RDF data consist of: • Event action data – the WordNet synsets and ontology types associated to all event mentions based on the intra-document event coreference sets – the labels used to make reference to the event action and the most frequent label as the preferred label – all the mentions in the documents • Participant data – the URI of the participant, either derived from DBpedia or created from the phrase as an entity or non-entity NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 62/148 – the labels used to make reference to the entity and the most frequent label as the preferred label – all the mentions in the documents • Time data – the URI of the time expression – the labels used to make reference to the time – all the mentions in the documents – the owl-time object that represents the normalised ISO value for the time expression, with a specification of at least the year and possibly the month and day • SEM relations – Any specific PropBank, FrameNet or ESO role between event actions and participants – Any sem:hasTime relations between event actions and time expressions When comparing the CompositeEvents across documents, we can use any of the above features to compare events. We experimented with a number of parameters to measure their impact on the performance of the system: Event matching Event mentions across documents need to match proportionally in terms of associated synsets, lemmas or combinations of these Temporal anchoring No temporal constraint, same year, same month or same day Participant match No participant constraint, at least one participant needs to match, either through the URI or through the preferred label For Event matching, we can apply different strategies: compare the WordNet synsets (program parameter: –match-type ili), lemmas of mentions (program parameter: –matchtype lemma) or a combination of synsets and lemmas (program parameter: –match-type ililemma). For using synsets associated with the coreference sets, we can choose the dominant senses, the lowest-common-subsumers (program parameter: –lcs) and the hypernyms of the dominant senses ((program parameter: –hypers). For the Temporal anchoring, we can set the granularity to years, months and days. This means that, depending on the amount of detail of the time anchoring, events need to have matching time according to these settings. When we leave out the –time value, the time anchoring is not considered. Finally for participants, we can define the precise roles to be matched (PropBank, VerbNet, FrameNet, ESO), any role to be matched or none. To match, the program NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 63/148 considers the role label and the role object separately. If the label is specified (e.g. program parameter: –roles a0,a1,a2), we need to match at least one participant object with that label. If no label is specified (program parameter: –roles anyrole), at least one participant needs to match regardless of the role label. Matching the objects is done on the basis of the URI. If the URIs do not match, we check of the preferred label of a participant is among the labels of the other participant. Event coreference depends on the detection of the events. The coreference scorer (Luo et al., 2014) also gives the scores for the detection of the mentions. Without adaptation, the event mention recall for the NewsReader output is 72,66% and the precision is 66,29% (F1 69,01%). For the results below, we cannot expect the coreference to perform above the results for the event mention detection. The event mention detection defines the upper bound. We first experimented with the proportion of WordNet synsets (program parameter: –match-type ili, –hypers, –lcs) that need to match across two CompositeEvents. Using these settings, events only match if they share some of the meanings as scored by the WSD system or through direct hyernyms or lowest-common-subsumers. Events are not matched through their lemma We set the time constraint to match the year and month and the participant have at least one match regardless of the role. We varied the proportion of WordNet synsets to match in steps of 10 between 0% to 100%. The threshold defines minimal the proportion of synsets that needs to be associated for both compared events to be merged. Likewise, events with excessive synsets cannot absorb events with few synsets. The results are shown in Table 12 and in Figure 36. Table 12: BLANC refererence results macro averaged over ECB+ topics in terms of recall (R), precision (P) and F1 (F) for NewsReader output with different proportions of WordNet synsets to match: S=only synset matches, SL= Synsets matches if synsets and lemma matches if no synsets associated, L=lemmas only. Different columns represent proportions in steps of 10% from 1% to 100%. R-SL R-S R-L P-SL P-S P-L F-SL F-S F-L 1 36.07% 36.04% 27.60% 30.57% 30.95% 46.12% 30.42% 30.78% 24.77% 10 36.14% 35.97% 20 35.98% 35.77% 37.60% 38.58% 39.40% 39.96% 34.22% 34.31% 34.50% 34.49% 30 35.74% 35.68% 27.52% 41.25% 41.35% 45.98% 34.79% 34.72% 24.63% 40 35.59% 35.40% 50 35.42% 35.17% 60 35.12% 35.09% 70 34.84% 34.82% 80 34.76% 34.70% 90 34.53% 34.41% 41.64% 41.72% 42.13% 42.07% 42.13% 42.18% 42.37% 42.44% 42.54% 42.54% 42.58% 42.55% 34.71% 34.56% 34.72% 34.47% 34.43% 34.40% 34.24% 34.22% 34.19% 34.13% 34.03% 33.88% 100 34.48% 34.33% 27.49% 42.83% 42.81% 45.61% 33.99% 33.82% 24.57% The highest recall is obtained using first synsets and lemmas in addition with 10% overlap (R-SL=36.14%). We can see that recall drops when more overlap is required. We see the opposite for precision but the highest precision is obtained using solely the lemmas, where 1% overlap is sufficient (R-L=46.12%).14 The highest f-measure is obtained using 14 We did not test all proportions of overlap for lemmas because there is 1.7 lemma per coreference set NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 64/148 Figure 36: Impact of more strict WordNet synset matching on the macro average BLANC for recall (R), precision (P) and F1 for NewsReader output first synsets and lemmas in addition with 30% overlap (F-SL=34.79%). Differences are however very small. We used the latter settings in the further experiments described below. Next, we varied the participant constraints and their roles, whether or not hypernyms and lowest-common-subsumers can be used to match actions and the temporal constraints. We used the following combinations of properties, where we kept the settings for WordNet synsets and lemmas to match for 30% proportionally (which gave the best f-measure so far): AR-H-L-M AR= a single participant match and role is not considered, H=Hypernyms, L=Lowest-common-subsumer, M = Month AR—M AR= a single participant match and role is not considered, Hypernyms and Lowest-common-subsumer are not considered, HM = Month AR–L-M AR= a single participant match and role is not considered, Hypernyms are not considered, L=Lowest-common-subsumer, M = Month AR-H–M AR= a single participant match and role is not considered, H=Hypernyms, Lowest-common-subsumer are not considered, M = Month AR-H-L- AR= a single participant match and role is not considered, H=Hypernyms, L=Lowest-common-subsumer, time is not considered on average. It makes little difference to have 1% or 100% overlap. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 65/148 AR-H-L-Y AR= a single participant match and role is not considered, H=Hypernyms, L=Lowest-common-subsumer, Y = Year AR-H-L-D AR= a single participant match and role is not considered, H=Hypernyms, L=Lowest-common-subsumer, D = Day -H-L-M participants are not considered, H=Hypernyms, L=Lowest-common-subsumer, M = Month A0-H-L-M A0= a single participant match and role should be A0, H=Hypernyms, L=Lowestcommon-subsumer, M = Month A1-H-L-M A1= a single participant match and role should be A1, H=Hypernyms, L=Lowestcommon-subsumer, M = Month A2-H-L-M A2= a single participant match and role should be A2, H=Hypernyms, L=Lowestcommon-subsumer, M = Month A0A1-H-L-M A0A1= a participant and role match for A0 and A1, H=Hypernyms, L=Lowest-common-subsumer, M = Month A0A2-H-L-M A0A2= a participant and role match for A0 and A2, H=Hypernyms, L=Lowest-common-subsumer, M = Month A1A2-H-L-M A1A2= a participant abd role match for A1 and A2, H=Hypernyms, L=Lowest-common-subsumer, M = Month A0A1A2-H-L-M A0A1A2= a participant and role match for A0, A1 and A2, H=Hypernyms, L=Lowest-common-subsumer, M = Month In Table 13, we first show the impact of using hypernyms and lowest-common-subsumbers for matching event actions. We kept the participant constraint stable to require a single participant to match regardless of the role. The time is first set to month matching. In the second part of the table, we maintained the participant constraint and also fixed the use of hypernyms and lowest-common-subsumers. In this case, we varied the time constraint to no time constraint, year, month and day. In all cases, we use WordNet hypernyms and lemmas for event action matching with a proportion of 30% (SL30). We first of all observe that the differences are small. The differences in recall and f-measure are not significant. Most notably, adding the day as time-constraint give the highest precision: 45.12% but als the lowest recall and f-measure. In Table 13, we kept the standard setting for hypernyms and lowest-common-subsumer with the time constraint to month matching but now varied the specification of the roles for which there should be a participant match. In principle, we can test roles from PropBank, FrameNet and ESO. However since the PropBank roles are more general and are always given, we restricted the testing to the most important PropBank roles A0, A1 and A2. We also tested combination of roles: A0A1, A0A2 and A1A2. The first column (NR) represents no role restriction. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 66/148 Table 13: BLANC refererence results macro averaged over ECB+ topics in terms of recall (R-SL30), precision (P-SL30) and F1 (F-SL30). AR is stable across the results, meaning that a single participant in any role needs to match. We varied the hypernyms (H) and lowest-common-subsumer (L) for action matches and the time constraints: no time constraint (NT), year (Y), month (M) and day (D) R-SL30 P-SL30 F-SL30 AR—M 35.77% 40.78% 34.68% AR–L-M 35.79% 41.00% 34.74% AR-H–M 35.72% 40.98% 34.71% AR-H-L-NT 35.78% 41.26% 34.84% AR-H-L-Y 35.74% 41.25% 34.80% AR-H-L-M 35.74% 41.25% 34.79% AR-H-L-D 30.40% 45.12% 29.35% Again in all cases, we use WordNet hypernyms and lemmas for event action matching with a proportion of 30% (SL30). Table 14: BLANC refererence results macro averaged over ECB+ topics in terms of recall (R-SL30), precision (P-SL30) and F1 (F-SL30). The hypernyms (H), lowest-commonsubsumer (L) and time constraint month (M) are kept stable. We varied the roleparticipant constraints: NR=no constraint, A0 role participant should match, A1 should match, A2 should match, A0 and A1 should match, A0 and A2 should match, A1 and A2 should match R-SL30 P-SL30 F-SL30 NR-H-L-M 41.54% 36.21% 37.58% A0-H-L-M 30.58% 42.57% 29.20% A1-H-L-M 31.53% 46.33% 30.76% A2-H-L-M 28.48% 43.62% 26.06% A0A1-H-L-M 28.66% 48.74% 26.60% A0A2-H-L-M 27.48% 46.09% 24.56% A1A2-H-L-M 27.48% 46.09% 24.56% Using no constraints on the participant and their role give the highest recall (41.54%) and f-measure (37.58%) so far. The results are even higher than for a single participant in any role (recall 35.74% and f-measure 34.79%), although the precision is lower. We can observe that adding more specific role constraints lowers the recall and increases the precision, with both prime participants required (A0A1) giving highest precision so far: 48.75%. Note that such a constraint can only be applied to semantic role structures where both participants have been detected in the sentence. There are many cases where either the A0 or A1 is not expressed or not recovered. Concluding In general, we can conclude that ECB+ is probably not rich enough to see any impact of constraints at a very specific level. Since the referential ambiguity is restricted to two seminal event, differentiating between them can be done using more global features such as any participant or just the year rather than the precise role and day. We expect that in realistic news data sets with thousands of sources reporting on the similar events, the details are needed to make a more fine-grained comparison. Nevertheless, we remain dependent on the quality of the NLP software to detect all these details correctly and sufficiently. We can also see that adding constraints increases the precision but that the low recall remains a problem. Finally, it is important to realise that over 95% of all event mentions are not coreferential in ECB+. Detecting coreference relations, even in an artificial data set such as ECB+, is a very delicate task. Since BLANC averages the results of noreference (singleton event mentions without a coreference link to any other NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 67/148 event mention) and coreference relations, any drastic approach on establish coreference relations will be penalised by the noreference results. 3.2.2 Maximizing the event detection To improve the performance of the event detection in terms of both precision and recall, we developed a Conditional Random Fields (CRF) classifier, Event Detection system, inspired by the TIPSem system (Llorens et al., 2010), a state-of-the-art system from the SemEval2010 TempEval task (Verhagen et al., 2010). We implemented the classifier using the SemEval 2013 - TempEval 3 data (UzZaman et al., 2013b), on which it performs with F1 scores of 82.6% and 85.9% using gold and silver training data respectively. We used this re-implemention to either confirm or disqualify predicates that were detected by the NewsReader SRL. The classifier also adds new events to the SRL output not detected by NewsReader. Note that for the latter, we only obtain the predicates and not the roles. When creating the intra-document coreference sets, we only consider predicates from the SRL that were not disqualified (status=”false”). We applied two versions of the Event Detection system to the NAF files to augment the SRL layer. One was trained with the gold data (EDg(old)) and one with the silver data (EDs(ilver)). In addition, we restricted the augmentation to those predicates that do not have an event class from VerbNet, FrameNet or ESO. We assumed that correct events are likely to have some typing from these resources through the PredicateMatrix, whereas wrong events are expected to have no typing. The Event detection systems that skip events with event classes are called EDg(old)EC and EDs(ilver)EC respectively. Finally, we report on event mention detection if we would consider only those tokens that we annotated as events in the key data (NWR-key in the last column). In this case, we can assume maximum precision of the predicates detected in relation to the real recall. In Table 15, the first column (NWR) shows the results for event mention detection using the NewsReader system. The other settings are ili30, hypers, lcs, month and anyrole Table 15: Macro averaged Mention identification for ECB+ topics. NWR=NewsReader pipeline v3.0 without adaptation, EDg(old)=NWR augmented with EventDetection trained with gold data, EDg(old)EC= same as EDg(old) but skipping predicates with an Event class, EDs(ilver)= NWR augmented with EventDetection trained with silver data, EDs(ilver)EC= same as EDs but skipping predicates with an Event class. NWR NWR-key EDg EDgEC EDs EDsEC recall 72.66% 72.66% 61.59% 72.64% 53.34% 72.43% precision 66.29% 99.91% 88.23% 76.25% 92.10% 76.37% f1 69.01% 83.80% 72.15% 74.01% 67.12% 73.94% First of all, we see in Table 15 that the Event Detection variants have very high precision (EDg(old) 88.23% and EDs(ilver) 92.10%) compared to NewsReader. Both variants score only less than 9 points lower than the NWR-key with a maximum precision of 99.91%. In NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 68/148 terms of recall, however, both variants score considerably lower: 61.59% and 34% respectively. The F1 scores are consequently not very different from NewsReader and significantly lower than for the SemEval-2013 task. The latter is not so surprising since it trained on a data set from the same task and thus annotated in a similar way. We see that skipping disqualifying predicates with Event classes, almost fully recovers the loss in recall while precision drops less with about 15 points. Maintaining predicates with event classes thus provides the highest F-measures (72.15% and 73.94% respectively), only 10 points lower when considering the key event tokens (83.80%). At the end of this section, we provide tables with the most frequent missed and invented predicates according to EDs(ilver)EC and also the full list of hapaxes, i.e. events only missed and invented once. The most frequently missed events (Table 16) are missed mostly due to their part-of-speech (nouns, adjectives, prepositions and names): dead, earthquake, Watergate, magnitude, guilty, according. Some tokenization errors are also frequent: Shooting, Checks were not down-cased and 1 is in all 22 cases actually 6.1 which was split into separate tokens and annotated as an event indicating the magnitude of the earthquake by the annotators. Especially Table 17 listing predicates missed once as an event makes clear that downcasing and lemmatising may solve many cases. Dealing with parts-of-speech other than verbs more extensively and more proper tokenization will solve the majority of the missed events. Tables 18 and 19 show the predicates for the invented events occurring more than once and only once respectively. In this case, the solution is less clear. Some of the more frequent predicates such as murder, patch, driving all seem correct events but have not been annotated for some reason. Others, such as mother, official, home, police clearly are unlikely to be events regardless off the context. Finally, store, camp, administration are ambiguous. Inspecting cases such as mother and police show that they appear to have eventual readings that were falsely assigned to these mentions. A better filtering of nominal predicates that do not have any eventual meaning (the second group) seems beneficial. The detection of event mentions defines a natural upper bound for the event detection. In Table 20, we give the reference results when maximizing the event detection compared to the standard NewsReader output and using different settings to maximize the recall, the precision and the f-measure. Results are reported in all different measures: MUC, BCUB, CEAFe and BLANC, while the F measure of the first 3 scores are averaged as CoNLL F. The table is divided into 3 sets of rows with different settings for running the system: ARM at least one participant should match regardless of the role (any role), time anchor match at granularity month and action match with phrases and concepts should overlap with 30%. mR maximizes recall: no matching of participants and time is required and action matches with phrases and concepts is set to 1%. mP maximizes precision: two participants should match with PropBank roles A0 and A1, time anchors should match with granularity of the day and actions should match 100% in terms of concepts and phrases associated. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 69/148 Table 16: Predicates missed more than once by NewsReader extended with EventDetection (silver) and Event class filter as events in ECB+ dead earthquake according Watergate fire magnitude guilty injured 1 Oscars quake security deal DUI party heart shooting playoff Shooting degree it murder playoffs pregnant record death first job sexual Checks scoring arson emergency drunken game piracy riots career business downtime drunk market outage problem role Run tsunami WWDC According Charged 42 37 35 30 28 27 23 23 22 21 20 18 17 17 17 16 16 14 14 13 13 13 13 13 13 12 12 12 12 11 11 10 10 9 9 9 9 8 7 7 7 7 7 7 7 7 7 7 6 6 double first-degree IR merger natural problems rehab went win acquisition be design Injured It Killed list Macworld MVC operation Oscar polygamy second-degree sequel stint suicide traffic Trial trial Valley accident basis blaze campaign Convicted custody damage Dead earthquakes fatally Fire going Murder news operations Playoff prison refresh rounds Sequel swimming NewsReader: ICT-316404 6 6 6 6 6 6 6 6 6 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 temblor top treatment triggered which ”spiritual Acquisition affected air arraignment bank Charges Clinch Conference congestive crash Cut cut damaging data dies DOUBLE Earthquake event go gone Guilty health hit Industry issues Magnitude manslaughter matters Missouri mortar move new NFL Nominee one Pregnant privacy quarterfinals Rehab safe Science Seacom season senior 4 4 4 4 4 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 SET striking technology that touchdown TRIAL Win worth 7.5 7.6 abortion Accident AFC announcement ARRESTED Arson attacks Attorney attorney Availability availability basket battery Beat bigamy Bombing Bombs bombs Burns Business California changes chase clashes clinch coma communications Consulting consulting Convicts crimes critical crossfire cuts Damaged deaths definitive democracy Direct done 3 3 3 3 3 3 3 3 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 door DWI Endangered energy engineering Fall fatal Financial financial Found Francisco furious games GUILTY Heist heist hijack history Hit Host Hurt incident Indonesia injures injury interception Internet journalism Killing landslides lawless lawyer leaner learning Leaves MAKE more musicals offer Overturn panic Placed policy Polygamy position pricing program public purchase quakes 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 Raid reasons rehires reigning remote repairs Reports reports rescue Returns Riot rioting San second Services sex Sexual sexually Shot smash specializes speculation Stolen strategy suspicion takeover telecommunications tensions unpatched unrest Unveils Victory vigil violence violent what February 1, 2016 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 Event Narrative Module, version 3 70/148 Table 17: Predicates missed once by NewsReader extended with EventDetection (silver) and Event class filter as events in ECB+ Riot ’This ’murder” #Oscars ’Knows ’bombs 2004 ACCIDENT ACCUSED ARRESTS Accuses Acquire Acquiring Affects Aftershocks Alzheimer Anger Arrest Arrested Attacks Awards BEAT BLAZE Begins Bombed Bumped CHARGES CLINCH Call Caught Celebrates Cloud Congrats Consultations Continue Cooling County Cruises DEAD DIRECT DNA Deal Death Defensive Dies Drunk ESCAPE Emergency Engineers Enraged Escape Expand Expo FALLOUT FIRE FIRED Fatal Feigned Fired Flagship Follow Football Game Goes Handcuffs Hire I.R Insanity Instated Interdicts Investigation Investment Jeopardy Jolts Lake Launches Lead Levels MIA MURDER MacWorld Make Marriage Musical NEGOTIATIONS Named Negotiation Negotiations New News Nominates Nomination OFFER Operation Oscars PLAYOFF PLAYOFFS POLYGAMY Pending Picked Plus Powder Pre Preorders Protection Protections Protests Pulls Quake REACH REMAINS Rampage Reasons Record Recorded Releases Rescue Restored Riots Rumors Rushes SAYSSEPTEMBER SEQUEL STATEMENT Sacked Semis Server Shelter Shoots Six Spree Strategy Strike Strikes Suspicious TOP Takes Testimony That Undisclosed Vigil Voting Vying WWDC12 War Winter Wounded ablaze access accidents achievement activities additional adultery affair aftershocks aid alert anger announcements any are armed arrest arrival artery assault attack attempted attractiveness available balloting banking barefoot basketball battleground behind betrayal bid bids NewsReader: ICT-316404 blow bout brak breadth break built-in bust cable-news casting cause ceasefire celebration chance chaotic cheaper checking chip chip-making circumstances code cold communication compared complete computer computing conference connectivity coronary count counts coverup coveted crazy credit credits culture dangerous daunting debut delighted die die’ disappointing disaster disguised dismissal disrupts do domestic double-team down drama drinkdrinking drive driving due effectiveness efficient eighth elections emotion equal era escalate escort experiences extinction failures fall famed feedback felony fighting finale fir fixes flagship flurry foils footing free frustrating fuels fun game-winning gaminggeothermal gettingready good graphics great green guide gunplay hacked hacking harbors has have health-care healthcare help herself home homicide homicides hostage hundreds hungry hurt impending implications incarnation indecency infected inferno ink insanity integration intoxicated investigation is isolation jail jumper keyboard keynote large largest latest lay-offs lead leading leaves lies life little-used living loan long longer lose lunacy made making manhunt mark markets matter measuring media-oriented menace mental microservers misdemeanor mixed modeling money murders musicals” needs negotiation negotiations neutrality next nice nine nominee normal office official opportunity outperform outraged overdue overturns payoffs pending percent pick-six pirate place playoffs playing pneumonia polygamous postpower practice pre prediction preorder presidency press profile protections published raids rare re-arrest reason recession recorded repair repercussions reserve resolve return returned returns review riot ripening robbery running sacks saga scenario scheduled school scientific screen secret sector seismic semifinals series service share shelter shocking shot show show’s significant situation sixth slain slow snafus snowballs sobriety sobs spate spending spiritual spiritual’ star stardom status stillness stopgap straight-talking stretch subject supported supremacy surgery-enhanced swoop system systems tackled talks task tech telecom telecoms teleconference telephone temblors terrestrial thwarts time time-travelling tour trade tradition transaction tremor turned underage underway underwent undisclosed update upgrade upping use vacant verdict vetting vigils war way weather weed wellness wet wheel whereabouts willing winter woes word worst years February 1, 2016 Event Narrative Module, version 3 71/148 Table 18: Predicates invented and occurring more than once by NewsReader extended with EventDetection (silver) and Event class filter as events in ECB+ mother murder patch store camp administration driving official home police star had pirates running season including run have host update left officials services assault attack head player team time building say source suspicion ’s coast director following life sequel show sources support users 28 26 23 20 19 17 16 16 15 14 14 13 13 13 13 12 12 11 11 11 10 10 10 9 9 9 9 9 9 8 8 8 8 7 7 7 7 7 7 7 7 7 7 branch causes center climber figure made portfolio receiver said seed sentences shot tackle age cables came coach death details employee estimated fix forces has part place Promises route shooting spot workers aged agent became belts berth bombs choice come contract denied deputy disclosed NewsReader: ICT-316404 6 6 6 6 6 6 6 6 6 6 6 6 6 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 4 4 4 4 4 4 4 4 4 4 4 4 end endangered failure group hit knee maker management manager operator products report reports robbers seems strike times tournament Catching affect aid appears authorities believed border boss Breaking candidate case caused centers committed connection convicted crimes date defenders execution factory film get got groups 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 help history image judge law leading lineup looks manufacturer name pass planned point processors refused reporter residents rule seeded service statement suspect term Voters warship weapons players presumed processor projects quake questions received record referee registration rest resulting reveal rival routes rules saw 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 sentence says section seen standings start statements steps stuck style suffered surrounding suspected suspects take target teams telecast terms territory toll trial trustee type unit used want watches waters wave where wholesaler worker parked partners parts pick pin ACCUSED actor agreement analyst assaulting 3 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 attackers attempt back backed began boats brazen carried cause chance charges couple crown deal display Employees employees evidence face failed feed Fix found generation government held helped hired homes house information leaving link marriages meaning measuring minister model nickname nominee number offender offerings opening operators 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 February 1, 2016 Event Narrative Module, version 3 72/148 Table 19: Predicates invented and occurring only once by NewsReader extended with EventDetection (silver) and Event class filter as events in ECB+ ‘undeserved accepted accused address advantage advocates aim aimed alive alleged analysts answer appeared armed arrest asked assurances attempted audience audiences backs band Based based battle become belt BERTH blown bomb born box breach break BREAKING bug call camps Case centred ceremony challenger change charge climbers close Coach coaching combination coming comment commented commit communications competition complications computing condition conditions conference confirmed consulting contain contention continuing control convention conversations convict Counts counts crews criminals critics cure daring data decision declined declining defender defending defense deficiency delays denying devastated did didn’t disclose disputed documents dodging duty earnings edge efforts employed employers ending entrance eWEEK excess exercise expected exploit extended fact fading father favorite feat feeding feuds finds fire flag followed followers Following NewsReader: ICT-316404 force gave given going grab growing guard guards guilt hand happened having he hold hole holiday husband impact indicated inmate input intended investigation issued killer kittens last launched leave leg let level live located look loss lost loved making managed managing manufacturers margin mark matter mean meant meetings memory merged mermaid moving murderer named names need network numbers occurred offense offer Officials onslaught operations opinion opposition Order order ordered organizers owner page park parking participants Parts passes past patrolled perform photographs pioneer pipeline plagued planning play portables portions poses present price priority prisoner prisoners producers product proposed prosecution prospect protected protection protesters provider providers put quarter range ranging reach reaction recommendation refugee related relations remain remained remaining repaired representative required researcher reserve resident resigning result Results results return revamped revenue review revolution revolving reward rewards risks rivals Robbers Ruling Running salesman saying scoring searching secret security see semifinal serve serves set setting shock shots sign signals signed skipping slide sorts sounds spasm specify speculate spoke spree stabbing stages stand standard standards stars started steering stock stockholders stop stores students success supplier survivors Take tapes tell tenant test tests thieves threatened throw tip total touched tourist touting transcripts transition treatment tribute tried try trying turned types user value valued vessel vessels view viewers violence visiting voters Wanted warehouse warming warning wasn’t way ways weapon weighing when whips winner wiretapped worry February 1, 2016 Event Narrative Module, version 3 73/148 Within each set of rows, the first row (kNWR) shows the results when we only use the key event annotations as markables for the system response. That means that we maximize the precision but not the recall. We thus expect higher results due to more precision. We can compare the real results against these maximized results, where NWR represents the NewsReader results without any adaptation. In case of EDg(old) and EDs(ilver), events are detected by CRF Event Detection system trained with gold data and silver data respectively. In case of EDg(old)EC and EDs(ilver)EC, the system only disqualifies predicates as events that do not have an event ontology type. Each of these are then combined with the above settings ARM, mR and mP. Best scores of the true systems per metric are in bold. Table 20: Reference results macro averaged over ECB+ topics with different options for event detection. kNWR=NewsReader event detection without invented mentions, maximizing precision, NWR=NewsReader pipeline v3.0 without adaptation, EDg=NWR augmented with EventDetection trained with gold data, EDgEC= same as EDg but skipping predicates with an Event class, EDs= NWR augmented with EventDetection trained with silver data, EDsEC= same as EDs but skipping predicates with an Event class. ARM= standard setting one participant in any role (AR), time month match and action concept and phrase match 30%, mR= maximizing recall by no constraints on participant match and time, action concept and phrase match 1%, mP=maximizing precision by participant roles A0A1, time day match and action concept and phrase match is set to 100%. kNWR-ARM NWR-ARM EDg-ARM EDgEC-ARM EDs-ARM EDsEC-ARM kNWR-mR NWR-mR EDg-mR EDgEC-mR EDs-mR EDsEC-mR kNWR-mP NWR-mP EDg-mP EDgEC-mP EDs-mP EDsEC-mP R 33.55 33.55 23.54 33.10 23.26 33.03 53.77 53.77 38.94 54.31 38.16 53.98 11.09 11.09 8.15 11.35 7.96 11.15 MUC P 74.63 53.78 69.78 59.78 70.54 59.91 68.75 48.59 63.47 54.60 64.15 54.58 70.31 54.42 62.77 57.45 63.02 57.97 F 45.54 40.64 34.45 41.92 34.12 41.91 59.73 50.54 47.58 53.84 47.10 53.65 18.77 18.02 14.12 18.54 13.82 18.31 R 41.10 41.10 28.49 39.98 28.63 40.15 52.05 52.05 36.26 51.83 36.01 51.75 31.59 31.59 22.34 30.99 22.45 31.09 BCUB P 85.21 55.68 78.50 63.94 79.44 64.10 62.70 39.68 58.55 45.27 59.31 45.35 95.93 63.18 87.09 72.52 88.11 72.65 F 54.83 46.73 41.25 48.42 41.55 48.61 56.14 44.55 44.11 47.62 44.18 47.62 46.97 41.44 35.19 42.75 35.40 42.85 R 60.59 60.23 43.61 58.14 43.94 58.38 45.62 44.41 33.53 42.71 34.01 42.90 62.71 62.54 44.47 60.60 44.96 60.85 CEAFe P 52.09 32.73 43.69 36.64 45.56 37.02 69.41 40.27 57.25 47.29 60.02 47.66 39.20 25.31 34.29 28.60 35.58 28.79 F 55.12 41.82 42.78 44.33 43.87 44.66 54.26 41.51 41.57 44.19 42.70 44.44 47.47 35.52 37.83 38.31 38.91 38.53 R 34.75 34.75 19.90 34.24 19.26 34.14 41.44 41.44 25.24 41.95 24.28 41.58 27.73 27.73 15.58 27.72 15.00 27.56 BLANC P 77.69 42.64 68.97 51.31 70.22 51.85 67.59 34.96 59.73 43.22 60.62 43.23 81.31 47.35 69.13 55.17 70.39 55.56 F 45.40 34.05 28.96 37.21 28.40 37.27 49.30 36.89 33.77 41.20 33.05 40.98 35.64 25.03 22.42 28.58 21.95 28.53 CoNLL F 51.83 43.06 39.49 44.89 39.85 45.06 56.71 45.53 44.42 48.55 44.66 48.57 37.73 31.66 29.04 33.20 29.38 33.23 We first discuss the ARM output which is supposed to give the most balanced results for precision and recall and thus the highest F measure. EDs(ilver) and EDs(ilver)EC in most cases have the highest score. The EDg(old) and EDg(old)EC results are slightly lower. This is in line with the difference in event mention detection observed earlier. The best CoNLL F score of 45.06% is obtained by EDs(ilver)EC, which is 5 points less than the kNWR version, with maximum precision. Overall the best scores are about 5 points lower than the maximum precision scores of kNWR. BLANC F scores are lower than CoNLL. This is mainly due to the low recall. The best precision of BLANC is 70.22% (EDs(ilver)), while the kNWR precision is 77.69%. These scores are high and comparable to the stateof-the-art. This shows that most improvement can be expected from improving the recall and especially the recall in the event detection. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 74/148 If we look at the results for maximizing recall (mR), we see that in most cases the recall is higher and the precision is lower, while when maximizing the precision (mP), the precision is significantly higher and the recall is lower. The exception is CEAFe where recall and precision are exactly the reversed. Ignoring the CEAFe results, we see that the maximum recall is 54.31% MUC for EDg(old)EC-mR and the maximum precision is 88.11% BCUB for EDs(ilver). As for ARM, most results are about 5 points below the key results kNWR. Highest F measures are obtained with maximized recall for EDs(ilver)EC (40.98% F-BLANC and 48.57% F-CoNLL). In the state-of-the-art literature, cross-document coreference is not only tested across documents within the same topic but also across the whole data set. To compare our results with the state-of-the-art, we abandoned the topic structure and ran the NAF2SEM program on all the 982 ECB+ files processed by NewsReader. This results in a single RDF file after comparing all events in the data set with each other. We compare our results with Yang et al. (2015), who report best results on ECB+ and compare their results to other systems that have so far only been tested on ECB and not on ECB+. Yang et al use a distance-dependent Chinese Restaurant Process (DDCRP (Blei and Frazier, 2011)), which is an infinite clustering model that can account for data dependencies. They define a hierarchical variant (HDDCRP) in which they first cluster event mentions and data within a document and next cluster the within document clusters across documents. Their hierarchical strategy is similar to our CompositeEvent approach, in the sense that event data can be scattered over multiple sentences in a document and needs to be gathered first. Our approach differs in that we use a semantic representation to capture all event properties and do a logical comparison, while Yang et al and all the other methods they report on are based on machine learning methods (both unsupervised clustering and supervised mention based comparison). Yang et al test their system on topics 23-43 while they used topics 1-20 as training data and topics 21-23 as development set. They do not report on topics 44 and 45. To compare our results with theirs, we also used topics 23-43 for testing. Since our system is fully unsupervised for the task itself, the training and development sets are irrelevant. In Table 21, we give the NewsReader results using the ARM settings that give the best F-measure. Table 22 is an exact copy of the results as reported by Yang et al. (2015). They follow a machine-learning approach to event-coreference, exploiting both clustering techniques and supervised techniques with rich feature sets. They also implemented a number of other state-of-the-art systems that use variations on clustering or supervised learning and applied them to the same data set within ECB+: LEMMA a heuristic method that groups all event mentions, either within or across documents, which have the same lemmatized head word. AGGLOMERATIVE a supervised clustering method for within-document event coreference following Chen and Ji (2009b). HDP-LEX an unsupervised Bayesian clustering model for within- and cross-document event coreference (Bejan and Harabagiu, 2010a). It is a hierarchical Dirichlet process NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 75/148 (HDP) model with the likelihood of all the lemmatized words observed in the event mentions. DDCRP a Distance-dependent Chinese Restaurant Process model that ignores document boundaries. HDDCRP* a variant of the proposed HDDCRP that only incorporates the withindocument dependencies but not the cross-document dependencies. HDDCRP their preferred HDDCRP system that also uses cross-document dependencies. Table 21: Reference results macro averaged over ECB+ corpus with different options for event detection. kNWR=NewsReader event detection without invented mentions, maximizing precision, NWR=NewsReader pipeline v3.0 without adaptation, EDg=NWR augmented with EventDetection trained with gold data, EDgEC= same as EDg but skipping predicates with an Event class, EDs= NWR augmented with EventDetection trained with silver data, EDsEC= same as EDs but skipping predicates with an Event class. ARM= standard setting one participant in any role (AR), time month match and action concept and phrase match 30%, mR= maximizing recall by no constraints on participant match and time, action concept and phrase match 1%, mP=maximizing precision by participant roles A0A1, time day match and action concept and phrase match is set to 100%. kNWR-ARM NWR-ARM EDg EDgEC EDs EDsEC R 33.21% 33.21% 23.56% 32.73% 24.73% 32.63% MUC P 73.38% 50.97% 69.09% 57.49% 69.35% 57.52% F 45.73% 40.22% 35.14% 41.71% 36.46% 41.64% R 40.05% 40.05% 27.05% 39.00% 28.49% 39.04% BCUB P 83.49% 53.28% 76.77% 61.98% 77.43% 62.00% F 54.13% 45.73% 40.00% 47.88% 41.66% 47.91% R 55.12% 54.76% 37.78% 52.74% 39.34% 52.54% CEAFe P 53.92% 33.30% 43.71% 37.99% 46.96% 38.48% F 54.52% 41.42% 40.53% 44.17% 42.81% 44.43% CoNLL F 51.46% 42.46% 38.56% 44.59% 40.31% 44.66% Table 22: Reference results macro averaged over ECB+ corpus as reported by Yang et al. (2015) for state-of-the-art machine learning systems MUC LEMMA HDP-LEX AGGLOMERATIVE DDCRP HDDCRP* HDDCRP R 55.40% 63.50% 59.20% 58.20% 66.40% 67.10% P 75.10% 75.50% 78.30% 79.60% 77.50% 80.30% F 63.80% 69% 67.40% 67.10% 71.50% 73.10% R 39.60% 43.70% 40.20% 39.60% 48.10% 40.60% BCUB P 71.70% 65.60% 73.20% 78.10% 69% 73.10% F 51% 52.50% 51.90% 52.60% 56.70% 53.50% R 61.10% 60.20% 65.60% 69.40% 63% 68.90% CEAFe P 36.20% 34.80% 30.20% 31.80% 38.20% 38.60% F 45.50% 44.10% 41.40% 43.60% 47.60% 49.50% CoNLL F 53.40% 55.20% 53.60% 54.40% 58.60% 58.70% In line with Yang et al, we averaged the different F-measures to obtain a CoNLL-F value. BLANC results are not reported by Yang et al. We can see that the best NewsReader CoNLL-F scores about 14 points lower (44.66% against 58.70%), while the NewsReader key version (kNWR) scores about 5 points lower. Looking more precisely at the recall and precision scores, we can see that NewsReader scores significantly lower in recall but often equal and in some cases higher in precision. In case of EDs(ilver), NewsReader scores 77.43% for BCUB precision, while HDDCRP scores 73.1% and the best precision is 78.1% by DDCRP. In the case of CEAFe, EDs(ilver) even has a precision of 46.96% and the NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 76/148 best score by Yang et al is 38.6% for HDDCRP. Yang et al noticed that event detection of a standard SRL system performs low (56% recall) and therefore trained a separate CRF event detection system for event detection using the ECB+ training documents.15 Their CRF classifier obtains 93.5%, 89.0%, 95.0%, and 72.8% recall of the annotated event mentions, participants, time and locations on the test data set. The NewsReader system scores at most 72.66% recall on event detection. This shows that there is a big potential for NewsReader to improve with respect to the above results, when specifically trained on detecting events. Increasing the recall of events by 20 points will have a big impact on the recall for event coreference as well. On the other hand, we can assume that the NewsReader results are more realistic as an indication of the performance on any data set since it has not been trained on a specific data set. The results reported by Yang et al are likely to drop when moving from ECB+ to another data set unless separate training data is provided. For comparison, our own CRF Even Detection system trained on SemEval 2013 TempEval 3 data performed with F-scores above 80% but performed much lower (more than 10 points) when applied to ECB+. Given the nature of the ECB+ data set, it makes sense to consider the actual distribution of predicates over documents, topics and the complete data set to measure the complexity of within-topic and across-topic comparison. The extend that predicates occur across topics can be seen as an indication for the referential ambiguity since the main referential events are separated by topics with a systematic ambiguity to two seminal events within a single topic. We thus expect that contextual events tend to occur only within a single topic, while for example source events and grammatical events are not restricted to a specific seminal event. Table 23 shows the division of mentions of three predicates across documents and across topics. We first give the distribution on the basis of the full news articles and next the distribution in the annotated part of the article, where on average 1.8 sentence has been annotated per article. Table 23: Distribution of tell, kill and election over all text and annotated text per mention, document and topic in ECB+ tell kill election mentions 397 420 77 documents 235 207 29 Full text ment/doc 1.69 2.03 2.66 topics 39 22 6 ment/top 10.18 19.09 12.83 mentions 23 141 17 Annotated text documents ment/doc 21 1.10 129 1.09 10 1.70 topics 8 14 1 ment/top 2.88 10.07 17.00 We can see in Table 23 that a source event such as tell occurs even less than a contextual event such as kill when considering the full article. Nevertheless, tell occurs in more documents and more topics than kill. We can see that a specific event such as election has a low frequency, the highest average document frequency and the lowest document frequency. These full text distributions confirm that a source event has a high dispersion compared to contextuals. However when we consider the annotated text, we see that there are hardly any mentions of tell left (5.7%) in comparison to kill (33.6%) and election (22.1%). Average mentions per document are lower and more equal for all three predicates 15 In fact they also noted this for the detection of participants: 76%, timex-expressions: 65% of times and locations: 13% NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 77/148 since there is on average 1.8 sentence per document that is annotated. However, the average mentions per topic ratio is much higher for election and dropped drastically for tell and substantially for kill. Given the distribution of the annotation in ECB+, we thus can see that source events are marginally annotated and this plays a minor role while some contextuals such as kill have a high dispersion across topics but others such as election do not. The seminal nature of the topics and the little overlap in events across topics, supports the fact that the results for within-topic and across-topic are relatively close. It also suggests that in real-life contexts when dealing with large volumes of news, it makes sense to apply some form of prior topic clustering to avoid excessive ambiguity for especially source events (and grammatical events) that are found in many texts regardless of the context of the event. 3.2.3 Conclusion NewsReader cross-document event coreference We have seen that NewsReader event coreference can be tuned towards either high recall (up to 62% at topic level and 41% at corpus level) or high precision (up to 81% at topic level and 72% at corpus level). We have seen that recall is still limited and that this is mainly due to the event detection. This is hopeful because it shows the validity of the approach and it is relatively easy to improve in comparison to the more complex event coreference process. We thus expect that further improving the event detection will also directly boost the quality of the event coreference. It is important to note that our state-of-the-art results are obtained using generic technology using logical comparison and processing without any domain adaptation and without any machine-learning on the specific data set that has been used for testing. This is an important feature of the system, since machine learning based systems tend to have lower results when applied to different data sets than trained on. We compared the performance of NewsReader against the latest state-of-the-art system by Yang et al. (2015). Although NewsReader performs lower for CoNLL-F and recall than the reported systems, it tends to have higher precision scores. The state-of-theart systems implemented by Yang et al benefit from training on ECB+ data, whereas NewsReader is not adapted to the annotations and the data set. We can thus expect that the NewsReader performance is more representative for other data sets than ECB+, whereas the methods reported by Yang et al will be expected to perform much lower on other data sets. Furthermore, Yang et al boosted the event mention detection from 56% to 95% (as well as the participants, locations and time detection) by training a separate classifier on ECB+, whereas the recall of the event detection of NewsReader is 72%. What events are annotated and what events are not is often dependent on the style of annotation and thus differs from data set to data set. We can assume that boosting the NewsReader event detection on a specific data set by training on annotated events will also lead to a significant boost in the event detection and consequently in the event coreference results. Finally, the ECB+ data set is an artificially created data set that does not represent a natural stream of news. Within natural daily news streams, there may not be two seminal events that compete for interpretation on the same day (e.g. two attacks in Paris) and there will be many more topics than the 42 topics in ECB+. The best settings for ECB+ NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 78/148 therefore may not be the best settings for dealing with daily news streams. Evaluating the best set up and best settings for event-coreference in a daily news stream is a lot of work and very complex. Within this project, we did not had the resources to carry out such an evaluation. Another problem is that there is no freely sharable data set that can be used. Such a data set needs to contain the news for a certain period (say one month) from many different sources so that we could follow the daily cumulation. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 4 79/148 Event Relations Event relation extraction, in particular temporal relation extraction, is a crucial step to anchor an event in time, to build event timelines and to reconstruct the plot of a story. In this section, we describe the task of detecting temporal relations, causal relations and predicate time anchor relations. The description of each task begins with the presentation of the annotation schema we have followed. 4.1 4.1.1 Temporal Relations Annotation Schema The annotation schema for temporal relations is based on the TimeML specification language (Pustejovsky et al. (2005)). In the TimeML annotation, temporal links are used to i) establish the temporal order of two events (event-event pair); ii) anchor an event to a time expression (event-timex pair); and iii) establish the temporal order of two time expressions (timex-timex pair). In TimeML, temporal links are annotated with the <TLINK> tag. The full set of temporal relations specified in TimeML version 1.2.1 (Saurı́ et al. (2006)) contains 14 types of relations, as illustrated in Table 24. Among them there are six paired relations (i.e. with one relation being the inverse of the paired one). These relations map one-to-one to 12 of Allen’s 13 basic relations.16 a |———| b |———| a |———| b |———| a |——| b |————| a |——| b |————| a |——| b |——————| a |——————| b |——| a |———| b |———| a |———| b a is before b b is after a a is ibefore b b is iafter a a begins b b is begun by a a ends b b is ended by a a is during b b is during inv a a includes b b is included in a a is simultaneous with b a is identity with b Table 24: Temporal relations in TimeML annotation According to the TimeML 1.2.1 annotation guidelines (Saurı́ et al. (2006)), the difference between during and is included (also their inverses) is that the during relation is specified when an event persists throughout a temporal duration (e.g. John drove for 5 16 Allen’s overlaps relation is not represented in TimeML. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 80/148 hours), while is included relation is specified when an event happens within a temporal expression (e.g. John arrived on Tuesday). In the NewsReader annotation guidelines (Tonelli et al. (2014)), we have simplified the set of relations by not considering the relation types during and identity which is a coreferential relation. A new relation has been added with respect to TimeML: the measure relation. It is used to connect an event and a timex of type duration which provides information on the duration of the related event. Example: The first A380 superjumbo, made pr1 by Airbus, was delivered pr2 today tmx2 to Singapore Airlines (SIA) 18 months tmx3 behind schedule. After the plane was delivered pr4 in Singapore, it was flown pr3 to Toulouse, France for the ceremony pr5 of about 500 guests. (DCT tmx1 : 2007-10-15) The NAF representation of a part of the temporal relations extracted from the sentences is as follows: <t e m p o r a l R e l a t i o n s > <!−−IS INCLUDED ( tmx1 , tmx2)−−> < t l i n k i d =” t l i n k 6 ” from=”tmx1” <!−−BEFORE( pr1 , p r 2)−−> < t l i n k i d =” t l i n k 2 2 ” from=”p r 1 ” <!−−BEFORE( pr4 , p r 3)−−> < t l i n k i d =” t l i n k 2 3 ” from=”p r 4 ” <!−−BEFORE( pr4 , p r 5)−−> < t l i n k i d =” t l i n k 3 1 ” from=”p r 4 ” <!−−IS INCLUDED ( pr2 , tmx2)−−> < t l i n k i d =” t l i n k 5 8 ” from=”p r 2 ” </ t e m p o r a l R e l a t i o n s > 4.1.2 t o=”tmx2” fromType=”t i m e x ” toType=”t i m e x ” r e l T y p e =”SIMULTANEOUS”/> t o=”p r 2 ” fromType=” e v e n t ” toType=” e v e n t ” r e l T y p e =”BEFORE”/> t o=”p r 3 ” fromType=” e v e n t ” toType=” e v e n t ” r e l T y p e =”BEFORE”/> t o=”p r 5 ” fromType=” e v e n t ” toType=” e v e n t ” r e l T y p e =”BEFORE”/> t o=”tmx2 ” fromType=” e v e n t ” toType=”t i m e x ” r e l T y p e =”IS INCLUDED”/> Temporal Relation Extraction The temporal relation extraction module extracts temporal relations holding between two events or between an event and a time expression or between two time expressions. Two methods are used to extract temporal relations: machine learning method based on SVM for classifying relations between two events or between an event and a time expression; rule based method for ordering two time expressions. Extraction of relations between two events or between an event and a timex. We consider all combinations of event/event and event/timex pairs within the same sentence (in a forward manner) as candidate temporal links. For example, if we have a sentence with entity order such as “...ev1 ...ev2 ...tmx1 ...ev3 ...”, the candidate pairs are (ev1 , ev2 ), (ev1 , tmx1 ), (ev1 , ev3 ), (ev2 , tmx1 ), (ev2 , ev3 ) and (ev3 , tmx1 ). We remove event pairs if the two events are part of the same verbal phrase. We also identify relations between verbal events and document creation times and between main events 17 of two consecutive sentences. The problem of determining the label (i.e. temporal relation type) of a given temporal link can be regarded as a classification problem. Given an ordered pair of entities (e1 , e2 ) that could be either event/event or event/timex pair, the classifier has to assign a certain label, namely one of the 14 TimeML temporal relation types. 17 Main events correspond to the ROOT element of the parsed sentence. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 81/148 A classification model is trained for each type of entity pair (event/event and event/timex), as suggested in several previous works (Mani et al. (2006); Chambers (2013)). We build our classification models (Mirza and Tonelli (2014b)) using the Support Vector Machine (SVM) implementation provided by YamCha 18 and train them with the TempEval3 training corpus UzZaman et al. (2013a). The feature vectors built for each pair of entities (e1 , e2 ) are as follows: • String and grammatical features. Tokens, lemmas, PoS tags and flat constituent (noun phrase or verbal phrase) of e1 and e2 , along with a binary feature indicating whether e1 and e2 have the same PoS tags (only for event/event pairs). • Textual context. Pair order (only for event/timex pairs, i.e. event/timex or timex/event), textual order (i.e. the appearance order of e1 and e2 in the text) and entity distance (i.e. the number of entities occurring between e1 and e2 ). • Entity attributes. Event attributes (class, tense, aspect and polarity), and timex type attribute19 of e1 and e2 as specified in TimeML annotation. Four binary features are used to represent whether e1 and e2 have the same event attributes or not (only for event/event pairs). • Dependency information. Dependency relation type existing between e1 and e2 , dependency order (i.e. governor-dependent or dependent-governor ), and binary features indicating whether e1 /e2 is the root of the sentence. • Temporal signals. We take the list of temporal signals extracted from the TimeBank 1.2 corpus into account. We found that the system performance benefit from distinguishing between event-related signals and timex-related signals, therefore we manually split the signals into two separate lists. Signals such as when, as and then are commonly used to temporally connect events, while signals such as at, for and within more likely occur with time expressions. There are also signals that are used in both cases such as before, after and until, and those kind of signals are added to both lists. Tokens of temporal signals occurring around e1 and e2 and and their positions with respect to e1 and e2 (i.e. between e1 and e2 , before e1 , or at the beginning of the sentence) are used as features. • Temporal discourse connectives. Consider the following sentences: i) “John has been taking that driving course since the accident that took place last week.” and ii) “John has been taking that driving course since he wants to drive better.” In order to label the temporal link holding between two events, it is important to know whether there are temporal connectives in the surrounding context, because they may contribute in identifying the relation type. For instance, it may be relevant to distinguish whether since is used as a temporal or a causal cue (example i) and ii) resp.). 18 http://chasen.org/~taku/software/yamcha/ The value attribute tends to decrease the classifier performance as shown in Mirza and Tonelli (2014b), and therefore, it is excluded from the feature set. 19 NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 82/148 This information about discourse connectives is acquired using the addDiscourse tool (Pitler and Nenkova (2009)), which identifies connectives and assigns them to one of four semantic classes in the framework of the Penn Discourse Treebank (The PDTB Research Group (2008)): Temporal, Expansion, Contingency and Comparison. We include as feature whether a discourse connective belonging to the Temporal class occurs in the textual context of e1 and e2 . Similar to temporal signals, we also include in the feature set the position of the discourse connective with respect to the events. The machine learning based module is available on github 20 and technical details about it can be found in the Deliverable 4.2.2 (Section 3.12). The result for relation classification (identification of the relation type given the relations) on TempEval3 test corpus (UzZaman et al., 2013a) is: 58.8% precision, 58.2% recall and 58.5% F1-measure. We compare the performance of tempRelPro to the other systems participating in the Tempeval-3 task in Table 25. According to the figures reported in UzZaman et al. (2013a), tempRelPro is the best performing system both in terms of precision and of recall. System tempRelPro UTTime-1, 4 UTTime-3, 5 UTTime-2 NavyTime-1 NavyTime-2 JU-CSE F1 58.48% 56.45% 54.70% 54.26% 46.83% 43.92% 34.77% Precision 58.80% 55.58% 53.85% 53.20% 46.59% 43.65% 35.07% Recall 58.17% 57.35% 55.58% 55.36% 47.07% 44.20% 34.48% Table 25: Tempeval-3 evaluation on temporal relation classification Our complete system (relation identification and classification) attempts to extract a document’s entire temporal graph, i.e. it extracts a high number of relations in a text. In the evaluation (see Deliverable 4.2.3) this leads to good performance in terms of recall but low in terms of precision due to the incompletness of the manually annotated corpora used as gold standard. Indeed annotating a corpus with all temporal relations between events and time expressions is a difficult and time consuming task. Consequently, in most of the available corpora only small portions of the temporal graph are annotated. For example in the NewsReader annotation guidelines five subtasks were defined to help annotators annotate the most important temporal relations, but many relations are not considered, such as relations between nominal events and document creation times. Cassidy et al. (2014) propose a new annotated corpus called TimeBank-Dense which is composed of files from the TimeBank corpus annotated with ten more temporal relations with respect to the original annotation. Currently, we are not able to evaluate our system 20 https://github.com/paramitamirza/TempCauseRelPro NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 83/148 on the TimeBank-Dense corpus because the set of relations annotated is slightly different from the one annotated by our system and because TimeBank is part of our training corpus. Relation extraction between two time expressions. The second step of the extraction of temporal relation in a document is the detection of timex/timex relations for all dates and times. This step is performed using rules depending on the normalized form of the value of time expressions. If the two time expressions are dates or times, we compare first the years, then the months, weeks, days, etc. In the case of fuzzy expressions with one of the following value PRESENT REF, PAST REF or FUTURE REF, then we use the relation between the second time expression of the pair and the Document Creation Time to order them. This step enables us to make the temporal relations between time expressions explicit. If the normalization of time expressions is correct, then the right order between them is extracted. Wrong relations are identified only if the normalization fails. Examples of timex/timex relations: Apple Computer announced today tmx1 another special event to be held on October 12 tmx2 . (DCT tmx0 : 2005-10-04) • Normalization: tmx1 : 2005-10-04; tmx2 : 2005-10-12 • Relations: tmx1 before tmx2 ; tmx0 simultaneous tmx1 ; tmx0 before tmx2 He will be repatriated to Cuba between now tmx3 and Feb. 10 tmx4 . (DCT tmx5 : 2000-01-07) • Normalization: tmx3 : PRESENT REF; tmx4 : 2000-02-10 • Relations: tmx3 before tmx4 ; tmx5 simultaneous tmx3 ; tmx5 before tmx4 4.2 4.2.1 Causal Relation Annotation Scheme The annotation scheme has been newly defined for the NewsReader project (see the NewsReader guidelines (Tonelli et al., 2014; Mirza et al., 2014)). Similar to the <TLINK> tag in TimeML for temporal relations, we introduce the <CLINK> tag to mark a causal relation between two events. Both TLINKs and CLINKs mark directional relations, i.e. they involve a source and a target event. However, while a list of relation types is part of the attributes for TLINKs (e.g. before, after, includes, etc.), for CLINKs only one relation type is foreseen, going from a source (the cause, indicated with s in the examples) to a target (the effect, indicated with t ). We also introduce the notion of causal signals through the <C-SIGNAL> tag. CSIGNALs are used to mark-up textual elements signalling the presence of causal relations, which include all causal uses of prepositions (e.g. because of, as a result of, due to), NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 84/148 conjunctions (e.g. because, since, so that), adverbial connectors (e.g. so, therefore, thus) and clause-integrated expressions (e.g. the reason why, the result is, that is why). Wolff (2007) claims that causation covers three main types of causal concepts, i.e. CAUSE, ENABLE and PREVENT. These causal concepts are lexicalized through three types of verbs listed in Wolff and Song (2003): i) CAUSE-type verbs, e.g. cause, prompt, force; ii) ENABLE-type verbs, e.g. allow, enable, help; and iii) PREVENT-type verbs, e.g. block, prevent, restrain. These categories of causation are taken into account as an attribute of CLINKs. Given two annotated events, a CLINK is annotated if there is an explicit causal construction linking them. Such construction can be expressed in one of the following ways: 1. Expressions containing affect verbs (affect, influence, determine, change, etc.), e.g. Ogun ACN crisis s influences the launch t of the All Progressive Congress. 2. Expressions containing link verbs (link, lead, depend on, etc.), e.g. An earthquake t in North America was linked to a tsunami s in Japan. 3. Basic constructions involving causative verbs of CAUSE, ENABLE and PREVENT type, e.g. The purchase s caused the creation t of the current building. 4. Periphrastic constructions involving causative verbs of CAUSE, ENABLE and PREVENT type, e.g. The blast s caused the boat to heel t violently. With “periphrastic” we mean constructions where a causative verb (caused ) takes an embedded clause or predicate as a complement expressing a particular result (heel ). 5. Expressions containing CSIGNALs, e.g. Its shipments declined t as a result of a reduction s in inventories by service centers. Example: The departure from France of the new Airbus A380 superjumbo airliner on a tour of Asia and Australia has been delayed pr4 , leading to a rearrangement pr8 of its public appearances. The NAF representation of the causal relation holding between delayed and rearrangement is as follows: <c a u s a l R e l a t i o n s > <!−−(pr4 , pr8)−−> <c l i n k i d=” c l i n k 5 ” from=”pr4 ” t o=”pr8”/> </ c a u s a l R e l a t i o n s > 4.2.2 Causal Relation Extraction We start with an assumption that causality may only occur between events in the same sentence and between events in two consecutive sentences. Therefore, every possible combination of events in the same sentence and in two consecutive sentences, in a forward manner, is considered as a candidate event pair. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 85/148 The problem of detecting causal relations (CLINKs) between events is taken as a supervised classification task. Given an ordered pair of events (e1 ,e2 ), the classifier has to decide whether there is a causal relation or not. However, since causality is a directional relation between a cause (source) and an effect (target), the classifier has to assign one of three possible labels: (i) clink (where e1 is the source and e2 is the target), (ii) clink-r (with the reverse order of source and target), and (iii) o for no relation. The classification model is built with YamCha21 (Kudo and Matsumoto (2003)), which implements Support Vector Machines (SVMs) algorithm. We employ one-vs-one strategy for multi-class classification, and use the polynomial kernel. The overall approach is inspired by an existing work for identifying causal relations between events (Mirza and Tonelli (2014a)), with some differences in the feature set. The implemented features are explained in the following sections. The module is available on github https: //github.com/paramitamirza/TempCauseRelPro. Event features We implement some morphological, syntactical and textual context features of e1 and e2 , such as: • lemma and part-of-speech (PoS) tags; • sentence distance (e.g. 0 if e1 and e2 are in the same sentence, 1 if they are in adjacent sentences); • entity distance (i.e. the number of entities occurring between e1 and e2 , which is only measured if e1 and e2 are in the same sentence); • dependency path existing between e1 and e2 ; • binary features indicating whether e1 /e2 is the root of the sentence; • event attributes of e1 /e2 , including tense, aspect and polarity; and • a binary feature indicating whether e1 and e2 co-refer.22 Causal marker features We consider three types of causal markers that can cue a causal relation between events: 1. Causal signals. We extracted a list of causal signals from the annotated C-SIGNALs in the Causal-TimeBank corpus.23 2. Causal connectives, i.e. the discourse connectives under the Contingency class according to the output of the addDiscourse tool (Pitler and Nenkova (2009)). 3. Causal verbs. The three types of verbs lexicalizing causal concepts as listed in Wolff and Song (2003): i) CAUSE-type verbs, e.g. cause, prompt, force; ii) ENABLE-type verbs, e.g. allow, enable, help; and iii) PREVENT-type verbs, e.g. block, prevent, restrain. 21 http://chasen.org/~taku/software/yamcha/ When two events co-refer, there is almost no chance that they hold a causal relation. 23 For some causal signals that can have some other tokens in between, e.g. due (mostly) to, we instead include their regular expression patterns, e.g. /due .*to/, in the list. 22 NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 86/148 We further enriched the list of causal signals and causal verbs with the Paraphrase Database (PPDB, Ganitkevitch et al. (2013)), using the initial list of signals and verbs as seeds. Based on the existence of causal markers around e1 and e2 , exactly in that priority order24 , we include as features: • causal marker string; • causal marker position, i.e. between e1 and e2 , before e1 , or at the beginning of the sentence where e1 /e2 is in; and • dependency path between the causal marker and e1 /e2 . TLINKs Mirza and Tonelli (2014a) showed that even though only 32% of the gold annotated causal links have the underlying temporal relations, the temporal relation type of an event pair (e1 , e2 ), if any, contributes in determining the direction of the causal relation (clink vs clink-r), if any. Therefore, we include the information of temporal relation types in the feature set. In building the causal relation extraction system, we use Causal-TimeBank25 (Mirza et al. (2014)) with the previously explained annotation scheme as our development dataset. Causal-TimeBank is the TimeBank corpus26 taken from the TempEval3 evaluation campaign, which is completed with causal information as well. There are 318 causal links (CLINKs), only around 6.2% of the total temporal links (TLINKs) found in the corpus, containing 183 documents in total. The developed causal relation (CLINK) relation system is then evaluated in a five-fold cross-validation setting. Table 26 shows the performance of the system, compared with the system of Mirza and Tonelli (2014a) as a baseline. Given the limited amount of data annotated with causality, the supervised systems still do not yield satisfactory results. Mirza and Tonelli (2014a) report issues with data sparseness, and suggest that other training data could be derived, for instance, from the Penn Discourse Treebank (Prasad et al. (2008)). We adopt a different approach by combining the small data set available with unlabelled dataset in a semi-supervised setting, specifically with the self-training method. We exploit the remaining of TempEval-3 corpus besides TimeBank, i.e. AQUAINT and TE3-platinum (TempEval-3 evaluation corpus), with gold events and TLINKs. The idea of using the corpus with gold standard events and TLINKs for semi-supervised learning is because event’s attributes (tense, aspect and polarity) and TLINKs are quite important as features. There are 90 additional documents in total. The self-training method is done with 9 iterations with 10 documents per iteration. Two different schemes of the self-training are explored: (1) adding all extracted CLINKs 24 We first look for causal signals. If we do not find any, then we continue looking for causal connectives. And so on. 25 http://hlt.fbk.eu/technologies/causal-timebank 26 Dataset annotated with temporal entities such as temporal expressions, events and temporal relations in TimeML annotation framework. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 87/148 as new training data (with imbalanced number of positive and negative examples) and (2) adding balanced number of positive and negative CLINKs. Self-training with the (1) scheme improves the precision but lowers the recall. Meanwhile, self-training with the (2) scheme reduces the precision but improves the recall, but increases the overal performance in terms of F1-score. System Mirza and Tonelli (2014a) CLINK extraction self-training (1) self-training (2) P 0.6729 0.6917 0.7167 0.6382 R 0.2264 0.2921 0.2730 0.3079 F1 0.3388 0.4107 0.3954 0.4154 Table 26: CLINK extraction system’s performance. 4.3 Predicate Time Anchors The amount of temporal relations extracted by the previously described Temporal Relation Extraction modules grows with the number of annotated events and temporal expressions. Some events are linked to a time expression with a relation of type simultaneous or is included, but some are only linked with relations of type after or before either to a time expression or to another event. With the main goal of structuring timelines from events in texts, we propose to use these relations and other textual information in order to build a relation “PredicateTimeAnchor” between all events that can be anchored in time and time expressions. 4.3.1 Annotation Scheme A narrative container is defined by Styler IV et al. (2014) as a temporal expression or an event explicitly mentioned in the text into which other events temporally fall. For the TimeLine shared task at SemEval 2015 (Minard et al. (2015)) we proposed the notion of temporal anchoring of an event, which is a specific type of temporal relation that links an event to the temporal expression to which it is anchored. The anchoring in time of an event can be realized in two ways: either the event is anchored in time through a time expression which can be a DATE or a DURATION, or the event is anchored to an interval through a begin point and an end point. Time expressions can be text consuming or not dependent on the fact that they are explicitly expressed in the text or that they were derived from another time expression. Examples: 27 Stock markets around the world have fallen pr1 dramatically today tmx1 . 27 In order to make examples more readable they all contain the time expression and the event in the same sentence. But the module also anchors in time events with time expressions that are in different sentences. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 88/148 PredicateTimeAnchor (pr1 ): time anchor: 2008-09-17 (tmx1 ) The U.S. dollar rose against the yen after six straight days tmx2 of losses pr2 . PredicateTimeAnchor (pr2 ): time anchor: P6D (tmx2 , begin point: 2008-10-21, end point: 2008-10-27) The Japanese economy contracted pr3 by 0.9% between April tmx3 and June tmx4 . PredicateTimeAnchor (pr3 ): begin point: 2008-04 (tmx3 ); end point: 2008-06 (tmx4 ) The Russian government has continued to hold all stock markets closed pr4 until Friday tmx5 . PredicateTimeAnchor (pr4 ): end point: 2008-09-19 (tmx5 ) In NAF PredicateTimeAnchor relation is described through three attributes which have as value a reference link to a time expression: anchorTime: indicate the point in time when the event occured beginPoint: indicate the begin of the interval in which the event occured endPoint: indicate the end of the interval in which the event occured The NAF representation is as follows: <t e m p o r a l R e l a t i o n s > <p r e d i c a t e A n c h o r i d=”an2 ” anchorTime=”tmx1”> <span> <t a r g e t i d=”pr1”/> </span> </p r e d i c a t e A n c h o r > <p r e d i c a t e A n c h o r i d=”an3 ” b e g i n P o i n t=”tmx3” endPoint=”tmx4”> <span> <t a r g e t i d=”pr3”/> </span> </p r e d i c a t e A n c h o r > </t e m p o r a l R e l a t i o n s > 4.3.2 Predicate Time Anchor Relation Extraction Following the definition of time anchoring of events in the TimeLine shared task at SemEval 2015 (Minard et al. (2015)), we have developed a system to extract “anchoring” relations between an event and a time expression in the text. The system is rule based and performs a kind of reasoning over the temporal information. It uses the temporal relations previously extracted, verb tenses, dependency trees and temporal signals. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 5 89/148 From TimeLines to StoryLines In the previous section, we discussed the detection of temporal expressions, temporal relations and causal relations within a single document. In this section, we describe large structures that go beyond the document level, such as TimeLines of events for entities across documents and StoryLines. Stories are seen as the most natural representation of changes over time that also provide explanatory information to these changes. Not all changes make a story. Repetitive changes without further impact, e.g. the rising and dawning of the sun, do not provide a story. We expect that news is typically focused on those changes that have a certain impact. We further assume that the news tries to explain these events (How did it get so far? Who is responsible? ) and describethe consequences of the event. Our starting point is therefore a key concept from narrative studies, namely that of plot structure (Ryan (1991); Herman et al. (2010)). A plot structure is a representational framework which underlies the way a narrative is presented to a receiver. Figure 37 shows a a graphical representation of a plot structure. Figure 37: General structure of a plot building up to a climax point We therefore seek to create structures of events selected from all extracted events that approximate such abstract plot structure. Contrary to what can be done in narrative studies, where the documents themselves normally provide a linear development of the plot structure, we aim at obtaining the plot structure from collections of news articles on the same topic and spanning over a period of time. We aim at identifying first the “climax”, which in our perspective will correspond to the most salient event in a news article. After the most salient event and its participants have been identified we will use event relations to identify the rising actions (i.e. how and why did the most salient event occur? ), if any28 , and the falling actions and consequences (i.e. what happened after the climax? what are the speculations linked to the climax event? . . . ). The first step towards the creation of StoryLines is to to establish the temporal ordering of events. In subsection 5.1.2 we report on a TimeLine extraction system. In subsection 5.2.1 we describe two computational models for StoryLines and their implementation, which to our knowledge are unique in their kind. Finally, in subsection 5.2.3, we report on the results and insights 28 Notice that (unforeseeable) natural events, like earthquakes, are to be considered as self-contained climax events. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 90/148 of the first workshop on “Computing News Storylines (CNewsStory 2015)” organised as a satellite event of the ACL-IJCNLP 2015 conference. 5.1 TimeLine extraction This section reports on the advancements in the development of the TimeLine extraction. TimeLine extraction aims at reconstructing the chronological temporal order of events form (large) collections of news spanning over different years. In the following subsection, we describe the task, the benchmark data which had been developed for the SemEval 2015 evaluation exercise Task 4: TimeLine: Cross-Document Event Ordering, and the new version of the TimeLine system. 5.1.1 TimeLines: task description The Task 4: TimeLine: Cross-Document Event Ordering was proposed as a new pilot task for the SemEval 2015 Evaluation Exercise. The task builds on previous temporal processing tasks organised in previous SemEval editions (TempEval-129 , TempEval-230 and TempEval331 ). The task aimed at advancing temporal processing by tackling for the first time with a common and public dataset the following issues: • cross-document and cross-temporal event detection and ordering. • entity-based temporal processing. Following the task guidelines, a TimeLine can be defined as a set of chronologically anchored and ordered events related to an entity of interest (i.e. a person, a commercial or financial product, an organization, and similar) obtained from a document collection spanning over a (large) period of time. Furthermore, not all events are eligible to enter a TimeLine. The task organisers have restricted the event mentions to specific parts-ofspeech and classes, as defined in the task Annotation Guidelines32 , in particular, an event can enter into a TimeLine only if the following conditions apply: • it is realised by a verb, a noun or a pronoun (anaphoric reference); • it semantically denotes the happening of something (e.g. the launch of a new product) or it describes the action of declaring something, narrating an event, informing about an event; • it is a factual or certain event, i.e. something which happened in the past or in the present, or for which there is evidence that will happen in the future. 29 http://www.timeml.org/tempeval/ http://timeml.org/tempeval2/ 31 http://www.cs.york.ac.uk/semeval-2013/task1/ 32 http://alt.qcri.org/semeval2015/task4/data/uploads/documentation/ manualannotationguidelines-task4.pdf 30 NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 91/148 No training data was provided. Only a trial dataset of 30 articles manually annotated from WikiNews, and associated TimeLines for six entities, were provided to the task participants. Each event in the TimeLine is associated with a time anchor of type “DATE” following the TimeML Annotation Guidelines (Saurı́ et al. (2006)). In case an event cannot be associated with a specific time anchor, an underspecified time anchor of the type “XXXX-XX-XX” is provided. The final TimeLine representation is composed by a tab field file containing three fields. The first field (ordering) contains a cardinal number which indicates the position of the event in the TimeLine. Simultaneous, but not coreferential, events are associated with the same ordering number. Events which cannot be reliably ordered, either because of a missing time anchor or underspecified temporal relations (e.g. an event which is associated with a generic date with value “PAST REF”) are to be put at the beginning of the TimeLine and associated with cardinal number 0. The second field (time anchor) contains the time anchor. The third column (event) consists of one event or a list of coreferential events. Each event must be represented by the file id, the sentence id and the extent of the event mention (i.e. the token). To clarify the representation of a TimeLine, in Figure 38 we report the output of SPINOZA VU 1 system for the target entity “Airbus”33 . Figure 38: Example of timeline output generated by the SPINOZA VU 1 system Two tracks were proposed. Track A aimed at extracting TimeLines for target entities from raw text. Track B aimed at extracting TimeLines for target entities by providing manually annotated data for the gold events. Both tracks have a subtrack, Subtrack A and Subtrack 33 The entity ”Airbus” was one of the entity provided by the SemEval task organisers NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 92/148 B, whose goals are to evaluate only the ordering of events, without taking into account the time anchoring. The test data consisted of three different corpora, each containing 30 articles, and an overall 37 target entities (12 for the first corpus, 12 for the second corpus and 13 for the third corpus). The evaluation is based on the TempEval-3 evaluation (UzZaman et al. (2012)) tool. All events associated with cardinal number 0 in a TimeLine are excluded from the evaluation. Results and ranking reports the micro average F1 score for temporal awareness. 5.1.2 System Description and Evaluation A detailed description of the first version of the TimeLine extraction system can be found in Rospocher et al. (2015) and Caselli et al. (2015a). Two different versions were developed, called SPINOZA VU 1 and SPINOZA VU 2 respectively. The systems took part only to the Track A (both main and subtask) of the SemEval 2015 Task 4. In Table 27 we report the results of both versions of the system for the Task A - Main, including the best performing system. Table 28 reports the results of both versions of the system for the Task A - Subtask. In Table 28 no other result is reported because only our system participated. The F1-scores ranges from 0 to 100. System Version SPINOZA VU 1 SPINOZA VU 2 WHUNLP 1 Corpus 1 4.07 2.67 8.31 Corpus 2 5.31 0.62 6.01 Corpus 3 0.42 0.00 6.86 Overall 3.15 1.05 7.28 Table 27: System Results (micro F1 score) for the SemEval 2015 Task 4 Task A - Main System Version SPINOZA VU 1 SPINOZA VU 2 Corpus 1 1.20 0.00 Corpus 2 1.70 0.92 Corpus 3 2.08 0.00 Overall 1.69 0.27 Table 28: System Results (micro F1 score) for the SemEval 2015 Task 4 Task A - Subtask Overall, the results are not satisfying. Out of 37 entity based TimeLines, we obtained results only for 31 of them. An error analysis showed three main sources of errors which affected both versions of our system: event detection, temporal relations, semantic role labelling and the connections between these three. This means that: i.) we may be able to identify the correct event with respect to the target entity, but we are lacking the temporal relation information for that event, thus failing to put it into the TimeLine, ii.) we fail in the identification of the target entity as an argument of an event. The temporal relation of an event is the main source of error from these three. A detailed report on the error analysis of the system can be found in Caselli et al. (2015c). Future work is directed towards detecting more temporal relations between events and expressions that are explicit in the text but also to use knowledge on temporal ordering of events that is implicit and NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 93/148 not expressed in the text. The latter can be learned from large text corpora. In the next subsection, we describe the first results for resolving implicit relations exploiting the whole document. 5.1.3 Document level time-anchoring for TimeLine extraction As seen previously, TimeLine extraction requires a quite complete time anchoring. We have shown that the temporal relations that explicitly connect events and time expressions are not enough to obtain a full time-anchor annotation and, consequently, produce incomplete TimeLines. For this reason, we propose that for a complete time-anchoring the temporal analysis must be performed at a document level in order to discover implicit temporal relations. We have developed a preliminary module based on other research lines involving the extraction of implicit information (Palmer et al., 1986; Whittemore et al., 1991; Tetreault, 2002). Particularly, we are inspired by recent works on Implicit Semantic Role Labelling (ISRL) (Gerber and Chai, 2012) and very specially on the work by Blanco and Moldovan (2014) who adapted the ideas about ISRL to focus on modifiers, including arguments of time, instead of core arguments or roles. We have developed a deterministic algorithm of the type of (Laparra and Rigau, 2013) for ISRL. Similarly to the module presented in 5.1.2, we implemented a system that builds TimeLines from events with explicit time-anchors. We defined a three step process to build TimeLines. Given a set of documents and a target entity, the system first obtains the events in which the entity is involved. Second, it obtains the time-anchors for each of these events. Finally, it sorts the events according to their time-anchors. For steps 1 and 2 we apply the NewsReader pipeline to obtain annotations at different levels. Specifically, we are interested in Named-Entity Recognition (NER) and Disambiguation (NED), Co-reference Resolution (CR), Semantic Role Labelling (SRL), Time Expressions Identification (TEI) and Normalization (TEN), and Temporal Relation Extraction (TRE). Named-Entity Recognition (NER) and Disambiguation (NED): We perform NER using the ixa-pipe-nerc that is part of IXA pipes (Agerri et al., 2014). The module provides very fast models with high performances, obtaining 84.53 in F1 on CoNLL tasks. Our NED module is based on DBpedia Spotlight (Daiber et al., 2013). We have created a NED client to query the DBpedia Spotlight server for the Named entities detected by the ixa-pipe-nerc module. Using the best parameter combination, the best results obtained by this module on the TAC 2011 dataset were 79.77 precision and 60.67 recall. The best performance on the AIDA dataset is 79.67 precision and 76.94 recall. Coreference Resolution (CR): In this case, we use a coreference module that is loosely based on the Stanford Multi Sieve Pass sytem (Lee et al., 2011). The system consists of a number of rule-based sieves that are applied in a deterministic manner. The system scores 56.4 F1 on CoNLL 2011 task, around 3 points worse than the system by Lee et al. (2011). NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 94/148 Semantic Role Labelling (SRL): SRL is performed using the system included in the MATE-tools (Björkelund et al., 2009). This system reported on the CoNLL 2009 Shared Task a labelled semantic F1 of 85.63 for English. Time Expression Identification (TEI) and Normalization (TEN): We use the time module from TextPro suite (Pianta et al., 2008) to capture the tokens corresponding to temporal expressions and to normalize them following TIDES specification. This module is trained on TempEval3 data. The average results for English is: 83.81% precision, 75.94% recall and 79.61% F1 values. Time Relation Extraction (TRE): We apply the temporal relation extractor module from TextPro to extract and classify temporal relations between an event and a time expression. This module is trained using yamcha tool on the TempEval3 data. The result for relation classification on the corpus of TempEval3 is: 58.8% precision, 58.2% recall and 58.5% F1. Our TimeLine extraction system uses the linguistic information provided by the pipeline. The process to extract the target entities, the events and time-anchors can be described as follows: (1) Target entity identification: The target entities are identified by the NED module. As they can be expressed in several forms, we use the redirect links contained in DBpedia to extend the search of the events involving those target entities. For example, if the target entity is Toyota the system would also include events involving the entities Toyota Motor Company or Toyota Motor Corp. In addition, as the NED does not always provide a link to DBpedia, we also consider the matching of the wordform of the head of the argument with the head of the target entity. (2) Event selection: We use the output of the SRL module to extract the events that occur in a document. Given a target entity, we combine the output of the NER, NED, CR and SRL to obtain those events that have the target entity as filler of their ARG0 or ARG1. We also set some constraints to select certain events according to the specification of the SemEval task. That is, we only return those events that are not negated and are not accompanied by modal verbs except will. (3) Time-anchoring: We extract the time-anchors from the output of the TRE and SRL. From the TRE, we extract as time-anchors those relations between events and timeexpressions identified as SIMULTANEOUS. From the SRL, we extract as time-anchors those ARG-TMP related to time expressions. In both cases we use the time-expression returned by the TEI module. The tests performed on the trial data show that the best choice for time-anchoring is combining both options. For each time anchor we normalize the time expression using the output of the TEN module. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 95/148 The TimeLine extraction process described following this approach builds TimeLines for events with explicit time-anchors. We call this system BTE and it can be seen as a baseline since we believe that the temporal analysis should be carried out at document level. The explicit time anchors provided by the NLP tools do not cover the full set of events involving a particular entity. That is, most of the events do not have an explicit time anchor and therefore are not captured as part of the TimeLine of that entity. Thus, we need to recover the time-anchors that appear implicitly in the text. In this preliminary work, we propose a simple strategy that tries to capture implicit time-anchors while maintaining the coherence of the temporal information in the document. This strategy follows previous works on Implicit Semantic Role Labelling. Figure 39: Example of document-level time-anchoring. The rationale behind algorithm 1 is that by default the events of an entity that appear in a document tend to occur at the same time as previous events involving the same entity, except stated explicitly. For example, in Figure 39 all the events involving Steve Jobs, like gave and announced, are anchored to the same time-expression Monday although this only happens explicitly for the first event gave. The example also shows how for other events that occur in different times the time-anchor is also mentioned explicitly, like for those events that involve the entities Tiger and Mac OS X Leopard. Algorithm 1 starts from the annotation obtained by the tools described above. For a particular entity a list of events (eventList) is created sorted by its occurrence in the text. Then, for each event in this list the system checks if that event has already a time-anchor (eAnchor). If this is the case, the time-anchor is included in the list of default timeanchors (def aultAnchor) for the following events of the entity with the same verb tense (eT ense). If the event does not have an explicit time-anchor but the system has found a time-anchor for a previous event belonging to the same tense (def aultAnchor[eT ense]), this time-anchor is also assigned to the current event (eAnchor). If none of the previous conditions satisfy, the algorithm anchors the event to the Document Creation Time (DCT) and sets this time-expression as the default time-anchor for the following events with the same tense. Note that algorithm 1 strongly depends on the tense of the events. As this information can be only recovered from verbal predicates, this strategy cannot be applied to events described by nominal predicates. For these cases just explicit time-anchors are taken into NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 96/148 Algorithm 1 Implicit Time-anchoring 1: eventList = sorted list of events of an entity 2: for event in eventList do 3: eAnchor = time anchor of event 4: eT ense = verb tense of event 5: if eAnchor not N U LL then 6: def aultAnchor[eT ense] = eAnchor 7: else if def aultAnchor[eT ense] not N U LL then 8: eAnchor = def aultAnchor[eT ense] 9: else 10: eAnchor = DCT 11: def aultAnchor[eT ense] = DCT 12: end if 13: end for account. The TimeLine is built ordering the events according to the time-anchors obtained both explicitly and implicitly. We call this system DLT. We evaluated our two TimeLine extractors on the main track of the SemEval 2015 task 4. Two systems participated in this track, WHUNLP and the module explained in 5.1.2, with three runs in total. Their performances in terms of Precision (P), Recall (R) and F1-score (F1) are presented in Table 29. We also present in italics additional results of both systems. On the one hand, the results of a corrected run of the WHUNLP system provided by the SemEval organizers. On the other hand, the results of an out-ofcompetition version of the SPINOZAVU module The best run is obtained by the corrected version of WHUNLP 1 with an F1 of 7.85%. The low figures obtained show the intrinsic difficulty of the task, especially in terms of Recall. Table 29 also contains the results obtained by our systems. We present two different runs. On the one hand, we present the results obtained using just the explicit time-anchors provided by BTE. As it can be seen, the results obtained by this run are similar to those obtained by WHUNLP 1. On the other hand, the results of the implicit time-anchoring approach (DLT) outperforms by far our baseline and all previous systems applied to the task. To check that these results are not biased by the time-relation extractor we use in our pipeline (TimePro), we reproduce the performances of BTE and DLT using another system to obtain the time-relations. For this purpose we used CAEVO by Chambers et al. (2014). The results obtained in this case show that the improvement obtained by our approach is quite similar, regardless of the time-relation extractor chosen. The figures in Table 29 seem to prove our hypothesis. In order to obtain a full timeanchoring annotation, the temporal analysis must be carried out at a document level. The TimeLine extractor almost doubles the performance by just including a straightforward strategy as the one described in this section. As expected, Table 29 shows that this improvement is much more significant in terms of Recall. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 System SPINOZAVU-RUN-1 SPINOZAVU-RUN-2 WHUNLP 1 OC SPINOZA VU WHUNLP 1 BTE DLT BTE caevo DLT caevo 97/148 P R F1 7.95 1.96 3.15 8.16 0.56 1.05 14.10 4.90 7.28 7.12 14.59 5.37 7.85 26.42 4.44 7.60 20.67 10.95 14.31 17.56 4.86 7.61 17.02 12.09 14.13 Table 29: Results on the SemEval-2015 task 5.2 Storylines The TimeLines discussed in the previous section form the basis for StoryLines. Stories are a pervasive phenomenon in human life. They are explanatory models of the world and of its happenings (Bruner, 1990). We make reference to the narratology framework of Bal (Bal, 1997) to identify the basic concepts which inform our model. Every story is a mention of a fabula, i.e., a sequence of chronologically ordered and logically connected events involving one or more actors. Actors are the agents, not necessarily humans, of a story that perform actions. In Bal’s framework “acting” refers both to performing and experiencing an event. Events are defined as transitions from one state to another. Furthermore, every story has a focalizer, a special actor from whom’s point of view the story is told. Under this framework, the term “story” is further defined as the particular way or style in which something is told. A story, thus, does not necessarily follow the chronological order of the events and may contain more than one fabula. Extending the basic framework and focusing on the internal components of the fabula, a kind of universal grammar can be identified which involves the following elements: • Exposition: the introduction of the actors and the settings (e.g. the location); • Predicament: it refers to the set of problems or struggles that the actors have to go through. It is composed by three elements: rising action, the event(s) that increases the tension created by the predicament, climax, the event(s) which creates the maximal level of tension , and, finally, falling action, the event(s) which resolve the climax and lower the tension; • Extrication: it refers to the “end” of the predicament and indicates the ending. The model allows to focus on each its the components, highlighting different, though connected, aspects: the internal components of the fabula are event-centered; the actors and the focalizer allows access to opinions, sentiments, emotions and world views; and, the medium to the specific genres and styles. We developed two different approaches to NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 98/148 Figure 40: Example of a StoryLine merging the TimeLines of the entities Steve Jobs and Iphone 4. create StoryLine structures that focus on different aspects of the fabula. The first approach aggregates stories from separate TimeLines for different actors through co-participation. The second approach aggregates stories from climax events and bridging relations with other events that precede and follow the climax. 5.2.1 StoryLines aggregated from entity-centered TimeLines Timelines as described in the previous section are built for single entities. However, stories usually involve more than one entity. In this section, we present a proposal to create StoryLines by merging the individual TimeLines of two or more different entities, provided that they are co-participants of at least one relevant event. In general, given a set of related documents, any entity appearing in the corpus is a candidate to take part in a StoryLine. Thus, a TimeLine for every entity should be extracted following the requirements described by the SemEval-2015 task. Then, those TimeLines that share at least one relevant event must be merged. Those entities that do not co-participate in any event with other entities are not considered participants of any StoryLine. The expected StoryLines should include both the events where the entities interact and the events where the entities selected for the StoryLines participate individNewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 99/148 ually. The events must be ordered and anchored in time in the same way as individual TimeLines, but it is also mandatory to include the entities that take part in each event. Figure 40 presents the task idea graphically. In the example, two TimeLines are extracted using 5 sentences from 3 different documents, one for the entity Steve Jobs and another one for the entity Iphone 4. As these two entities are co-participants of the events introducing and introduced, the TimeLines are merged in a single StoryLine. As a result, the StoryLine contains the events of both entities. The events are represented by the ID of the file, the ID of the sentence, the extent of the event mention and the participants (i.e. entities) of the event. timelines from SemEval storylines events events / storyline interacting-events interacting-events / storyline entities entities / storyline Apple Inc. Airbus GM Stock Total 6 13 11 13 43 1 2 1 3 7 129 135 97 188 549 129 67.5 97 62.7 78.4 5 12 2 11 30 5 6 2 3.7 4.3 4 9 4 9 26 4 4.5 4 3 3.7 Table 30: Figures of the StoryLine gold dataset. Dataset As a proof-of-concept, we start from the dataset provided in SemEval-2015. It is composed of 120 Wikinews articles grouped in four different corpora about Apple Inc.; Airbus and Boeing; General Motors, Chrysler and Ford; and Stock Market. The Apple Inc. set of 30 documents serve as trial data and the remaining 90 documents as the test set. We consider each corpus a topic to extract StoryLines. Thus, for each corpus, we have merged the interacting individual TimeLines to create a gold standard for StoryLines. As a result of this process, from a total of 43 TimeLines we obtained 7 gold-standard StoryLines spread over 4 topics. Table 30 shows the distribution of the StoryLines and some additional figures about them. Airbus, GM and Stock corpora are similar in terms of size but the number of gold StoryLines varies from 1 to 3. We also obtain 1 StoryLine from the Apple Inc. corpus, but in this case the number of TimeLines is lower. The number of events per StoryLine is quite high in every corpus, but the number of interacting events is very low. Finally, 26 out of 43 target entities in SemEval-2015 belong to a gold StoryLine. Note that in real StoryLines all interacting entities should be annotated whereas now we only use those already selected by the TimeLines task. Evaluation The evaluation methodology proposed in SemEval-2015 is based on the evaluation metric used for TempEval-3 (UzZaman et al., 2013a) which captures the temporal awareness of NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 100/148 an annotation (UzZaman and Allen, 2011). For that, they first transform the TimeLines into a set of temporal relations. More specifically, each time anchor is represented as a TIMEX3 so that each event is related to the corresponding TIMEX3 by means of the SIMULTANEOUS relation. In addition, SIMULTANEOUS and BEFORE relation types are used to connect the events. As a result, the TimeLine is represented as a graph and evaluated in terms of recall, precision and F1-score. As a first approach, the same graph representation can be used to characterize the StoryLines. Thus, for this trial we reuse the same evaluation metric as the one proposed in SemEval-2015. However, we already foresee some issues that need to be addressed for a proper StoryLines evaluation. For example, when evaluating TimeLines, given a set of target entities, the gold standard and the output of the systems are compared based on the F1 micro average scores. In contrast, when evaluating StoryLines, any entity appearing in the corpus is a candidate to take part in a StoryLine, and several StoryLines can be built given a set of related documents. Thus, we cannot compute the micro-average of the individual F1-scores for each StoryLine because the number of StoryLines is not set in advance. In addition, we also consider necessary to capture the cases in which having one gold standard StoryLine a system obtains more than one StoryLine. This could happen when a system is not able to detect all the entities interacting in events but only some of them. We consider necessary to offer a metric which takes into account this type of outputs and also scores partial StoryLines. Obviously, a deeper study of the StoryLines casuistry will lead to a more complete and detailed evaluation metric. Example of a system-run In order to show that the dataset and evaluation strategy proposed are ready to be used on StoryLines, we follow the strategy described to build the gold annotations to implement an automatic system. This way, we create a simple system which merges automatically extracted TimeLines. To build the TimeLines, we use the system explained in section 5.1.3. For each target entity, we first obtain the corresponding Timeline. Then, we check which TimeLines share the same events. In other words, which entities are co-participants of the same event and we build StoryLines from the TimeLines sharing events. This implies that more than two TimeLines can be merged into one single StoryLine. The system builds 2 StoryLines in the Airbus corpus. One StoryLine is derived from the merging of the TimeLines of 2 target entities and the other one from the merging of 4 TimeLines. In the case of the GM corpus, the system extracts 1 StoryLine where 2 target entities participate. For the Stock corpus, one StoryLine is built merging 3 TimeLines. In contrast, in the Apple corpus, the system does not obtain any StoryLine. We evaluated our StoryLine extractor system in the cases where it builts StoryLines. The evaluation results are presented in Table 31. Based on the corpus, the results of our strategy vary. The system is able to create StoryLines which share data with the gold-standard in the Airbus corpus, but it fails to create comparable StoryLines in the GM and Stock corpora. Finding the interacting events is crucial for the extraction of the StoryLines. If these events are not detected for all their participant entities, their corresponding TimeLines cannot be merged. For that reason, NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 Corpus Airbus GM Stock 101/148 Precision 6.92 0.00 0.00 Recall Micro-F 14.29 4.56 0.00 0.00 0.00 0.00 Table 31: Results of the StoryLine extraction process. our dummy system obtains null results for the GM and Stock corpus. However, this is an example of a system capable of creating StoryLines. Of course, more sophisticated approaches or approaches that do not follow the TimeLine extraction approach could obtain better results. 5.2.2 Storylines aggregated from climax events The above method aggregates StoryLines across TimeLines. However, it still does not provide an explanatory notion for the sequences of events. In this section, we present a model that starts from a climax event that motivates the selection of events. In our model we use the term StoryLine to refer to an abstract structured index of connected events which provides a representation matching the internal components of the fabula (rising action(s), climax, falling action(s) and resolution). On the other hand, we reserve the term Story for the textual expression of such an abstract structure34 . Our model, thus, does not represent texts but event data from which different textual representations could be generated. The basic elements of a StoryLine are: • A definition of events, participants (actors), locations and time-points (settings) • Anchoring of events to time • A TimeLine (or basic fabula): a set of events ordered for time (chronological order) • Bridging relations: a set of relations between events with explanatory and predictive value(s) (rising action, climax and falling action) Storylines are built on top of the instance level of representation, as illustrated in 2, and TimeLines. Given a TimeLine for a specific period of time, we define a StoryLine S as n-tuples T, E, R such that: Timepoints = (t1 , t2 , , ..., tn ) Events = (e1 , e2 , ..., en ) Relations = (r1 , r1 , ..., rn ) 34 Note that a StoryLine can be used to generate a textual summary as a story, comparable to (cross)document text summarization. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 102/148 T consists of an ordered set of points in time, E is a set of events and R is a set of bridging relations between these events. Each e in E is related to a t in T. Furthermore, for any pair of events ei and ej , where ei precedes ej there holds a bridging relation [r, ei , ej ] in R. We assume that there is a set of TimeLines L for every E, which is any possible sequence of events temporally ordered. Not every temporal sequence l of events out of L makes a good StoryLine. We want to approximate a StoryLine that people value by defining a function that maximizes the set of bridging relations across different sequences of events l in L. We therefore assume that there is one sequence l that maximizes the values for R and that people will appreciate this sequence as a story. For each l in L, we therefore assume that there is a bridging function B over l that sums the strength of the relations and that the news StoryLine S is the sequence l with the highest score for B : S(E) = M AX(B(l)) B(l)) = n X C(r, ei , ej ) i,j=1 Our bridging function B sums the connectivity strength C of the bridging relations between all time-ordered pairs of events from the set of temporally ordered events l. The kind of bridging relation r and the calculation of the connectivity strength C can be filled in many ways: co-participation, expectation, causality, enablement, and entailment, among others. In our model, we leave open what type of bridging relations people value. This needs to be determined empirically in future research. The set L for E can be very large. However, narratology models state that stories are structured around climax events. The climax event makes the story worthwhile to tell. Other preceding and following events are grouped around the climax to explain it. It thus makes sense to consider only those sequences l that include at least one salient event as a climax and relate other events to this climax event. Instead of calculating the score B for all l in L, we thus only need to build event sequences around events that are most salient as a climax event and select the other events on the basis of the strength of their bridging relation with that climax or with each other. For any climax event ec , we can therefore define: n M AX(B(ec E)) = max C(r, ei , ec ) i=1 The climax value for an event can be defined on the basis of salience features, such as: • prominent position in a source; • number of mentions; • strength of sentiment or opinion; • salience of the involved actors with respect to the source. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 103/148 An implementation should thus start from the event with the highest climax score. Next, it can select the preceding event el with the strongest value for r. Note that this is not necessarily the event that is most close in time. After that, the event el with the strongest connectivity is taken as a new starting point to find any event ek preceding this event with the highest value for r. This is repeated until there are no preceding events in the TimeLine l. The result is a sequence of events up to ec with the strongest values for r. The same process is repeated forward in time starting from ec and adding em with the strongest connectivity value for r, followed by en with the strongest connectivity score r to em . The result is a sequence of events with local maxima spreading from ec : ...ek , rmax , el , rmax , ec , rmax , em , rmax , en ... This schema models the optimized StoryLine starting from a climax event. By ranking the events also for their climax score, the climax events will occupy the highest position and the preceding and following events the lower positions approximating the fabula or plot graph shown in Figure 37. Storyline Extraction System The StoryLine extraction system is composed by three components: a.) TimeLine extraction; b.) climax event identification; c.) rising and falling actions identification. The TimeLine structures are obtained from the system described in Caselli et al. (2015a). Although, all events may enter in a TimeLine, including speech-acts such as say, not every sequence of ordered events makes a StoryLine. Within the total set of events in a TimeLine, we compute for each event its prominence on the basis of the mention sentence number and the number of mentions in the source documents. We currently sum the inverse sentence P number of each mention of an event in the source documents: P (e) = (1/S(em )). em =1→N This formula combines the number of references made to an event with the position in the text that a word is mentioned: early mentions counts more than late mentions and more mentions make it more prominent. All event instances are then ranked according to the degree of prominence P. We implemented a greedy algorithm in which the most prominent event will become the climax event.35 Next, we determine the events with the strongest bridging relation preceding and following the climax event in an iterative way until there are no preceding and following events with a bridging relation. Once an event is claimed for a StoryLine, we prevent it from being re-used for another StoryLine. For all remaining events (not connected to the event with the highest climax score), we again select the event with the highest climax score of the remaining events and repeat the above process. Remaining events thus can create parallel StoryLines although with a lower score. When descending the climax scores, we ultimately are left with events with low climax score that are not added to any StoryLine and do not constitute StoryLines themselves. 35 Future versions of the system can include other properties such as emotion or salience of actors NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 104/148 For determining the value of the bridging relations we use various features and resources, where we make a distinction between structural and implicit relations: • Structural relations: – co-participation; – explicit causal relations; – explicit temporal relations; • Implicit relations: – expectation based on corpus co-occurrence data; – causal WordNet relation; – frame relatedness in FrameNet; – proximity of mentions; – entailment; – enablement. Our system can in principle use any of the above relations and resources. However, in the current version, we have limited ourselves to co-participation and FrameNet frame relations. Co-participation is the case when two events share at least one participant URI which has a PropBank relation A0, A1 or A2. The participant does not need to have the same relation in the two events. Events are related to FrameNet frames if there is any relation between their frames in FrameNet up to a distance of 3. Below we show an example of a larger StoryLine extracted from the corpus used in the SemEval 2015 TimeLine task. :Airbus 29 3 [C]61 23 6 1 15 22 39 20040101 20041001 20040301 20050613 20050613 20050613 20061005 20070228 20070319 21 12 3 21 13 4 20 20070319 20070609 20070708 20080201 20090730 20041124 20141213 ["manufacturer","factory","manufacture"] :Boeing:European_aircraft_manufacturer_Airbus:Airbus ["charge","level","kill"] :United_States_Department_of_Defense:the_deal ["purchase"] :People_s_Republic_of_China:Airbus_aircraft ["win"] :European_aircraft_manufacturer_Airbus:Boeing ["aid","assistance","assist"] :Airbus:Boeing:for_the_new_aircraft ["spark"] :Airbus ["compensate"] :Airbus:of_its_new_superjumbo_A380s ["cut","have","reduction","make"] :Airbus:the_company ["supply","provide","resource","supplier","fund","tube"] :European_Aeronautic_Defence_and_Space_Company_EADS_N.V. :Airbus:United_States_Department_of_Defense ["carry","carrier"]:the_airplane:Airbus_will ["jet"]:Airbus:Airbus_A320_family ["write","letter"]:Airbus:Boeing ["ink","contract","sign"]:Royal_Air_Force:Airbus ["lead","give","offer"] :France:Airbus ["personnel","employee"] :Airbus:Former_military_personnel ["carry","flight","fly"] :The_plane:Airbus Figure 41: Storyline for Airbus and Boeing from the SemEval 2015 Task 4 dataset. The StoryLine is created from a climax event ["purchase"] involving Airbus with a score of 61. The climax event is marked with C at the beginning of the line. After connecting the other events, they are sorted according to their time anchor. Each line NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 Figure 42: Airbus StoryLines order per climax event 105/148 Figure 43: Airbus StoryLine for climax event [61] “purchase” is a unique event instance (between square brackets) anchored in time, preceded by the climax score and followed by major actors involved.36 We can see that all events reflect the commercial struggle between Airbus and Boeing and some role played by governments. In Figure 42, we visualise the extracted StoryLines ordered per climax event. Every row in the visualisation is a StoryLine grouped per climax event, ordered by the climax score. The label and weight of the climax event are reported in the vertical axis together with the label of the first participant with an A1 Propbank role, which is considered to be most informative. Within a single row each dot presents an event in time. The size of the dot represents the climax score. Currently, the bridging relations are not scored. A bridging relation is either present or absent. If there is no bridging relation, the event is not included in the StoryLine. When clicking on a target StoryLine a pop up window opens showing the StoryLine events ordered in time (see Figure 43). Since we present events at the instance level across different mentions, we provide a semantic class grouping these mentions based on WordNet which is shown on the first line. Thus the climax event “purchase” is represented with the label more general label “buy” that represents a hypernym synset. If a StoryLine is well structured, the temporal order and climax weights mimic the fabula internal structure, as in this case. We expect that events close to the climax have larger dots than more distant events in time.37 Stories can be selected per target entity through the drop-down menu on top of the graph. In the Figure 42, all stories concerning Airbus are marked in red. An online version of this visualisation can be found on the project website at http://ic.vupr.nl/timeline. You can upload NAF files from which StoryLines are extracted or JSON files extracted from a collection of NAF files. Comparing the StoryLine representation with the TimeLine (see Figure 38) some differences can be easily observed. In a StoryLine, events are ordered in time and per climax weight. The selection of events in the StoryLine is motivated by the bridging relations which exclude non-relevant events, such as say. We used the visualisation to inspect the results. We observed that some events were missed because of metonymic relations between participants, e.g. Airbus and Airbus 380 are not considered as standing in a coparticipation relation by our system because they have different URIs. In other cases, we 36 37 We manually cleaned and reduced the actors for reasons of space. In future work, we will combine prominence with a score for the strength of the bridging. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 106/148 see more or less the opposite: a StoryLine reporting on journeys by Boeing is interrupted by a plane crash from Airbus due to overgenerated bridging relations. What is the optimal combination of features still needs to be determined empirically. Storyline Evaluation: Unsolved Issues At this stage phase we are not able to provide an extensive evaluation of the system yet. Evaluation methods for StoryLines are not trivial. Most importantly, they cannot be evaluated with respect to standard measures such as Precision and Recall. The value of a story depends a lot on the interest of a user. Evaluation of StoryLines thus should be based on relevance rather than precision and recall. In this section, we describe and propose a set of evaluation methods to be used as a standard reference method for this kind of tasks. The evaluation of a StoryLine must be based, at least, on two aspects: informativeness and interest. A good StoryLine is a StoryLine which interests the user, provides all relevant and necessary information with respect to a target entity, and is coherent. We envisage two types of evaluation: direct and indirect. Direct evaluation necessarily needs human interaction. This can be achieved in two ways: using experts and using crowdsourcing techniques. Experts can evaluate the data provided with the StoryLines with respect to a set of reference documents and check the informativeness and coherence parameters. Following Xu et al. (2013), two types of questions can be addressed at the micro-level and at the macro-level of knowledge. Both evaluation types address the quality of the generated StoryLines. The former addresses the efficiency of the StoryLines in retrieving the information while the latter addresses the quality of the StoryLines with respect to a certain topic (e.g. the commercial “war” between Boeing and Airbus). Concerning metrics, micro-knowledge can be measured by the time the users need to gather the information, while the macroknowledge can be measured as text proportion, i.e. how many sentences of the source documents composing the StoryLine are used to write a short summary. Crowdsourcing can be used to evaluate the StoryLines by means of simplified tasks. One task can ask the crowd to identify salient events in a corpus and then validate if the identified events correlate with the climax events of the StoryLines. Indirect evaluation can be based on a cross-document Summarization tasks. The ideal situation is the one in which the StoryLine contains the most salient and related events and nothing else. These data sets can be used either to recover the sentences in a collection of documents and generate an extractive summary (story) or used to produce an abstractive summary. Summarization measures such as ROUGE can then be used to evaluate the quality of summaries and, indirectly, of the StoryLines (Nguyen et al., 2014; Huang and Huang, 2013; Erkan and Radev, 2004). NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 5.2.3 107/148 Workshop on Computing News Storylines The notion of computational StoryLines for streams of news articles is new. We organised a workshop (Caselli et al., 2015b)38 at ACL in 2015 to discuss this as a new paradigm for research. The workshop brought together researchers from different communities working on representing and extracting narrative structures in news, a text genre which is highly used in NLP but which received little attention with respect to narrative structure, representation and analysis. Currently, advances in NLP technology have made it feasible to look beyond scenario-driven, atomic extraction of events from single documents and work towards extracting story structures from multiple documents, while these documents are published over time as news streams. Policy makers, NGOs, information specialists (such as journalists and librarians) and others are increasingly in need of tools that support them in finding salient stories in large amounts of information to more effectively implement policies, monitor actions of ”big players” in the society and check facts. Their tasks often revolve around reconstructing cases either with respect to specific entities (e.g. persons or organizations) or events (e.g. hurricane Katrina). We received 12 submissions and accepted 9. Overall, we had 20 participants to the workshop. The two approaches developed in NewsReader were also presented at the workshop. In final discussion additional lines of research around questions such as: • which properties do make a sequence of events a story? • how can we identify the salience of events and differentiate it from more subjective notions such as interestingness and importance? • is there a “pattern of type of events” which guides the writing of stories in news? • how can we move from an entity-based approach to StoryLines to coarser-grained representations? • what is the best granularity of representation of a StoryLine? • is a gold standard dataset feasible and useful for evaluating StoryLines? We submitted a follow-up workshop proposal for ACL2016 in Berlin. 38 https://sites.google.com/site/computingnewsstorylines2015/ NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 6 108/148 Perspectives In this section, we describe the design and first steps towards the implementation of a perspective and attribution module. The implementation so far is preliminary, and will be updated in the coming period. The attribution module described in 6.2 is currently restricted to factuality. Other attribution values will be derived in further updates. 6.1 Basic Perspective Module Events are distinguished into contextual events about the domain and source events; speech-acts and cognitive events that relate sources to the statements about the contextual events. In NewsReader, we represent both types of events as instances. They are both events involving participants and bound by time and possibly a place. The perspective layer is generated on top of this initial event representation. Perspectives are complex: they consist of what someone says (the information they choose to provide) and how he/she sais it (choices in expression, opinions and sentiment). A complete representation of perspectives thus includes the basic information provided, who it is provided by (the author or a quoted source), the certainty with which the source makes a statement, whether the source is speculating about the future, confirming or denying something, uses plain language or expressions that carry sentiment, expresses an explicit opinion, etc. Within NewsReader, we focus on identifying the source and establishing attribution values relating to factuality (how certain, future or not, and confirmed or denied) and the sentiment. This information is obtained using information from various layers of the NAF representation. Perspectives are expressed through mentions in text. We find both cases where various sources confirm a perspective and other cases where a different perspective is expressed on the same contextual statement: for example, this would be the case if one source denies a statement while another source confirms it. Moreover, the same source can express different perspectives over time. Because we (1) only aim to represent the perspective a source expresses and do not wish to make a claim about the perspective (s)he has and (2) for this reason feel that perspectives are linked to mentions, we also model perspectives at a mention level. In our current setup and model, a perspective consists of: • The instance representation of the source of a statement, e.g. the CEO of Porsche Wiedeking. • The statement made by the source in the form of triples representing an instance of a contextual event, e.g. Porsche increasing sales in 2004. • A mention of the statement (which is linked to the original statement) • The attribution values that define the relation between the source and the contextual statement: denial/confirmation, positive/negative, future/current, certain/probable/possible, etc. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 109/148 • A link to the source of the mention which can either be the document it is situated in (which in turn is linked to an author or publisher) or a quoted source. According to this specification, we abstract from how the source made the statement, e.g. saying, shouting, thinking, which is represented by the source event itself that is the basis for deriving the perspective. However, we do represent implications that can be derived from the way in which the source is made, e.g. promise to implies future, claim implies certainty, guess implies that the source thinks something is probable and hope that it is possible, therefore is positive. To derive the perspective relations between sources and the contextual statement, we implemented a perspective module that takes several layers as input: • source events with entities that have roles like prop:A0, fn:Statement@Speaker, fn:Expectation@Cognizer, fn:Opinion@Cognizer. • the contextual events that can be related to the source event through a message or topic role as defined in FrameNet or PropBank. • the attribution layer in NAF that indicates the future/non-future tense, the confirmation or denial and/or the certainty of the contextual event according to the source. • the opinion layer in NAF that indicates whether the source has a positive or negative opinion with respect to some target. • the NAF header which provides information about the magazine, publisher and author (if known). In order to combine these pieces of information, we need to intersect the layers. In this first version of the module, we use a very basic and simple algorithm to do this: 1. For all source events in the triples we check if it has a valid relation to an object of the proper semantic type such as person or organization. Valid relations are e.g. pb:A0, fn:Statement@Speaker, fn:Expectation@Cognizer, fn:Opinion@Cognizer. If not we ignore it. 2. We access the SRL layer in NAF to check if the event mention also has message role, e.g. fn:Statement@Message or fn:Statement@Topic. If no such role is found we ignore the source event. 3. We take the span of the message role and we check for any triples in the contextual data to see if they have mentions equal to or embedded within the span of the message. For all these triples, we create a perspective relation between the source and contextual triples. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 110/148 4. We check the attribution layer of each NAF file from which the source events and selected contextual events originate to see of there are attribution properties assigned to the event mention. If so, we copy them to the perspective relation. 5. We check the opinion layer in NAF to find opinion-targets. Next, we check if the spans of the above contextual triples also match or are embedded in the span of the opinion-target. If there is a match, we take over the opinion value (positive/negative) as a value for the perspective relation. The module can be adapted by defining the role constraints in the algorithm. After applying the above algorithms there will be event mentionsz with an explicit perspective value assigned to an explicit source and another set of event mentions for which no perspective is represented. The latter mentions will be assigned to the document as a source. We use the document URI as the source URI. If the event mention is within an opinion-target we represent the sentiment as a value for the attribution relation. If there is a factuality value associated with the mention, we also assign it as a value of the attribution relation. The author, publisher and magazine are meta-properties of the document URI. Finally, all event mentions are either assigned to the document using prov:wasAttributedToor to an explicit source in the document using gaf:wasAttributedTo. Let us consider the following example to illustrate the algorithm: Chrysler expects to sell 5,000 diesel Liberty SUVs, President Dieter Zetsche says at a DaimlerChrysler Innovation Symposium in New York. There are two perspectives expressed with respect to the selling of 5,000 diesel Liberty SUVs. First, the fact that this is an expectation of Chrysler and, secondly, that this is a statement made by the President Dieter Zetsche. Our text processing generates the following 3 predicate-role structures for this sentence as shown in Figure 44 starting from line 1, 26 and 50 respectively, where we abbreviated some of the lists of span elements. Next, the NAF2SEM modules generates two source events (line 1 and line 30) from this data involving the entities Chrysler (line 6) and Zetsche (line 26), as shown in Figure 45. The relations between the events and the entities are expressed starting from line 15 for expect and line 40 for say. Both source events expect and say meet the first constraints that they have an entity of the proper type with a role of the type source: fn:Expectation@Cognizer and fn:Statement@Speaker. Within the set of contextual triples, we find the event sell and its corresponding triples as shown in Figure 46. Next, we intersect the mentions of the contextual event (line 4) with the role layer to see if they can be connected to the source events in the proper way. The SRL has roles for Expectation@Topic, Statement@Topic and Statement@Message. Their spans are defined as a list of term identifiers that need to be matched with tokens that can be matched with their offsets. In this case, we can conclude that the offsets for sell, Chrysler and Livery SUVs match with these roles. Therefore, the software decides that these triples fall within the scope of the perspective. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 111/148 <!−−t 3 6 e x p e c t s : A0 [ t 3 5 C h r y s l e r ] A1 [ t 3 7 t o]−−> <p r e d i c a t e i d =”p r 6”> <!−−e x p e c t s −−> <span><t a r g e t i d =”t 3 6 ”/></span> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =” E x p e c t a t i o n ”/> <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =”O p i n i o n ”/> <e x t e r n a l R e f r e s o u r c e =”EventType ” r e f e r e n c e =” c o g n i t i o n ”/> </ e x t e r n a l R e f e r e n c e s > < r o l e i d =” r l 1 6 ” semRole=”A0”> <!−−C h r y s l e r −−> <span><t a r g e t i d =”t 3 5 ” head=” y e s ”/></span> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =”E x p e c t a t i o n @ C o g n i z e r ”/> <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =”O p i n i o n @ C o g n i z e r ”/> </ e x t e r n a l R e f e r e n c e s > </ r o l e > < r o l e i d =” r l 1 7 ” semRole=”A1”> <!−−t o s e l l 5 , 0 0 0 d i e s e l L i b e r t y SUVs−−> <span><t a r g e t i d =”t 3 7 ” head=” y e s ”/ >... < t a r g e t i d =”t 4 2 ”/></span> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =”Expectation@Phenomenon”/> <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =”E x p e c t a t i o n @ T o p i c ”/> <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =”Opinion@Topic”/> </ e x t e r n a l R e f e r e n c e s > </ r o l e > </ p r e d i c a t e > <!−−t 3 8 s e l l : A0 [ t 3 5 C h r y s l e r ] A1 [ t 3 9 5,000]−−> <p r e d i c a t e i d =”p r 7”> <!−− s e l l −−> <span><t a r g e t i d =”t 3 8 ”/></span> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =”C o m m e r c e s e l l ”/> <e x t e r n a l R e f r e s o u r c e =”ESO” r e f e r e n c e =” S e l l i n g ”/> </ e x t e r n a l R e f e r e n c e s > < r o l e i d =” r l 1 8 ” semRole=”A0”> <!−−C h r y s l e r −−> <span><t a r g e t i d =”t 3 5 ” head=” y e s ”/></span> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =” C o m m e r c e s e l l @ S e l l e r ”/> <e x t e r n a l R e f r e s o u r c e =”ESO” r e f e r e n c e =” S e l l i n g @ p o s s e s s i o n −o w n e r 1 ”/> </ e x t e r n a l R e f e r e n c e s > </ r o l e > < r o l e i d =” r l 1 9 ” semRole=”A1”> <!−−5,000 d i e s e l L i b e r t y SUVs−−> <span><t a r g e t i d =”t 3 9 ”/ >... < t a r g e t i d =”t 4 2 ” head=” y e s ”/></span> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =”VerbNet ” r e f e r e n c e =” g i v e −13.1@Theme”/> <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =”Commerce sell@Goods”/> <e x t e r n a l R e f r e s o u r c e =”PropBank ” r e f e r e n c e =” s e l l . 0 1 @1”/> </ e x t e r n a l R e f e r e n c e s > </ r o l e > </ p r e d i c a t e > <!−−t 4 7 s a y s : A1 [ t 3 5 C h r y s l e r ] A0 [ t 4 4 P r e s i d e n t ] AM−LOC [ t 4 8 a t]−−> <p r e d i c a t e i d =”p r 8”> <!−−s a y s −−> <span><t a r g e t i d =”t 4 7 ”/></span> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =”S t a t e m e n t ”/> <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =” T e x t c r e a t i o n ”/> </ e x t e r n a l R e f e r e n c e s > < r o l e i d =” r l 2 0 ” semRole=”A1”> <!−− C h r y s l e r e x p e c t s t o s e l l 5 , 0 0 0 d i e s e l L i b e r t y SUVs−−> <span><t a r g e t i d =”t 3 5 ”/> <t a r g e t i d =”t 3 6 ” head=” y e s ”/ >... < t a r g e t i d =”t 4 2 ”/></span> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =”Statement@Message”/> <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =”Statement@Topic”/> <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =” T e x t c r e a t i o n @ T e x t ”/> <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =”Choosing@Chosen”/> </ e x t e r n a l R e f e r e n c e s > </ r o l e > < r o l e i d =” r l 2 1 ” semRole=”A0”> <!−− P r e s i d e n t D i e t e r Z e t s c h e −−> <span><t a r g e t i d =”t 4 4 ”/ >... < t a r g e t i d =”t 4 6 ” head=” y e s ”/></span> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =”S t a t e m e n t @ S p e a k e r ”/> <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =”T e x t c r e a t i o n @ A u t h o r ”/> <e x t e r n a l R e f r e s o u r c e =”FrameNet ” r e f e r e n c e =”C h o o s i n g @ C o g n i z e r ”/> </ e x t e r n a l R e f e r e n c e s > </ r o l e > < r o l e i d =” r l 2 2 ” semRole=”AM−LOC”> <!−−a t a D a i m l e r C h r y s l e r I n n o v a t i o n Symposium i n New York−−> <span><t a r g e t i d =”t 4 8 ” head=” y e s ”/ >... < t a r g e t i d =”t 5 5 ”/></span> </ r o l e > </ p r e d i c a t e > Figure 44: Semantic Role elements in NAF for expect, say and sell NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 112/148 nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#e v 1 6 E x p e c t a sem : Event , f n : E x p e c t a t i o n , f n : O p i n i o n , n w r o n t o l o g y : SPEECH COGNITIVE ; rdfs : label ” expect ” ; g a f : denotedBy nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r = 2 0 8 , 2 1 5 . dbp : r e s o u r c e / C h r y s l e r rdfs : label ” C h r y s l e r ” , ” C h r y s l e r Group ” ; g a f : denotedBy nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r = 3 6 , 5 0 , nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r =740 ,748 , nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r =199 ,207 , nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r =1114 ,1122 > nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r =130 ,132 . , nwr : / d a t a / c a r s / e n t i t i e s / L i b e r t y S U V s a n w r o n t o l o g y : MISC ; rdfs : label ” Liberty suvs ” ; g a f : denotedBy nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r =237 ,249 nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#e v 1 6 E x p e c t fn : Expectation@Cognizer dbp : r e s o u r c e / C h r y s l e r ; f n : Expectation@Phenomenon nwr : / d a t a / c a r s / e n t i t i e s / L i b e r t y S U V s ; fn : Expectation@Topic nwr : / d a t a / c a r s / e n t i t i e s / L i b e r t y S U V s ; f n : Opinion@Topic nwr : / d a t a / c a r s / e n t i t i e s / L i b e r t y S U V s . . dbp : r e s o u r c e / D i e t e r Z e t s c h e rdfs : label ” Dieter Zetsche ” ; g a f : denotedBy nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r =261 ,275 . nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#ev11Say a sem : Event , n w r o n t o l o g y : SPEECH COGNITIVE , f n : S t a t e m e n t ; rdfs : label ” say ” ; g a f : denotedBy nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r =276 ,280 . nwr : / d a t a / c a r s / e n t i t i e s / D a i m l e r C h r y s l e r I n n o v a t i o n S y m p o s i u m a n w r o n t o l o g y : ORGANIZATION ; rdfs : label ” D a i m l e r C h r y s l e r I n n o v a t i o n Symposium ” ; g a f : denotedBy nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r =286 ,322 . nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#ev11Say f n : Statement@Speaker dbp : r e s o u r c e / D i e t e r Z e t s c h e ; sem : h a s P l a c e nwr : / d a t a / c a r s / e n t i t i e s / D a i m l e r C h r y s l e r I n n o v a t i o n S y m p o s i u m . Figure 45: SEM-RDF extracted from in NAF for expect, say and sell 1 2 3 4 5 6 7 8 9 10 11 12 13 nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#e v 1 7 S e l l a sem : Event , f n : C o m m e r c e s e l l , e s o : S e l l i n g ; rdfs : label ”sell” ; g a f : denotedBy nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r =219 ,223 . nwr : / d a t a / c a r s / e n t i t i e s / L i b e r t y S U V s rdfs : label ” Liberty suvs ” ; g a f : denotedBy nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r = 2 3 7 , 2 4 9 . nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#e v 1 7 S e l l e s o : p o s s e s s i o n −o w n e r 1 dbp : r e s o u r c e / C h r y s l e r ; fn : Commerce sell@Seller dbp : r e s o u r c e / C h r y s l e r ; f n : Commerce sell@Goods nwr : d a t a / c a r s / e n t i t i e s / L i b e r t y S U V s . Figure 46: SEM-RDF extracted from in NAF for expect, say and sell NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 113/148 <o p i n i o n i d =”o1”> <o p i n i o n e x p r e s s i o n p o l a r i t y =” p o s i t i v e ” s t r e n g t h =”1”> <!−− P r e s i d e n t D i e t e r Z e t s c h e s a y s a t a D a i m l e r C h r y s l e r I n n o v a t i o n Symposium i n New York .−−> <span> <t a r g e t i d =”t 4 4 ” / > . . . . <t a r g e t i d =”t 5 6 ”/> </span> </ o p i n i o n e x p r e s s i o n > </ o p i n i o n > Figure 47: Opinion element in NAF In a similar way, we check for opinions and attribution values to fill in further details of the perspective relation. For this example, we find the following opinion information as shown in Figure 47. The span of the opinion expression matches event say. From this we derive a positive value for the attitude of the source towards the triples in its scope. 6.2 Factuality module The attribution values between a source and the target events are derived from the opinion layer and the factuality layer. In this section, we describe the factuality module that was developed for NewsReader. The description of the opinion module can be found in Agerri et al. (2015). We first describe how factuality needs to be modeled within the attribution module, then describe the current implementation of the module that identifies these values and conclude this section with an outline of future work. 6.2.1 Event factuality Event factuality is a property of the events expressed in a (written or oral) text. We follow Saurı́ (2008) conception of event factuality, which is understood as the level of information expressing the commitment of relevant sources towards the factual nature of eventualities in text. That is, it is in charge of conveying whether eventualities are characterised as corresponding to a fact, to a possibility, or to a situation that does not hold in the world. The term eventualities is used here to refer to events, which can be processes or states. The main characteristics of events are that they have a temporal structure and a set of participants. Factuality is not an absolute property, but it is always relative to a source, since events are always presented from the point of view of someone. The source does not need to be the author of a text, several sources can be reporting about the same event and the same source can assign different factuality values to an event along different points in time. Additionally, we assume that factuality value assignments are made at a specific point in time. We also follow Saurı́ in considering three factuality components: source, time and factuality value. Source refers to the entity that commits to the factuality value of a certain event. Time is relevant because the factuality values of an event can change not only depending on the source but also along time. Furthermore, we assume that any statement made about the future is speculation to a certain extent. It should be noted that, NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 114/148 even though time of the statement and source are considered inherent parts of factuality components, the current implementation only focuses on factuality values and relative tense (i.e. did the source talk about the future or not). Time expressions and source identification are handled by separate components. The previous section explained how we determine the source of given information. The factuality values will be characterized across three dimensions: polarity, certainty and tense. The certainty dimension measures to which extent the source commits to the correlation of an event with a situation in the world, whereas the polarity dimension encodes whether the source is making a positive or a negative statement about an event happening in the world. The certainty dimension can be described as a continuum ranging from absolute certain to absolute uncertain. For the sake of simplicity, here we will consider it as a discrete category with three values, certain, probable and possible, following Saurı́. Polarity is a discrete category which can have two values, positive or negative. Additionally both categories have also an underspecified value, for cases in which there is not enough information to assign a value. Events will be assigned one value per dimension. Finally, our “tense” dimension simply indicates whether a statement is about the future or not. Certainty certain (CERT) probably (PROB) possible (POSS) unknown (U) Polarity positive (POS) negative (NEG) unknown (U) Tense non-future (NONFUT) future (FUT) unknown (U) Table 32: Certainty, polarity and tense values 6.2.2 Identifying factualities Factuality is often expressed by multiple elements in an expression which do not necessarily stand right next to the event the factuality values apply to. A (highly) summarized overview of potentially relevant information is provided below: • Tense, aspect • Lexical markers – Polarity markers: no, nobody, never, fail, ... . They can act at different structural levels. At the clausal level they scope over the event-referring expression; at the subclausal level they affect one of the arguments of the event; at the lexical level by means of affixes. Polarity markers can negate the predicate expressing the event, the subject, the direct or indirect object. – Modality markers (epistemic or deontic) include verbal auxiliaries, adverbials and adjectives: may, might, .perhaps, possible, likely, hopefully, hopeful, ... NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 115/148 – Commissive and volitional predicates (offered, decided) assign the value underspecified to the subordinated event. – Event selecting predicates (ESP) are predicates that select for an argument denoting an event. Syntactically they subcategorise for a that-, gerundive or infinitival clause: claim, suggest, promise, request, manage, finish, decide, offer, want, etc. ESP project factuality information on the event denoted by its argument. Depending on the type of ESP they project different factuality values. For example, prevent projects a counterfactual value, while manage projects a factual value. Some ESP such as claim, are special in that they assign a factuality value to the subcategorised event and at the same time they express who is the source that commits to that factuality value, without the author of the statement committing to it. ESP can be source (SIP) or non-source introducing predicates (NSIP). SIP such as suspect or know are ESP that contribute an additional source relative to which the factuality of the subcategorised event is assessed. • Some syntactic constructions can introduce a factuality value. – In participial adverbial clauses, the event in the subordinated clause is presupposed as true (e.g. Having won Slovenia’s elections, political newcomer Miro Cerar will have to make tough decisions if he is to bring stability to a new government). – In purpose clauses, the main event is presented as underspecifed (e.g. Government mobilizes public and private powers to solve unemployment in the country). – In conditional constructions, the factuality value of the main event in the consequent clause is dependent on the factuality of the main event in the antecedent clause (e.g. If this sentence is not true then it is true). – In temporal clauses, the event is presupposed to be certain in most cases (e.g. While the main building was closed for renovation, the architects completed the Asian Pavilion). Additionally, some syntactic constructions act as scope islands, which means that the events in that construction cannot be affected by markers which are outside the constructions and at the same time the markers in the construction cannot scope over the events which are outside the construction. – Non-restrictive relative clauses (e.g. The new law might affect the house owners who bought their house before 2002 ). – Cleft sentences (e.g. It could have been the owner who replaced the main entrance door ). NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 116/148 An event can be within the scope of one or more lexical markers, which means that in order to compute its factuality, the values of all the markers that scope over it have to be considered, as well as the ways in which markers interact with each other. It is also necessary to take into account that some syntactic contexts establish scope boundaries. The next subsection describes our current factuality module and explains how relevant factors are integrated in the model. 6.2.3 Factuality module The factuality module consists of two main components. The first component is a machine learning component that establishes the certainty and polarity of the event based on a model trained on FactBank. The second component provides a rule-based interpretation of the tense of the event indicating whether it is future, past or present or unknown. The rule based component simply checks the explicit tense marking on the verb or, in case of a nominal event, the tense marking on the verb that governs it. If no tense values are found this way, the value is set to ‘unknown’. The machine learning component improves on the previously existing module in three ways. FactBank contains several layers of annotation for factuality values. The factuality values assigned by the direct source making the statement and all factuality values assigned by any other source. This means that when someone is quoted in an article, the factuality values from the quoted source are provided as well as the factuality values of the author of the article. Because the author typically does not provide explicit indication of whether he or she agrees with their source, the values associated with the author tend to be underspecified for both polarity and certainty. As a result, factuality values at the highest level (those attributed to the author) are almost exclusively certain-positive and underspecifiedunderspecified. The first difference between the new module and the old one is that we train on the most embedded layer. This provides more variety in factuality values. The second difference is that the values are translated from the joint certainty-polarity values found in FactBank to the individual dimensions used in NewsReader annotations. This allows us to experiment with training each dimension separately. The third difference is that we use a much more elaborate set of features based on the relevant elements outlined above. The following features are used in the current system: • lemma and surface form of event and words in direct context • lemma of the event’s head • dependency relation • POS and morphological information about the event and the head word • dependency chain NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 117/148 < f a c t u a l i t y i d =” f 1 ”> <span> <t a r g e t i d =”t 3 1 0 ”/> </span> <f a c t V a l r e s o u r c e =”nwr : a t t r i b u t i o n T e n s e ” v a l u e =”UNDERSPECIFIED”/> <f a c t V a l r e s o u r c e =” f a c t b a n k ” v a l u e =”NONE”/> <f a c t V a l r e s o u r c e =”nwr : a t t r i b u t i o n C e r t a i n t y ” v a l u e =”CERTAIN”/> <f a c t V a l r e s o u r c e =”nwr : a t t r i b u t i o n P o l a r i t y ” v a l u e =”POS”/> </ f a c t u a l i t y > < f a c t u a l i t y i d =” f 2 ”> <span> <t a r g e t i d =”t 2 1 4 ”/> </span> <f a c t V a l r e s o u r c e =”nwr : a t t r i b u t i o n T e n s e ” v a l u e =”NON FUTURE”/> <f a c t V a l r e s o u r c e =” f a c t b a n k ” v a l u e =”CT+”/> <f a c t V a l r e s o u r c e =”nwr : a t t r i b u t i o n C e r t a i n t y ” v a l u e =”CERTAIN”/> <f a c t V a l r e s o u r c e =”nwr : a t t r i b u t i o n P o l a r i t y ” v a l u e =”POS”/> </ f a c t u a l i t y > < f a c t u a l i t y i d =” f 2 0 ”> <span> <t a r g e t i d =”t 8 1 ”/></span> <f a c t V a l r e s o u r c e =”nwr : a t t r i b u t i o n T e n s e ” v a l u e =”FUTURE”/> <f a c t V a l r e s o u r c e =” f a c t b a n k ” v a l u e =”CT+”/> <f a c t V a l r e s o u r c e =”nwr : a t t r i b u t i o n C e r t a i n t y ” v a l u e =”CERTAIN”/> <f a c t V a l r e s o u r c e =”nwr : a t t r i b u t i o n P o l a r i t y ” v a l u e =”POS”/> </ f a c t u a l i t y > Figure 48: NAF examples for factuality It should furthermore be noted that the system was trained on the events that the NewsReader pipeline identified in the FactBank data. Events that are not marked as such in FactBank receive ’NONE’ as a value, since the gold values for these are unknown. These events form a significant part of the data and the ’NONE’ value is regularly found by the classifier. We add the interpretation certain and positive to these events, because this combination forms a strong majority class. Our pipeline identifies more nominal event references than found in FactBank and the majority class is even stronger for these nominal references, which further justifies this decision. The output does still indicate that the original value found by the classifier was ‘NONE’, so that one can distinguish between these default interpretations and cases where the value was assigned by the classifier. Figure 48 shows some output examples. The span elements need to be matched with the event spans to decide on the events to which the factuality applies. 6.2.4 Future work The module described above covers all main factuality values we are interested in and takes the most relevant information that can influence factuality into account. It thus forms a solid basis for investigating factuality detection in text. It is, however, only the first version of this new approach. In future work, we will work in two directions to improve the model. First, we will experiment with various forms of training data. This involves not considering events that are not in FactBank and thus avoiding the ‘NONE’ class as well as extending the set so that the machine learner can also identify tense features.39 Second, we will experiment with a variety of features aiming for information that is more specifically linked to factuality markers. Currently, the factuality module only handles English. In order to adapt the module to other languages new factuality lexicons will be needed and the syntactic rules to find 39 Note that excluding the NONE-class needs to be taken into account when evaluating on FactBank: for events that are identified by the pipeline, but that are not in FactBank, we have no idea how well the classification behaves. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 118/148 nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#e v 1 6 E x p e c t rdfs : label ” expect ” ; g a f : denotedBy nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r = 2 0 8 , 2 1 5 . fn : Expectation@Cognizer dbp : r e s o u r c e / C h r y s l e r . nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#ev11Say rdfs : label ” say ” ; g a f : denotedBy nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r =276 ,280 f n : Statement@Speaker dbp : r e s o u r c e / D i e t e r Z e t s c h e ; sem : h a s P l a c e nwr : / d a t a / c a r s / e n t i t i e s / D a i m l e r C h r y s l e r I n n o v a t i o n S y m p o s i u m . nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#e v 1 7 S e l l rdfs : label ”sell” ; g a f : denotedBy nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r =219 ,223 dbp : r e s o u r c e / C h r y s l e r ; fn : Commerce sell@Seller f n : Commerce sell@Goods nwr : d a t a / c a r s / e n t i t i e s / L i b e r t y S U V s . . . Figure 49: Simplified SEM-RDF triples for events as named-graphs the scopes of factuality markers will have to be adapted. The lexicons can be probably translated from the English lexicons and manually revised by a linguistics expert in order to check whether the same factuality behaviour applies. More costly might be to adapt the syntactic rules, though part of the cost can be reduced by preprocessing the documents with parsers that have models for several languages (such as MaltParser). We will make a predevelopment analysis of the cost of adapting the processor to other languages in order to design it in such a way that the cost can be maximally reduced. 6.3 A perspective model The previous sections described how we extract the source of events and how we identify factuality attributes of events. In this subsection, we explain how we combine this information to represent perspectives in RDF. As mentioned, we associate perspective information with mentions. After all, people can change their perspective so that a single source may make incompatible statements about an event. To capture this fact, we represent all information related to perspectives (for now: source, factuality and sentiment) in the mention layer. The triples below illustrate how this looks like for the example from Section 6.1, repeated here for convenience: Chrysler expects to sell 5,000 diesel Liberty SUVs, President Dieter Zetsche says at a DaimlerChrysler Innovation Symposium in New York. Consider the following triples associated with these statements in Figure 49 for the events expect, say and sell, which is a reduced representation according to the SEM-RDF format. They represent the labels, the SEM relations and the mentions of the three events. We know from our perspective interpretation algorithm that the source of the statement about selling SUVs (line 13) is Chrysler. Chrysler introduced the statement through its expectation (line 7, 8 and 9). The source of this expectation is Dieter Zetsche (lines 1 till 4). The factuality module should tell us that Zetsche’s statement (saying) is CERTAIN, POSITIVE and NON-FUTURE according to the article. Zetsche assigns the same factuality values to Chrysler’s expecting SUV sales. On the other hand, the selling is NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 119/148 # meta d a t a p r o p e r t i e s on t h e document : a u t h o r , magazine , p u b l i s h e r nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml p r o v : w a s A t t r i b u t e d T o <h t t p : / /www. n e w s r e a d e r −p r o j e c t . eu / p r o v e n a n c e / m a g a z i n e / autoweek . com> . #a t t r i b u t i o n : ev11Say of Dieter Zetsche g a f : denotedBy says nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r =276 ,280 nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r =276 ,280 gaf : hasAttribution nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . x m l d A t t r 1 . . nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . x m l d A t t r 1 rdf : value g a f : CERTAIN NON−FUTURE POS ; prov : wasAttributedTo nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml . #a t t r i b u t i o n o f C h r y s l e r e x p e c t s : e v 1 6 E x p e c t g a f : denotedBy nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r =208 ,215 . nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r =208 ,215 gaf : generatedBy nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r =276 ,280 gaf : hasAttribution nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . x m l s A t t r 2 nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . x m l s A t t r 2 rdf : value g a f : CERTAIN NON−FUTURE POS g a f : wasAttributedTo dbp : r e s o u r c e / D i e t e r Z e t s c h e . ; . ; #a t t r i b u t i o n o f s e l l : e v S e l l g a f : denotedBy nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r = 2 1 9 , 2 2 3 . nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r =219 ,223 gaf : hasAttribution nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . x m l s A t t 3 ; gaf : generatedBy nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . xml#c h a r =208 ,215 nwr : / d a t a / c a r s / 2 0 0 3 / 0 1 / 0 1 / 4 7VH−FG30−010D−Y3YG . x m l s A t t 3 rdf : value g a f : PROBABLE FUTURE POS, g a f : wasAttributedTo dbp : r e s o u r c e / C h r y s l e r <p o s i t i v e > . . ; Figure 50: Perspective RDF triples for event mentions PROBABLE, POSITIVE and FUTURE (according to Chrysler). In principle, each statement comes from a source and has factuality values.40 In some cases, we have multiple pieces of information that constitute the source (e.g. author, magazine and publisher of an article). In order to avoid massive multiplications of triples, we attribute statements and factuality values that come directly from the article to the article itself. In turn, the article can be attributed to an author, publisher or magazine. This way, information is not repeated for each mention. Figure 50 shows the output triples for the provenance and factuality of the statements about Zetsche and Chrysler from our example, attached to the mentions of the events given in Figure 49. Starting with line 1, we first give the document source properties through the prov:wasAttributed predicate. The sentence originates from the autoweek.com website which is given as the magazine of the source text. The source text is represented through its NewsReader URL. Next, we find attribution relations for each event mention. For convenience, we repeat the gaf:denotedBy triple that links the mention to the event URI: lines 6, 16, 27. Each mention is then associated with specific factuality values. These factuality values also come from the source of the mention. We therefore use an intermediate node related to rdf:value to model factuality and sentiment values and provenance of the statement. This means that each event mention is linked to its own attribution node. This attribution node is linked to the factuality/sentiment values of the 40 Note that unknown factuality values are also factuality values. For instance, the sentence I do not know whether Zetsche said that forms a case where polarity and certainty are ’unknown’ by the source. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 120/148 statement and to the source of the mention. Specifically, line 9 shows the attribution triple for say, line 20 for expect and line 30 for sell. The attribution itself is a URI for which we can define any set of properties. For example, we can see in lines 11, 12, 13 that :dAttr1 has rdf:value gaf:CERTAIN NONFUTURE POS and that it is attributed (gaf:hasAttribution) to the document. The document properties then point to the magazine autoweek.com. We can also see (lines 18, 19, 20) that :sAttr2 has the same rdf:value but is assigned to the source Dieter Zetsche. Finally, :sAtt3 is attributed to the source Chrysler, has factuality values gaf:PROBABLE FUTURE POS and the sentiment value positive. Note that we distinguish between attribution to the source that printed the statement from attribution to a source mentioned in text, where we use prov:wasAttributedTo to indicate the provenance of the article and gaf:wasAttributedTo to indicate the provenance of a quoted source. The attribution relation defined in the grounded annotation framework establishes the same relation between subject and object as the original relation from the PROV-O (Moreau et al., 2012). The difference is that in the case of prov:wasAttributedTo, we are modelling the fact that we pulled the information from a specific source. In the case of gaf:wasAttributedTo we are modelling the fact that the information is attributed to a specific source by someone else. The factuality component indicates which factuality value is associated with the event or statement. We use composed values that contain the three factuality elements described above: certainty, polarity and tense. The ontology defines each of these values separately providing the components of the complex values used in the example above (e.g. CERTAIN NON FUTURE POS has properties certainty CERTAIN, polarity POSITIVE and tense NON-FUTURE). As the information on perspectives grows by adding information such as specific emotions, the ontology will be extended to contain not only these new values, but also more combined values so that we can continue using the compact perspective representation that is presented here. However, the model is flexible enough to also allow separated values for different perspective values, so that we can separate sentiment and emotion from factuality in our representation. Finally, the current implementation creates a perspective RDF file in addition to the SEM RDF output. To connect the mentions of events in the perspective RDF with the event instances in the SEM RDF, a query needs to be formulated that matches the perspective mentions with the gaf:denotedBy triples in SEM RDF. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 7 121/148 Cross-lingual extraction The processing of text in NewsReader takes place in two steps. First, language-specific processors interpret text in terms of mentions of entities, events, time-expressions and relations between them. Secondly, we create a language-independent representation of the instances of entities, events, time-expressions and their relations in RDF-TRiG, as shown schematically in Figure 51: Figure 51: Semantic interoperability across NAF representations Although the first step is specific for each language, the representation of the output in NAF is the same for all the 4 NewsReader languages. Differences are restricted to the tokens that make up the mentions of things. In the case of entities, we normalize these tokens by using the DBpedia URIs, where Dutch, Spanish and Italian DBpedia URIs are mapped to English URIs. In the case of time-expressions, such as Thursday and yesterday, we interpret them to ISO dates. Events are however still represented by the tokens of the predicates. To make the events interoperable across languages, we use the GlobalWordnetGrid to map each predicate to concepts in the InterLingualIndex (ILI, (Vossen et al., 2016; Bond et al., 2016)). As explained in section 2, coreference and therefore identity of events is defined as a function of the identity of its components: the similarity of the action, the participants, the place and time. The latter 3 are semantically interoperable across the NAF representations in the four languages. The events can be compared through their ILI mapping based on the wordnets in their language, as we have been doing so far for English as well. In Figures 52 and 53, we show two fragments in English and Spanish respectively for the same Wikinews article with the representation of an entity, a predicate with roles and a time-expression. The English representation of the entity Boeing 787 has an external reference to the English DBpedia, where the Spanish entity maps to both the Spanish and English DBpedia entries. By taking the English reference for Spanish, we can map both entity references to each other across the documents. We can see the same for the time-expressions in both languages that are normalized to the same value: 2005-01-29. In NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 122/148 case of the predicates, both examples show external references to various ontologies among which FrameNet and ESO and also to WordNet ILI-concepts. The predicates of these two examples can therefore be mapped into each other through the reference ili-30-02207206-v. A similar example is shown in Figure 54 for Dutch NAF where the verb kopen is mapped to the same ILI record and in 55 for the Italian NAF in which acquisterá is mapped to the same concept. To compare the capacity of the different language pipelines to extract the same information, we use the NewsReader MEANTIME corpus (van Erp et al. (2015)). The MEANTIME corpus consists of 120 English news articles taken from Wikinews on 4 topics: Airbus, Apple, automotive industry (GM, Chrysler and Ford), stock market news. Each topic has 30 articles. We translated the 120 English articles to Spanish, Dutch and Italian. The texts in all languages have been annotated according to the same NewsReader annotation scheme. We processed the English, Spanish and Dutch Wikinews articles through the respective pipelines, as described in Agerri et al. (2015). After generating the NAF files, we applied the NAF2SEM process to each language data set. We made a small modification to the original algorithm that compares events on the basis of their lemmas. To be able to compare events across languages, we used the ILI-concept references of the predicates to represent events. Spanish, Italian and Dutch events can thus be mapped to English events provided they were linked to the same concept.41 We generated RDF-TRiG files for the 4 Wikinews corpora for English, Spanish, Italian and Dutch. We implemented a triple counter to compare the data created for each language data set. The triple counter generates statistics for the following information: 1. Entities with a DBpedia URI with the number of mentions 2. Entities without a DBpedia URI with the number of mentions 3. Events represented by their ILI-reference with the number of mentions 4. Frequency count of the roles relating events and entities 5. Frequency counts of the triples relating events with ILI-references and entities through these roles We compared the Spanish, Italian and Dutch data against the English data, calculating the coverage of the English mentions by the other languages. Note that we cannot calculate recall through this method. If the software detects a DBpedia entity E in English in sentence A and not in the translated Spanish sentence T(A) but the same entity E is detected in another Spanish sentence T(B) while it is not detected in the English sentence B, then our current method counts 1 mention in English and 1 mention in Spanish with 100% coverage. This may still count as coverage but not as recall because we do not know 41 This also means that we lump together event-instances that are normally kept separate because they do not share the same time and participant. The comparison therefore does not tell us anything about the precise cross-lingual event-coreference but merely gives a rough indication NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 123/148 <e n t i t y i d =”e 1 ” t y p e=”MISC”> <r e f e r e n c e s > <!−−B o e i n g 787−−> <span><t a r g e t i d =”t 8”/>< t a r g e t i d =”t 9 ”/></span> </ r e f e r e n c e s > <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f c o n f i d e n c e =”1.0” r e f e r e n c e =”h t t p : / DBpedia . o r g / r e s o u r c e / B o e i n g 7 8 7 D r e a m l i n e r ” r e f t y p e =”en ” r e s o u r c e =” s p o t l i g h t v 1 ””/> </ e x t e r n a l R e f > </ e x t e r n a l R e f e r e n c e s > </ e n t i t y > <p r e d i c a t e i d =”p r 6”> <!−−p u r c h a s e −−> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e f e r e n c e =” p u r c h a s e . 0 1 ” r e s o u r c e =”PropBank”/> <e x t e r n a l R e f r e f e r e n c e =”o b t a i n − 1 3 . 5 . 2 ” r e s o u r c e =”VerbNet”/> <e x t e r n a l R e f r e f e r e n c e =”o b t a i n −13.5.2 −1” r e s o u r c e =”VerbNet”/> <e x t e r n a l R e f r e f e r e n c e =”Commerce buy ” r e s o u r c e =”FrameNet”/> <e x t e r n a l R e f r e f e r e n c e =” p u r c h a s e . 0 1 ” r e s o u r c e =”PropBank”/> <e x t e r n a l R e f r e f e r e n c e =”Buying ” r e s o u r c e =”ESO”/> <e x t e r n a l R e f r e f e r e n c e =” c o n t e x t u a l ” r e s o u r c e =”EventType”/> <e x t e r n a l R e f r e f e r e n c e =” i l i −30−02207206−v ” r e s o u r c e =”WordNet”/> </ e x t e r n a l R e f e r e n c e s > <span><t a r g e t i d =”t 2 8 ”/></span> < r o l e i d =” r l 9 ” semRole=”A0”> <!−− O f f i c i a l s from t h e P e o p l e ’ s R e p u b l i c o f China−−> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e f e r e n c e =”o b t a i n − 1 3 . 5 . 2 @Agent” r e s o u r c e =”VerbNet”/> <e x t e r n a l R e f r e f e r e n c e =”Commerce buy@Buyer ” r e s o u r c e =”FrameNet”/> <e x t e r n a l R e f r e f e r e n c e =” p u r c h a s e . 0 1 @0” r e s o u r c e =”PropBank”/> <e x t e r n a l R e f r e f e r e n c e =”B u y i n g @ p o s s e s s i o n −o w n e r 2 ” r e s o u r c e =”ESO”/> </ e x t e r n a l R e f e r e n c e s > <span> <t a r g e t head=” y e s ” i d =”t 1 7 ”/ >... < t a r g e t i d =”t 2 4 ”/></span> </ r o l e > < r o l e i d =” r l 1 0 ” semRole=”A1”> <!−−60 B o e i n g 787 D r e a m l i n e r a i r c r a f t −−> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e f e r e n c e =”o b t a i n − 1 3 . 5 . 2 @Theme” r e s o u r c e =”VerbNet”/> <e x t e r n a l R e f r e f e r e n c e =”Commerce buy@Goods ” r e s o u r c e =”FrameNet”/> <e x t e r n a l R e f r e f e r e n c e =” p u r c h a s e . 0 1 @1” r e s o u r c e =”PropBank”/> </ e x t e r n a l R e f e r e n c e s > <span><t a r g e t i d =”t 2 9 ” / > . . . <t a r g e t head=” y e s ” i d =”t 3 3 ”/></span> </ r o l e > < r o l e i d =” r l 1 1 ” semRole=”AM−LOC”> <!−− i n a d e a l worth US$ 7 . 2 bn−−> <span> <t a r g e t head=” y e s ” i d =”t 3 4 ”/ >... < t a r g e t i d =”t 4 0 ”/> </span> </ r o l e > </ p r e d i c a t e > <t i m e x 3 i d =”tmx5” t y p e=”DATE” v a l u e =”2005−01−29”> <!−−today−−> <span><t a r g e t i d =”w183”/> </span> </timex3> Figure 52: Example of representation of entities, events and roles from an English Wikinews fragment NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 124/148 <e n t i t y i d =”e 1 ” t y p e=”MISC”> <r e f e r e n c e s > <span> <!−−B o e i n g 787−−> <t a r g e t i d =”t 8 ”/> <t a r g e t i d =”t 9 ”/> </span> </ r e f e r e n c e s > <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f c o n f i d e n c e =”1.0” r e f e r e n c e =”h t t p : / e s . DBpedia . o r g / r e s o u r c e / B o e i n g 7 8 7 ” r e f t y p e =” e s ” r e s o u r c e =” s p o t l i g h t v 1 ”> <e x t e r n a l R e f c o n f i d e n c e =”1.0” r e f e r e n c e =”h t t p : / DBpedia . o r g / r e s o u r c e / B o e i n g 7 8 7 D r e a m l i n e r ” r e f t y p e =”en ” r e s o u r c e =” w i k i p e d i a −db−esEn”/> </ e x t e r n a l R e f > </ e x t e r n a l R e f e r e n c e s > </ e n t i t y > <p r e d i c a t e i d =”p r 3”> <!−−comprar−−> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e f e r e n c e =”comprar . 1 . b e n e f a c t i v e ” r e s o u r c e =”AnCora”/> <e x t e r n a l R e f r e f e r e n c e =”g e t − 1 3 . 5 . 1 ” r e s o u r c e =”VerbNet”/> <e x t e r n a l R e f r e f e r e n c e =”o b t a i n − 1 3 . 5 . 2 ” r e s o u r c e =”VerbNet”/> <e x t e r n a l R e f r e f e r e n c e =”o b t a i n −13.5.2 −1” r e s o u r c e =”VerbNet”/> <e x t e r n a l R e f r e f e r e n c e =”Commerce buy ” r e s o u r c e =”FrameNet”/> <e x t e r n a l R e f r e f e r e n c e =”buy . 0 1 ” r e s o u r c e =”PropBank”/> <e x t e r n a l R e f r e f e r e n c e =” p u r c h a s e . 0 1 ” r e s o u r c e =”PropBank”/> <e x t e r n a l R e f r e f e r e n c e =”Buying ” r e s o u r c e =”ESO”/> <e x t e r n a l R e f r e f e r e n c e =” c o n t e x t u a l ” r e s o u r c e =”EventType”/> <e x t e r n a l R e f r e f e r e n c e =” i l i −30−02207206−v ” r e s o u r c e =”WordNet”/> <e x t e r n a l R e f r e f e r e n c e =” i l i −30−02646757−v ” r e s o u r c e =”WordNet”/> </ e x t e r n a l R e f e r e n c e s > <span> <t a r g e t i d =”t 3 4 ”/> </span> < r o l e i d =” r l 5 ” semRole=”a r g 1”> <!−−60 a v i o n e s B o e i n g 787 D r e a m l i n e r −−> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e f e r e n c e =”g e t − 1 3 . 5 . 1 @Theme” r e s o u r c e =”VerbNet”/> <e x t e r n a l R e f r e f e r e n c e =”o b t a i n − 1 3 . 5 . 2 @Theme” r e s o u r c e =”VerbNet”/> <e x t e r n a l R e f r e f e r e n c e =”Commerce buy@Goods ” r e s o u r c e =”FrameNet”/> <e x t e r n a l R e f r e f e r e n c e =”buy . 0 1 @1” r e s o u r c e =”PropBank”/> <e x t e r n a l R e f r e f e r e n c e =” p u r c h a s e . 0 1 @1” r e s o u r c e =”PropBank”/> </ e x t e r n a l R e f e r e n c e s > <span> <t a r g e t i d =”t 3 5 ”/> <t a r g e t head=” y e s ” i d =”t 3 6 ”/ >... < t a r g e t i d =”t 3 9 ”/> </span> </ r o l e > < r o l e i d =” r l 6 ” semRole=”argM”> <!−−en un a c u e r d o p o r v a l o r de 7 . 2 0 0 m i l l o n e s de US\$−−> <span> <t a r g e t head=” y e s ” i d =”t 4 0 ”/ >... < t a r g e t i d =”t 4 9 ”/> </span> </ r o l e > </ p r e d i c a t e > <t i m e x 3 i d =”t x 2 ” t y p e=”DATE” v a l u e =”2005−01−29”> <!−−29 de e n e r o d e l 2005−−> <span> <t a r g e t i d =”w20 ” / > . . . <t a r g e t i d =”w24”/> </span> </timex3> Figure 53: Example of representation of entities, events and roles from a Spanish Wikinews fragment NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 125/148 <e n t i t y i d =”e 2 ” t y p e=”MISC”> <r e f e r e n c e s > <span> <!−−A i r b u s A320−−> <t a r g e t i d =” t 1 3 ”/> <t a r g e t i d =” t 1 4 ”/> </span> </ r e f e r e n c e s > <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f c o n f i d e n c e =”1.0” r e f e r e n c e =”h t t p : / n l . DBpedia . o r g / r e s o u r c e / A i r b u s A 3 2 0 ” r e f t y p e =” n l ” r e s o u r c e =” s p o t l i g h t v 1 ”> <e x t e r n a l R e f c o n f i d e n c e =”1.0” r e f e r e n c e =”h t t p : / DBpedia . o r g / r e s o u r c e / A i r b u s A 3 2 0 f a m i l y ” r e f t y p e =”en ” r e s o u r c e =” w i k i p e d i a −db−nlEn”/> </ e x t e r n a l R e f > </ e x t e r n a l R e f e r e n c e s > </ e n t i t y > <p r e d i c a t e i d =”p r 6”> <!−−kopen−−> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e f e r e n c e =”r v −4101” r e s o u r c e =”C o r n e t t o ”/> <e x t e r n a l R e f r e f e r e n c e =” i l i −30−02207206−v ” r e s o u r c e =”WordNet”/> </ e x t e r n a l R e f e r e n c e s > <span> <t a r g e t i d =” t 1 7 ”/> </span> < r o l e i d =”r 8 ” semRole=”Arg1”> <!−− t w i n t i g A i r b u s A320 p a s s a g i e r s v l i e g t u i g e n −−> <span> <t a r g e t i d =” t 1 2 ” / > . . . <t a r g e t head=” y e s ” i d =” t 1 5 ”/> </span> </ r o l e > < r o l e i d =”r 1 0 ” semRole=”ArgM−PNC”> <!−−v o o r een−−> <span> <t a r g e t head=” y e s ” i d =” t 1 8 ”/> <t a r g e t i d =” t 1 9 ”/> </span> </ r o l e > </ p r e d i c a t e > <t i m e x 3 i d =”tmx3” t y p e=”DATE” v a l u e =”2009−06−18”> <!−−donderdag−−> <span> <t a r g e t i d =”w5”/> </span> </timex3> Figure 54: Example of representation of entities, events and roles from a Dutch Wikinews fragment NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 126/148 <e n t i t y i d =”e 4 ” t y p e=”ORGANIZATION”> <r e f e r e n c e s > <!−−A320−−> <span> <t a r g e t i d =”t 9 ” /> </span> </ r e f e r e n c e s > <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =” s p o t l i g h t v 1 ” r e f e r e n c e =”h t t p : / i t . d b p e d i a . o r g / r e s o u r c e / A i r b u s A 3 2 0 f a m i l y ” c o n f i d e n c e =”1.0” r e f t y p e =” i t ” s o u r c e =” i t ”> <e x t e r n a l R e f r e s o u r c e =” w i k i p e d i a −db−i t E n ” r e f e r e n c e =”h t t p : / d b p e d i a . o r g / r e s o u r c e / A i r b u s A 3 2 0 f a m i l y ” c o n f i d e n c e =”1.0” r e f t y p e =”en ” s o u r c e =” i t ” /> </ e x t e r n a l R e f > </ e x t e r n a l R e f e r e n c e s > </ e n t i t y > <!−−t 4 a c q u i s t e r \ ’ { a } : A0 [ t 1 China ] A1 [ t 5 v e n t i ]−−> <p r e d i c a t e i d =”p r 1”> <!−− a c q u i s t e r \ ’ { a}−−> <span> <t a r g e t i d =”t 4 ” /> </span> <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =”EventType ” r e f e r e n c e =”OCCURRENCE” /> <e x t e r n a l R e f r e s o u r c e =”PropBank ” r e f e r e n c e =” a c q u i s t a r e . 0 1 ” /> <e x t e r n a l R e f r e s o u r c e =”WordNet” r e f e r e n c e =” i l i −30−00079018−n” /> <e x t e r n a l R e f r e s o u r c e =”WordNet” r e f e r e n c e =” i l i −30−02207206−v ” /> </ e x t e r n a l R e f e r e n c e s > < r o l e i d =” r l 1 ” semRole=”A0”> <!−−China E a s t e r n A i r l i n e s −−> <span> <t a r g e t i d =”t 1 ” head=” y e s ” /> <t a r g e t i d =”t 2 ” /> <t a r g e t i d =”t 3 ” /> </span> </ r o l e > < r o l e i d =” r l 2 ” semRole=”A1”> <!−− v e n t i n u o v i j e t A i r b u s A320−−> <span> <t a r g e t i d =”t 5 ” /> <t a r g e t i d =”t 6 ” /> <t a r g e t i d =”t 7 ” head=” y e s ” /> <t a r g e t i d =”t 8 ” /> <t a r g e t i d =”t 9 ” /> </span> </ r o l e > </ p r e d i c a t e > <t i m e x 3 i d =”tmx1” t y p e=”DATE” v a l u e =”2009−06−18”> <!−−18 g i u g n o 2009−−> <span> <t a r g e t i d =”w10” /> <t a r g e t i d =”w11” /> <t a r g e t i d =”w12” /> </span> </timex3> Figure 55: Example of representation of entities, events and roles from an Italian Wikinews fragment NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 127/148 whether the Spanish mention in T(B) corresponds with the English mention in A.42 The coverage scores we give below are thus just a rough approximation. A true comparison requires a cross-lingual annotation that also aligns the mentions of entities and events with respect to the same instances. The NAF2SEM module is ignorant of the language of the NAF file. This means that we can process English, Spanish, Italian and Dutch NAF files as if they are different sources, just as we processed multiple English NAF files. The module will merge all entities and events according to the implemented heuristics and generate a single RDFTRiG for all the sources. In the ideal case, the same data should be extracted across the languages as would be extracted for any of the languages separately due to their translation relation. Therefore, merging the NAF files across languages should result in exactly the same numbers of entities, events and triples. Merging the data set can be seen as an extreme test for the cross-lingual extraction and compatibility of the different NLPpipelines. Not all extracted information can be compared. Events that are not mapped to WordNet concepts (ILI-records) are represented by their labels, which are different across the languages. The same holds for so-called dark entities that are not mapped to DBpedia. They are represented by their linguistic form which is usually different. For non-entities, i.e. expressions not detected as entities but that play an important role in the event, string matches across languages are very unlikely. In Figures 56, we show the RDF-TRiG result of merging English, Spanish, Italian and Dutch NAF files for some entities. The entity Airbus was found in NAF files for 3 languages with 6 mentions in the English source, 7 mentions in the Spanish source and 4 mentions in the Dutch source. The entity Airbus A380 on the other hand was only detected by the Italian (7) and Dutch (21 mentions) pipelines. In Figure 57, we see events detected across the languages. The first event, represented through the ILI ili-30-00634472-v, is matched across all 4 languages, the other events across different subsets of languages. In the next subsections, we show the statistics for the 4 corpora and 4 languages. We also provide some statistics for the merging of data against the English results. 7.1 Crosslingual extraction of entities In Table 33, we give the totals of DBpedia entities extracted for English and the other languages: Spanish, Italian and Dutch. For English, we give the unique instances (I) for each corpus and the mentions (M). For the other languages, we give the same but in addition the overlapping mentions (O) and the macro (per document) and micro average of coverage. The English figures show that there is some variation across the 4 corpora. The stock market corpus contains only few instances and mentions and the airbus and gm corpus contain most instances and mentions. Ratios between instances and mentions differ slightly across the corpora. Spanish has a similar amount and ratio for airbus but different 42 In some cases a single English sentence has been translated in more than one sentence in another language NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 128/148 <h t t p : / DBpedia . o r g / r e s o u r c e / A i r b u s> rdfs : label ” Airbus ” , ” Airbus ,” ; g a f : denotedBy nwr : dutch−w i k i n e w s / 1 8 1 6 A i r b u s w i n s Q a t a r A i r w a y s o r d e r w o r t h 1 5 b n#c h a r =31 ,37 , nwr : dutch−w i k i n e w s /816 A i r b u s w i n s Q a t a r A i r w a y s o r d e r w o r t h 1 5 b n#c h a r =564 ,570 , nwr : dutch−w i k i n e w s / 1 8 1 6 A i r b u s w i n s Q a t a r A i r w a y s o r d e r w o r t h 1 5 b n#c h a r =655 ,661 , nwr : dutch−w i k i n e w s / 1 8 1 6 A i r b u s w i n s Q a t a r A i r w a y s o r d e r w o r t h 1 5 b n#c h a r =911 ,917 , nwr : s p a n i s h −w i k i n e w s / 1 8 1 6 A i r b u s w i n s#c h a r =0 ,6 , nwr : s p a n i s h −w i k i n e w s / 1 8 1 6 A i r b u s w i n s#c h a r =615 ,621 , nwr : s p a n i s h −w i k i n e w s / 1 8 1 6 A i r b u s w i n s#c h a r =716 ,722 , nwr : s p a n i s h −w i k i e n e w s / 1 8 1 6 A i r b u s w i n s#c h a r =945 ,951 , nwr : s p a n i s h −w i k i e n e w s / 1 8 1 6 A i r b u s w i n s#c h a r =450 ,456 , nwr : s p a n i s h −w i k i e n e w s / 1 8 1 6 A i r b u s w i n s#c h a r =381 ,387 , nwr : s p a n i s h −w i k i e n e w s / 1 8 1 6 A i r b u s w i n s#c h a r =135 ,141 , nwr : e n g l i s h −w i k i e n e w s / A i r b u s w i n s Q a t a r A i r w a y s o r d e r w o r t h $ 1 5 b n#c h a r =1 ,7 , nwr : e n g l i s h −w i k i e n e w s / A i r b u s w i n s Q a t a r A i r w a y s o r d e r w o r t h $ 1 5 b n#c h a r =93 ,100 , nwr : e n g l i s h −w i k i e n e w s / A i r b u s w i n s Q a t a r A i r w a y s o r d e r w o r t h $ 1 5 b n#c h a r =356 ,362 , nwr : e n g l i s h −w i k i e n e w s / A i r b u s w i n s Q a t a r A i r w a y s o r d e r w o r t h $ 1 5 b n#c h a r =641 ,647 , nwr : e n g l i s h −w i k i e n e w s / A i r b u s w i n s Q a t a r A i r w a y s o r d e r w o r t h $ 1 5 b n#c h a r =153 ,156 , nwr : e n g l i s h −w i k i e n e w s / A i r b u s w i n s Q a t a r A i r w a y s o r d e r w o r t h $ 1 5 b n#c h a r =872 ,878 . <h t t p : / d b p e d i a . o r g / r e s o u r c e / A i r b u s A 3 8 0> rdfs : label ” A380 ” , ” A i r b u s A380 ” , ” A i r b u s 380” ; g a f : denotedBy nwr : i t a l i a n −w i k i n e w s /20583 − t e x t p r o . t x t . txp#c h a r =88 ,99 , nwr : i t a l i a n −w i k i n e w s /3828− t e x t p r o . t x t . txp#c h a r =112 ,123 , nwr : i t a l i a n −w i k i n e w s /3828− t e x t p r o . t x t . txp#c h a r =205 ,216 , nwr : i t a l i a n −w i k i n e w s /31769 − t e x t p r o . t x t . txp#c h a r =109 ,120 > , nwr : i t a l i a n −w i k i n e w s /23242 − t e x t p r o . t x t . txp#c h a r =17,28> , nwr : i t a l i a n −w i k i n e w s /23242 − t e x t p r o . t x t . txp#c h a r =122 ,133 > , nwr : i t a l i a n −w i k i n e w s /25115 − t e x t p r o . t x t . txp#c h a r =2011 ,2022 > , , nwr : dutch−w i k i n e w s / 1 0 0 2 6 F i r s t A 3 8 0 e n t e r s c o m m e r c i a l s e r v i c e#c h a r =117 ,128 > nwr : dutch−w i k i n e w s / 1 0 0 2 6 F i r s t A 3 8 0 e n t e r s c o m m e r c i a l s e r v i c e#c h a r =840 ,844 , nwr : dutch−w i k i n e w s / 1 0 0 2 6 F i r s t A 3 8 0 e n t e r s c o m m e r c i a l s e r v i c e#c h a r =954 ,958 , nwr : dutch−w i k i n e w s / 6 4 7 5 S i n g a p o r e A i r l i n e s t o b e c o m p e n s a t e d f o r A 3 8 0 d e l a y s#c h a r =53,57> , , nwr : dutch−w i k i n e w s / 6 4 7 5 S i n g a p o r e A i r l i n e s t o b e c o m p e n s a t e d f o r A 3 8 0 d e l a y s#c h a r =190 ,194 > nwr : dutch−w i k i n e w s / 6 4 7 5 S i n g a p o r e A i r l i n e s t o b e c o m p e n s a t e d f o r A 3 8 0 d e l a y s#c h a r =1193 ,1197 > , nwr : dutch−w i k i n e w s /555 B o e i n g u n v e i l s l o n g −r a n g e 7 7 7#c h a r =1742 ,1746 > ; nwr : dutch−w i k i n e w s / 7 9 2 4 A 3 8 0 m a k e s m a i d e n f l i g h t t o U S#c h a r =48 ,59 , nwr : dutch−w i k i n e w s / 7 9 2 4 A 3 8 0 m a k e s m a i d e n f l i g h t t o U S#c h a r =512 ,516 , nwr : dutch−w i k i n e w s / 7 9 2 4 A 3 8 0 m a k e s m a i d e n f l i g h t t o U S#c h a r =1188 ,1192 , nwr : dutch−w i k i n e w s /260 A i r b u s l a u n c h e s w o r l d l a r g e s t p a s s e n g e r p l a n e#c h a r =785 ,789 , nwr : dutch−w i k i n e w s / 1 3 8 0 W o r l d l a r g e s t p a s s e n g e r a i r l i n e r m a k e s f i r s t f l i g h t#c h a r =84 ,88 , nwr : dutch−w i k i n e w s / 1 3 8 0 W o r l d l a r g e s t p a s s e n g e r a i r l i n e r m a k e s f i r s t f l i g h t#c h a r =166 ,177 , , nwr : dutch−w i k i n e w s / 1 3 8 0 W o r l d l a r g e s t p a s s e n g e r a i r l i n e r m a k e s f i r s t f l i g h t#c h a r =841 ,845 > nwr : dutch−w i k i n e w s / 1 3 8 0 W o r l d l a r g e s t p a s s e n g e r a i r l i n e r m a k e s f i r s t f l i g h t#c h a r =1037 ,1041 > , , nwr : dutch−w i k i n e w s / 8 9 3 5 B o e i n g u n v e i l s n e w 7 8 7 D r e a m l i n e r#c h a r =1709 ,1713 > nwr : dutch−w i k i n e w s / 2 0 0 7 / 7 / 9 / 8 9 3 5 B o e i n g u n v e i l s n e w 7 8 7 D r e a m l i n e r#c h a r =1871 ,1882 > , , nwr : dutch−w i k i n e w s / 1 0 0 2 1 F i r s t A i r b u s A 3 8 0 d e l i v e r e d#c h a r =52,56> nwr : dutch−w i k i n e w s / 1 0 0 2 1 F i r s t A i r b u s A 3 8 0 d e l i v e r e d#c h a r =617 ,621 > , nwr : dutch−w i k i n e w s / 3 2 3 5 E n g i n e t r o u b l e s d e l a y A i r b u s s u p e r j u m b o t o u r#c h a r =624 ,634 > , nwr : dutch−w i k i n e w s / 7 7 4 2 A i r b u s a n n o u n c e s j o b c u t s o f 1 0 ,000# c h a r =1403 ,1407 > , skos : prefLabel ” A380 ” , ” A i r b u s A380 ” , ” A i r b u s 380” . <h t t p : / d b p e d i a . o r g / r e s o u r c e / White House> rdfs : label ” Witte H u i s ” , ” Casa B i a n c a ” ; g a f : denotedBy nwr : i t a l i a n −w i k i n e w s /16014 − t e x t p r o . t x t . txp#c h a r =3,14> , nwr : i t a l i a n −w i k i n e w s /16014 − t e x t p r o . t x t . txp#c h a r =195 ,206 > , nwr : i t a l i a n −w i k i n e w s /16014 − t e x t p r o . t x t . txp#c h a r =387 ,398 > , nwr : i t a l i a n −w i k i n e w s /16014 − t e x t p r o . t x t . txp#c h a r =981 ,992 > , nwr : dutch−w i k i n e w s / 1 4 0 8 3 B a r a c k O b a m a p r e s e n t s r e s c u e p l a n a f t e r G M d e c l a r a t i o n o f b a n k r u p t c y# c h a r =383 ,393 > , nwr : dutch−w i k i n e w s / 1 3 1 4 3 W h i t e H o u s e c o n s i d e r i n g a u t o r e s c u e p l a n#c h a r =0,10> , nwr : dutch−w i k i n e w s / 1 3 1 4 3 W h i t e H o u s e c o n s i d e r i n g a u t o r e s c u e p l a n#c h a r =175 ,185 > , nwr : dutch−w i k i n e w s / 1 3 1 4 3 W h i t e H o u s e c o n s i d e r i n g a u t o r e s c u e p l a n#c h a r =359 ,369 > , nwr : dutch−w i k i n e w s / 1 3 1 4 3 W h i t e H o u s e c o n s i d e r i n g a u t o r e s c u e p l a n#c h a r =864 ,874 > ; skos : prefLabel ” Witte H u i s ” , ” Casa B i a n c a ” . Figure 56: RDF-TRiG representation of entities merged from English, Spanish, Italian and Dutch Wikinews NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 129/148 i l i : i l i −30−00634472−v a sem : Event , n w r o n t o l o g y : s o u r c e E v e n t , f n : C o m i n g t o b e l i e v e , f n : R e a s o n i n g ; rdfs : label ” c o n c l u i r ” , ” concluderen ” , ” concludere ” , ” reason ” , ” conclude ” ; g a f : denotedBy nwr : s p a n i s h −w i k i n e w s / 1 2 0 4 7 G o v e r n m e n t A c c o u n t a b i l i t y . t x t#c h a r =875 ,883 > , nwr : i t a l i a n −w i k i n e w s /25429 − t e x t p r o . t x t . txp#c h a r =52,60> , nwr : i t a l i a n −w i k i n e w s /25429 − t e x t p r o . t x t . txp#c h a r =239 ,250 > , nwr : i t a l i a n −w i k i n e w s /10246 − t e x t p r o . t x t . txp#c h a r =1025 ,1035 > , nwr : dutch−w i k i n e w s / 1 2 0 4 7 G o v e r n m e n t A c c o u n t a b i l i t y O f f i c e r e q u e s t s r e r u n o f U S A i r F o r c e t a n k e r b i d# c h a r =850 ,863 > , nwr : e n g l i s h −w i k i n e w s / I n d o n e s i a ’ s t r a n s p o r t m i n i s t e r t e l l s a i r l i n e s n o t t o b u y E u r o p e a n a i r c r a f t d u e t o E U b a n#c h a r =1801 ,1807 > , nwr : e n g l i s h −w i k i n e w s / G o v e r n m e n t A c c o u n t a b i l i t y O f f i c e r e q u e s t s r e r u n o f U S A i r F o r c e t a n k e r b i d# c h a r =724 ,732 > . i l i : i l i −30−01656788−v a sem : Event , f n : B u i l d i n g , f n : C r e a t i n g , n w r o n t o l o g y : c o n t e x t u a l E v e n t ; rdfs : label ” ensamblar ” , ” assemblare ” , ” s a m e n s t e l l e n ” ; g a f : denotedBy nwr : s p a n i s h −w i k i n e w s / 1 2 0 4 7 G o v e r n m e n t A c c o u n t a b i l i t y . t x t#c h a r =481 ,490 > , nwr : s p a n i s h −w i k i n e w s / 1 1 1 6 9 Northrop Grumman . t x t#c h a r =1845 ,1856 > , nwr : i t a l i a n −w i k i n e w s /21058 − t e x t p r o . t x t . txp#c h a r =1979 ,1989 > , nwr : i t a l i a n −w i k i n e w s /25429 − t e x t p r o . t x t . txp#c h a r =461 ,471 > , nwr : dutch−w i k i n e w s / 5 1 3 5 B o e i n g d e l i v e r s f i n a l 7 1 7 t o A i r T r a n , e n d i n g D o u g l a s e r a#c h a r =487 ,494 > . i l i : i l i −30−02680814−v a sem : Event , n w r o n t o l o g y : c o n t e x t u a l E v e n t , f n : A c t i v i t y s t o p , e s o : S t o p p i n g A n A c t i v i t y , fn : P r o c e s s s t o p , fn : Quitting , eso : LeavingAnOrganization , fn : Halt ; rdfs : label ” c e a s e ” , ” d i s c o n t i n u e ” , ” o p h e f f e n ” , ” c e s a r ” , ” op doeken ” ; g a f : denotedBy nwr : s p a n i s h −w i k i n e w s / 1 4 0 8 4 C E O o f G M o u t l i n e s p l a n . t x t#c h a r =3049 ,3054 , nwr : e n g l i s h −w i k i n e w s / C E O o f G M o u t l i n e s p l a n f o r \%22New GM\%22 a f t e r a u t o c o m p a n y d e c l a r e d b a n k r u p t c y#c h a r =2795 ,2800 , nwr : e n g l i s h −w i k i n e w s / F o r d T a u r u s t o b e r e v i v e d#c h a r =930 ,942 , nwr : e n g l i s h −w i k i n e w s / P e n s k e A u t o s e l e c t e d t o b u y G e n e r a l M o t o r s ’ S a t u r n u n i t#c h a r =444 ,456 , nwr : dutch−w i k i n e w s / 3 9 7 1 F o r d M o t o r C o m p a n y c u t t i n g 3 0 , 0 0 0 j o b s b y 2 0 1 2#c h a r =1309 ,1318 , nwr : dutch−w i k i n e w s / 1 3 7 7 4 G M a n d C h r y s l e r r e c e i v e C a n a d i a n l o a n s a m i d U S r e s t r u c t u r i n g u l t i m a t a# char =1784 ,1793. i l i : i l i −30−00156601−v−and− i l i −30−00153263−v a sem : Event , n w r o n t o l o g y : c o n t e x t u a l E v e n t , f n : C h a n g e p o s i t i o n o n a s c a l e , f n : C a u s e c h a n g e o f p o s i t i o n o n a s c a l e , e s o : I n c r e a s i n g , e s o : QuantityChange ; rdfs : label ” aumentare ” , ” i n c r e a s e ” , ” i n c r e m e n t a r ” ; g a f : denotedBy nwr : i t a l i a n −w i k i n e w s /12718 − t e x t p r o . t x t . txp#c h a r =1095 ,1104 , nwr : s p a n i s h −w i k i n e w s / 1 2 6 6 7 M a r k e t s r a l l y a s w o r l d c e n t r a l b a n k s i n f u s e c a s h . t x t#c h a r =912 ,923 , nwr : e n g l i s h −w i k i n e w s / S h a r e s w o r l d w i d e s u r g e d u e t o U S g o v e r n m e n t p l a n#c h a r =129 ,138 , nwr : e n g l i s h −w i k i n e w s / S h a r e s w o r l d w i d e s u r g e d u e t o U S g o v e r n m e n t p l a n#c h a r =503 ,511 , nwr : e n g l i s h −w i k i n e w s / M a r k e t s d o w n a c r o s s t h e w o r l d ; D o w J o n e s f a l l s b e l o w 9 ,000# c h a r =1356 ,1364 nwr : e n g l i s h −w i k i n e w s / B a n k o f A m e r i c a r e p o r t s l o s s e s o f o v e r U S $ 2 . 2 b i l l i o n#c h a r =244 ,253 , nwr : e n g l i s h −w i k i n e w s / B a n k o f A m e r i c a r e p o r t s l o s s e s o f o v e r U S $ 2 . 2 b i l l i o n#c h a r =816 ,824 , nwr : e n g l i s h −w i k i n e w s / S t o c k m a r k e t s w o r l d w i d e f a l l d r a m a t i c a l l y#c h a r =478 ,486 , nwr : e n g l i s h −w i k i n e w s / S t o c k m a r k e t s w o r l d w i d e f a l l d r a m a t i c a l l y#c h a r =1574 ,1582 , nwr : e n g l i s h −w i k i n e w s / U S s t o c k m a r k e t s h a v e t h e i r b e s t w e e k s i n c e N o v e m b e r#c h a r =446 ,454 , nwr : e n g l i s h −w i k i n e w s / U S s t o c k m a r k e t s h a v e t h e i r b e s t w e e k s i n c e N o v e m b e r#c h a r =1266 ,1275 , nwr : e n g l i s h −w i k i n e w s / R u s s i a n s t o c k m a r k e t s s u s p e n d e d a m i d m a r k e t t u r m o i l#c h a r =688 ,696 , nwr : e n g l i s h −w i k i n e w s / R u s s i a n s t o c k m a r k e t s s u s p e n d e d a m i d m a r k e t t u r m o i l#c h a r =1046 ,1054 , nwr : e n g l i s h −w i k i n e w s / W o r l d w i d e m a r k e t s f a l l p r e c i p i t o u s l y#c h a r =461 ,469 , nwr : e n g l i s h −w i k i n e w s / W o r l d w i d e m a r k e t s f a l l p r e c i p i t o u s l y#c h a r =800 ,809 , nwr : e n g l i s h −w i k i n e w s / M a r k e t s r a l l y a s w o r l d ’ s c e n t r a l b a n k s i n f u s e c a s h#c h a r = 7 1 7 , 7 2 5 . , i l i : i l i −30−01128193−v a sem : Event , n w r o n t o l o g y : c o n t e x t u a l E v e n t , i l i : i 3 9 7 0 2 , f n : P r o t e c t i n g ; rdfs : label ” p r o t e c t i o n ” , ” p r o t e c t ” , ” p r o t e g e r ” , ” beschermen ” ; g a f : denotedBy nwr : s p a n i s h −w i k i n e w s / 1 3 7 7 4 G M a n d C h r y s l e r r e c e i v e C a n a d i a n l o a n s a m i d U S r e s t r u c t u r i n g u l t i m a t a . t x t# c h a r =411 ,419 , nwr : e n g l i s h −w i k i n e w s / C E O o f G M o u t l i n e s p l a n f o r \%22New GM\%22 a f t e r a u t o c o m p a n y d e c l a r e d b a n k r u p t c y#c h a r =252 ,262 , nwr : e n g l i s h −w i k i n e w s / U S a u t o m a k e r G M r e p o r t s l o s s e s o f $ 6 b i l l i o n#c h a r =841 ,851 , nwr : e n g l i s h −w i k i n e w s / G M a n d C h r y s l e r r e c e i v e C a n a d i a n l o a n s a m i d U S r e s t r u c t u r i n g u l t i m a t a#c h a r =384 ,391 , nwr : e n g l i s h −w i k i n e w s / G M a n d C h r y s l e r r e c e i v e C a n a d i a n l o a n s a m i d U S r e s t r u c t u r i n g u l t i m a t a#c h a r =2545 ,2555 nwr : e n g l i s h −w i k i n e w s / B a r a c k O b a m a p r e s e n t s r e s c u e p l a n a f t e r G M d e c l a r a t i o n o f b a n k r u p t c y#c h a r =326 ,336 , nwr : e n g l i s h −w i k i n e w s /U . S . m a n u f a c t u r e r G e n e r a l M o t o r s s e e k s b a n k r u p t c y p r o t e c t i o n#c h a r =50 ,60 , nwr : e n g l i s h −w i k i n e w s /U . S . m a n u f a c t u r e r G e n e r a l M o t o r s s e e k s b a n k r u p t c y p r o t e c t i o n#c h a r =171 ,181 , nwr : e n g l i s h −w i k i n e w s / P e n s k e A u t o s e l e c t e d t o b u y G e n e r a l M o t o r s ’ S a t u r n u n i t#c h a r =157 ,167 , nwr : dutch−w i k i n e w s / 1 3 7 7 4 G M a n d C h r y s l e r r e c e i v e C a n a d i a n l o a n s a m i d U S r e s t r u c t u r i n g u l t i m a t a# c h a r =416 ,426 > . Figure 57: RDF-TRiG representation of events merged from English, Spanish, Italian and Dutch Wikinews NewsReader: ICT-316404 February 1, 2016 , Event Narrative Module, version 3 130/148 numbers for the other data sets, although also here the stock market has least instances and mentions. Overlap is highest for airbus and apple, up to 60%, 10 points less for gm and very low for the stock market. If we look at the Italian and Dutch results, we see that they perform very similar but for all data sets lower than Spanish, in terms of instances, mentions and coverage, except for the stock market. When averaged over the data sets, the coverage across the languages is very close. This means that the pipelines are reasonably compatible and interoperable across the languages for entity detection and linking. Table 33: DBpedia entities extracted for English, Spanish, Italian and Dutch Wikinews with proportion of coverage, measured as macro and micro coverage. I=instances, M=mentions, O=overlap, maC=macro-average over all document results, miC=microAverage over all mentions airbus apple gm stock Total English I M 157 795 96 680 118 757 5 61 376 2293 I 142 124 93 12 371 M 756 644 627 42 2069 Spanish O maC 489 51.6 424 52.1 393 35.7 2 2.8 1308 35.5 miC 61.5 62.4 51.9 3.3 44.8 I 110 91 76 77 354 M 446 490 369 202 1507 Italian O 352 344 244 23 963 maC 32.8 31.7 24.1 60.0 37.1 miC 44.3 50.6 32.2 37.7 41.2 I 121 91 82 100 394 M 557 445 540 380 1922 Dutch O 360 321 337 23 1041 maC 35.6 34.6 31.0 60.0 40.3 miC 45.3 47.2 44.5 37.7 43.7 The Tables 34, 35, 36, 37 show the top 15 entities most frequent in English with the number in Spanish, Italian and Dutch for all 4 corpora. For each entity, we show the number of mentions and the proportion of English mentions covered. If there are more mentions in Spanish, Italian or Dutch, the coverage is maximized to 100%. There are a two interesting observations to be made. First of all, United States dollar with 35, 16, 59 and 36 mentions in English across the data sets, turned out to be a systematic error in the English pipeline that is not mirrored by the other languages. The English pipeline erroneously linked mentions of the US to the dollar instead of the country. The second observation relates to the granularity of the mapping. For example in the case of the airbus data, Boeing is the most frequent entity in all 4 languages. The more specific entity Boeing Commercial Airplanes is however only detected in English and not in any of the other languages. This is due to the fact that the mappings across Wikipedia from the other language to English are at a more coarse-grained level. The example in Figure 58 shows that this is partly due to the absence of the specific link in the DBpedias of the specific languages (Italian and Spanish) or to the absence of a link from the specific page in a language to English (the Dutch example). 7.2 Crosslingual extraction of events As explained above, we represent events through the ILI-concepts that are associated with their lemmas. This approximates the representation of the concept. Furthermore in some cases, more than one concept is assigned to a single lemma. To compare such lists of concepts, we checked if there was at least one intersecting ILI-concept across events to decide on a match. We can see in Table 38 that there is a broader set of instances (I) than for entities. The proportions of matched event mentions from Spanish, Italian and Dutch to English NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 131/148 <e n t i t y i d =”e 1 4 ” t y p e=”ORGANIZATION”> <r e f e r e n c e s > <!−−B o e i n g Commercial A i r p l a n e s −−> <span> <t a r g e t i d =”t 2 3 3 ” /> <t a r g e t i d =”t 2 3 4 ” /> <t a r g e t i d =”t 2 3 5 ” /> </span> </ r e f e r e n c e s > <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f r e s o u r c e =” s p o t l i g h t v 1 ” r e f e r e n c e =”h t t p : / / i t . d b p e d i a . o r g / r e s o u r c e / B o e i n g ” c o n f i d e n c e =”1.0” r e f t y p e =” i t ” s o u r c e =” i t ”> <e x t e r n a l R e f r e s o u r c e =” w i k i p e d i a −db−i t E n ” r e f e r e n c e =”h t t p : / / d b p e d i a . o r g / r e s o u r c e / B o e i n g ” c o n f i d e n c e =”1.0” r e f t y p e =”en ” s o u r c e =” i t ” /> </ e x t e r n a l R e f > </ e x t e r n a l R e f e r e n c e s > </ e n t i t y > <e n t i t y i d =”e 2 1 ” t y p e=”ORG”> <r e f e r e n c e s > <span> <!−−B o e i n g Commercial A i r p l a n e s −−> <t a r g e t i d =”t 2 2 9 ”/> <t a r g e t i d =”t 2 3 0 ”/> <t a r g e t i d =”t 2 3 1 ”/> </span> </ r e f e r e n c e s > <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f c o n f i d e n c e =”1.0” r e f e r e n c e =”h t t p : / / e s . d b p e d i a . o r g / r e s o u r c e / B o e i n g ” r e f t y p e =” e s ” r e s o u r c e =” s p o t l i g h t v 1 ” s o u r c e =” e s ”> <e x t e r n a l R e f c o n f i d e n c e =”1.0” r e f e r e n c e =”h t t p : / / d b p e d i a . o r g / r e s o u r c e / B o e i n g ” r e f t y p e =”en ” r e s o u r c e =” w i k i p e d i a −db−esEn ” s o u r c e =” e s ”/> </ e x t e r n a l R e f > </ e x t e r n a l R e f e r e n c e s > </ e n t i t y > <e n t i t y i d =”e 2 2 ” t y p e=”ORG”> <r e f e r e n c e s > <span> <!−−B o e i n g Commercial A i r p l a n e s −−> <t a r g e t i d =” t 2 1 0 ”/> <t a r g e t i d =” t 2 1 1 ”/> <t a r g e t i d =” t 2 1 2 ”/> </span> </ r e f e r e n c e s > <e x t e r n a l R e f e r e n c e s > <e x t e r n a l R e f c o n f i d e n c e = ” 0 . 9 9 9 9 9 9 6 4 ” r e f e r e n c e =”h t t p : / / n l . d b p e d i a . o r g / r e s o u r c e / B o e i n g C o m m e r c i a l A i r p l a n e s ” r e f t y p e =” n l ” r e s o u r c e =” s p o t l i g h t v 1 ” s o u r c e =” n l ”/> </ e x t e r n a l R e f e r e n c e s > </ e n t i t y > Figure 58: Cross-lingual entity linking NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 132/148 Table 34: DBpedia entities in the Wikinews Airbus corpus most frequent in English with Spanish, Italian and Dutch frequencies Airbus Boeing Airbus United States dollar European Union Boeing Commercial Airplanes Boeing 787 Dreamliner United States Air Force Singapore France Airbus A320 family Ryanair Aeroflot Government Accountability Office Aer Lingus Boeing 747 English 131 85 35 34 33 29 29 18 16 15 14 13 12 12 11 Spanish 136 100.00 116 100.00 0 0.00 17 50.00 0 0.00 12 41.38 20 68.97 8 44.44 14 87.50 3 20.00 12 85.71 15 100.00 0 0.00 9 75.00 0 0.00 Italian 103 78.63 75 88.24 0 0.00 12 35.29 0 0.00 3 10.34 4 13.79 9 50.00 10 62.50 3 20.00 10 71.43 9 69.23 1 8.33 11 91.67 2 18.18 Dutch 78 59.54 74 87.06 0 0.00 20 58.82 0 0.00 10 34.48 18 62.07 8 44.44 10 62.50 4 26.67 10 71.43 8 61.54 0 0.00 9 75.00 0 0.00 Table 35: DBpedia entities in the Wikinews Apple corpus most frequent in English with Spanish, Italian and Dutch frequencies Apple Apple Inc. Steve Jobs Steve Waugh The Beatles United States dollar Intel Cisco Systems Microsoft James Cook Mac OS X Lion United Kingdom Motorola Software development kit IBM Apple Worldwide Developers Conference NewsReader: ICT-316404 English 312 49 27 22 16 16 14 12 10 10 8 8 8 8 7 Spanish 240 76.92 21 42.86 0 0.00 9 40.91 0 0.00 11 68.75 9 64.29 3 25.00 4 40.00 9 90.00 6 75.00 4 50.00 4 50.00 10 100.00 5 71.43 Italian 218 69.87 35 71.43 0 0.00 2 9.09 0 0.00 10 62.50 7 50.00 0 0.00 3 30.00 2 20.00 3 37.50 1 12.50 0 0.00 6 75.00 0 0.00 Dutch 179 57.37 31 63.27 0 0.00 7 31.82 1 6.25 4 25.00 8 57.14 4 33.33 1 10.00 0 0.00 8 100.00 1 12.50 5 62.50 5 62.50 2 28.57 February 1, 2016 Event Narrative Module, version 3 133/148 Table 36: DBpedia entities in the Wikinews GM, Chrysler, Ford corpus most frequent in English with Spanish, Italian and Dutch frequencies GM General Motors Ford Motor Company Chrysler United States dollar Fiat Ford Motor Company of Australia United Auto Workers Daimler AG Barack Obama United States Henderson, Nevada Federal government of the United States Canada Toyota Clarence Thomas English 155 81 76 59 30 22 21 21 16 15 14 13 12 11 9 Spanish 143 92.26 71 87.65 31 40.79 0 0.00 21 70.00 0 0.00 0 0.00 3 14.29 8 50.00 70 100.00 0 0.00 0 0.00 10 83.33 11 100.00 0 0.00 Italian 107 69.03 49 60.49 0 0.00 0 0.00 0 0.00 0 0.00 0 0.00 3 14.29 3 18.75 54 100.00 0 0.00 0 0.00 6 50.00 6 54.55 0 0.00 Dutch 119 76.77 50 61.73 43 56.58 0 0.00 0 0.00 0 0.00 0 0.00 13 61.90 10 62.50 107 100.00 0 0.00 0 0.00 21 100.00 8 72.73 0 0.00 Table 37: DBpedia entities in the Wikinews stock market corpus most frequent in English with Spanish, Italian and Dutch frequencies Stock United States dollar United States United Kingdom FTSE 100 Index Andy Kaufman Washington, D.C. Buenos Aires United States House of Representatives Dow Jones Industrial Average Reuters JPMorgan Chase Afghanistan Ben Bernanke State (polity) France NewsReader: ICT-316404 English 36 14 5 4 2 0 0 0 0 0 0 0 0 0 0 Spanish 0 0.00 2 14.29 0 0.00 0 0.00 0 0.00 0 0.00 0 0.00 1 0.00 0 0.00 0 0.00 0 0.00 0 0.00 0 0.00 0 0.00 0 0.00 0 29 11 4 0 2 0 0 9 2 1 1 0 1 0 Italian 0.00 100.00 100.00 100.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0 90 11 12 0 1 1 1 29 1 0 1 1 0 3 Dutch 0.00 100.00 100.00 100.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 February 1, 2016 Event Narrative Module, version 3 134/148 are just a little bit lower than for DBpedia entities but not much. This is promising since matching events is more difficult than matching entities. We see again that Dutch scores a bit lower than Spanish and Italian. This is due to the fact that the Spanish and Italian wordnets are a direct extension of the English wordnet and have been developed for many years, whereas the Open Dutch wordnet is recently built and partly built independently. Across the different data sets, the results are very similar. Table 38: ILI-based events extracted for English, Spanish, Italian and Dutch Wikinews with proportion of coverage, measured as macro and micro coverage. I=instances, M=mentions, O=overlap, maC=macro-average over all document results, miC=microAverage over all mentions airbus apple gm stock Total English I M 365 848 342 1007 319 1140 283 673 1309 3668 I 166 152 142 140 600 M 484 476 387 325 1672 Spanish O 217 242 257 163 879 maC 26.4 25.8 24.7 28.6 26.4 miC 25.6 24.0 22.5 24.2 24.1 I 535 500 504 450 1989 M 984 1090 1055 895 4024 Italian O 248 202 209 163 822 maC 33.3 29.8 33.1 34.0 32.6 miC 29.3 20.1 18.3 24.2 23.0 I 199 170 170 147 686 M 483 498 622 362 1965 Dutch O 164 170 192 92 618 maC 20.0 19.1 20.4 17.9 19.3 miC 19.3 16.9 16.8 13.7 16.7 In the Tables 39, 40, 41, 42 we show the ILI-based events that are most frequent in English for the 4 corpora. If there is more than one ILI-record assigned, we only list the synonyms for the first synset. For individual events, the results vary a lot across the different languages. There does not appear to be any pattern in this. A typical case is represented by ili-30-02207206-v[buy] in Table 42, which has a good match in Spanish, only 1 in Italian and 0 in Dutch. The Dutch equivalent kopen is linked to a hypernym of buy and the Italian equivalent acquistare is linked to another meaning. Table 39: ILI-based events in the Wikinews Airbus corpus most frequent in English with Spanish, Italian and Dutch frequencies Airbus ili-30-01438304-v[deliver] ili-30-02204692-v[have] ili-30-00764222-v;ili-30-02657219-v;ili-30-00805376-v[agree] ili-30-00974367-v;ili-30-00975427-v[announce] ili-30-02207206-v[buy] ili-30-01653442-v[construct] ili-30-00755745-v;ili-30-00719734-v[ask;expect] ili-30-02413480-v;ili-30-02410855-v[work] ili-30-00705227-v[be after] ili-30-02257767-v;ili-30-00162688-v[interchange;replace] ili-30-02244956-v;ili-30-02242464-v[deal;sell] ili-30-00998399-v[record] ili-30-02641957-v;ili-30-00459776-v[delay] ili-30-01955984-v;ili-30-01957529-v;ili-30-02102398-v;ili-30-01847676-v[ride] ili-30-01583142-v;ili-30-01654628-v[construct;build] 7.3 English 17 17 16 15 14 14 13 12 12 10 9 8 8 8 8 3 0 0 2 16 2 0 0 3 7 5 0 7 0 4 Spanish 17.65 0.00 0.00 13.33 100.00 14.29 0.00 0.00 25.00 70.00 55.56 0.00 87.50 0.00 50.00 7 3 1 3 3 0 1 1 0 0 1 1 5 0 6 Italian 41.18 17.65 6.25 20.00 21.43 0.00 7.69 8.33 0.00 0.00 11.11 12.50 62.50 0.00 75.00 9 0 0 14 4 0 0 2 0 0 5 0 0 0 0 Dutch 52.94 0.00 0.00 93.33 28.57 0.00 0.00 16.67 0.00 0.00 55.56 0.00 0.00 0.00 0.00 Crosslingual extraction of relations Finally, we compared the actual triples extracted across the languages. The triples represent the actual statements, where we only consider triples where the ILI-based event is the subject. Table 43 gives the predicates that are most frequent in the English data. We limited ourselves here to the generic SEM predicates (hasActor, hasTime and hasPlace) as well as the more specific temporal relations added in NewsReader and the most -frequent NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 135/148 Table 40: ILI-based events in the Wikinews Apple corpus most frequent in English with Spanish, Italian and Dutch frequencies Apple ili-30-01224744-v;ili-30-01525666-v[control;function] ili-30-00674607-v[choose] ili-30-02204692-v[have] ili-30-02421374-v[free] ili-30-00974367-v;ili-30-00975427-v[announce] ili-30-02244956-v;ili-30-02242464-v[deal;sell] ili-30-00721889-v;ili-30-02351010-v[price] ili-30-02630189-v[feature] ili-30-00933821-v[break] ili-30-01642437-v[innovate] ili-30-02735282-v;ili-30-02501278-v;ili-30-01486312-v[suit;adjudicate;case] ili-30-00341917-v;ili-30-02743921-v;ili-30-01849221-v[come;come up] ili-30-00802318-v[allow] ili-30-00756338-v[claim] ili-30-00515154-v[process] English 64 48 35 34 34 20 19 16 15 15 15 14 12 12 12 0 1 2 0 30 22 0 0 0 0 0 6 10 0 0 Spanish 0.00 2.08 5.71 0.00 88.24 100.00 0.00 0.00 0.00 0.00 0.00 42.86 83.33 0.00 0.00 2 0 2 3 10 3 0 45 1 0 0 1 0 0 0 Italian 3.13 0.00 5.71 8.82 29.41 15.00 0.00 100.00 6.67 0.00 0.00 7.14 0.00 0.00 0.00 2 0 0 0 17 17 0 0 14 2 0 0 0 0 1 Dutch 3.13 0.00 0.00 0.00 50.00 85.00 0.00 0.00 93.33 13.33 0.00 0.00 0.00 0.00 8.33 Table 41: ILI-based events in the Wikinews GM, Chrysler, Ford corpus most frequent in English with Spanish, Italian and Dutch frequencies GM ili-30-00674607-v;ili-30-00679389-v[choose] ili-30-02244956-v;ili-30-02242464-v[deal;sell] ili-30-01621555-v;ili-30-01640207-v;ili-30-01753788-v;ili-30-01617192-v[create] ili-30-00705227-v[be after] ili-30-02204692-v[have] ili-30-02511551-v[order] ili-30-00561090-v[cut] ili-30-00974367-v;ili-30-00975427-v[announce] ili-30-02410175-v[keep on] ili-30-02324182-v[lend] ili-30-02547586-v[aid] ili-30-00358431-v;ili-30-00354845-v[buy the farm;die] ili-30-01182709-v;ili-30-02327200-v[provide;furnish] ili-30-02613487-v;ili-30-02297142-v[offer up;proffer] ili-30-02207206-v[buy] English 153 60 36 32 26 25 24 24 23 18 16 15 12 12 11 2 13 1 8 0 0 45 19 0 0 4 0 4 7 8 Spanish 1.31 21.67 2.78 25.00 0.00 0.00 100.00 79.17 0.00 0.00 25.00 0.00 33.33 58.33 72.73 2 3 1 0 7 0 0 4 0 0 4 0 6 2 1 Italian 1.31 5.00 2.78 0.00 26.92 0.00 0.00 16.67 0.00 0.00 25.00 0.00 50.00 16.67 9.09 2 29 2 0 0 0 0 22 0 0 3 0 5 0 0 Dutch 1.31 48.33 5.56 0.00 0.00 0.00 0.00 91.67 0.00 0.00 18.75 0.00 41.67 0.00 0.00 Table 42: ILI-based events in the Wikinews stock market corpus most frequent in English with Spanish and Dutch frequencies Stock ili-30-02244956-v;ili-30-02242464-v[deal;sell] ili-30-01307142-v;ili-30-00356649-v[even out;level off] ili-30-02204692-v[have] ili-30-00658052-v;ili-30-00660971-v[grade;rate] ili-30-00153263-v;ili-30-00156601-v[increase] ili-30-02324182-v[lend] ili-30-00974367-v;ili-30-00975427-v[announce] ili-30-02000868-v;ili-30-00589738-v;ili-30-02445925-v;ili-30-01998432-v[follow;be] ili-30-01645601-v[cause] ili-30-00721889-v;ili-30-02351010-v[price] ili-30-00998399-v[record] ili-30-02678438-v[concern] ili-30-02259005-v;ili-30-02260085-v[swap;trade in] ili-30-01778568-v;ili-30-01780434-v;ili-30-01780202-v[fear;dread] ili-30-00352826-v]ili-30-01620854-v[end] ili-30-02421374-v[free] NewsReader: ICT-316404 English 40 26 21 18 14 13 12 10 10 10 9 9 9 8 8 7 4 0 1 0 5 0 12 0 2 0 12 0 1 4 2 1 Spanish 10.00 0.00 4.76 0.00 35.71 0.00 100.00 0.00 20.00 0.00 100.00 0.00 11.11 50.00 25.00 14.29 1 0 0 0 3 0 3 1 1 0 6 0 0 6 2 0 Italian 2.50 0.00 0.00 0.00 21.43 0.00 25.00 10.00 10.00 0.00 66.67 0.00 0.00 75.00 25.00 0.00 5 0 0 0 2 0 13 0 0 0 0 0 0 3 1 0 Dutch 12.50 0.00 0.00 0.00 14.29 0.00 100.00 0.00 0.00 0.00 0.00 0.00 0.00 37.50 12.50 0.00 February 1, 2016 Event Narrative Module, version 3 136/148 PropBank relations. The hasActor, hasTime and hasPlace generalize over the others. We can see that Spanish scores a little better for hasActor than Italian and Dutch except for the stock market data set where the Italian system has a coverage of almost 84 which is twice as high as for the other data sets. In the case of airbus and apple, the Italian pipeline scores high for A2 in comparison to the others but for stock market it is A0 and A1 that score high. For Spanish, we see that results are more consistent but with a high score for A1 in the apple data set. Dutch score lower on the actors overall except for A0 in the stock market data set. The Dutch pipeline is apparently very successful in recovering locations compared to the others, where the Italian pipeline is successful in recovering temporal relations. Table 43: Triple predicates that are most frequent in the English Wikinews corpus with coverage in Spanish, Italian and Dutch Role Airbus Apple GM Stock A0 A1 A2 AM-LOC hasActor hasAtTime hasFutureTime hasPlace hasTime A0 A1 A2 AM-LOC hasActor hasAtTime hasFutureTime hasPlace hasTime A0 A1 A2 AM-LOC hasActor hasAtTime hasFutureTime hasPlace hasTime A0 A1 A2 AM-LOC hasActor hasAtTime hasPlace hasTime English 343 388 97 37 857 1494 74 51 1568 282 248 68 21 608 1809 51 22 1860 307 330 87 29 734 1580 101 29 1681 337 1043 283 40 1714 1700 42 1732 Spanish 112 32.65 216 55.67 64 65.98 0 0.00 528 61.61 852 57.03 0 0.00 0 0.00 852 54.34 113 40.07 207 83.47 48 70.59 0 0.00 471 77.47 1021 56.44 0 0.00 0 0.00 1021 54.89 105 34.20 182 55.15 51 58.62 0 0.00 421 57.36 768 48.61 0 0.00 0 0.00 768 45.69 162 48.07 575 55.13 89 31.45 0 0.00 950 55.43 977 57.47 0 0.00 977 56.41 Italian 140 40.82 178 45.88 91 93.81 2 5.41 410 47.84 1326 88.76 170 100.00 2 3.92 1496 95.41 122 43.26 165 66.53 55 80.88 7 33.33 342 56.25 1321 73.02 144 100.00 7 31.82 1465 78.76 101 32.90 140 42.42 46 52.87 3 10.34 287 39.10 1371 86.77 196 100.00 3 10.34 1567 93.22 435 100.00 851 81.59 153 54.06 4 10.00 1439 83.96 1335 78.53 4 9.52 1443 83.31 151 193 40 38 403 921 0 42 921 90 162 32 29 299 968 0 32 968 119 145 22 33 303 916 0 34 916 371 470 63 52 967 875 56 875 Dutch 44.02 49.74 41.24 100.00 47.02 61.65 0.00 82.35 58.74 31.91 65.32 47.06 100.00 49.18 53.51 0.00 100.00 52.04 38.76 43.94 25.29 100.00 41.28 57.97 0.00 100.00 54.49 100.00 45.06 22.26 100.00 56.42 51.47 100.00 50.52 Table 44 shows the coverage results for the actual triples that relate ILI-based events with entities through the above predicates. We only considered the hasActor and hasPlace relations. Obviously, the coverage is low since this is a very difficult task: all 3 elements need to match exactly. Spanish results are on average 3%, and Italian and Dutch less than 1%. Since triples usually are mentioned only once and at most a few times in the corpora (30 articles only), it makes no sense to show frequency tables of triples. In Figure 59, we give some examples of triples shared by all 4 languages. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 137/148 Table 44: ILI-based Triples extracted for English, Spanish, Italian and Dutch Wikinews with proportion of coverage, measured as macro and micro coverage. I=instances, M=mentions, O=overlap, maC=macro-average over all document results, miC=microAverage over all mentions airbus apple gm stock Total 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 English I M 775 775 525 525 647 647 1463 1473 853 3410 I 369 292 262 497 355 M 369 292 262 500 1420 Spanish O maC 24 3.1 32 6.1 20 3.1 0 0 1423 3.1 miC 3.1 6.1 3.1 0 3.1 I 390 312 273 1312 572 M 390 312 273 1326 2287 Italian O 7 7 1 0 2301 maC 1.0 1.3 0.2 0 0.6 miC 0.9 1.3 0.2 0 0.6 I 381 251 273 773 420 M 381 251 273 801 1678 Dutch O 6 8 2 0 1706 maC 0.8 1.5 0.3 0 0.7 Triples in a l l 4 languages i l i −30−00975427−v ; i l i −30−00974367−v [ announce ] : h a s A c t o r : B o e i n g i l i −30−00975427−v ; i l i −30−00974367−v [ announce ] : h a s A c t o r : A i r b u s i l i −30−02646757−v ; i l i −30−02207206−v [ buy ] : h a s A c t o r : E u r o p e a n U n i o n i l i −30−00761713−v [ n e g o c i a t e ] : h a s A c t o r : A e r o f l o t i l i −30−02244956−v ; i l i −30−02242464−v [ d e a l ; s e l l ] : h a s A c t o r : A i r b u s i l i −30−00634472−v [ c o n c l u d e ] : h a s A c t o r : B o e i n g i l i −30−00882948−v ; i l i −30−00875141−v [ commend ; a d v o c a t e ] : h a s A c t o r : A i r b u s i l i −30−00354845−v ; i l i −30−00358431−v [ d i e ; b u y t h e f a r m ] : h a s A c t o r : S t e v e J o b s i l i −30−01734502−v ; i l i −30−00246217−v [ d u p l i c a t e ; d o u b l e ] : h a s A c t o r : A p p l e I n c . i l i −30−00975427−v ; i l i −30−00974367−v [ announce ] : h a s A c t o r : S t a r b u c k s i l i −30−00975427−v ; i l i −30−00974367−v ; i l i −30−00820801−v ; i l i −30−01010118−v [ announce ; d e c l a r e ] : hasActor : U n i t e d S t a t e s i l i −30−01182709−v ; i l i −30−02327200−v ; i l i −30−02479323−v [ p r o v i d e ; f u r n i s h ; i s s u e ] : hasActor : General Motors i l i −30−00975427−v ; i l i −30−00974367−v ; i l i −30−00820801−v ; i l i −30−01010118−v [ announce ; d e c l a r e ] : h a s A c t o r : Ford Motor Company i l i −30−02244956−v ; i l i −30−02242464−v [ d e a l ; s e l l ] : h a s A c t o r : Opel i l i −30−00975427−v ; i l i −30−00974367−v ; i l i −30−00820801−v ; i l i −30−01010118−v [ announce ; d e c l a r e ] : hasActor : General Motors Figure 59: Identical triples across different languages 7.4 Conclusions We described the results of cross-lingual semantic processing of text. To our knowledge, there is no other system that can perform such a task. Being able to merge the interpretation of text across language is a big achievement and it shows the opportunities for interoperability of the NewsReader system. We have also seen that for most data types coverage still leaves room for improvement. We have also seen that differences in implementation have an impact on the comparability. Spanish results are more closer to English because most of the NLP modules for English and Spanish are developed by the same group, whereas the Dutch and Italian pipelines are mostly based on different software. That does not mean that the output of the Spanish software is better than the Dutch software. It only means that it is more compatible. For a qualitative evaluation, we need to use the cross-lingual annotation of the Wikinews. This is reported in Agerri et al. (2015). NewsReader: ICT-316404 February 1, 2016 miC 0.8 1.5 0.3 0 0.7 Event Narrative Module, version 3 8 138/148 Conclusions In this deliverable, we described the final project results on event modelling, as part of WP5 activities. We explained in detail the conversion process from NAF to SEM representation according to a batch and streaming architecture. This process explains how we get from text to RDF specifications of textual content. The core problem here is event-coreference. We describe the different approaches implemented and the evaluations on cross-document coreference on the ECB+ data set. The NAF2SEM process resolving event-coreference has been applied to over 3 million car documents generating more than half a billion triples. To move beyond event structures, we need to relate events to time and to each other. We describe our modules for extracting event relations, one for temporal relations and one for causal relations. The output of these modules can be used to create timelines, for which we organised a SemEval task in 2015 with evaluation results. Timelines form the basis for creating Storylines. Our approach has been presented at the ACL workshop on this topic that we organized as well in 2015. Not all events are real-world events. Many expressions in news reflect perspectives on real-world events. We explained our perspective module that takes various NAF layers as input to model the attribution relation of sources with respect to their beliefs and opinions. Finally, we reported the results on the cross-lingual processing of news document, obtained by comparing generated RDF-TRiG files for the Wikinews corpora for English, Spanish, Italian and Dutch. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 9 139/148 Appendix Table 45: FrameNet frames for contextualEvents Absorb heat Abundance Abusing Adding up Adjusting Adorning Aging Amalgamation Amounting to Apply heat Arranging Arriving Assemble Assistance Attaching Attack Avoiding Becoming a member Becoming detached Behind the scenes Being attached Being employed Being in category Being in operation Being located Body movement Breathing Bringing Building Bungling Catastrophe Cause change Cause change of consistency Cause change of phase Cause change of position on a scale Cause change of strength Cause expansion Cause fluidic motion Cause harm Cause impact Cause motion Cause temperature change Cause to amalgamate Cause to be dry Cause to be sharp Cause to be wet Cause to experience Cause to fragment Cause to make noise Cause to make progress NewsReader: ICT-316404 Cause to move in place Cause to start Cause to wake Change direction Change event duration Change event time Change of consistency Change of leadership Change of phase Change operational state Change position on a scale Change posture Change tool Closure Collaboration Colonization Come together Coming up with Commerce buy Commerce collect Commerce pay Commerce sell Compatibility Competition Compliance Conquering Cooking creation Corroding Corroding caused Cotheme Create physical artwork Create representation Creating Cure Cutting Damaging Daring Death Defend Delivery Departing Destroying Detaching Dimension Dispersal Dodging Dressing Duplication Earnings and losses Eclipse Education teaching Elusive goal Emitting Employing Emptying Escaping Evading Examination Exchange Exchange currency Exclude member Excreting Expansion Expensiveness Experience bodily harm Experiencer obj Filling Fining Firing Fleeing Fluidic motion Forging Forming relationships Friction Frugality Gathering up Getting Getting up Giving Grinding Grooming Hiding objects Hiring Hit target Holding off on Hostile encounter Imitating Immobilization Impact Imprisonment Inchoative attaching Inchoative change of temperature Ingest substance Ingestion Inspecting Installing Institutionalization Intentional traversing Intentionally create Kidnapping Killing Knot creation Leadership Light movement Limiting Location of light Locative relation Make acquaintance Make noise Manipulate into doing Manipulation Manufacturing Mass motion Motion Motion directional Motion noise Moving in place Operate vehicle Operational testing Path shape Perception Personal relationship Piracy Placing Posture Precipitation Preserving Processing materials Prohibiting Provide lodging Quarreling Quitting Quitting a place Reading Receiving Recording Recovery Rejuvenation Releasing Removing Render nonfunctional Renting Renting out Replacing Reshaping Residence Resolve problem Resurrection Revenge Rewards and punishments Ride vehicle Robbery Rope manipulation Rotting Scouring Scrutiny Seeking Self motion Sending Separating Setting fire Shoot projectiles Shopping Sign agreement Similarity Sleep Smuggling Soaking Social event Sound movement Storing Supply Surpassing Surviving Take place of Taking Text creation Theft Translating Travel Traversing Undergo change Undressing Use firearm Visiting Waiting Waking up Wearing Weather Execution Inhibit movement Proliferating in number February 1, 2016 Event Narrative Module, version 3 140/148 Table 46: FrameNet from for sourceEvents Achieving first Adding up Adducing Agree or refuse to act Appointing Attempt suasion Bail decision Be in agreement on assessment Be translation equivalent Become silent Behind the scenes Being named Body movement Bragging Categorization Chatting Choosing Claim ownership Coming up with Commitment Communicate categorization Communication Communication manner Communication means Communication noise Communication response Compatibility Complaining Compliance Confronting problem Contacting Criminal investigation Deny permission Deserving Discussion Distinctiveness Encoding Eventive cognizer affecting Evidence Experiencer obj Expressing publicly Forgiveness Gesture Grant permission Have as translation equivalent Heralding Imposing obligation Judgment Judgment communication Judgment direct address Justifying Labeling Linguistic meaning Make agreement on action Make noise Making faces Manipulate into doing Motion noise Name conferral Notification of charges Omen Pardon Predicting Prevarication Prohibiting Questioning Referring by name Regard Reporting Request Respond to proposal Reveal secret Rite Seeking Sign Silencing Simple naming Speak on topic Spelling and pronouncing Statement Suasion Subjective influence Successfully communicate message Talking into Telling Text creation Verdict Appearance Categorization Chemical-sense description Locating Perception active Perception body Perception experience Seeking Trust Adopt selection Assessing Awareness Becoming aware Categorization Cause emotion Certainty Choosing Cogitation Coming to believe Daring Desiring Differentiation Emotion active Estimating Expectation Experiencer focus Experiencer obj Familiarity Feeling Feigning Grasp Importance Judgment Occupy rank Opinion Partiality Place weight on Preference Purpose Reliance Scrutiny Seeking Taking sides Topic Table 47: FrameNet frames for grammaticalEvents Accomplishment Achieving first Activity finish Activity ongoing Activity prepare Activity start Activity stop Amassing Arriving Assistance Attempt Avoiding Becoming Birth Causation Cause change Cause to continue Cause to end Coming to be Containing Cooking creation Cotheme Creating Departing Detaining Dough rising Emanating Event Evidence Execute plan NewsReader: ICT-316404 Existence Experiencer obj Grant permission Halt Have as requirement Hindering Holding off on Inclusion Influence of event on cognizer Intentionally act Intentionally affect Launch process Left to do Manipulate into doing Manufacturing Motion Operating a system Permitting Possession Preventing Process continue Process end Process resume Process start Process stop Reasoning Relative time Remainder Ride vehicle Self motion Setting fire Setting out Sidereal appearance State continue Storing Success or failure Successful action Taking Taking time Thriving Thwarting Topic Undergo change Using February 1, 2016 Event Narrative Module, version 3 141/148 References Rodrigo Agerri, Josu Bermudez, and German Rigau. IXA pipeline: Efficient and Ready to Use Multilingual NLP tools. In Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC-2014), 2014. 00013. Rodrigo Agerri, Itziar Aldabe, Zuhaitz Beloki, Egoitz Laparra, German Rigau, Aitor Soroa, Marieke van Erp, Antske Fokkens, Filip Ilievski, Ruben Izquierdo, Roser Morante, and Piek Vossen. Event detection, version 2. NewsReader Deliverable 4.2.3, 2015. Amit Bagga and Breck Baldwin. Algorithms for scoring coreference chains. In Proceedings of the International Conference on Language Resources and Evaluation (LREC), 1998. Collin F. Baker, Charles J. Fillmore, and John B. Lowe. The Berkeley FrameNet project. In COLING-ACL ’98: Proceedings of the Conference, pages 86–90, Montreal, Canada, 1998. Mieke Bal. Narratology: Introduction to the theory of narrative. University of Toronto Press, 1997. Cosmin Adrian Bejan and Sanda Harabagiu. Unsupervised event coreference resolution with rich linguistic features. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics, Uppsala, Sweden, 2010. Cosmin Adrian Bejan and Sanda Harabagiu. Unsupervised event coreference resolution with rich linguistic features. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics, Uppsala, Sweden, 2010. Steven Bird, Ewan Klein, and Edward Loper. Natural Language Processing with Python: http://nltk.org/book. O’Reilly Media Inc., 2009. Anders Björkelund, Love Hafdell, and Pierre Nugues. Multilingual semantic role labeling. In Proceedings of the Thirteenth Conference on Computational Natural Language Learning: Shared Task, CoNLL ’09, pages 43–48, Boulder, Colorado, USA, 2009. Eduardo Blanco and Dan Moldovan. Leveraging verb-argument structures to infer semantic relations. In Proceedings of the 14th Conference of the European Chapter of the Association for Computational Linguistics, pages 145–154, Gothenburg, Sweden, 2014. David M Blei and Peter I Frazier. Distance dependent chinese restaurant processes. The Journal of Machine Learning Research, 12:2461–2488, 2011. Francis Bond, Piek Vossen, John P. McCrae, and Christiane Fellbaum. Cili: the collaborative interlingual index. Proceedings of the Eighth meeting of the Global WordNet Conference (GWC 2016), Bucharest, 2016. Jerome S Bruner. Acts of meaning. Harvard University Press, 1990. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 142/148 Tommaso Caselli, Antske Fokkens, Roser Morante, and Piek Vossen. Spinoza vu: An nlp pipeline for cross document timelines. In Proceedings of the 9th International Workshop on Semantic Evaluation (SemEval 2015), pages 787–791, Denver, Colorado, June 2015. Association for Computational Linguistics. Tommaso Caselli, Marieke van Erp, Anne-Lyse Minard, Mark Finlayson, Ben Miller, Jordi Atserias, Alexandra Balahur, and Piek Vossen, editors. Proceedings of the First Workshop on Computing News Storylines. Association for Computational Linguistics, Beijing, China, July 2015. Tommaso Caselli, Piek Vossen, Marieke van Erp, Antske Fokkens, Filip Ilievski, Ruben Izquierdo Bevia, Minh Le, Roser Morante, and Marten Postma. When it’s all piling up: investigating error propagation in an nlp pipeline. In WNACP2015, 2015. Taylor Cassidy, Bill McDowell, Nathanael Chambers, and Steven Bethard. An annotation framework for dense event ordering. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 501–506, Baltimore, Maryland, June 2014. Association for Computational Linguistics. Nathanael Chambers, Taylor Cassidy, Bill McDowell, and Steven Bethard. Dense event ordering with a multi-pass architecture. Transactions of the Association for Computational Linguistics, 2:273–284, 2014. Nate Chambers. Navytime: Event and time ordering from raw text. In Proceedings of the Seventh International Workshop on Semantic Evaluation, SemEval ’13, pages 73–77, Atlanta, Georgia, USA, 2013. Zheng Chen and Heng Ji. Event coreference resolution: Feature impact and evaluation. In Proceedings of Events in Emerging Text Types (eETTs) Workshop, 2009. Zheng Chen and Heng Ji. Graph-based event coreference resolution. In TextGraphs-4 Proceedings of the 2009 Workshop on Graph-based Methods for Natural Language Processing, pages 54–57, 2009. Bin Chen, Jian Su, Sinno Jialin Pan, and Chew Lim Tan. A unified event coreference resolution by integrating multiple resolvers. In Proceedings of the 5th International Joint Conference on Natural Language Processing, Chiang Mai, Thailand, 2011. Francesco Corcoglioniti, Marco Rospocher, Roldano Cattoni, Bernardo Magnini, and Luciano Serafini. Interlinking unstructured and structured knowledge in an integrated framework. In Proc. of 7th IEEE International Conference on Semantic Computing (ICSC), Irvine, CA, USA, 2013. (to appear). Agata Cybulska and Piek Vossen. Semantic relations between events and their time, locations and participants for event coreference resolution. In Proceedings of Recent Advances in Natural Language Processing (RANLP-2013), pages 156–163, 2013. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 143/148 Agata Cybulska and Piek Vossen. Semantic relations between events and their time, locations and participants for event coreference resolution. In Proceedings of recent advances in natural language processing, 2013. Agata Cybulska and Piek Vossen. Guidelines for ecb+ annotation of events and their coreference, 2014. Agata Cybulska and Piek Vossen. Using a sledgehammer to crack a nut? lexical diversity and event coreference resolution. In Proceedings of the International Conference on Language Resources and Evaluation (LREC 2014), 2014b. Agata Cybulska and Piek Vossen. ”bag of events” approach to event coreference resolution. supervised classification of event templates. In proceedings of the 16th Cicling 2015 (colocated: 1st International Arabic Computational Linguistics Conference), Cairo, Egypt, April 14–20 2015. Joachim Daiber, Max Jakob, Chris Hokamp, and Pablo N. Mendes. Improving efficiency and accuracy in multilingual entity extraction. In Proceedings of the 9th International Conference on Semantic Systems (I-Semantics), 2013. Günes Erkan and Dragomir R Radev. Lexrank: graph-based lexical centrality as salience in text summarization. Journal of Artificial Intelligence Research, pages 457–479, 2004. Antske Fokkens, Marieke van Erp, Piek Vossen, Sara Tonelli, Willem Robert van Hage, Luciano Serafini, Rachele Sprugnoli, and Jesper Hoeksema. GAF: A grounded annotation framework for events. In Proceedings of the first Workshop on Events: Definition, Dectection, Coreference and Representation, Atlanta, USA, 2013. Antske Fokkens, Aitor Soroa, Zuhaitz Beloki, Niels Ockeloen, German Rigau, Willem Robert van Hage, and Piek Vossen. Naf and gaf: Linking linguistic annotations. In Proceedings 10th Joint ISO-ACL SIGSEM Workshop on Interoperable Semantic Annotation, page 9, 2014. Juri Ganitkevitch, Benjamin Van Durme, and Chris Callison-Burch. PPDB: The paraphrase database. In Proceedings of NAACL-HLT, pages 758–764, Atlanta, Georgia, June 2013. Association for Computational Linguistics. Matthew Gerber and Joyce Chai. Semantic role labeling of implicit arguments for nominal predicates. Computational Linguistics, 38(4):755–798, December 2012. Paul Grice. Logic and conversation. Syntax and semantics. 3: Speech acts, pages 41–58, 1975. David Herman, Manfred Jahn, and Marie-Laure Ryan, editors. Routlege Encyclopedia of Narrative Theory. Routledge, 2010. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 144/148 Johannes Hoffart, Fabian M. Suchanek, Klaus Berberich, and Gerhard Weikum. YAGO2: A Spatially and Temporally Enhanced Knowledge Base from Wikipedia. Artif. Intell., 194:28–61, 2013. Lifu Huang and Lian’en Huang. Optimized event storyline generation based on mixtureevent-aspect model. In Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, pages 726–735, Seattle, Washington, USA, October 2013. Association for Computational Linguistics. Kevin Humphreys, Robert Gaizauskas, and Saliha Azzam. Event coreference for information extraction. In Proceedings of a Workshop on Operational Factors in Practical, Robust Anaphora Resolution for Unrestricted Texts, 1997. Taku Kudo and Yuji Matsumoto. Fast Methods for Kernel-based Text Analysis. In Proceedings of the 41st Annual Meeting on Association for Computational Linguistics - Volume 1, ACL ’03, pages 24–31, Stroudsburg, PA, USA, 2003. Egoitz Laparra and German Rigau. Impar: A deterministic algorithm for implicit semantic role labelling. In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (ACL 2013), pages 33–41, 2013. LDC. Ldc. ace (automatic content extraction) english annotation guidelines for events ver. 5.4.3 2005.07.01., 2005. Claudia Leacock and Martin Chodorow. Combining local context with wordnet similarity for word sense identification, 1998. Heeyoung Lee, Yves Peirsman, Angel Chang, Nathanael Chambers, Mihai Surdeanu, and Dan Jurafsky. Stanford’s multi-pass sieve coreference resolution system at the conll2011 shared task. In Proceedings of the Fifteenth Conference on Computational Natural Language Learning: Shared Task, CONLL Shared Task ’11, Portland, Oregon, 2011. Heeyoung Lee, Marta Recasens, Angel Chang, Mihai Surdeanu, and Dan Jurafsky. Joint entity and event coreference resolution across documents. In Proceedings of the 2012 Conference on Empirical Methods in Natural Language Processing and Natural Language Learning (EMNLPCoNLL), 2012. Zhengzhong Liu, Jun Araki, Eduard Hovy, and Teruko Mitamura. Supervised withindocument event coreference using information propagation. In Proceedings of the International Conference on Language Resources and Evaluation, 2014. Hector Llorens, Estela Saquete, and Borja Navarro. Tipsem (english and spanish): Evaluating crfs and semantic roles in tempeval-2. In Proceedings of the 5th International Workshop on Semantic Evaluation, pages 284–291. Association for Computational Linguistics, 2010. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 145/148 A. Xiaoqiang Luo, Sameer Pradhan, Marta Recasens, and Eduard Hovy. Scoring coreference partitions of predicted mentions: A reference implementation. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics, Baltimore, MD, June 2014. Xiaoqiang Luo. On coreference resolution performance metrics. In Proceedings of the Human Language Technology Conference and Conference on Empirical Methods in Natural Language Processing (EMNLP-2005), 2005. Inderjeet Mani, Marc Verhagen, Ben Wellner, Chong Min Lee, and James Pustejovsky. Machine learning of temporal relations. In Proceedings of the 21st International Conference on Computational Linguistics and the 44th Annual Meeting of the Association for Computational Linguistics, ACL-44, pages 753–760, Stroudsburg, PA, USA, 2006. Anne-Lyse Minard, Manuela Speranza, Eneko Agirre, Itziar Aldabe, Marieke van Erp, Bernardo Magnini, German Rigau, Rubén Urizar, and Fondazione Bruno Kessler. SemEval-2015 Task 4: TimeLine: Cross-Document Event Ordering. In Proceedings of the 9th International Workshop on Semantic Evaluation (SemEval 2015). Association for Computational Linguistics, 2015. Paramita Mirza and Sara Tonelli. An analysis of causality between events and its relation to temporal information. In Proceedings of COLING 2014, the 25th International Conference on Computational Linguistics: Technical Papers, pages 2097–2106, Dublin, Ireland, August 2014. Dublin City University and Association for Computational Linguistics. Paramita Mirza and Sara Tonelli. Classifying Temporal Relations with Simple Features. In Proceedings of the 14th Conference of the European Chapter of the Association for Computational Linguistics, pages 308–317, Gothenburg, Sweden, 2014. Paramita Mirza, Rachele Sprugnoli, Sara Tonelli, and Manuela Speranza. Annotating causality in the tempeval-3 corpus. In Proceedings of the EACL 2014 Workshop on Computational Approaches to Causality in Language (CAtoCL), pages 10–19, Gothenburg, Sweden, April 2014. Association for Computational Linguistics. Luc Moreau, Paolo Missier, Khalid Belhajjame, Reza B’Far, James Cheney, Sam Coppens, Stephen Cresswell, Yolanda Gil, Paul Groth, Graham Klyne, Timothy Lebo, Jim McCusker, Simon Miles, James Myers, Satya Sahoo, and Curt Tilmes. PROV-DM: The PROV Data Model. Technical report, W3C, 2012. Kiem-Hieu Nguyen, Xavier Tannier, and Veronique Moriceau. Ranking multidocument event descriptions for building thematic timelines. In Proceedings of COLING‘14, pages 1208–1217, 2014. Martha S. Palmer, Deborah A. Dahl, Rebecca J. Schiffman, Lynette Hirschman, Marcia Linebarger, and John Dowding. Recovering implicit information. In Proceedings of NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 146/148 the 24th annual meeting on Association for Computational Linguistics, ACL ’86, pages 10–19, New York, New York, USA, 1986. F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay. Scikit-learn: Machine learning in python. In Journal of Machine Learning Research, 12: 2825–2830, 2011. Emanuele Pianta, Christian Girardi, and Roberto Zanoli. The textpro tool suite. In Proceedings of the Sixth International Conference on Language Resources and Evaluation (LREC’08), Marrakech, Morocco, may 2008. Emily Pitler and Ani Nenkova. Using syntax to disambiguate explicit discourse connectives in text. In Proceedings of the ACL-IJCNLP 2009 Conference Short Papers, ACLShort ’09, pages 13–16, Stroudsburg, PA, USA, 2009. Association for Computational Linguistics. Sameer Pradhan, Lance Ramshaw, Mitchell Marcus, Martha Palmer, Ralph Weischedel, and Nianwen Xue. Conll-2011 shared task: Modeling unrestricted coreference in ontonotes. In Proceedings of CoNLL 2011: Shared Task, 2011. Rashmi Prasad, Nikhil Dinesh, Alan Lee, Eleni Miltsakaki, Livio Robaldo, Aravind Joshi, and Bonnie Webber. The penn discourse treebank 2.0. In In Proceedings of LREC, 2008. James Pustejovsky, Bob Ingria, Roser Sauri, Jose Castano, Jessica Littman, Rob Gaizauskas, Andrea Setzer, Graham Katz, and Inderjeet Mani. The specification language timeml. The language of time: A reader, pages 545–557, 2005. Willard V. Quine. Events and reification. In Actions and Events: Perspectives on the Philosophy of Davidson, pages 162–71. Blackwell, 1985. Marta Recasens and Eduard Hovy. Blanc: Implementing the rand index for coreference evaluation. In Natural Language Engineering,17, (4), pages 485–510, 2011. Marco Rospocher, Anne-Lyse Minard, Paramita Mirza, Piek Vossen, Tommaso Caselli, Agata Cybulska, Roser Morante, and Itziar Aldabe. Event narrative module, version 2. NewsReader Deliverable 5.1.1, 2015. Marco Rospocher, Marieke van Erp, Piek Vossen, Antske Fokkens, Itziar Aldabe, German Rigau, Aitor Soroa, Thomas Ploeger, and Tessel Bogaard. Building event-centric knowledge graphs from news. Journal of Web Semantics, 2016. Marie-Laure Ryan. Possible Worlds, Artificial Intelligence and Narrative Theory. Bloomington: Indian University Press, 1991. Roser Saurı́, Jessica Littman, Robert Gaizauskas, Andrea Setzer, and James Pustejovsky. TimeML Annotation Guidelines, Version 1.2.1, 2006. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 147/148 Roser Saurı́. A factuality profiler for eventualities in text. PhD thesis, Brandeis University, Waltham, MA, USA, 2008. William F Styler IV, Steven Bethard, Sean Finan, Martha Palmer, Sameer Pradhan, Piet C de Groen, Brad Erickson, Timothy Miller, Chen Lin, Guergana Savova, et al. Temporal annotation in the clinical domain. Transactions of the Association for Computational Linguistics, 2:143–154, 2014. Joel R. Tetreault. Implicit role reference. In International Symposium on Reference Resolution for Natural Language Processing, pages 109–115, Alicante, Spain, 2002. The PDTB Research Group. The PDTB 2.0. Annotation Manual. Technical Report IRCS08-01, Institute for Research in Cognitive Science, University of Pennsylvania, 2008. Sara Tonelli, Rachele Sprugnoli, Manuela Speranza, and Anne-Lyse Minard. NewsReader Guidelines for Annotation at Document Level. Technical Report NWR2014-2-2, Fondazione Bruno Kessler, 2014. http://www.newsreader-project.eu/files/2014/12/ NWR-2014-2-2.pdf. Naushad UzZaman and James Allen. Temporal evaluation. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, pages 351–356, Portland, Oregon, USA, 2011. Naushad UzZaman, Hector Llorens, James Allen, Leon Derczynski, Marc Verhagen, and James Pustejovsky. Tempeval-3: Evaluating events, time expressions, and temporal relations. arXiv preprint arXiv:1206.5333, 2012. Naushad UzZaman, Hector Llorens, Leon Derczynski, James Allen, Marc Verhagen, and James Pustejovsky. Semeval-2013 task 1: Tempeval-3: Evaluating time expressions, events, and temporal relations. In Proceedings of the Seventh International Workshop on Semantic Evaluation, SemEval ’13, pages 1–9, Atlanta, Georgia, USA, 2013. Naushad UzZaman, Hector Llorens, Leon Derczynski, Marc Verhagen, James Allen, and James Pustejovsky. Semeval-2013 task 1: Tempeval-3: Evaluating time expressions, events, and temporal relations, 2013. Teun A. van Dijk. News As Discourse. Routledge, 1988. Marieke van Erp, Piek Vossen, Rodrigo Agerri, Anne-Lyse Minard, Manuela Speranza, Ruben Urizar, and Egoitz Laparra. Annotated data, version 2. NewsReader Deliverable 3.3.2, 2015. Willem Robert van Hage, Véronique Malaisé, Roxane Segers, Laura Hollink, and Guus Schreiber. Design and use of the Simple Event Model (SEM). J. Web Sem., 9(2):128– 136, 2011. http://dx.doi.org/10.1016/j.websem.2011.03.003. NewsReader: ICT-316404 February 1, 2016 Event Narrative Module, version 3 148/148 Marc Verhagen, Roser Sauri, Tommaso Caselli, and James Pustejovsky. Semeval-2010 task 13: Tempeval-2. In Proceedings of the 5th international workshop on semantic evaluation, pages 57–62. Association for Computational Linguistics, 2010. Marc Vilain, John Burger, John Aberdeen, Dennis Connolly, and Lynette Hirschman. A model theoretic coreference scoring scheme. In Proceedings of MUC-6, 1995. Piek Vossen, Francis Bond, and John P. McCrae. Toward a truly multilingual global wordnet grid. Proceedings of the Eighth meeting of the Global WordNet Conference (GWC 2016), Bucharest, 2016. Greg Whittemore, Melissa Macpherson, and Greg Carlson. Event-building through rolefilling and anaphora resolution. In Proceedings of the 29th annual meeting on Association for Computational Linguistics, ACL ’91, pages 17–24, Berkeley, California, USA, 1991. Phillip Wolff and Grace Song. Models of causation and the semantics of causal verbs. Cognitive Psychology, 47(3):276–332, 2003. Phillip Wolff. Representing causation. Journal of experimental psychology: General, 136(1):82, 2007. Shize Xu, Shanshan Wang, and Yan Zhang. Summarizing complex events: a cross-modal solution of storylines extraction and reconstruction. In Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, pages 1281–1291, Seattle, Washington, USA, October 2013. Association for Computational Linguistics. Bishan Yang, Claire Cardie, and Peter I. Frazier. A hierarchical distance-dependent bayesian model for event coreference resolution. CoRR, abs/1504.05929, 2015. NewsReader: ICT-316404 February 1, 2016