<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Big Data</journal-id>
<journal-title>Frontiers in Big Data</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Big Data</abbrev-journal-title>
<issn pub-type="epub">2624-909X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">779792</article-id>
<article-id pub-id-type="doi">10.3389/fdata.2021.779792</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Big Data</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Link Prediction Between Structured Geopolitical Events: Models and Experiments</article-title>
<alt-title alt-title-type="left-running-head">Kejriwal</alt-title>
<alt-title alt-title-type="right-running-head">Linking Geopolitical Events</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Kejriwal</surname>
<given-names>Mayank</given-names>
</name>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1076225/overview"/>
</contrib>
</contrib-group>
<aff>Viterbi School of Engineering, Information Sciences Institute, University of Southern California, <addr-line>Los Angeles</addr-line>, <addr-line>CA</addr-line>, <country>United&#x20;States</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1164642/overview">Philippe Giabbanelli</ext-link>, Miami University, United&#x20;States</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1490174/overview">Soumajyoti Sarkar</ext-link>, Arizona State University, United&#x20;States</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1491231/overview">Arunkumar Bagavathi</ext-link>, Oklahoma State University, United&#x20;States</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1180061/overview">Sridhar Nerur</ext-link>, University of Texas at Arlington, United&#x20;States</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Mayank Kejriwal, <email>kejriwal@isi.edu</email>
</corresp>
<fn fn-type="other">
<p>This article was submitted to Big Data Networks, a section of the journal Frontiers in Big&#x20;Data</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>30</day>
<month>11</month>
<year>2021</year>
</pub-date>
<pub-date pub-type="collection">
<year>2021</year>
</pub-date>
<volume>4</volume>
<elocation-id>779792</elocation-id>
<history>
<date date-type="received">
<day>19</day>
<month>09</month>
<year>2021</year>
</date>
<date date-type="accepted">
<day>12</day>
<month>11</month>
<year>2021</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2021 Kejriwal.</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Kejriwal</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these&#x20;terms.</p>
</license>
</permissions>
<abstract>
<p>Often thought of as higher-order entities, events have recently become important subjects of research in the computational sciences, including within complex systems and natural language processing (NLP). One such application is event link prediction. Given an input event, event link prediction is the problem of retrieving a <italic>relevant</italic> set of events, similar to the problem of retrieving relevant documents on the Web in response to keyword queries. Since geopolitical events have complex semantics, it is an open question as to how to best model and represent events within the framework of event link prediction. In this paper, we formalize the problem and discuss how established representation learning algorithms from the machine learning community could potentially be applied to it. We then conduct a detailed empirical study on the Global Terrorism Database (GTD) using a set of metrics inspired by the information retrieval community. Our results show that, while there is considerable signal in both network-theoretic and text-centric models of the problem, classic text-only models such as bag-of-words prove surprisingly difficult to outperform. Our results establish both a baseline for event link prediction on GTD, and currently outstanding challenges for the research community to tackle in this&#x20;space.</p>
</abstract>
<kwd-group>
<kwd>event representations</kwd>
<kwd>representation learning</kwd>
<kwd>geopolitical event link prediction</kwd>
<kwd>word embeddings</kwd>
<kwd>multi-partite networks</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="s1">
<title>1 Introduction</title>
<p>In recent years, there has been an increasing focus on representing, reasoning over and doing inference on, <italic>events</italic> <xref ref-type="bibr" rid="B14">Lorenzini et&#x20;al. (2021)</xref>, <xref ref-type="bibr" rid="B1">Battistelli et&#x20;al. (2013)</xref>. Unlike ordinary named entities, events are complex data structures, embodied by artifacts such as triggers, actors, locations, descriptions, and spatiotemporal cues. In the case of events with geopolitical consequences, such as terrorist attacks, assassinations, or bombings, automatically and accurately predicting links between events is an important research application that can be used to populate and enrich geopolitical, sparse (and proprietarily gathered) knowledge bases with global scope. <xref ref-type="fig" rid="F1">Figure&#x20;1</xref> provides a simplified illustration, based on real data, of a linked set of events.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>Illustrations of three geopolitical event fragments that are linked.</p>
</caption>
<graphic xlink:href="fdata-04-779792-g001.tif"/>
</fig>
<p>For such inferential tasks, Representation Learning (RL), an important sub-area of deep learning research, has emerged as extremely influential in both graph- and text-centric communities. In the general case, latent space embeddings (dense, real-valued vectors) are learned on graphs with simple structures, or [in the case of text embedding algorithms like GloVe and word2vec <xref ref-type="bibr" rid="B20">Pennington et&#x20;al. (2014)</xref>; <xref ref-type="bibr" rid="B16">Mikolov et&#x20;al. (2013)</xref>] on word or character sequences. On the other hand, structured event data contains rich structure and semantics that can be exploited to learn better representations.</p>
<p>As one important event-centric application, consider <italic>geopolitical forecasting</italic>, which continues to be an important and relevant problem, especially for policy institutes, think tanks, analysts and pollsters <xref ref-type="bibr" rid="B22">Tetlock (2017)</xref>, <xref ref-type="bibr" rid="B24">Tetlock (1992)</xref>. The problem is also known to be difficult, although the geographic region and complexity involved in making a forecast for an IFP (individual forecasting problem) can make some forecasts more challenging than others <xref ref-type="bibr" rid="B25">Tetlock (2014)</xref>. One of the critical tasks of forecasting is to distinguish relevant material from the irrelevant. This is especially true at the level of events e.g., when one is trying to forecast political instability in Nigeria, it is helpful to consider &#x201c;prototypical political instability&#x201d; events like riots or protests, and retrieve both recent such events, as well as links to other events that have a connection to the prototypical event <xref ref-type="bibr" rid="B4">Esser and Str&#xf6;mb&#xe4;ck (2013)</xref>, <xref ref-type="bibr" rid="B28">Zinken (2003)</xref>.</p>
<p>Obtaining and reasoning over such contextual and background knowledge is ever more important also because (arguably) it is becoming harder to make accurate forecasts, even for events that are being globally scrutinized and studied. In 2016, several incidents occurred globally that went against the predictions of famous (i.e.,&#x20;traditionally accurate) polls<xref ref-type="fn" rid="FN1">
<sup>1</sup>
</xref>, including the outcomes of Brexit and the 2016 US Presidential Election. While these examples may arguably be described as representing extremes [&#x201c;Black Swan&#x201d; incidents <xref ref-type="bibr" rid="B17">Nassim (2007)</xref>], <italic>consistently</italic> making correct forecasts is a valued skill that several individuals (and by extension, organizations) have been known to possess <xref ref-type="bibr" rid="B23">Tetlock and Gardner (2016)</xref>. A key differentiator between successful (on average) forecasters and ordinary forecasters is the ability to consider contextual and linked information when researching the forecasting problem. Events, such as COVID-19 and the US Capitol Hill riots following the Presidential 2020 election, only serve to highlight the severity and suddenness of Black Swan events.</p>
<p>We also note that while, on the surface, tasks like event link prediction may sound similar to link prediction as studied in social networks, the complex structure of an event makes the former much more difficult and ill-defined. This is also true for other event-centric problems like event resolution <xref ref-type="bibr" rid="B10">Kejriwal et&#x20;al. (2018b)</xref>, event extraction (in the natural language processing community) and event co-referencing <xref ref-type="bibr" rid="B19">Ng (2017)</xref>, <xref ref-type="bibr" rid="B15">Lu and Ng (2017)</xref>, for which special techniques have now been developed, as well as situational awareness and visualization <xref ref-type="bibr" rid="B9">Kejriwal et&#x20;al. (2018a)</xref>, <xref ref-type="bibr" rid="B11">Kejriwal and Zhou (2019)</xref>.</p>
<p>In this paper, we address the research problem of what features make for good event representations, both when text summaries are available, or unavailable. We propose and consider several models of events, including models that just use a short text description, a combination of text and locations, paths in a multi-layer semantic network, or in the most general case, novel models that rely on various statistical-semantic cues in both text- and graph-theoretic frameworks. Using both classic methods, such as cosine similarity applied on bag-of-words vectors, as well as deep embedding methods, we study and contrast these representations by conducting a full set of event link prediction experiments on the Global Terrorism Database (GTD) <xref ref-type="bibr" rid="B12">LaFree and Dugan (2007)</xref>. Our goal here is not to present novel algorithms but to introduce and present a rigorous methodology (including data and evaluation metrics) for studying event link prediction as a fundamental application area in multi-relational networks and complex systems.</p>
<p>Using various metrics inspired by the information retrieval and traditional link prediction communities <xref ref-type="bibr" rid="B13">Liben-Nowell and Kleinberg (2007)</xref>, we quantify the most salient aspects in learning good event representations, especially when a combination of structured and unstructured information sets may be available. To the best of our knowledge, this is the first such study to rigorously model, formalize and quantify event representation learning.</p>
</sec>
<sec id="s2">
<title>2 Problem Description</title>
<p>We begin by first defining and scoping the notion of an <italic>event</italic> as assumed by this article. If events were completely arbitrary, it would be sufficient to assume an event ontology (EO), and declare instances defined in terms of certain classes (event types and sub-types) in the EO to be &#x201c;events&#x201d;. In the real world, however, events, despite exhibiting heterogeneity across sources, domains and datasets, do obey some near-universal restrictions. First, events are generally <italic>typed</italic>, whether automatically or not. For example, the Global Terrorism Database (GTD) <xref ref-type="bibr" rid="B12">LaFree and Dugan (2007)</xref> describes terrorism events, as the name suggests, while datasets like the Armed Conflict Location and Event Data Project (ACLED)<xref ref-type="fn" rid="fn2">
<sup>2</sup>
</xref> or Political Instability Task Force (PITF)<xref ref-type="fn" rid="fn3">
<sup>3</sup>
</xref> contain a mix of events that are useful to geopolitical analysts. A good example of an ontology describing many event types, and that has been extensively used in the real world, is Conflict and Mediation Event Observations (CAMEO)<xref ref-type="fn" rid="fn4">
<sup>4</sup>
</xref>.</p>
<p>Second, a commonality between databases that describe geopolitical events is some notion of space and time. Although the granularity can differ (e.g., some highly proprietary event datasets may be associated with very precise geo-coordinates, while many others contain coarser information, like region and country names), some spatial and temporal information is almost always included. Third, depending on the event type, some structured information could be encoded using a highly controlled vocabulary. In ACLED, very specific (and consistent) terminology is used to indicate event modalities like riots or protests, for example, while in GTD (as we later describe), information like the attack type and target type obey a controlled vocabulary that is given by a codebook.</p>
<p>At the same time, events that have different modalities or provenance can also be very heterogeneous. Thus, it is important to be flexible in an event formalism to accommodate the &#x2018;quirks&#x2019; of individual datasets. With this caveat in mind, we can use the three observations above to define a geopolitical event <italic>E</italic> in the following way. Given an event ontology <italic>O</italic>, a geolocation ontology <italic>G</italic> and a temporal ontology <italic>T</italic>, a <italic>geopolitical event</italic> <italic>E</italic> may be defined as an <italic>instance</italic> of <italic>O</italic> with a spatiotemporal span <inline-formula id="inf1">
<mml:math id="m1">
<mml:mo>&#x3c;</mml:mo>
<mml:mi>g</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x3e;</mml:mo>
</mml:math>
</inline-formula>, with <italic>g</italic> and <italic>t</italic> being instances of <italic>G</italic> and <italic>T</italic> respectively. A good example of <italic>G</italic> is the GeoNames ontology <xref ref-type="bibr" rid="B26">Wick (2006)</xref>, which is widely used in spatial sciences and geography. However, <italic>G</italic> can also be an ontology that is extremely fine-grained such as the underlying ontology behind systems like Google Maps and OpenStreetMap (OSM) <xref ref-type="bibr" rid="B6">Haklay and Weber (2008)</xref>. In contrast, temporal ontologies are usually simple, although sophisticated options have been proposed in the literature <xref ref-type="bibr" rid="B7">Hobbs and Pan (2006)</xref>.</p>
<p>We note that, while this description abstracts the full scope of event databases (which can contain tens, if not hundreds, of fields in their schemas), it does not abstract away the fact that events are complex data structures that can contain a combination of free text, structured elements (such as date and location) and elements from controlled vocabularies (such as attack types for terrorist events). Unlike natural language text, or RDF graphs, it is not clear how to <italic>model</italic>, and <italic>learn representations</italic> for, events in a way that makes them amenable to advanced machine learning-centric analytics like link prediction, event classification or anomaly detection<xref ref-type="fn" rid="fn5">
<sup>5</sup>
</xref>. In keeping with established terminology, the learned representation of an event intuitively corresponds to a &#x201c;feature vector&#x201d; that can be used in (either supervised or unsupervised) machine learning systems for various classification and clustering problems.</p>
<p>With the advent of deep learning and embedding methods, modeling and representation have become linked. Intuitively, modeling an event (for the purposes described above) defines which <italic>information sets</italic> of an event are relevant, and what the relationships are between these information sets. We consider some models and information sets in a subsequent section. The representation learning is the application of an algorithm (whether developed from scratch, or from the literature, like word2vec) on a modeled set of events. In Natural Language Processing (NLP) terminology, modeling determines the definition of a context<xref ref-type="fn" rid="fn6">
<sup>6</sup>
</xref>, while representation learning uses the context to embed events into a vector space, which is usually (but not always, as we explain later) dense and real-valued.</p>
</sec>
<sec id="s3">
<title>3 Materials and Methods</title>
<sec id="s3-1">
<title>3.1 Global Terrorism Database (GTD)</title>
<p>Before describing the models, we start by describing an important dataset called Global Terrorism Database (GTD) that contains thousands of structured terrorism events spanning the globe, typed according to an expansive schema. As described on the project page<xref ref-type="fn" rid="fn7">
<sup>7</sup>
</xref>, GTD is an &#x201c;open-source database including information on terrorist events around the world from 1970 through 2016 (with annual updates planned for the future). Unlike many other event databases, the GTD includes systematic data on domestic as well as international terrorist incidents that have occurred during this time period and now includes more than 170,000 cases&#x201d;.</p>
<p>Because GTD is relatively clean, it allows us to quantify many of the (subsequently discussed) models without being concerned about bias being caused by specific kinds of noise that are usually non-random and caused by imperfect information extraction algorithms (that are still active areas of research). The dataset profile is provided in <xref ref-type="table" rid="T1">Table&#x20;1</xref>, categorized by <italic>attack types</italic> (<xref ref-type="table" rid="T2">Table&#x20;2</xref>). The profile shows that GTD is quite diverse, and spans multiple world regions and countries, a range of target types, and covers events that are fairly recent, allowing us to relevantly extend the conclusions drawn in this article to modern-day events.</p>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>A profile of GTD events, categorized by attack types. The number in the first column is an attack type code; see <xref ref-type="table" rid="T2">Table&#x20;2</xref> for the codebook.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Attack type</th>
<th align="center">Num. events</th>
<th align="center">Num. unique regions</th>
<th align="center">Num. unique countries</th>
<th align="center">Num. unique target types</th>
<th align="center">Date range</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">0</td>
<td align="center">41</td>
<td align="center">9</td>
<td align="center">21</td>
<td align="center">4</td>
<td align="center">2013/01/21- 2016/12/01</td>
</tr>
<tr>
<td align="left">1</td>
<td align="center">3,478</td>
<td align="center">11</td>
<td align="center">79</td>
<td align="center">18</td>
<td align="center">2013/01/01- 2016/12/31</td>
</tr>
<tr>
<td align="left">2</td>
<td align="center">12,856</td>
<td align="center">12</td>
<td align="center">98</td>
<td align="center">21</td>
<td align="center">2013/01/01- 2016/12/31</td>
</tr>
<tr>
<td align="left">3</td>
<td align="center">29,683</td>
<td align="center">12</td>
<td align="center">101</td>
<td align="center">21</td>
<td align="center">2013/01/01- 2016/12/31</td>
</tr>
<tr>
<td align="left">4</td>
<td align="center">144</td>
<td align="center">9</td>
<td align="center">28</td>
<td align="center">15</td>
<td align="center">2013/02/06- 2016/12/23</td>
</tr>
<tr>
<td align="left">5</td>
<td align="center">205</td>
<td align="center">10</td>
<td align="center">39</td>
<td align="center">18</td>
<td align="center">2013/01/16- 2016/12/23</td>
</tr>
<tr>
<td align="left">6</td>
<td align="center">4,239</td>
<td align="center">10</td>
<td align="center">62</td>
<td align="center">20</td>
<td align="center">2013/01/01- 2016/12/31</td>
</tr>
<tr>
<td align="left">7</td>
<td align="center">2,675</td>
<td align="center">12</td>
<td align="center">89</td>
<td align="center">22</td>
<td align="center">2013/01/01- 2016/12/28</td>
</tr>
<tr>
<td align="left">8</td>
<td align="center">251</td>
<td align="center">10</td>
<td align="center">41</td>
<td align="center">13</td>
<td align="center">2013/01/08- 2016/12/26</td>
</tr>
<tr>
<td align="left">9</td>
<td align="center">2,791</td>
<td align="center">9</td>
<td align="center">55</td>
<td align="center">19</td>
<td align="center">2013/01/10- 2016/12/31</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap id="T2" position="float">
<label>TABLE 2</label>
<caption>
<p>Attack type codebook employed in GTD.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Attack type code</th>
<th align="center">Description</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">0</td>
<td>Description unavailable</td>
</tr>
<tr>
<td align="left">1</td>
<td>Assassination</td>
</tr>
<tr>
<td align="left">2</td>
<td>Armed Assault</td>
</tr>
<tr>
<td align="left">3</td>
<td>Bombing/Explosion</td>
</tr>
<tr>
<td align="left">4</td>
<td>Hijacking</td>
</tr>
<tr>
<td align="left">5</td>
<td>Hostage Taking (Barricade Incident)</td>
</tr>
<tr>
<td align="left">6</td>
<td>Hostage Taking (Kidnapping)</td>
</tr>
<tr>
<td align="left">7</td>
<td>Facility/Infrastructure Attack</td>
</tr>
<tr>
<td align="left">8</td>
<td>Unarmed Assault</td>
</tr>
<tr>
<td align="left">9</td>
<td>Unknown</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>As with any study and set of experiments, it is important to keep in mind the limitations of GTD, including data coverage. Currently, it is unknown if GTD is biased toward events of a specific type (whether involving a specific attack type, target type, geographical region, number of individuals involved, and so on), since published research on event databases and their analysis continues to be sparse. We do not claim that GTD is perfect; our goal in using it is to ensure that, caveats notwithstanding, we are able to conduct sufficiently controlled experiments and draw suitable conclusions. Future work will attempt to add more degrees of freedom to these studies.</p>
<sec id="s3-1-1">
<title>3.1.1 Link Prediction Ground Truth</title>
<p>GTD includes a column that states the event IDs linked with a given event ID. More than one event ID can be linked to a given event ID. In this ground truth, we found that the number of <italic>reference</italic> events<xref ref-type="fn" rid="fn8">
<sup>8</sup>
</xref> was 10,259&#x20;i.e.,&#x20;most of the events in GTD are isolated and are not linked with any other (at least to the extent that it is currently known). The average number of linked events per reference event was found to be 5.204 and the number of <italic>ordered</italic> linked event pairs is 53,392. A frequency distribution is illustrated in <xref ref-type="fig" rid="F2">Figure&#x20;2</xref>. By ordered, we mean that a linked pair (<italic>event</italic>
<sub>1</sub>, <italic>event</italic>
<sub>2</sub>) is considered distinct from (<italic>event</italic>
<sub>2</sub>, <italic>event</italic>
<sub>1</sub>). There is a reason for this methodology. Usually, in event retrieval scenarios, analysts have an event in mind already and are executing a <italic>more-like-this</italic> task whereby the goal is to retrieve linked events from a database of events. As we describe later, given such a reference event, the correct way to evaluate a retrieval system is to produce a ranked list of candidate events and then quantify the performance of the ranking using information retrieval metrics. Because the rankings are with respect to a reference event, and can be asymmetric (it is possible for <italic>event</italic>
<sub>1</sub> to be ranked high when <italic>event</italic>
<sub>2</sub> is the reference vector, but not the other way around), it is appropriate to consider pairs to be ordered rather than unordered.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>GTD link prediction ground-truth frequency&#x20;plot.</p>
</caption>
<graphic xlink:href="fdata-04-779792-g002.tif"/>
</fig>
</sec>
</sec>
<sec id="s3-2">
<title>3.2 Models</title>
<p>Given the event definition in <xref ref-type="sec" rid="s2">Section 2</xref>, we explore in this section the information sets that are expected to be useful for representation learning, especially as applied to a downstream task like link prediction. One reason to consider several possible information sets is that there is a natural tradeoff between including more information in the model, which could lead to richer and finer-grained representations, but that may not generalize as well (especially if some of the information is missing in some records). On the other hand, a model that is too coarse (e.g., that only considers the region in which the event takes place) will likely not be able to distinguish between sufficiently many events and will have poor retrieval performance.</p>
<p>Given that events are usually accompanied by text descriptions in databases such as GTD, the simplest possible information set is <italic>text</italic>. A <italic>text-centric model</italic> can be constructed by simply taking the description (also called the &#x2018;summary&#x2019; in GTD) and not assuming or using any other structure.</p>
<p>At the other extreme is a model that only takes the graph structure into account. We refer to such a model as a <italic>multi-partite semantic network (MPSN)</italic> model, illustrated in <xref ref-type="fig" rid="F3">Figure&#x20;3</xref>. Assuming the model is represented as an edge-list, each event is represented using an &#x201c;event ID&#x201d; vertex, with edges linking the vertex to any other vertices that characterize the event. An important point to note here is that the different &#x201c;semantic layers&#x201d; in the network must form a closed set i.e.,&#x20;the vertices must be pre-specified in advance. This implies that we know the regions, attack types (and so on) that are in our domain. Constructing multi-partite semantic networks over open sets of nodes is not a well-defined problem at the present moment for the purposes of specifying and learning representations on a network.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>Events modeled as nodes in a multi-partite semantic network (MPSN). We only consider events that have at least one piece of information associated with it (besides an event ID).</p>
</caption>
<graphic xlink:href="fdata-04-779792-g003.tif"/>
</fig>
<p>However, although the text-based and MPSN have their respective advantages, one could presumably aim to take advantage of both. On the one hand, it is difficult to frame the text as a graph without losing its natural language structure, and to frame the graph as text. On the other hand, representation learning (and downstream link prediction) can accept heterogeneous information sets as input. Keeping this in mind, and with a view towards simplicity in the modeling stage, we model an event with both semantic and text information sets as <italic>E</italic>&#x20;&#x3d; &#x3c; <italic>E</italic>
<sub>
<italic>T</italic>
</sub>, <italic>E</italic>
<sub>
<italic>N</italic>
</sub> &#x3e;, where <italic>E</italic>
<sub>
<italic>T</italic>
</sub> and <italic>E</italic>
<sub>
<italic>N</italic>
</sub> are the text-centric (expressed simply as a string) and MPSN (expressed as a sub-graph of a network such as the one in <xref ref-type="fig" rid="F3">Figure&#x20;3</xref>) representations of the event. We refer to each component of the tuple above as an <italic>event&#x20;facet</italic>.</p>
<p>One can even generalize the notion above, where an event has multiple facets, and each facet captures a unique combination of information sets. For example, we described earlier how multiple text-centric and MPSNs are possible. Rather than pick one or the other, one could consider &#x201c;bags&#x201d; of models by including each model as an event facet. Whether it is worthwhile to do so would depend both on how the representation learning processes this data. We provide a brief set of results exploring such &#x201c;hybrid&#x201d; models in <xref ref-type="sec" rid="s5">Section&#x20;5</xref>.</p>
</sec>
<sec id="s3-3">
<title>3.3 Representation Learning on Models</title>
<p>Over the last decade, neural network models like skip-gram and continuous bag of words (CBOW) have been used with great effect<xref ref-type="fn" rid="fn9">
<sup>9</sup>
</xref> to embed sequences and structured data (like graphs) into a dense, real-valued vector space <xref ref-type="bibr" rid="B16">Mikolov et&#x20;al. (2013)</xref>. The vectors can be used as features for link prediction. An important question arises as to how to embed each of the event models described in the previous section. For text-only models, an appropriate neural text embedding such as the word2vec, or even classic methods like the &#x201c;bag-of-words&#x201d;, or term-frequency inverse document frequency model (TF-IDF) can be applied. For graph-only models such as the MPSN, a node embedding representation learning model such as DeepWalk or node2vec can be applied <xref ref-type="bibr" rid="B21">Perozzi et&#x20;al. (2014)</xref>, <xref ref-type="bibr" rid="B5">Grover and Leskovec (2016)</xref>. In other cases, such as the hybrid models, it is not clear what the best way to proceed is e.g., one could independently embed the text and graph components and concatenate the feature vectors, or train some kind of joint model. We detail these possibilities next, with more details on &#x201c;hybrid&#x201d; models and results in <xref ref-type="sec" rid="s5">Section&#x20;5</xref>.</p>
<p>First, concerning the text-centric models, two such models can be considered. In the first model, an event <italic>E</italic> was simply represented by its description or &#x201c;summary&#x201d; field. In the second model, arbitrary string field values (which have non-numeric primitive types, thereby precluding the use of dates and integers that may cause noise in such representations) can be &#x201c;concatenated&#x201d; together to form a single text field. The idea is to supplement the description where it is sparse, or not distinctive enough between events. For either model, the final &#x201c;value&#x201d; for the event is a single text string, and the event database is exactly like a document corpus in an NLP setting.</p>
<p>Representation learning on such a corpus at the document level can be achieved in two ways, one adaptive and the other non-adaptive. The non-adaptive model is the bag-of-words model (also called tf-idf). It has been used prolifically in information retrieval and document classification settings over the decade, and remains both scalable and interpretable. However, one issue with its usage is that the dimensions of the tf-idf vectors are sparse, large and not able to easily generalize to unseen words, or even words with similar meaning. Statistically, these issues were not problematic when the corpus was extremely large, such as search engines can often avail&#x20;of.</p>
<p>More recently, however, adaptive models such as word embeddings have achieved scale, while addressing the dimensionality and sparsity issues, and thereby achieving better performance without necessarily requiring more data. Specifically, the adaptive model &#x201c;learns&#x201d; representations of individual words using a neural network like skip-gram. A second model is the continuous bag of words (CBOW) model, but it has typically found to be outperformed by skip-gram. For more details on CBOW, and also other representation learning methods, we refer the reader to an earlier paper on word2vec <xref ref-type="bibr" rid="B16">Mikolov et&#x20;al. (2013)</xref>. Our text embedding relies on a more recent version of word2vec called fastText or &#x201c;bag-of-tricks embedding&#x201d; that was released openly by Facebook AI research and is able to more gracefully handle misspelled and unusual words <xref ref-type="bibr" rid="B8">Joulin et&#x20;al. (2016)</xref>.</p>
<p>Concerning representation learning on the MPSN model, we note that it is more complex than the text, but can be embedded using a graph embedding algorithm. Most graph embedding algorithms are technically designed for uni-partite or uni-type graphs, as is often observed for social networks where every entity is (for testing purposes) an individual, and every link is akin to a friendship (or follower/followee) link. While one could (in theory) devise complex schemes from scratch for multi-partite graphs, a consensus on such representation learning models has yet to emerge in the machine learning community. Therefore, for the purposes of the empirical study and methodology in this paper, we &#x201c;treat&#x201d; the MPSN as an ordinary, undirected network. In turn, this implies that all nodes (and not just <italic>event_id</italic> nodes in the event partition) will get embedded by the algorithm; however, only <italic>event_id</italic> node embeddings will get used during the ranking phase. For the embedding algorithm itself, we use DeepWalk <xref ref-type="bibr" rid="B21">Perozzi et&#x20;al. (2014)</xref>, although algorithms like node2vec <xref ref-type="bibr" rid="B5">Grover and Leskovec (2016)</xref> could also be considered by future researchers to obtain potential performance increases.</p>
<p>An important point to note about all methods presented thus far is that they work in an unsupervised fashion i.e.,&#x20;no &#x201c;training&#x201d; links need to be observed for the system to predict them during test time. Initially, therefore, no two event_ids in the network are directly linked, though many have indirect links (e.g., if they share a location). The reason that unsupervised link prediction between events is important was mentioned earlier, namely, geopolitical events such as terrorist attacks tend to be special, irregular and &#x201c;black swan&#x201d; events by definition <xref ref-type="bibr" rid="B17">Nassim (2007)</xref>, and it is not practical to assume that a machine (especially, deep) learning system can be trained on a sparse set of events and still be expected to generalize well. Although investigating the issue of transfer learning (as applicable to this task) is a promising direction <xref ref-type="bibr" rid="B27">Zhuang et&#x20;al. (2020)</xref>, it is beyond the scope of this current work and we leave it for future research to pursue.</p>
</sec>
<sec id="s3-4">
<title>3.4 Quality Evaluation Metrics</title>
<p>We consider several metrics that are regularly employed both in the link prediction as well as in the knowledge graph embeddings literature. These metrics include <italic>Hits@10</italic>, <italic>Precision@k</italic>, <italic>Recall@k</italic>, and <italic>Mean Rank</italic>. These metrics are defined more completely below, but a common aspect of the metrics is that they are used to evaluate mechanisms that take an event as input, and output a ranked list of events as output. When events are represented as vectors, cosine similarity is used to generate the rankings. Namely, given an input (event) vector <italic>e</italic>
<sub>
<italic>in</italic>
</sub>, the ranked list <italic>E</italic>
<sub>
<italic>out</italic>
</sub> &#x3d; [<italic>e</italic>
<sub>1</sub>, &#x2026;, <italic>e</italic>
<sub>&#x7c;<italic>E</italic>&#x7c;&#x2212;1</sub>] is generated<xref ref-type="fn" rid="fn10">
<sup>10</sup>
</xref> by computing the cosine similarity between <italic>e</italic>
<sub>
<italic>in</italic>
</sub> and each event vector in <italic>E</italic>&#x20;&#x2212; <italic>e</italic>
<sub>
<italic>in</italic>
</sub>, where <italic>E</italic> is the set of all events. The ranked list obeys the rule <italic>cosineSim</italic>(<italic>e</italic>
<sub>1</sub>, <italic>e</italic>
<sub>
<italic>in</italic>
</sub>) &#x2265;&#x2026;&#x2265; <italic>cosineSim</italic>(<italic>e</italic>
<sub>&#x7c;<italic>E</italic>&#x7c;&#x2212;1</sub>, <italic>e</italic>
<sub>
<italic>in</italic>
</sub>). However, for some metrics (Hits@10 and Mean Rank) it is standard to &#x2018;filter&#x2019; the set <italic>E</italic> for a linked event pair (in the ground truth) (<italic>e</italic>
<sub>
<italic>in</italic>
</sub>, <italic>e</italic>
<sub>
<italic>i</italic>
</sub>) by removing from <italic>E</italic> all events <italic>except</italic> <italic>e</italic>
<sub>
<italic>i</italic>
</sub> that are <italic>also</italic> linked with <italic>e</italic>
<sub>
<italic>in</italic>
</sub>. This ensures that the ideal rank for <italic>e</italic>
<sub>
<italic>i</italic>
</sub> (given <italic>e</italic>
<sub>
<italic>in</italic>
</sub> as input) should always be 1, since there is no danger that another &#x2018;relevant&#x2019; entry is above it in the ranked list. We now define the metrics:</p>
<p>
<italic>Hits@10</italic>: Given an event <italic>e</italic>
<sub>
<italic>in</italic>
</sub> as input, and a (with-held) ground-truth linked pair (<italic>e</italic>
<sub>
<italic>in</italic>
</sub>, <italic>e</italic>
<sub>
<italic>i</italic>
</sub>), the Hits@10 metric measures whether <italic>e</italic>
<sub>
<italic>i</italic>
</sub> is in the top-10. It is important to note that Hits@10 is evaluated independently for each pair of events linked in the ground truth. As we noted in the example above, an event can be linked with more than one event, which necessitates removing true positives (except the true positive that is in the pair) from the full set of events before evaluating the ranking.</p>
<p>
<italic>Mean Reciprocal Rank (MRR)</italic>: The MRR is the reciprocal of the rank at which <italic>e</italic>
<sub>
<italic>i</italic>
</sub> occurs. Unlike Hits@10, it can be non-zero if <italic>e</italic>
<sub>
<italic>i</italic>
</sub> is not in the top-10 though it declines very quickly. MRR is evaluated in a similar way to Hits@10 in that the event set has to be filtered prior to ranking for a given input event and a withheld linked pair of events. Because of the event filtering, the optimal MRR is always&#x20;1.</p>
<p>
<italic>Normalized Discounted Cumulative Gain (NDCG)</italic>: The MRR has several issues, the most important of which is that it is designed to work for only one relevant item per input, and declines quickly the further away that relevant item is from the top of the list. The NDCG is widely used in the information retrieval community as a more robust measure. Unlike the previously described metrics, NDCG does not assume that a given event is only linked to one other event, and hence, filtering is unnecessary.</p>
<p>To compute the NDCG, we first have to calculate the DCG for input event <italic>e</italic>
<sub>
<italic>in</italic>
</sub>, defined by the following equation:<disp-formula id="e1">
<mml:math id="m2">
<mml:mi>D</mml:mi>
<mml:mi>C</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>e</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>r</mml:mi>
<mml:mi>e</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>l</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:mfrac>
<mml:mrow>
<mml:mi>r</mml:mi>
<mml:mi>e</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>l</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">log</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mfrac>
</mml:math>
<label>(1)</label>
</disp-formula>
</p>
<p>Here, <italic>rel</italic>
<sub>
<italic>i</italic>
</sub> is the relevance of the <italic>i</italic>
<sup>
<italic>th</italic>
</sup> item in a ranked list of size <italic>n</italic>. In our case, this is either a 1 (if the event in <italic>E</italic>&#x20;&#x2212; <italic>e</italic>
<sub>
<italic>in</italic>
</sub> at that rank is paired with <italic>e</italic>
<sub>
<italic>in</italic>
</sub> in the GTD link prediction ground truth) or a 0. We can compute the DCG of both the actual ranking and of an <italic>ideal</italic> ranking (where all relevant items are ranked at the top), the latter denoted as the IDCG (Ideal DCG). The NDCG is then given by:<disp-formula id="e2">
<mml:math id="m3">
<mml:mi>N</mml:mi>
<mml:mi>D</mml:mi>
<mml:mi>C</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>e</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mi>C</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>e</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi>I</mml:mi>
<mml:mi>D</mml:mi>
<mml:mi>C</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>e</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:math>
<label>(2)</label>
</disp-formula>
</p>
<p>Note that the NDCG is between 0.0 and 1.0, since the DCG is always less than the IDCG. Similar to MRR and Hits@10, to obtain performance over the entire set of input events in the ground-truth, we average the NDCG obtained per input&#x20;event.</p>
<p>We note finally that for each of the models described in the previous section, the evaluation is on a uniform footing because 1) each method is <italic>unsupervised</italic>; 2) each method <italic>represents</italic> an event as a vector<xref ref-type="fn" rid="fn11">
<sup>11</sup>
</xref>; 3) the ranked list (for an input event) for each method is generated in an identical way, namely using cosine similarity. In turn, this implies that, within the scope of the event link prediction task, we can use the results to evaluate the power of the representation (and where applicable, its &#x2018;learning&#x2019; using contexts and neural networks) itself.</p>
</sec>
</sec>
<sec id="s4">
<title>4 Results</title>
<p>
<xref ref-type="table" rid="T3">Table&#x20;3</xref> reports results for the text-centric models introduced in <xref ref-type="sec" rid="s3-3">Section 3.2</xref>. We consider using only the &#x201c;summary&#x201d; or description field, as discussed therein, as well as the concatenation of all text-based fields, which includes both the summary, as well as the &#x201c;location&#x201d; field. Note that other fields, such as attack type, date, and so on, are categorical or numerical. Two important things stand out from the table. First, text-based methods generally do quite well, as long as the summary is included. As might be expected, using location alone leads to&#x20;very noisy results<xref ref-type="fn" rid="fn12">
<sup>12</sup>
</xref>. Second, we find that the classic tf-idf method is difficult to outperform, with the embedding-based method doing significantly worse no matter the experimental setting. In other work, the embedding-based method usually outperforms the tf-idf, although the margin depends significantly on the dataset. It is possible that transformer-based models such as BERT may end up outperforming the tf-idf but we leave an evaluation of this hypothesis for future research. Overall, the results are quite promising: an MRR of 57.88% (the best result, using the simplest possible combination of tf-idf on the summary field) implies that, on average, given an input event, the best method is able to retrieve a relevant result between ranks 1 and 2. The NDCG suggests that the performance gets even better once&#x20;we consider the unfiltered version of the dataset wherein an input event can have multiple relevant events linked to it in the ground-truth.</p>
<table-wrap id="T3" position="float">
<label>TABLE 3</label>
<caption>
<p>Results of text-rich systems on the event link prediction task. Metrics are described in <xref ref-type="sec" rid="s3-4">Section 3.4</xref>. In all cases below, the ranking is generated using the cosine similarity function between the vectors.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Representation method</th>
<th align="center">Field(s) being represented</th>
<th align="center">MRR</th>
<th align="center">Hits@10</th>
<th align="center">NDCG</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">tf-idf</td>
<td>Summary</td>
<td align="char" char=".">0.5788</td>
<td align="char" char=".">0.9821</td>
<td align="char" char=".">0.7482</td>
</tr>
<tr>
<td align="left">Bag-of-tricks embeddings</td>
<td>Summary</td>
<td align="char" char=".">0.5247</td>
<td align="char" char=".">0.9043</td>
<td align="char" char=".">0.6883</td>
</tr>
<tr>
<td align="left">tf-idf</td>
<td>Summary &#x2b; Location</td>
<td align="char" char=".">0.5593</td>
<td align="char" char=".">0.9657</td>
<td align="char" char=".">0.7339</td>
</tr>
<tr>
<td align="left">Bag-of-tricks embeddings</td>
<td>Summary &#x2b; Location</td>
<td align="char" char=".">0.5149</td>
<td align="char" char=".">0.8944</td>
<td align="char" char=".">0.6838</td>
</tr>
<tr>
<td align="left">tf-idf</td>
<td>Location</td>
<td align="char" char=".">0.0406</td>
<td align="char" char=".">0.0659</td>
<td align="char" char=".">0.1111</td>
</tr>
<tr>
<td align="left">Bag-of-tricks embeddings</td>
<td>Location</td>
<td align="char" char=".">0.0377</td>
<td align="char" char=".">0.05995</td>
<td align="char" char=".">0.1017</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>
<xref ref-type="table" rid="T4">Table&#x20;4</xref> reports results for the MPSN methods. We find that performance is significantly worse than the text-based methods; however, as more information is included in the MPSN model, the performance starts increasing. This suggests that the problem is one of information scarcity, not faults with the model or representation learning itself. It also provides some guidance on the &#x2018;information gap&#x2019; between the structured attributes, such as attack type (AT) and target type (TT) compared to the text. Indeed, in comparing the <italic>AT &#x2b; TT &#x2b; Country &#x2b; Region</italic> results to the next two information sets in the table, we find the critical importance of the temporal component of the event. Unfortunately, many NLP algorithms that extract events from text focus more on actors and triggers rather than on temporal prediction. For such extractions to have value in graph-based link prediction tasks, more attention needs to be given to accurately populating spatiotemporal spans of events, and on acquiring sufficiently robust descriptions, perhaps by applying a summarization algorithm on event field reports <xref ref-type="bibr" rid="B18">Nenkova and McKeown (2012)</xref>.</p>
<table-wrap id="T4" position="float">
<label>TABLE 4</label>
<caption>
<p>Results of graph-based methods on the event link prediction task. The representation learning in all cases was the DeepWalk algorithm.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Layers used</th>
<th align="center">MRR</th>
<th align="center">Hits@10</th>
<th align="center">NDCG</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">Attack Type (AT)&#x2b;Target Type (TT)</td>
<td align="char" char=".">0.0054</td>
<td align="char" char=".">0.0112</td>
<td align="char" char=".">0.0575</td>
</tr>
<tr>
<td align="left">AT &#x2b; TT &#x2b; Country &#x2b; Region</td>
<td align="char" char=".">0.0474</td>
<td align="char" char=".">0.1375</td>
<td align="char" char=".">0.2032</td>
</tr>
<tr>
<td align="left">AT &#x2b; TT &#x2b; Date</td>
<td align="char" char=".">0.2936</td>
<td align="char" char=".">0.7428</td>
<td align="char" char=".">0.5342</td>
</tr>
<tr>
<td align="left">AT &#x2b; TT &#x2b; Country &#x2b; Region &#x2b; Date</td>
<td align="char" char=".">0.3678</td>
<td align="char" char=".">0.8637</td>
<td align="char" char=".">0.6278</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s5">
<title>5 Discussion</title>
<p>Results in the previous section showed that there is obvious value in both structured attributes, and in the text descriptions, although the latter seems to contain more information than the former for machine learning purposes. In <xref ref-type="sec" rid="s3-3">Section 3.2</xref>, we suggested the possibility of &#x201c;combining&#x201d; the text-centric model with an MPSN in what was referred to as a &#x201c;hybrid&#x201d; model. Herein, we briefly illustrate two separate results, one that uses a <italic>joint</italic> model and another that uses an <italic>ensemble</italic>&#x20;model.</p>
<p>Each of these two models accepts as input one of the structured information sets in <xref ref-type="table" rid="T4">Table&#x20;4</xref> and the &#x2018;Summary&#x2019; field. The ensemble model concatenates the bag-of-tricks embedding (which is set to have the same dimensionality as the DeepWalk network embedding applied on the MPSN) with the MPSN network embedding. We then apply the cosine similarity, as earlier, except that the vector is a concatenation of two vectors<xref ref-type="fn" rid="fn13">
<sup>13</sup>
</xref>. The joint model, in contrast, adds the summary field as another &#x201c;information set&#x201d; layer in the MPSN model illustrated in <xref ref-type="fig" rid="F3">Figure&#x20;3</xref>. Because the field is text-based, rather than structured, we combine its sentences with the random walks output by DeepWalk prior to the algorithm applying the classic word2vec on the random walks. The joint model is therefore hierarchical: it combines the sentences in the summary field with the random walks, thereby embedding words and vertices in a joint setting. The results for both models are illustrated in <xref ref-type="table" rid="T5">Table&#x20;5</xref>.</p>
<table-wrap id="T5" position="float">
<label>TABLE 5</label>
<caption>
<p>Results of hybrid methods on the event link prediction task. E and J respectively indicate whether the method is an &#x201c;ensemble&#x201d; or &#x201c;joint&#x201d; method.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Graph method</th>
<th align="center">E</th>
<th align="center">J</th>
<th align="center">MRR</th>
<th align="center">Hits@10</th>
<th align="center">NDCG</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">AT &#x2b; TT</td>
<td align="center">&#x2014;</td>
<td align="center">X</td>
<td align="char" char=".">0.1864</td>
<td align="char" char=".">0.3761</td>
<td align="char" char=".">0.2366</td>
</tr>
<tr>
<td align="left">AT &#x2b; TT &#x2b; Country &#x2b; Region</td>
<td align="center">&#x2014;</td>
<td align="center">X</td>
<td align="char" char=".">0.2077</td>
<td align="char" char=".">0.4412</td>
<td align="char" char=".">0.3294</td>
</tr>
<tr>
<td align="left">AT &#x2b; TT &#x2b; Date</td>
<td align="center">&#x2014;</td>
<td align="center">X</td>
<td align="char" char=".">0.3184</td>
<td align="char" char=".">0.7341</td>
<td align="char" char=".">0.5530</td>
</tr>
<tr>
<td align="left">AT &#x2b; TT &#x2b; Country &#x2b; Region &#x2b; Date</td>
<td align="center">&#x2014;</td>
<td align="center">X</td>
<td align="char" char=".">0.3772</td>
<td align="char" char=".">0.8467</td>
<td align="char" char=".">0.6282</td>
</tr>
<tr>
<td align="left">AT &#x2b; TT</td>
<td align="center">X</td>
<td align="center">&#x2014;</td>
<td align="char" char=".">0.0901</td>
<td align="char" char=".">0.2244</td>
<td align="char" char=".">0.1717</td>
</tr>
<tr>
<td align="left">AT &#x2b; TT &#x2b; Country &#x2b; Region</td>
<td align="center">X</td>
<td align="center">&#x2014;</td>
<td align="char" char=".">0.1221</td>
<td align="char" char=".">0.3262</td>
<td align="char" char=".">0.2827</td>
</tr>
<tr>
<td align="left">AT &#x2b; TT &#x2b; Date</td>
<td align="center">X</td>
<td align="center">&#x2014;</td>
<td align="char" char=".">0.3786</td>
<td align="char" char=".">0.8343</td>
<td align="char" char=".">0.6156</td>
</tr>
<tr>
<td align="left">AT &#x2b; TT &#x2b; Country &#x2b; Region &#x2b; Date</td>
<td align="center">X</td>
<td align="center">&#x2014;</td>
<td align="char" char=".">0.4168</td>
<td align="char" char=".">0.9238</td>
<td align="char" char=".">0.6623</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>We find again (analogous to the results in <xref ref-type="table" rid="T4">Table&#x20;4</xref>) that the best results are achieved when the full information set is used, with the ensemble model achieving an almost 4% improvement on NDCG and MRR compared to the joint model. While both text and graph attributes have value, combining them in a single embedding framework clearly requires more thought, and an independent summing (as the ensemble model achieves) may be a safer approach in the absence of a large training dataset that could be used to fine-tune such a model. We also find that none of the methods, even in the hybrid setting, outperforms the classic tf-idf using just the summary field, illustrating that, on difficult problems like geopolitical event link prediction that are sufficiently different from benchmark link prediction problems often encountered in the literature, the utility of classic methods cannot be discounted. However, there is still much work to be done on the performance front since no method exceeds an NDCG of 80%. Another promising line of future work is to consider a <italic>supervised</italic> version of the problem wherein, in lieu of using cosine similarity on embeddings in an unsupervised framework, a classifier would be trained using a limited quantity of provided linked events (training data), with the embeddings as <italic>features</italic>. Such a classifier should yield better performance than the unsupervised methods presented in this article as initial approaches. Within the supervised learning paradigm, metrics such as accuracy, precision, recall and F1-Measure could also be applied to assess linking quality.</p>
</sec>
<sec id="s6">
<title>6 Conclusion</title>
<p>In this paper, we introduced and presented an empirical study on the problem of event link prediction. We presented various viable models for addressing the problem, derived from established literature on representation learning, followed by a detailed set of results using metrics inspired by the information retrieval community that has previously been applied mostly to Web retrieval and social networks.</p>
<p>Many questions still remain and constitute valuable opportunities for future research. First, it is not clear if the superior performance of text representations (even using simple bag-of-words approach) is fundamentally because the text contains much more information than the graph attributes do or because we have not designed or applied a sufficiently powerful representational model. For instance, it may very well be the case that the particular multi-partite semantic network model that we considered for representing an event-record is unsuitable, and a different model may prove to be more suitable. Similar concerns may apply to the representation learning algorithm used. Teasing apart these various effects is an empirical exercise. Theoretically, much work remains to be done on understanding how various algorithms and models in the machine learning and NLP communities apply differently to events rather than to entities.</p>
</sec>
</body>
<back>
<sec id="s7">
<title>Data Availability Statement</title>
<p>The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author.</p>
</sec>
<sec id="s8">
<title>Author Contributions</title>
<p>The author confirms being the sole contributor of this work and has approved it for publication.</p>
</sec>
<sec sec-type="COI-statement" id="s9">
<title>Conflict of Interest</title>
<p>The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s10">
<title>Publisher&#x2019;s Note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<fn-group>
<fn id="FN1">
<label>1</label>
<p>Two examples being Nate Silver&#x2019;s FiveThirtyEight, and the New York Times&#x2019;s forecasting interface.</p>
</fn>
<fn id="fn2">
<label>2</label>
<p>
<ext-link ext-link-type="uri" xlink:href="https://acleddata.com/">https://acleddata.com/</ext-link>
</p>
</fn>
<fn id="fn3">
<label>3</label>
<p>
<ext-link ext-link-type="uri" xlink:href="https://dss.princeton.edu/catalog/resource1507">https://dss.princeton.edu/catalog/resource1507</ext-link>
</p>
</fn>
<fn id="fn4">
<label>4</label>
<p>
<ext-link ext-link-type="uri" xlink:href="http://eventdata.parusanalytics.com/data.dir/cameo.html">http://eventdata.parusanalytics.com/data.dir/cameo.html</ext-link>
</p>
</fn>
<fn id="fn5">
<label>5</label>
<p>Although in this article, we primarily focus on link prediction as the primary use&#x20;case.</p>
</fn>
<fn id="fn6">
<label>6</label>
<p>In much of the NLP work, this statement is obvious enough that it is never stated explicitly: e.g., word2vec assumes that the document corpus is a large set of word &#x2018;sequences&#x2019;, with the skip-gram model used for representation learning, while the &#x2018;bag-of-words&#x2019; approach models each document as a multi-set of words, with vector weights derived using the famous tf-idf formula.</p>
</fn>
<fn id="fn7">
<label>7</label>
<p>
<ext-link ext-link-type="uri" xlink:href="http://www.start.umd.edu/gtd/">http://www.start.umd.edu/gtd/</ext-link>
</p>
</fn>
<fn id="fn8">
<label>8</label>
<p>Events for which at one link was available.</p>
</fn>
<fn id="fn9">
<label>9</label>
<p>Even more recently, so-called <italic>transformer-based</italic> neural networks such as BERT and GPT-3&#x20;<xref ref-type="bibr" rid="B3">Devlin et&#x20;al. (2018)</xref>, <xref ref-type="bibr" rid="B2">Brown et&#x20;al. (2020)</xref> have also been used with great effect in NLP tasks. Their application to graph applications and event link prediction remains untested, and they are expensive to fine-tune (requiring supervision). We leave their exploration for this problem domain for future research.</p>
</fn>
<fn id="fn10">
<label>10</label>
<p>The index of the last event in the ranked list if &#x7c;<italic>E</italic>&#x7c; &#x2212; 1 because the input event is not included in the ranked&#x20;list.</p>
</fn>
<fn id="fn11">
<label>11</label>
<p>This is also true of the tf-idf model, even though the vector is not derived using a neural network, and for that reason (in keeping with terminology as it is commonly used today), is not considered as an &#x2018;embedding&#x2019;.</p>
</fn>
<fn id="fn12">
<label>12</label>
<p>However, there is some signal in location clearly, since even a performance of 4% on MRR, as the tf-idf achieves using location alone is far better than what would be achieved through a random ranking of events in <italic>E</italic>&#x20;&#x2212; <italic>e</italic>
<sub>
<italic>i</italic>
</sub>
<italic>n</italic> per input event&#x20;<italic>e</italic>
<sub>
<italic>i</italic>
</sub>
<italic>n</italic>.</p>
</fn>
<fn id="fn13">
<label>13</label>
<p>Mathematically, this is equivalent to summing the cosine similarities obtained independently from the text embedding and graph embedding components; hence, the model is referred to as an ensemble&#x20;model.</p>
</fn>
</fn-group>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Battistelli</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Charnois</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Minel</surname>
<given-names>J.-L.</given-names>
</name>
<name>
<surname>Teiss&#xe8;dre</surname>
<given-names>C.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Detecting Salient Events in Large Corpora by a Combination of Nlp and Data Mining Techniques</article-title>. <source>Comp. y Sist.</source> <volume>17</volume>, <fpage>229</fpage>&#x2013;<lpage>237</lpage>. </citation>
</ref>
<ref id="B2">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Brown</surname>
<given-names>T. B.</given-names>
</name>
<name>
<surname>Mann</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Ryder</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Subbiah</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Kaplan</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Dhariwal</surname>
<given-names>P.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Language Models Are Few-Shot Learners</article-title>. <comment>arXiv preprint arXiv:2005.14165</comment>. </citation>
</ref>
<ref id="B3">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Devlin</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Chang</surname>
<given-names>M.-W.</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Toutanova</surname>
<given-names>K.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Bert: Pre-training of Deep Bidirectional Transformers for Language Understanding</article-title>. <comment>arXiv preprint arXiv:1810.04805</comment>. </citation>
</ref>
<ref id="B4">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Esser</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Str&#xf6;mb&#xe4;ck</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2013</year>). &#x201c;<article-title>Comparing News on National Elections</article-title>,&#x201d; in <source>The Handbook of Comparative Communication Research</source> (<publisher-loc>New York</publisher-loc>: <publisher-name>Routledge</publisher-name>), <fpage>330</fpage>&#x2013;<lpage>348</lpage>. </citation>
</ref>
<ref id="B5">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Grover</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Leskovec</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>node2vec: Scalable Feature Learning for Networks</article-title>,&#x201d; in <conf-name>Proceedings of the 22nd ACM SIGKDD international conference on Knowledge discovery and data mining</conf-name>, <fpage>855</fpage>&#x2013;<lpage>864</lpage>. </citation>
</ref>
<ref id="B6">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Haklay</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Weber</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2008</year>). <article-title>Openstreetmap: User-Generated Street Maps</article-title>. <source>IEEE Pervasive Comput.</source> <volume>7</volume>, <fpage>12</fpage>&#x2013;<lpage>18</lpage>. <pub-id pub-id-type="doi">10.1109/mprv.2008.80</pub-id> </citation>
</ref>
<ref id="B7">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hobbs</surname>
<given-names>J.&#x20;R.</given-names>
</name>
<name>
<surname>Pan</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2006</year>). <article-title>Time Ontology in Owl</article-title>. <comment>Technical report, Ontology Engineering Patterns Task Force of the Semantic Web Best Practices and Deployment Working Group, World Wide Web Consortium (W3C)</comment> <comment>Available at: <ext-link ext-link-type="uri" xlink:href="http://www.w3.org/TR/owl-time/">http://www.w3.org/TR/owl-time/</ext-link>
</comment>. </citation>
</ref>
<ref id="B8">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Joulin</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Grave</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Bojanowski</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Mikolov</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Bag of Tricks for Efficient Text Classification</article-title>. <comment>arXiv preprint arXiv:1607.01759</comment>. </citation>
</ref>
<ref id="B9">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Kejriwal</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Gilley</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Szekely</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Crisman</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2018a2018</year>). &#x201c;<article-title>Thor: Text-Enabled Analytics for Humanitarian Operations</article-title>,&#x201d; in <conf-name>Companion Proceedings of the The Web Conference</conf-name>, <fpage>147</fpage>&#x2013;<lpage>150</lpage>. </citation>
</ref>
<ref id="B10">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Kejriwal</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Peng</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Szekely</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2018b</year>). &#x201c;<article-title>Structured Event Entity Resolution in Humanitarian Domains</article-title>,&#x201d; in <conf-name>International Semantic Web Conference</conf-name> (<publisher-name>Springer</publisher-name>), <fpage>233</fpage>&#x2013;<lpage>249</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-00671-6_14</pub-id> </citation>
</ref>
<ref id="B11">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Kejriwal</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2019</year>). &#x201c;<article-title>Saviz: Interactive Exploration and Visualization of Situation Labeling Classifiers over Crisis Social media Data</article-title>,&#x201d; in <conf-name>Proceedings of the 2019 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining</conf-name>, <fpage>705</fpage>&#x2013;<lpage>708</lpage>. </citation>
</ref>
<ref id="B12">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>LaFree</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Dugan</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2007</year>). <article-title>Introducing the Global Terrorism Database</article-title>. <source>Terrorism Polit. violence</source> <volume>19</volume>, <fpage>181</fpage>&#x2013;<lpage>204</lpage>. <pub-id pub-id-type="doi">10.1080/09546550701246817</pub-id> </citation>
</ref>
<ref id="B13">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liben-Nowell</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Kleinberg</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2007</year>). <article-title>The Link-Prediction Problem for Social Networks</article-title>. <source>J.&#x20;Am. Soc. Inf. Sci.</source> <volume>58</volume>, <fpage>1019</fpage>&#x2013;<lpage>1031</lpage>. <pub-id pub-id-type="doi">10.1002/asi.20591</pub-id> </citation>
</ref>
<ref id="B14">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Lorenzini</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Kriesi</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Makarov</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>W&#xfc;est</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2021</year>). <source>Protest Event Analysis: Developing a Semiautomated Nlp Approach</source>. <publisher-name>American Behavioral Scientist</publisher-name>, <fpage>00027642211021650</fpage>. </citation>
</ref>
<ref id="B15">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Lu</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Ng</surname>
<given-names>V.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Joint Learning for Event Coreference Resolution</article-title>,&#x201d; in <conf-name>Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics</conf-name>, <conf-loc>Vancouver, Canada</conf-loc> (<publisher-name>Long Papers</publisher-name>) <volume>Vol. 11</volume>, <fpage>90</fpage>&#x2013;<lpage>101</lpage>. <pub-id pub-id-type="doi">10.18653/v1/p17-1009</pub-id> </citation>
</ref>
<ref id="B16">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Mikolov</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Sutskever</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Corrado</surname>
<given-names>G. S.</given-names>
</name>
<name>
<surname>Dean</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2013</year>). &#x201c;<article-title>Distributed Representations of Words and Phrases and Their Compositionality</article-title>,&#x201d; in <source>Advances in Neural Information Processing Systems</source>, <fpage>3111</fpage>&#x2013;<lpage>3119</lpage>. </citation>
</ref>
<ref id="B17">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Nassim</surname>
<given-names>N. T.</given-names>
</name>
</person-group> (<year>2007</year>). <source>The Black Swan: The Impact of the Highly Improbable</source>. <publisher-loc>NY</publisher-loc>: <publisher-name>Random House</publisher-name>. </citation>
</ref>
<ref id="B18">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Nenkova</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>McKeown</surname>
<given-names>K.</given-names>
</name>
</person-group> (<year>2012</year>). &#x201c;<article-title>A Survey of Text Summarization Techniques</article-title>,&#x201d; in <source>Mining Text Data</source> (<publisher-name>Springer</publisher-name>), <fpage>43</fpage>&#x2013;<lpage>76</lpage>. <pub-id pub-id-type="doi">10.1007/978-1-4614-3223-4_3</pub-id> </citation>
</ref>
<ref id="B19">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Ng</surname>
<given-names>V.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Machine Learning for Entity Coreference Resolution: A Retrospective Look at Two Decades of Research</article-title>,&#x201d; in <conf-name>AAAI</conf-name>, <fpage>4877</fpage>&#x2013;<lpage>4884</lpage>. </citation>
</ref>
<ref id="B20">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Pennington</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Socher</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Manning</surname>
<given-names>C.</given-names>
</name>
</person-group> (<year>2014</year>). &#x201c;<article-title>Glove: Global Vectors for Word Representation</article-title>,&#x201d; in <conf-name>Proceedings of the 2014 conference on empirical methods in natural language processing</conf-name>, <conf-loc>Doha, Qatar</conf-loc> (<publisher-name>EMNLP</publisher-name>), <fpage>1532</fpage>&#x2013;<lpage>1543</lpage>. <pub-id pub-id-type="doi">10.3115/v1/d14-1162</pub-id> </citation>
</ref>
<ref id="B21">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Perozzi</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Al-Rfou</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Skiena</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2014</year>). &#x201c;<article-title>Deepwalk: Online Learning of Social Representations</article-title>,&#x201d; in <conf-name>Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining</conf-name>, <fpage>701</fpage>&#x2013;<lpage>710</lpage>. </citation>
</ref>
<ref id="B22">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Tetlock</surname>
<given-names>P. E.</given-names>
</name>
</person-group> (<year>2017</year>). <source>Expert Political Judgment: How Good Is it? How Can We Know?</source> <publisher-name>Princeton University Press</publisher-name>. </citation>
</ref>
<ref id="B23">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Tetlock</surname>
<given-names>P. E.</given-names>
</name>
<name>
<surname>Gardner</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2016</year>). <source>Superforecasting: The Art and Science of Prediction</source>. <publisher-loc>New York</publisher-loc>: <publisher-name>Random House</publisher-name>. </citation>
</ref>
<ref id="B24">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Tetlock</surname>
<given-names>P. E.</given-names>
</name>
</person-group> (<year>1992</year>). <article-title>Good Judgment in International Politics: Three Psychological Perspectives</article-title>. <source>Polit. Psychol.</source> <volume>13</volume>, <fpage>517</fpage>&#x2013;<lpage>539</lpage>. <pub-id pub-id-type="doi">10.2307/3791611</pub-id> </citation>
</ref>
<ref id="B25">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Tetlock</surname>
<given-names>P. E.</given-names>
</name>
</person-group> (<year>2014</year>). &#x201c;<article-title>The Good Judgment Project</article-title>,&#x201d; in <source>Talk at Credit Suisse Thought Leader Forum</source>. </citation>
</ref>
<ref id="B26">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Wick</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2006</year>). <source>
<italic>GeoNames</italic> (GeoNames)</source>. </citation>
</ref>
<ref id="B27">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhuang</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Qi</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Duan</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Xi</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>H.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>A Comprehensive Survey on Transfer Learning</article-title>. <source>Proc. IEEE</source> <volume>109</volume>, <fpage>43</fpage>&#x2013;<lpage>76</lpage>. <pub-id pub-id-type="doi">10.1109/JPROC.2020.3004555</pub-id> </citation>
</ref>
<ref id="B28">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zinken</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2003</year>). <article-title>Ideological Imagination: Intertextual and Correlational Metaphors in Political Discourse</article-title>. <source>Discourse Soc.</source> <volume>14</volume>, <fpage>507</fpage>&#x2013;<lpage>523</lpage>. <pub-id pub-id-type="doi">10.1177/0957926503014004005</pub-id> </citation>
</ref>
</ref-list>
</back>
</article>