Horn Affairs አፍሪካ ቀንድ

Semantic Knowledge Graphs for the News: A Review

By Esleman Abay

May 19, 2023

Abstract

ICT platforms for news production, distribution, and consumption must exploit the ever-growing availability of digital data. These data originate from different sources and in different formats; they arrive at different velocities and in different volumes. Semantic knowledge graphs (KGs) is an established technique for integrating such heterogeneous information. It is therefore well-aligned with the needs of news producers and distributors, and it is likely to become increasingly important for the news industry. This article reviews the research on using semantic knowledge graphs for production, distribution, and consumption of news. The purpose is to present an overview of the field; to investigate what it means; and to suggest opportunities and needs for further research and development.Skip 1INTRODUCTION Section

INTRODUCTION

Journalism relies increasingly on computers and the Internet [114]. Central drivers are the big and open data sources that have become available on the Web. For example, researchers have investigated how news events can be extracted from big-data sources such as tweets [108] and other texts [107] and how big and open data can benefit journalistic creativity during the early phases of news production [115].

Semantic knowledge graphs and other semantic technologies [83] offer a way to make big and open data sources more readily available for journalistic and other news-related purposes. They offer a standard model and supporting resources for sharing, processing, and storing factual knowledge on both the syntactic and semantic level. Such knowledge graphs thus offer a way to make big, open, and other data sources better integrated and more meaningful. They make it possible to integrate the highly heterogeneous information available on the Internet and to make it more readily available for journalistic and other news-related purposes.

This article will systematically review the research literature on semantic knowledge graphs in the past two decades, from the time when the Semantic Web—an important precursor to semantic knowledge graphs—was first proposed [86]. The purpose is to present an overview of the fieldto investigate what it means; and to suggest opportunities and needs for further research and development. We understand both semantic knowledge graphs and the news in a broad sense. Along with semantic knowledge graphs, we include facilitating semantic technologies such as RDF, OWL, and SPARQL and their uses for semantically Linked (Open) Data and Semantic Web. We also include all aspects of production, distribution, and consumption of news. More precise inclusion and exclusion criteria will follow in Section 2.

To the best of our knowledge, no literature review has previously attempted to cover this increasingly important area in depth. Several reviews have been published recently on computational journalism in its various guises (e.g., References [9297129134]), but none of them go deeply into the technology in general nor into semantic knowledge graphs in particular. Also, recent overviews of knowledge graphs (e.g., References [9399104106121]) do not consider the specific challenges and opportunities for journalism or the news domain. Among the few papers that discuss the relation between semantic technologies and news, Reference [125] discusses how Linked Data can be integrated into and add value to news production processes and value chains in a non-disruptive way. It presents use cases from dynamic semantic publishing at BBC with attention to professional scepticism towards technology-driven innovation. More recently, Newsroom 3.0 [120] builds on an international field study of three newsrooms—in Brazil, Costa Rica, and the UK—to propose a framework for managing technological and media convergence in newsrooms. The framework uses semantic technologies to manage news knowledge, attempting to support interdisciplinary teams in their coordination of journalistic activities, cooperative production of content, and communication between professionals and news prosumers. Transitions in Journalism [124] discusses how new technologies are constantly challenging well-established journalistic norms and practices, discussing ways in which semantic journalism can exploit semantic technologies for everyday journalism.

Compared to these targeted efforts, this article presents the first systematic review of semantic knowledge graphs for news-related purposes in a broad sense. We ask the following research questions:

To answer these questions, the rest of the article is organised as follows: Section 2 outlines the literature-review process. Section 3 reviews the main papers. Section 4 discusses the main papers, answers the research questions, and offers many paths for further work. Section 5 concludes the article. The article is supported by an online Addendum that: describes our systematic review method in further detail; provides additional analyses of the main papers and related papers; and offers further readings about the resources and tools that are mentioned in the papers we review. These further readings are marked with an “A” in the main text, for example, “RDF”.Skip 2METHOD Section

METHOD

To answer our research questions, we conduct a systematic literature review (SLR) [111]. In line with our aim to present an overview of the field, we review the research literature in breadth to cover as many salient research problems, approaches, and potential solutions as possible. A detailed description of our systematic review method is available in the online Addendum (Section A).

Our review covers research on semantic knowledge graphs for the news understood in a wide sense. We include papers that use semantic technologies such as RDF, OWL, and SPARQL, [83] and practices such as Linked (Open) Data [87] and Semantic Web [86], but we exclude papers that use graph structures only for computational purposes isolated from the semantically linked Web of Data. We also include all aspects of production, distribution, and consumption of news, but we exclude research that uses news corpora only for evaluation purposes.

We search for literature through the five search engines ACM Digital Library, Elsevier ScienceDirect, IEEE Xplore, SpringerLink, and Clarivate Analytics’ Web of Science. We also conduct supplementary searches using Google Scholar. We search using variations of the phrases “knowledge graph,” “semantic technology,” “linked data,” “linked open data,” and “semantic web” combined with variations of “news” and “journalism” adapted to each search engine’s syntax. We select peer-reviewed and archival papers published in esteemed English-language journals or in high-quality conferences and workshops.

The search results are screened in three stages, so each selected paper is in the end considered by at least three co-authors. In the first stage, we screen search results based on title, abstract, and keywords. In the second stage, we skim the full papers and also consider the length, type, language, and source of each paper. In the third stage, we analyse the selected papers in detail according to the framework described below (Table 1). When several papers describe the same line of work, we select the most recent and comprehensive report. In the end, more than 6,000 search results are narrowed down to 80 fully analysed main papers. They are listed near the end of this article, right before the Reference list, and we distinguish them from other references by the letter “M,” e.g., [37].

Table 1.

 Show More

Table 1. Analysis Framework

Through a pilot study, we establish an analysis framework that we continue to revise and refine as the analysis progresses [90]. Table 1 lists the 10 top-level themes in the final framework, along with examples of sub-themes that we use to describe and compare the main papers in Section 3. For example, many main papers address specific groups of intended users. Intended users therefore becomes a top-level theme in our framework, with more specific groups of users, such as journalistsarchivists, and fact checkers, as sub-themes.

We make the detailed paper analyses along with their metadata available as a semantic knowledge graph through a SPARQL endpoint at http://bg.newsangler.uib.no. To support impact analysis, the metadata includes all incoming and outgoing citations of and by our main papers. The complete graph contains information about 4,238 papers, 9,712 authors, and 699 topics from Semantic Scholar. The online Addendum (Section A) provides further details and presents examples of SPARQL queries that can be used to explore the graph (Table 10).Skip 3REVIEW OF MAIN PAPERS Section

REVIEW OF MAIN PAPERS

This section reviews the 80 main papers according to the themes of Table 1. Our review and discussion is based on careful manual reading, analysis, marking, and discussion of the main papers, organised by the evolving themes and sub-themes in our analysis framework.

3.1 Technical Result Types

As shown in Figure 1(a), the main papers present a wide variety of technical research results. Further details are available in Table 6.

Fig. 1. The most frequent (a) technical and (b) empirical result types.

Pipelines and prototypes: A clear majority of main papers develop ICT architectures and tools for supporting news-related information processing with semantic knowledge graphs and related techniques. Most common are research prototypes and experimental pipelines. For example, the Knowledge and Information Management (KIM) platform [37] is an early and much-cited information extraction system that annotates and indexes named entities found in news documents semantically and makes them available for retrieval. To allow precise ontology-based retrieval, each identified entity is annotated with both a specific instance in an extensive knowledge base and a class defined in the associated KIM Ontology (KIMO). Which defines around 250 classes and 100 attributes and relations. The platform offers a graphical user interface for viewing, browsing and performing complex searches in collections of annotated news articles. Another early initiative is the News Engine Web Services (NEWS) project [15], which presents a prototype that automatically annotates published news items in several languages. The aim is to help news agencies provide fresh, relevant, and high-quality information to their customers. NEWS uses a dedicated ontology (the NEWS Ontology [17]) to facilitate semantic search, subscription-based services, and news creation through a web-based user interface. Hermes [4] supplies news-based evidence to decision-makers. To facilitate semantic retrieval, it automatically identifies topics in news articles and classifies them. The topics and classes are defined in an ontology that has been extended with synonyms and hypernyms from WordNet [98117] to improve recall.

Production systems: Some main papers take one step further and present industrial platforms that have run in news organisations, either experimentally or in production. The earliest example is AnnoTerra [57], a system developed by NASA to enhance earth-science news feeds with content from relevant multimedia data sources. The system matches ontology concepts with keywords found in the news texts to identify data sources and support semantic searches. Also, Reference [130] reports industrial experience with NEWS at EFE, a Spanish international news agency. The most recent example is VLX-Stories [14], a commercial, multilingual system for event detection and information retrieval from media feeds. The system harvests information from online news sites; aggregates them into events; labels them semantically; and represents them in a knowledge graph. The system is also able to detect emerging entities in online news. VLX-Stories is deployed in production in several organisations in several countries. Each month, it detects over 9,000 events from over 4,000 news feeds from seven different countries and in three different languages, extending its knowledge graph with 1,300 new entities as a side result.

System architectures: Whether oriented towards research or industry, another group of papers proposes system architectures. The World News Finder [34] presents an architecture that is representative of many systems that exploit KGs for managing news content. Online news articles in HTML format are parsed and analysed using GATE (General Architecture for Text Engineering) and ANNIE (A Nearly New Information Extraction system) with the support of JAPE (Java Annotations Pattern Engine) rules and ontology gazetteering lists. A domain ontology is then used in combination with heuristic rules to annotate the analysed news texts semantically. The annotated news articles are represented in a metadata repository and made available for semantic search through a GUI.

Algorithms: Another group of papers focuses on developing algorithms that exploit semantic knowledge graphs and related techniques, usually supported by proof-of-concept prototypes that are also used for evaluation. Inspired by Google’s PageRank algorithm, Reference [16] proposes the IdentityRank algorithm for named entity disambiguation in the NEWS project [15]. IdentityRank dynamically adjusts its weights for ranking candidate instances based on news trends (the frequency of each instance in a period of time) and semantic coherence (the frequency of the instance in a certain context), and it can be retrained based on user feedback and corrections. Reference [54] takes trending entities in news streams as its starting point and attempts to identify and rank other entities in their context. The purpose is to represent trends more richly and understand them better. One unsupervised and one supervised algorithm are compared. The unsupervised approach uses a personalised version of the PageRank algorithm over a graph of trending and contextual entities. The edges encode directional similarities between the entities using embeddings from a background knowledge graph. The supervised, and better performing, approach uses a selection of hand-crafted features along with a learning-to-rank (LTR) model, LightGBM. The selected features include positions and frequencies of the entities in the input texts, their co-occurrences and popularity, coherence measures based on TagMe and on entity embeddings, and the entities’ local importance in the text (or salience). NewsLink [75] processes news articles and natural-language (NL) queries from users in the same way, using standard natural-language processing (NLP) techniques. Co-occurrence between entities in a news article or query is used to divide it into segments, for example, corresponding to sentences. The entities in each segment are mapped to an open KG from which a connected sub-graph is extracted to represent the segment. The sub-graphs are then merged to represent the articles and queries as KGs that can be compared for similarity to support more robust and explainable query answering. Hermes also provides an algorithm (to be presented later) for ranking semantic search results [25].

Neural-network architectures: Rather than proposing algorithms, many recent main papers instead exploit semantic knowledge graphs for news purposes using deep neural-network (NN) architectures. These papers, too, are supported by proof-of-concept prototypes, which are usually evaluated using gold-standard datasets and information retrieval (IR) metrics. Heterogeneous graph Embedding framework for Emerging Relation detection (HEER) [79] detects emerging entities and relations from text reports, i.e., new entities and relations in the news that have so far not been included in a knowledge graph. The challenges addressed are that new entities and relations appear at high speed, with little available information at first and without negative examples to learn from. HEER represents incoming news texts as graphs based on entity co-occurrence and incrementally maintains joint embeddings of the news graphs and an open knowledge graph. The result is positive and unlabelled (PU) entity embeddings that are used to train and maintain a PU classifier that detects emerging relations incrementally.

Context-Aware Graph Embedding (CAGE) [66] is an approach for session-based news recommendation. Entities are extracted from input texts and used to extract a sub-knowledge graph from an open knowledge graph (the paper uses Wikidata). Knowledge-graph embeddings are calculated from the sub-knowledge graph, whereas pre-trained word embeddings and Convolutional Neural Networks (CNNs) [102] are used to derive content embeddings from the corresponding input texts. The knowledge-graph and content embeddings are concatenated and combined with user embeddings and refined further using CNNs. Finally, an Attention Neural Network (ANN) [135] on top of Gated Recurrent Units (GRUs) [102] are used to recommend articles from the resulting embeddings, taking short-term user preferences into account.

Deep Triple Networks (DTN) [42] use a deep-network architecture for topic-specific fake news detection. News texts are analysed in two ways in parallel: The first way is to use word2vec [116] embeddings and self-attention [135] on the raw input text. The second way is to extract triples from the text and analyse them using TransD [110] graph embeddings, attention and a bi-directional LSTM (Long Short-Term Memory) [102]. A CNN is used to combine the results of the two parallel analyses into a single output vector. Background knowledge has been infused into the second way by training the TransD graph embeddings, not only on the triples extracted from the input text, but also on related triples from a 4-hop DBpedia [85] extract. Maximum and average biases from the graph triples are concatenated with the CNN output vector and used to classify news texts as real or fake. The intuition behind this and other bias-based approaches to fake news detection is that, if the input text is false, triples learned only from the input text will have smaller bias than triples learned from the same text infused with true (and thus conflicting) real-world knowledge.

Ontologies: Almost half the papers include a general or domain-specific ontology for creating and managing other semantic knowledge graphs. For example, the NEWS project uses OWL to represent the NEWS Ontology [17], which standardises and interconnects the semantic labels used to annotate and disseminate news content. The Semantics-based Pipeline for Economic Event Detection (SPEED) [24] uses a finance ontology represented in OWL to ensure interoperability between and reuse of existing semantic and NLP solutions. Reference [71] represents the IPTC (International Press Telecommunications Council) News Codes as SKOS concepts in an OWL ontology and discusses its uses for semantic enrichment and search. The Evolutionary Event Ontology Knowledge (EEOK) ontology [45] represents how different types of news events tend to unfold over time. The ontology is supported by a pipeline that mines event-evolution patterns from natural-language news texts that report different stages of the same macro event (or storyline). The patterns are represented in OWL and used to extract and predict further events in developing storylines more precisely.

Knowledge graphs: A few papers even present a populated, instance-level semantic knowledge graph or other linked knowledge base as a central result. For example, K-Pop [36] populates a semantic knowledge graph for enriching news about Korean pop artists. The purpose is to provide comprehensive profiles for singers and groups, their activities, organisations, and catalogues. As an example application, the resulting entertainment KG is used to power Gnosis, a mobile application for recommending K-Pop news articles. CrimeBase [67] presents a knowledge graph that integrates crime-related information from popular Indian online newspapers. The purpose is to help law enforcement agencies analyse and prevent criminal activities by gathering and integrating crime entities from text and images and making them available in machine-readable form. ClaimsKG [69] is a live knowledge graph that represents more than 28,000 fact-checked claims published since 1996, totalling over 6 million triples. It uses a semi-automatic pipeline to harvest fact checks from popular fact-checking websites; annotate them with entities from DBpedia; represent them in RDF according to a semantic data model in RDFS; normalise the validity ratings; and resolve co-references across claims. Reference [12] uses hashtags and other metadata associated with tweets and tweeters to build an RDF model of over 900,000 French political tweets, totalling more than 20 million triples that describe facts, statements, and beliefs in time. The purpose is to trace how actors propagate knowledge—as well as misinformation and hearsay—over time.

Formal models: A small final group of papers proposes formal models of various types and for different purposes. For example, Reference [22] presents a formal model for managing inconsistencies that arise when live news streams are represented incrementally using description logic. A trust-based algorithm for belief-base revision is presented that takes users’ trust in information sources into account when choosing which inconsistent information to discard.

Summary: Our review suggests that the most common types of results are pipelines and prototypes. In addition, many papers propose ontologies, system architectures, algorithms, and neural-network architectures. A few papers also introduce new knowledge graphs. There has been a shift in recent years from research on algorithms and system architectures towards papers that propose deep neural-network architectures. A few of those recent papers also mention explainability.

3.2 Empirical Result Types

As shown in Figure 1(b), a large majority of the papers include an empirical evaluation of their technical proposals.

Experiments: As shown in the previous section, a majority of papers develop pipelines or prototypes, which are then evaluated empirically. The most common evaluation method is controlled experiments using gold-standard datasets and information retrieval (IR) measures such as precision (P), recall (R), and accuracy (A). For example, KOPRA [70] is a deep-learning approach that uses a Graph Convolutional Network (GCN) [91103] for news recommendation. An initial entity graph (called interest graph) is created for each user from entities mentioned in the news titles and abstracts of that user’s short- and long-term click histories. A joint knowledge pruning and Recurrent Graph Convolution (RGC) mechanism is then used to augment the entities in the interest graph with related entities from an open KG. Finally, entities extracted from candidate news texts are compared with entities in the interest graphs to predict articles a user may find interesting. The approach is evaluated experimentally with Wikidata as the open KG and using two standard datasets (MIND and Adressa). RDFLiveNews [21] aims to represent RSS data streams as RDF triples in real time. Candidate triples are extracted from individual RSS items and clustered to suggested output triples. Components of the approach are evaluated in two ways. The first way measures RDFLiveNews’ ability to disambiguate alternative URIs for named entities detected in the input items. Disambiguation results are evaluated against a manually crafted gold standard using precision, recall, and F1 metrics and by comparing them to the outputs of a state-of-art NED tool (AIDA). The second way measures RDFLiveNews’ ability to cluster similar triples extracted from different RSS items. The clusters are evaluated against the manually crafted gold standard using sensitivity (S)positive predictive value (PPV), and their geometric mean.

Performance evaluation: A smaller number of experimental papers collect performance measures such as execution times and throughput in addition to or instead of IR measures. For example, the scalability of RDFLiveNews [21] is also measured using run times for different components of the approach on three test corpora. The results suggest that, with some parallelisation, it is able to handle at least 1,500 parallel RSS feeds. The performance of KnowledgeSeeker [39], an ontology-based agent system for recommending Chinese news articles, is measured through execution times on three datasets for a given computer configuration and using the performance of a vanilla TF-IDF-based approach as comparison baseline. The throughput of SPEED [24] is benchmarked on a corpus of 200 news messages extracted from Yahoo!’s business and technology news feeds.

Ablation, explainability, and parameter studies: Many recent papers also include ablation studies [54667075], explainability studies [457075], and parameter and sensitivity studies [79]. A common theme is that they all use deep or other machine learning techniques. We will present more examples later (e.g., References [194073747880]).

Industrial testing: A few papers present case studies or experience reports from industry. We have already mentioned the commercial VLX-Stories [14] system. Reference [44] extends the news production workflow at VRT (Vlaamse Radio- en Televisieomroep), a national Belgian broadcaster, to support personalised news recommendation and dissemination via RSS feeds. A semantic version of the IPTC’s NewsML-G2 standard is proposed as a unifying (meta-)data model for dynamic distributed news event information. As a result, RDF/OWL and NewsML-G2 can be used in combination to automatically categorise, link, and enrich news-event metadata. The system has been hooked into the VRT’s workflow engine, facilitating automatic recommendation of developing news stories to individual news users. Reference [68] semantically enriches the content of archival news texts. The proposed system identifies mentions of named entities along with their contexts; links the contextualised mentions to entities in a knowledge base; and uses the links to retrieve further relevant information from the knowledge base. The system has been deployed and applied to 10 years of archival news in a local Italian newspaper. And as already mentioned, a prototype of the NEWS system [15] has run experimentally at EFE, alongside their legacy production system, introducing a semi-automatic workflow that lets journalists validate the annotations suggested by the system [130].

Case studies and examples: Other papers present realistic examples based on industrial experience. For example, the MediaLoep project [10] (involving many of the authors behind Reference [44], and Reference [9], to be presented later) discusses how to improve retrieval and increase reuse of previously broadcast multimedia news items at VRT, the national Belgian broadcaster, both as background information and as reusable footage. The paper reports experiences with collecting descriptive metadata from different news production systems; integrating the metadata using a semantic data model; and connecting the data model to other semantic data sets to enable more powerful semantic search.

Proof-of-concept demonstrations and use cases: Similar types of qualitative evaluations, but with less focus in industrial-scale examples, are proof-of-concept demonstrations and hypothetical use cases (e.g., Reference [65]).

User studies: A final group of papers presents user studies and usability tests. Reference [76] represents news articles as small knowledge graphs enriched with word similarities from WordNet [98117]. Overlaps between the sub-graphs of new articles and of articles a user has found interesting in the past are used to recommend new articles to the user. Sub-graphs are compared using Jaccard similarity. The approach is evaluated on a collection of Japanese news articles. Twenty users were asked to rate suggested articles in terms of relevance and of interest, breaking the latter down into curiosity and serendipity.

Summary: Our review shows that experimental evaluation of proposed pipelines/prototypes is the most used research method. Experiments most often use information retrieval measures, but usability and performance measures are also employed. In recent years, experiments are increasingly often supplemented by studies of ablation, explainability, and parameter selection. Other used research methods are industrial testing, case studies and examples, proof-of-concept demos, use cases, and user studies.

3.3 Intended Users

The most frequent types of intended users—or immediate beneficiaries—of the results from our main papers are shown in Figure 2(a).

Fig. 2. The most frequently (a) intended users and (b) their tasks.

News users: More than half the main papers aim to offer news services to the general public. An early example is Rich News [11], a system that automatically transcribes and segments radio and TV streams. Key phrases extracted from each segment are used to retrieve web pages that report the same news event. The web pages are annotated semantically using the KIM platform [37], whose web interface is used to support searching and browsing news stories semantically by topic and playing the corresponding segments of the associated media files.

Journalists, newsrooms, and news agencies: The second largest group of papers aims to support journalists and other professionals in newsrooms and news agencies. Several projects mentioned already belong to this type, including the NEWS project [15]. The proposals in References [10455471] also target journalists and other news professionals. The ambition of the News Angler project [46] is to enable automatic detection of newsworthy events from a dynamically evolving knowledge graph by representing news angles, such as “proximity,” “nepotism,” or “fall from grace” [123], formally using Common Logic.

Knowledge-base maintainers: Rather than supporting news users directly, some papers support knowledge-base maintainers on a technical level. For example, Reference [17] presents a plugin for maintaining the NEWS ontology. Aethalides [58] extends the Hermes framework [4] with a pipeline for semantic classification using concepts defined in a domain ontology.

Archivists: A smaller group of papers targets archivists, who maintain knowledge bases on the content level. For example, Neptuno [7] is an early semantic newspaper archive system that aims to give archivists and reporters richer ways to describe and annotate news materials and to give reporters and news readers better search and browsing capabilities. It uses an ontology for classifying archive content along with modules for semantic search, browsing, and visualisation. The purpose of the formal model for belief-base revision [22] presented earlier is also to maintain knowledge bases by detecting and resolving inconsistencies.

Fake-news detectors and fact checkers: Several recent papers focus on supporting fake-news detectors and fact checkers. We have already mentioned Deep Triple Networks (DTN) [42]. Reference [5] detects fake news through a hybrid approach that assesses sentiments, entities, and facts extracted from news texts. ClaimsKG [69], the large knowledge graph of French political tweets, can be used to trace how knowledge—along with misinformation and hearsay—is propagated over time. Several of the recent deep-NN approaches we will present later also target fake-news detection and fact checking.

Knowledge workers: A smaller group of papers targets general knowledge workers and information professionals outside the news profession. For example, KIM [37] aims to improve news browsing and searching for knowledge workers in general. Other papers aim to support specific information professions. The Automatic Georeferencing Video (AGV) pipeline [13] makes news videos from the RAI archives available for geography education. Audio is extracted from video using ffmpeg and transcribed using Ants. Apache OpenNLP is used to extract named entities mentioned in the video segment. Google’s Knowledge Graph is used to add representative images and facts about related people and places. The places are in turn used to make the videos and their metadata available through Google Street Map-based user interfaces. The pipeline is tested on a dataset of 10-minute excerpts from 6,600 videos from a thematic RAI newscast (Leonardo TGR). AnnoTerra [57] uses ontologies and semantic search to improve NASA’s earth-science news feeds, targeting both experts and inexperienced users of earth-science data. CrimeBase [67] uses rules to extract entities from text and associated image captions in multimodal crime-related online news. The extracted entities are correlated using contextual and semantic similarity measures, whereas image entities are correlated using image features. The resulting knowledge base uses an OWL ontology to integrate crime-related information from popular Indian online newspapers. Other main papers (to be presented later) target professionals in domains such as economy and finance [51], environmental communication [63], and medicine [23].

Summary: Our review indicates that the most frequently intended users (or beneficiaries) of the main-paper proposals are general news users and journalists. Other intended users/beneficiaries are newsrooms, knowledge-base maintainers, archivists, fake-news detectors and fact checkers, and different types of knowledge workers.

3.4 Tasks

As shown in Figure 2(b), the main papers target a wide range of news production, dissemination, and consumption activities, such as search, recommendation, categorisation, and event detection.

Semantic annotation: Many of the earliest approaches focus on adding semantic labels to entities and topics mentioned in published news texts. We have already introduced KIM [37], which labels named entities found in news items with links to instances in a knowledge base and to classes defined in the KIM Ontology (KIMO). We have also introduced NEWS [15], which annotates news items with named entities linked to external sources such as Wikipedia, ISO country codes, NASDAQ company codes (e.g., Reference [34]), the CIA World Factbook, and SUMO/MILO. It also categorises the news items by content and represents news metadata using standards and vocabularies such as the Dublin Core (DC) vocabulary, the IPTC’s News Codes, the News Industry Text Format (NITF), NewsML,, and PRISM—the Publishing Requirements for Industry Standard Metadata.

Enrichment: A smaller group of papers instead focuses on enriching annotated news items with Linked Open Data or information from other semantically labelled sources. For example, Reference [2] extends the life of TV content by integrating heterogeneous data from sources such as broadcast archives, newspapers, blogs, social media, and encyclopedia and by aligning semantic content metadata with the users’ evolving interests. AGV [13] annotates TV news programs with geographical entities to make archival video content available through a map-based user interface for educational purposes. In addition to representing the IPTC News Codes using SKOS, Reference [71] discusses how multimedia news metadata can be augmented using natural-language and multimedia analysis techniques and enriched with Linked Data, such as facts from DBpedia [85] and GeoNames. Contributions that represent news texts as sub-graphs of open KGs such as Wikidata (e.g., CAGE [66], KOPRA [70], and NewsLink [75]) can also be considered enrichment approaches. We will present a few similar approaches later [4080].

Content retrieval: Other papers use semantic annotations (or “semantic footprints”) to support on-demand (“pull”) or proactive (“push”) dissemination of news content. On the retrieval (on-demandpull) side, a clear majority of the main papers support tasks such as searching for and otherwise retrieving news items. Projects such as KIM [37], NEWS [15], and Hermes [4] all have content provision as central tasks. The Hermes Graphical Query Language (HGQL) [25] makes it simpler for non-expert users to search semantically for content available in the Hermes framework. It is based on RDF-GL, a SPARQL-based graphical query language for RDF, and also provides an algorithm for ranking search results. The World News Finder [34] uses a World News Ontology along with heuristic rules to automatically create metadata files from HTML news documents to support semantic user queries. The aim of NewsLink [75] is to support more robust as well as explainable query answering.

Content provision: On the provision (proactivepush) side, another large group of papers focuses on actively propagating news to users. For example, Reference [33] aims to provide more accurate content-based recommendations. It uses existing tools for entity discovery and linking to represent news messages as sub-graphs by adding edges from Freebase. A new human-annotated data set (CNREC) for evaluating content-based news recommendation systems is made available and used to evaluate the approach. Reference [6] aims to deal with data sparsity and cold-start issues in news recommender systems. It enriches semantic representations of news items and of users with Linked Data to provide more input to recommendation algorithms. Focusing on the user-profiling (or personalisation) side of news recommendation, Reference [26] uses semantic annotations of news videos to profile users’ evolving information needs and interests to recommend the most suitable news stories.Context-Aware Graph Embedding (CAGE) [66] focuses on providing session-based recommendations, whereas KOPRA [70] aims to take both users’ short- and long-term behaviours into account.

Event detection: Several more recent approaches go beyond semantic labelling and enrichment of news content, attempting to extract events or relations (triples, facts) from news items to represent their meaning on a fine-grained level. NewsReader [72] is a cross-lingual system (or “reading machine”) that is designed to ingest high volumes of news articles and represent them as Event-Centric Knowledge Graphs (ECKGs) [59]. Each graph describes an event, and perhaps how it develops over time, along with the actors and other entities involved in the event. The graphs are connected through shared entities and temporal overlaps, and the entities are linked to background information in knowledge bases such as DBpedia. The ASRAEL project [60] maps events described in unstructured news articles to structured event representations in Wikidata, which are used to enrich the representations of the articles. Because Wikidata’s event hierarchy is considered too fine-grained for use in search engines, a hierarchical clustering step follows, after which the more coarsely categorised events are made available for querying and navigation through an event-oriented knowledge graph. To keep the Hermes [4] knowledge base up to date, Reference [64] represents lexico-semantic patterns and associated actions as rules that are used to semi-automatically detect and semantically describe news events. The approach is implemented in the Hermes News Portal (HNP), a realisation of the Hermes framework that lets news users browse and query for relevant news items. The Evolutionary Event Ontology Knowledge (EEOK) ontology [45] aims to support event detection by suggesting which event types to look for next in a developing storyline. Reference [38] identifies and reconciles named events from news articles and represents them in a semantic knowledge graph according to textual contents, entities, and temporal ordering. The commercial tool VLX-Stories [14] also detects events in media feeds.

Relation extraction: Other papers instead focus on relation extraction, detecting triples (or facts) that can be used to build new or update existing RDF graphs. An early proposal for deeper text analysis is SemNews [30], which extracts textual-meaning representations (TMRs) from RSS news items using the OntoSem tool (see, e.g., Reference [33]), which represents each text as a set of facts about: which actions that are described in the text; which agents, locations, and themes each action involves; and any temporal relations between the actions. The SemNews tool transforms the TMRs into OWL to support semantic searching, browsing, and indexing of RSS news items. It also powers an experimental web service that provides semantically annotated news items along with news summaries to human users. BKSport [49] automatically annotates sports news using language-pattern rules in combination with a domain ontology and a knowledge base built on top of the KIM platform [37]. The tool extracts links and typed entities as well as semantic relations between them. It also uses pronoun recognition to resolve co-references. Reference [55] represents the sentences in a news item as triples, analysing not only top-level but also subordinate clauses. The triples are run through a pipeline of natural language tools that fuse and prioritise them. Finally, selected triples are used to summarise the underlying event reported in the news item. Reference [18] identifies novel statements in the news, building on ClausIE and DBpedia to propose a semantic novelty measure that takes individual user-relevance into account.

Sub-graph extraction: An alternative to extracting relations from news texts is to represent texts by sub-graphs extracted from open knowledge graphs. An early example is Reference [33], which uses standard techniques to discover and link entities and adds edges from Freebase to represent news messages as sub-graphs to support content-based news recommendation. AnchorKG [40] represents news articles as small anchor graphs, which consist of entities that are prominently mentioned in the news text, along with relations between those entities taken from an open knowledge graph, and along with those entities’ k-hop neighbourhoods in the graph. One aim is to improve news recommendation by making real-time knowledge reasoning scalable to large open knowledge graphs. Another aim is to support explainable reasoning about similarity. Reinforcement learning is used to train an anchor-graph extractor jointly with a news recommender, using already recognised and linked named entities as inputs. The approach is evaluated using the MIND dataset and a private dataset extracted from Bing News with Wikidata as reference graph. CAGE [66] represents news texts as sub-graphs extracted from an open reference knowledge graph to support session-based news recommendation. KOPRA [70] extracts an entity graph (called interest graph) for each user from seed entities that are mentioned in the news titles and abstracts in the user’s short- and long-term click histories. NewsLink [75] represents both news articles and user queries as small KGs that can be compared for similarity.

KG updating: Several recent contributions use deep and other machine-learning techniques to keep evolving knowledge graphs up-to-date by identifying new (emerging, dark) entities and new (or emerging) relations between (the new or existing) entities. We have already mentioned HEER [79]. PolarisX [77] automatically expands language-independent knowledge graphs in real time with representations of new events reported by news sites and on social media. It uses a relation extraction model based on pre-trained multilingual BERT [96] to detect new relations. Challenges addressed are that available reference knowledge graphs have limited size and scope and that existing techniques are not able to deal with neologisms based on human common sense. Text-Aware MUlti-RElational learning method (TAMURE) [78] also extends a knowledge graph with relations that emerge in the news. It addresses the source heterogeneity of structured knowledge graphs and unstructured news texts by learning joint embeddings of entities, relations, and texts using tensor factorisation implemented in TensorFlow. TAMURE is linear in the number of parameters, making it suitable for large-scale KGs and live news streams. Reference [61] empirically investigates the prevalence of entities in online news feeds that cannot be identified by DBpedia Spotlight or by Google’s Knowledge Graph API. Out of 13,456 named entities in an RSS sample, 378 were missing from DBpedia, 488 were missing from Google’s Knowledge Graph, and 297 were missing from both.

Ontology development: In various ways, several main papers support ontology development. Early projects such as KIM [37] and NEWS [15] focus on developing new domain ontologies, whereas Reference [37] integrates existing IPTC standards and vocabularies into the LOD cloud. More recent efforts, such as Reference [45], use machine learning techniques to automate ontology creation and maintenance.

Fake-news detection and fact checking: Several recent papers focus on the detection of fake news, such as Reference [5]. Another proposal is Reference [52], which uses graph embeddings of news texts to identify fake news. Reference [48] presents a multimodal approach to quantify whether real-world news texts and their associated images represent the same or connected entities, suggesting that low coherence is a possible indicator of fake news. Reference [23] lifts medical information from non-trusted sources into semantic form using FRED [101] and reasons over the resulting description logic representations using Racer and HermiT. Reasoning inconsistencies are taken to indicate potential “medical myths” that are verbalised and presented to human agents along with an explanation of the inconsistency. KLG-GAT [80] uses an open knowledge graph to enhance fact checking and verification. Constituency parsing is used to find entity mentions in the claims, which are used to retrieve relevant Wikipedia articles as potential evidence. A BERT-based sentence retrieval model is then used to select the most relevant evidence for the claim. TagMe is used to link entities in the claims and in the evidence sentences to the Wikidata5M subset of Wikidata and extract triples whose entities are mentioned in the claim and/or evidence. The triples are further ranked using a BERT-based learning-to-rank (LTR) model. High-ranked triples are used to construct a graph of the central claim, its potential evidence sentences, and triples that connect the claim to the evidence sentences. A two-level multi-head graph attention network is used to propagate information between the claim, evidence, and knowledge (triple) nodes in the graph as input to a claim classification layer.

Content generation: Targeting news content generation, Tweet2News [3] extracts RDF triples from documentary (headline-like) tweets using the IPTC’s rNews vocabulary, organises them into storylines, and enriches them with Linked Open Data to facilitate news generation in addition to retrieval. The Pundit algorithm [56] even suggests plausible future events based on descriptions of current events. Structured representations of news titles are extracted from a large historical news archive that covers more than 150 years, and a machine-learning algorithm is used to extract causal relations from the structured representations. Although the authors do not propose specific journalistic uses of Pundit, their algorithm might be used in newsrooms to anticipate alternative continuations of developing events. Reference [31] aims to auto-generate human-quality news image captions based on a corpus of news texts with associated images and captions. Each news image is represented as a feature vector using a pre-trained CNN, and each corresponding article text is split into sentences containing named entities that are processed further in two ways. One line of analysis enriches the sentences and entities with related information from DBpedia. Another line instead replaces the named entities with type placeholders, such as PERSON, NORP, LOC, ORG, and GPE, producing generic sentences that are compressed using dependency parsing and represented as TF-IDF weighted bags-of-words. Correlations are then established between the generic-sentence representations and the features of the associated images in the corpus. An LSTM model is trained to generate matching caption templates for images on top of the pre-trained CNN. Finally, the semantically enriched original sentences are used to fill in individual entities for the type placeholders. The approach is evaluated on two public datasets, Good News (466K examples) and Breaking News (110K examples), that include news images and captions along with article texts. Reference [55] (presented earlier) uses the triples that have been extracted, fused, and prioritised from news sentences to generate new sentences that summarise the underlying news events. The News Angler project [46] represents news angles to support automatic detection of newsworthy events from a knowledge graph.

Prediction:Prediction is the focus of a small group of papers that includes Pundit [56] and EEOK [45]. To predict stock prices, EKGStock [41] uses named-entity recognition and relation extraction to represent news about Chinese enterprises as knowledge graphs. Embeddings of the enterprise-specific graphs are then used to estimate connectedness between enterprises. Sentiments of news reports that mention an enterprise are then fed into a Gated Recurrent Unit (GRU) [102] model that predicts stock prices, not only for the mentioned enterprise, but also for its semantically related ones. Recent predictive approaches include deep-neural network-based recommendation papers, such as Reference [73] (more later), that are trained to predict click-through rates (CTR).

Other tasks: In addition to these most frequent uses of knowledge graphs for news, several main papers address semantic similarity. For example, Reference [35] uses information extraction techniques to automatically annotate news documents semantically to facilitate cross-lingual retrieval of documents with similar annotations. Reference [9] clusters semantic representations to detect how news items are derived from one another, using the PROV-O ontology to represent the results semantically. Supporting visualisation, the Visualizing Relations Between Objects (VRBO) framework [32] uses semantic and statistical methods to identify temporal patterns between entities mentioned in economic news. It uses the patterns to create and visualise news alerts that can be formalised and used to manage equity portfolios. Neptuno [7] uses visualisation on the ontology level to show and publish how knowledge-base concepts are organised. Archiving and general information organisation is a central task of Neptuno [7] and several other main papers. Interoperability and data integration is the focus in MediaLoep [10]. Focusing on multimedia and other metadata, Reference [20] (more later) also has interoperability as a central task, along with contributions such as References [2245357].

Summary: Our review shows that the research on semantic knowledge graphs for the news support a broad variety of tasks, such as semantic annotation, enrichment, content retrieval and provision, event detection, relation and sub-graph extraction, KG updating, ontology development, fake-news detection and fact checking, content generation, and prediction. The past few years have seen a rapidly growing interest in KGs for fake-news identification. Support for factual journalism is a related area that is growing. Automatic news detection is another emerging area that is becoming increasingly important.

3.5 Input Data

As shown in Figure 3(a), the proposed approaches rely on a variety of sources and types of primary input data. Note that this section discusses the data used as input by the solutions proposed or discussed in each main paper, and not the data used for evaluation.

Fig. 3. The most frequent types of (a) input data used and (b) news life-cycle phases targeted.

News articles: The most common input data are textual news articles in digital form. For example, Reference [47] reads template-based HTML pages and exploits semantic regularities in the templates to automatically annotate HTML elements with semantic labels according to their DOM paths. Online news articles are also used as examples and for evaluation.

RSS and other news feeds: Other main papers take their inputs via RSS feeds or other news feeds. The Ontology-based Personalised Web-Feed Platform (OPWFP) [28] inputs RSS news streams and uses an ontology to provide more precisely customised web feeds. User profiles are expressed using the semantic Composite Capability/Preference Profiles (CC/PP) and FOAF vocabularies along with a domain ontology. The three vocabularies and ontologies are used in combination to select appropriate search topics for the RSS search engine.

Social media and the Web: Several main papers use social media and other web resources as input, such as Twitter, Wikinews, Wikipedia, and regular HTML-based web sites. To support personalised news recommendation and dissemination, the extension of VRT’s news workflow mentioned earlier [44] uses OpenID and OAuth for identification and authentication. In this way, the system can compile user profiles based on data from multiple social-media accounts, using ontologies such as FOAF and SIOC to interoperate user data. Focusing on geo-hashtagged tweets, Location Tagging News Feed (LTNF) [8] is a semantics-based system that extracts geographical hashtags from social media and uses a geographical domain ontology to establish relations between the hashtags and the messages they occur in. Wikipedia is also used as a direct source of input in a few papers [136].

Multimedia news: Several papers use multimedia data as input. Reference [31] analyses news texts in combination with associated images to suggest human-level image captions. To extend the lifetime of TV news, the AGV pipeline [13] makes news videos from the RAI archives available for geography education.

News metadata: Focusing on multimedia metadata, Reference [20] inputs metadata embedded in formats such as MPEG-7 for content description and MPEG-21 for delivery and consumption. The approach uses semantic mappings from XML Schema to OWL and from XML to RDF to integrate administrative multimedia metadata in newspaper organisations. As already explained, MediaLoep [10] also integrates descriptive multimedia news metadata from news production systems semantically.

Knowledge graphs: Many papers use existing knowledge graphs as inputs. The number has risen in the past few years due to the appearance of deep-NN architectures that infuse triples from open KGs to enhance learning from news texts. Indeed, almost all the recent deep learning papers exploit open KGs in this way, e.g., References [314042667080].

User histories: A smaller group of deep-NN papers inputs user histories, for example, in the form of click logs, to train recommenders [19667073].

Summary: Our review shows that the research on semantic knowledge graphs for the news exploits a broad range of input sources. Textual news articles in digital form is the most important source. Other frequently used types of input data are RSS and other news feeds, social media and the Web, multimedia news, news metadata, knowledge graphs, and user histories. Multimedia, including TV news, were popular in first years of the study period and have seen a rebound in the deep-learning era. RSS and other news feeds were popular for many years, but have recently been overtaken by social media, including Twitter. In recent years, KGs are being used increasingly often to infuse world knowledge into deep NNs for news analysis. User histories have also been used in recent recommendation papers.

3.6 News Life Cycle

The main papers also target different phases of the news life cycle, as shown in Figure 3(b). The largest group of papers focuses on organising and managing already published news. For example, Neptuno [7] extends the life of published news by annotating reports with keywords and IPTC codes, thereby relating past news reports to current ones that share the same keywords or code. It thus re-contextualises old news in light of more recent events. The MediaLoep data model [10] supports managing information generated by news production and publishing processes. AGV [13] makes archival news videos available for geography education.

Focus in recent years has shifted from focusing on already published news to also targeting earlier phases of the news life cycle. As already mentioned, the Pundit algorithm [56] predicts likely future news events based on short textual descriptions of current events. A small group of mostly Twitter-based papers deals with detecting emerging news, or potentially newsworthy events or situations that are not yet reported as news but that may be circulating in social media or elsewhere. For example, Tweet2News [3] identifies emerging news from documentary (or headline-like) tweets and lifts them into RDF graphs, which are then enriched with triples from the LOD cloud and arranged into storylines to generate news reports in real time.

Focusing on breaking news, the Semantics-based Pipeline for Economic Event Detection (SPEED) [24] uses a domain ontology to detect and annotate economic events. The approach combines ontology-based word and event-phrase gazetteers; a word-phrase look-up component; a word-sense disambiguator; and an event detector that recognises event-patterns described in a domain ontology. The Evolutionary Event Ontology Knowledge (EEOK) [45] ontology presented earlier represents the typical evolution of developing news stories as patterns. It can thereby be used to predict the most likely next events in a developing story and to train dedicated detectors for different event types and phases (such as “investigation,” “arrest,” “court hearing”) in a complex storyline (“fire outbreak”). RDFLiveNews [21] also follows developing news by combining statistical and other machine-learning techniques to represent news events as knowledge graphs in real time by extracting RDF triples from RSS data streams.

Summary: Our review shows that all the different phases of the news-life cycle are covered by the research, from predicting future news, through detecting and monitoring emerging, breaking, and developing news, to managing and exploiting already published news. Many main papers attempt to cover several of these life-cycle phases.

3.7 Semantic Techniques and Tools

The main papers use a broad variety of semantic techniques and tools. For the purpose of this review, we separate them into exchange formatsontologies and vocabulariesinformation resources, and processing and storage techniques.

Semantic exchange formats: By semantic exchange formats, we mean standards for exchanging and storing semantic data. As shown in Figure 4(a), RDF, OWL, and SPARQL are most common. More than half of the papers use RDF to manage information. The earliest example is Neptuno [7], which uses RDF to represent the IPTC’s hierarchical subject reference system. More than a third of the main papers use OWL for ontology representation. For example, the MediaLoep data model [10] is represented in OWL (using the SKOS vocabulary), and its concepts are linked to standard knowledge bases such as DBpedia [85] and GeoNames. And we have already mentioned the NEWS Ontology [17], which is represented in OWL-DL, the description logic subset of OWL. SPARQL is also common. It is central in the Hermes project [4] and in the News Articles Platform [53]. RDFS is also widely used, including in the NEWS [15] project.

Fig. 4. The most frequently used semantic (a) exchange formats and (b) vocabularies and ontologies.

Ontologies and vocabularies: By ontologies and vocabularies, we mean formal terminologies for semantic information exchange. As shown in Figure 4(b), Dublin Core (DC) and Friend of a Friend (FOAF) are the most used general vocabularies, starting already with KIM [37], whose ontology is designed to be aligned with them both. The NEWS project [15] also uses DC, whereas FOAF plays a prominent role in a few approaches that deal with personalisation, in particular in a social context [2844]. Another much-used ontology is the Simple Knowledge Organization System (SKOS). It is used by the NEWS Ontology [17] to align and interoperate concepts from different annotation standards, including the IPTC News Codes. It is also used for personalised multimedia recommendation in Reference [26] and for integrating news-production metadata in Reference [10]. The OWL-representation of IPTC’s News Codes in Reference [71] links to Dublin Core and SKOS concepts to increase precision and facilitate content enrichment. The Simple Event Model (SEM) and OWL Time are used in the NewsReader [72] and News Angler projects [46]. SUMO/MILO is used in the NEWS project [15]. SUMO and ESO are used in NewsReader [5972]. Other general ontologies include schema.org, used to contextualise the ClaimsKG [69] and KIM’s PROTON ontology [37]. The Provenance Data Model (PROV-DM) is used to discover high-level provenance using semantic similarity in [9]. Although several other papers mention provenance, too, they do not explicitly refer to or use PROV-DM, nor its OWL formulation, PROV-O. However, the NewsReader project [72] uses the Grounded Representation and Source Perspective (GRaSP) framework, which has at least been designed to be compatible with PROV-DM.

On the news side, the rNews vocabulary is used for semantic mark-up of web news resources in several papers. Whereas most of the papers in this review rely on the older versions of rNews, the ASRAEL project [60] uses its newer schema.org-based rNews vocabulary. The Internationalization Tag Set (ITS) is also used in a few papers, for example, to unify claims in ClaimsKG [69], The IPTC’s General Architecture Framework (GAF) is used in NewsReader [5972].

On the natural-language side, several proposals [216972] use the RDF/OWL-based NLP Interchance Format (NIF) to exchange semantic data between NLP components. In addition, more than a third of the papers propose their own domain ontologies.

Semantic information resources: By semantic information resources, we mean open knowledge graphs, or openly available semantic datasets expressed as triples. As shown in Figure 5(a), semantic encyclopedias are most frequently used. More than a quarter of the main papers somehow exploit DBpedia. It is, for example, used by NewsReader [5972] for semantic linking and enrichment. Wikidata is an alternative that is used in several recent approaches. It is used by ASRAEL [60] and VLX-Stories [14] to support semantic labelling, enrichment, and search, and it used to detect fake news in Reference [5]. There is also increasing uptake of Google’s KG, which is used by VLX-Stories [14] to detect emerging entities, in Reference [61] to separate emerging from already-known entities, and by AGV [13] to provide additional information about entities extracted from educational TV programs. Although it has been seeded into Google’s knowledge graph and is no longer maintained, Freebase is still being used for external linking in K-Pop [36], for content-based recommendation in Reference [33], for evaluation of TAMURE [78], and for enriching government data in Reference [62] (more later). GeoNames is used as the reference graph for geographical information in many papers, such as References [10444662637172]. With the availability of large one-stop KGs like these, fewer papers than before rely on the LOD cloud in general. An exception is Reference [26], which exploits the LOD cloud to identify news stories that match users’ interests.

Fig. 5. The most frequently used semantic (a) information resources and (b) processing techniques.

Beyond general semantic encyclopedias and other LOD resources, YAGO2 and its integration of WordNet event classes is used in Reference [38] to classify named news events. The initial version of YAGO is used by Pundit [56] to build a world entity graph for mining causal relationship between news events, and in Reference [74] to infuse world knowledge into a Knowledge-driven Multimodal Graph Convolutional Network (KMGCN) for fake news detection. Common-sense knowledge from the Cyc project [112] is used, too, for example, to augment reasoning over semantic representations mined from financial news texts [51] and to predict future events [56]. PolarisX [77] uses ConceptNet 5.5 as a development case and for evaluating its approach to automatically expand knowledge graphs with new news events. ConceptNet is also used by Pundit [56]. Several of the general semantic information resources, such as DBpedia, ConceptNet, Cyc, Wikidata, and YAGO, come with their own resource-specific ontologies and vocabularies in addition to the ones mentioned in the previous section.

On the natural language side, WordNet [98117] is not natively semantic, but it is used in a third of the main papers—more than any of the natively semantic resources—including Hermes [4], NewsReader [5972], and SPEED [24]—although only a single paper [58] explicitly mentions WordNet’s RDF version.

Semantic processing techniques: By semantic processing technique, we mean programming techniques and tools used to create and exploit semantic information resources. As shown in Figure 5(b), entity linking [82] is the most frequently used technique by far. The most used entity linkers are DBpedia Spotlight, Thomson-Reuters’ OpenCalais, and TagMe. Beyond entity linking, seven papers employ logical reasoning. Description logic and OWL-DL is used for trust-based resolution of inconsistent KBs in Reference [22] and for managing the NEWS Ontology [17]. Other papers mention general ontology-enabled reasoning without OWL-DL. For example, PWFF [28] and Reference [51], which uses Cyc to answer questions about business news. Rule-based inference is also used, e.g., in References [615203557]. The most common programming API for semantic data processing is Apache’s Java-based Jena framework, used in 12 main papers. Only 4 papers mention Python’s RDFlib, most of them from recent years. Protégé is used in 4 papers, for example, for ontology development in Neptuno [7] and in the NEWS project [15].

Semantic storage techniques: Although almost all the main papers mention ontologies or knowledge graphs, few of them discuss storage and none of them focus primarily on the storage side. The two most frequently used triple stores are RDF4J (formerly Sesame) used by four papers and OpenLink Virtuoso also used by four papers. AllegroGraph is employed in two papers [4449]. Used by NewsReader [5972], the KnowledgeStore is designed to store large collections of documents and link them to RDF triples that are extracted from the documents or collected from the LOD cloud. It uses a big-data ready file system (Hadoop Distributed File System, HDFS) and databases (Apache HBase and Virtuoso) to store unstructured (e.g., news articles) and structured information (e.g., RDF triples) together.

Summary: Our review demonstrates that the research on KG for news exploits a broad variety of available semantic resources, techniques, and tools. The research on KGs for news differs from the mainstream research on KGs mainly in its stronger focus on language (e.g., the ITS and NIF vocabularies), on events (e.g., the SEM ontology), and, of course, on news (the rNews vocabulary). The border between semantic and non-semantic computing techniques is not always sharp. For example, although WordNet is not natively semantic, it is available as RDF and is used as a semantic information resource in many proposals. A recent trend is that Wikidata is becoming more popular compared to DBpedia.

3.8 Other Techniques and Tools

Most main papers use semantic knowledge graphs in combination with other techniques and tools. Similar to the previous section, we separate them into exchange formatsinformation resources, and processing and storage techniques. The online Addendum (Section B.1) presents a detailed review, which shows that the research on semantic knowledge graphs for the news is technologically diverse. We find examples of research that exploit most of the popular news-related standards and most of the popular techniques for NLP, machine learning, deep learning, and computing in general.

On the news side, the IPTC family of standards and resources is central. On the NLP side, entity extraction, NL pre-processing, co-reference resolution, morphological analysis, and semantic-role labelling are common, whereas GATE, Lucene, spaCy, JAPE, and StanfordNER are the most used tools.

On the ML side, the past decade has seen more and more proposals that exploit machine-learning techniques, as illustrated by three early examples from 2012: Reference [9] uses greedy clustering to automatically detect provenance relations between news articles. The Hermes framework [29] uses a pattern-language and rule-based approach to learn ontology instances and event relations from text, combining lexico-semantic patterns with semantic information. It is used to analyse financial and political news articles, splitting its corpus of news articles into a training and a test set. Pundit [56] mines text patterns from news headlines to predict potential future events based on textual descriptions of current events. It uses machine learning to automatically induce a causality function based on examples of causality pairs mined from a large collection of archival news headlines. Whereas these early approaches rely on hand-crafted rules and dedicated learning algorithms, more recent proposals use standard machine-learning techniques for word, graph, and entity embeddings, such as TransE [89], TransR [113], TransD [110], and word2vec [116].

On the DL side, there has been a sharp rise since 2019 in deep learning [102] approaches. PolarisX [77] uses pre-trained multilingual BERT model to detect new relations, with the aim of updating its underlying knowledge graph in real time. TAMURE [78] uses tensor factorisation implemented in TensorFlow to learn joint embedding representations of entities and relation types. Focusing on click-through rate (CTR) prediction in online news sites, DKN [73] uses a Convolutional Neural Network [102] with separate channels for words and entities and an attention module to dynamically aggregate user histories. Reference [19] proposes a deep neural network model that employs multiple self-attention modules for words, entities, and users for news recommendation. Reference [52] proposes the B-TransE model to detect fake news based on content. The most used deep learning techniques and tools are CNN [102], GRU [102], GCN [91103], LSTM [102], BERT [96], and attention [135].

The focus on news standards is strongest in the first part of the study period, up to around 2014, when many approaches incorporate existing news standards into the emerging LOD cloud [87]. The second part, from around 2015, sees a shift towards machine learning approaches [119], first focusing on NLP and embedding techniques and, since around 2019, on deep learning [102].

3.9 News Domain

As shown in Figure 6(a), most of the main papers do not focus on a particular news domain, except as examples or for evaluation. Among the domain-specific papers, economy/finance is most common. For example, Reference [43] presents a semantic search engine for financial news that uses an automatically populated knowledge graph that is kept up to date with semantically annotated financial news items, and we have already mentioned the SPEED pipeline for economic event detection and annotation in real time [24].

Fig. 6. The most frequently targeted (a) news domains and (b) languages.

Politics is the theme of over 900,000 French tweets collected to trace propagation of knowledge, misinformation, and hearsay [12]. Reference [62] uses named entity linking to contextualise open government data, making them available in online news portals alongside related news items that match each user’s interests. In the sports news domain, Reference [50] proposes a recommender system based on BKSport [49] that combines semantic and content-based similarity measures to suggest relevant news items. Other domains targeted by multiple papers include science [1357], business [4151], health [1323], and the stock market [441].

Targeting entertainment news, K-Pop [36] builds on an entertainer ontology to compile a semantic knowledge graph that represents the profiles and activities of Korean pop artists. The artists’ profiles in the graph are based on information from Wikipedia and enriched with content from DBpedia [85], Freebase, LinkedMDB, and MusicBrainz. They are also linked to other sources that represent not only the artist, but also their activities, business, albums, and schedules. The graph and ontology are used in the Gnosis app to enhance K-Pop entertainment news with information about artists retrieved from the knowledge graph.

WebLyzard [63] identifies topics and entities in news about the environment and uses visualisations to present lexical, geo-spatial, and other contextual information to gain overview of perceptions of and reactions to environmental threats and options. AGV [13] targets education, in particular in science and technology. Other domains include medicine [23], crime [67], and earth science [57].

Summary: Our review suggests that semantic knowledge graphs and related semantic techniques are useful in a broad range of news domains. Most investigated so far are economy and finance. There is little domain-specificity in the research so far: Most architectures and techniques proposed for one news domain appear readily transferable to others. The higher interest in the financial and business domains may result from economic opportunities combined with the availability of both quantitative and qualitative data streams in real time.

3.10 Language

As shown in Figure 6(b), the most frequently covered languages beside English are Italian and Spanish, but neither is supported by more than 10 papers. Support for French and German in the main papers appear only in combination with English. Many papers deal with a combination of several languages, such as English, Italian, and Spanish in the NEWS project [15], and a few recent approaches explicitly aim to be multi-lingual (or language-agnostic). For example, NewsReader [72] mentions Dutch, Italian, and Spanish in addition to English, whereas PolarisX [77] aims to cover Chinese, Japanese, and Korean.

Summary: Our review suggests that English is the best supported language by far but, of course, this may be because we use English language as an inclusion criterion. Additional papers addressing other major languages, such as Chinese, French, German, Hindi, and Spanish, may instead be written and published in their own languages. The other most frequently supported languages are Spanish, Italian, French, Chinese, Dutch, German, and Japanese, with many of the Chinese and Japanese papers published in recent years. Many approaches also support more than one language, exploiting the inherent language-neutrality of ontologies and knowledge graphs. There is a growing interest in offering multi-language and language-agnostic solutions.

3.11 Important Papers

Our main papers reference 1,842 earlier papers and are themselves cited 2,381 times according to Semantic Scholar. Table 2 shows that the most cited of our main papers is the one about KIM [37] from 2004, with the much more recent DKN paper [73] from 2018 second. The paper about Pundit [56] from 2012 is also frequently cited. To account for recency, Table 2 is therefore ordered by citation numbers that are weighted against the expected number of citations of a main paper from the same year. Just outside the top five, References [2124386972] are also frequently cited.

Table 2.

 Show More

Table 2. The Five Most Frequently Cited Main Papers (Recency Weighted)

Table 3 shows the papers that are referenced most frequently by our main papers. Among the outgoing citations, seminal papers on the Semantic Web [86] and on WordNet [98] are most frequently cited. Also much cited are the central papers on GATE, TransE, and word2vec. Just outside the top five, other frequently referenced papers are References [8588113132136], confirming the importance of LOD resources and embedding models for the research on semantic KGs for news. Closely related to our main papers, another paper on KIM [128] is cited six times, and Reference [109], a precursor to the SemNews paper [30], is also cited several times.

Table 3.

 Show More

Table 3. The Five Papers Most Frequently Referenced by Our Main Papers

Summary: No paper yet stands out as seminal for the research area. With the exception of the KIM project, none of the main papers or projects are frequently cited by other main papers, suggesting that research on semantic KGs has not yet matured into a clearly defined research area that is recognised by the larger research community.

3.12 Frequent Authors and Projects

Table 4(a) shows the most frequent main-paper authors, along with their most centrally related projects. Table 5 also shows co-authorship cliques, defined by chains of at least two co-authored papers. The table shows that repeated co-authorship among the frequent authors occurs exclusively within a small number of research projects (or persistent collaborations), such as NEWS [15], Hermes [4], and NewsReader [72].

Table 4.

Table 4. (a) Authors with Three Main Papers or More and (b) Projects with Multiple Papers

Table 5.

 Show More

Table 5. Groups of Authors Connected by Chains of Two or More Co-authored Papers

The seven cliques cover all the repeated collaborations, we have found. Table 4(b) also shows the cumulative citation counts for each project or collaboration. Hermes and NewsReader are the most frequently cited projects, but there are very few citations to these and to the other projects/collaborations from other main papers. Indeed, none of the main papers from the seven listed projects and collaborations are citing one another (although, of course, such cross-references may still exist between papers from the same projects that we have not included as main papers). Only 43 references in total are from one main paper to another (the online Addendum (Figure 10) presents a citation graph).

Summary: The analysis underpins that the research on semantic KGs has not yet matured into a distinct research area. The research is carried out mostly by independently working researchers and groups, although the NewsReader project has involved several institutions located in different countries. There is so far little collaboration and accumulation of knowledge in the area, although the early KIM [37] proposal has been used in later research.

3.13 Evolution over Time

The research on semantic knowledge graphs for news can be divided into four eras that broadly follow the evolution of knowledge graphs and related technologies in general: the Semantic Web (–2009), Linked Open Data (LOD, 2010–2014), knowledge graphs (KGs, 2015–2018), and deep learning (DL, 2019–) eras. Figure 7 presents corresponding timelines that show the percentage of main papers from each year that match each theme. To underpin the separation into four eras further, the online Addendum (Section B.2) presents additional timelines that show typical sub-themes from each era.

Fig. 7. Timeline for the percentages of main papers from each year that match the sub-themes Semantic Web, LOD, KGs, and deep learning.

The first era (until around 2009) is inspired by the Semantic-Web idea and early ontology work. Almost all the main papers from this era mention the Semantic Web or Semantic-Web technologies prominently in their introductions. They combine basic natural-language processing with central Semantic-Web ideas such as semantic annotation, domain ontologies, and semantic search applied to the news domain. Many of the papers bring existing news and multimedia publishing standards into the Semantic-Web world, and the IPTC Media Topics are therefore important. Central semantic techniques are RDF, RDFS, OWL, and SPARQL, and important tasks are archiving and browsing. There is also an early interest in multimedia. Figure 8(a) shows a word cloud of the most prominent sub-themes for papers published during this era.

Fig. 8. Word clouds for (a) the Semantic-Web era (until around 2009), (b) the Linked Open Data (LOD) era (around 2010–2014), (c) the knowledge-graph era (around 2015–2018), and (d) the deep-learning era (from around 2019).

The main papers in second era (2010–2014, but starting with Reference [71] already in 2008) trails the emergence of the LOD cloud [87], which many of the papers use to motivate their contributions. Contextualisation and other types of semantic enrichment of news texts is central, aiming to support more precise search and recommendation. Although some papers use Wikipedia and DBpedia for enrichment, the most used information resource is WordNet. To link news texts precisely to existing semantic resources, more advanced pre-processing of news texts is used along with techniques such as morphological analysis and vector spaces. GATE is a much-used NLP tool in this era, as is OpenCalais for entity linking and Jena for managing RDF data.

The third era (2015–2018), reflects Google’s adoption of the term “Knowledge Graphs” in 2012 and the growing importance of machine learning [119]. One of the first main papers to mention knowledge graphs is Reference [2] already in 2013, but most of the main papers are published starting in 2015. The research increasingly considers knowledge graphs independently of semantic standards such as RDF and OWL, and uses machine learning and related techniques to analyse news texts more deeply, for example, extracting events and facts (relations). DBpedia and entity linking become more frequently used, along with word and graph embeddings. On the NLP side, co-reference resolution and dependency parsing become more important, along with StanfordNER.

Since around 2019, a fourth and final era starts to emerge. Typical approaches analyse news articles using deep neural network (NN) architectures that combine text- and graph-embedding approaches and that infuse triples from open KGs into graph representations of news texts. Central emerging tasks are fact checking, fake-news detection, and click-through rate (CTR) prediction. Deep-learning techniques such as CNN, LSTM, and attention become important, and spaCy is used for NLP. On the back of deep image-analysis techniques, multimedia data also makes a return. Because the boundary between this and the KG era is not sharp, the word cloud in Figure 8(d) has many similarities to Figure 8(c).Skip 4DISCUSSION Section

DISCUSSION

Based on the analysis, this section will discuss each main theme in our analysis framework (Table 1). We will then answer the four research questions posed in the Introduction and discuss the limitations of our article.

4.1 Conceptual Framework

Table 6 shows the conceptual framework that results from populating our analysis framework in Table 1 with the most frequently used sub-themes from the analysis. It is organised in a hierarchy of depth up to 4 (e.g., Other techniques: → Other resources → language → WordNet). The framework shows which areas and aspects of semantic knowledge graphs for news that have so far been most explored in the literature. It can be used both as an overview of the research area, as grounds for further theory building, and as a guide for further research.

Table 6.

 Show More

Table 6. Conceptual Framework

The earliest versions of our framework also contained geographical region as a top-level theme, alongside news domain and language, but very few of our main papers were specific to a region, and never exclusively so. For example, although the contextualisation of open government data in Reference [62] focuses on Colombian politics, the proposed solution is straightforwardly adaptable to other regions.

4.2 Implications for Practice

For each main theme, this section suggests implications for practice, before the next section proposes paths for further research.

Technical result types: There are already many tools and techniques available that are sufficiently developed to be tested in industrial workflows. Commercial tools such as VLX-Stories [14] and ViewerPro are also starting to emerge. But most research proposals are either research pipelines/prototypes or standalone components that require considerable effort to integrate into existing workflows before they can become productive. Pilot projects that match high-benefit tasks with low-risk technologies and tools are therefore essential to successfully introduce semantic KGs in newsrooms.

Empirical result types: Although there are examples of tools and techniques that have been deployed in real news production workflows, they are the exception rather than the rule. This poses a double challenge for newsrooms that want to use KGs for news: It is usually not known how robust the proposed techniques and tools are in practice, and it is usually not known how well they fit actual industrial needs. Introducing KGs into newsrooms must therefore focus on continuous evaluation both of the technology itself and of its consequences, opening possibilities for collaboration between industry (which wants its projects evaluated) and researchers (who want access to industrial cases and data).

Intended users: The most mature solutions support journalists through tools and techniques for searching, archiving, and content recommendation. The general news user is supported by proposals for news recommendation and to some extent searching.

Tasks: The most mature research proposals target long-researched tasks such as semantic annotation, searching, and recommendation, both for content retrieval (pull) and provision (push). In particular, annotation of news texts with links to mentioned entities and concepts is already used in practice and will become even more useful as the underlying language models continue to improve. Semantic searching and browsing are also well-understood areas. Semantic enrichment with information from open KGs and other sources is a maturing area that builds on a long line of research, but suffers from the danger of creating information overload. Rising areas that are becoming available for pilot projects are automatic news detection and automatic provision of background information.

Input data: The most mature tools and techniques are text-based. When multimedia is supported, it is often done indirectly by first converting speech to text or by using image captions only. Newer approaches that exploit native audio and image analysis techniques in combination with semantic KGs may soon become ready for industrial trials. Many newsrooms already have experience with robots [118] that exploit input data from sensors, the Internet of Things (IoT), and open APIs [84]. This creates opportunities to explore new uses of semantic KGs that augment existing robot-journalism tools and techniques. Much of the research that exploits social media is based on Twitter. This poses a challenge, because Twitter use is dwindling in some parts of the world, sometimes with traffic moving to more closed platforms, such as Instagram, Snapchat, Telegram, TikTok, WhatsApp, and so on. In response, news organisations could attempt to host more social reader interactions inside their own distribution platforms, where they retain access to the user-generated content. Semantic KGs offer opportunities through their support for personalisation, recommendation, and networking.

News life cycle: Low-risk starting points for industrial trials are the mature research areas based on already-published news, such as archive management, recommendation, and semantically enriched search. Automated detection of emerging news events and live monitoring of breaking news situations are higher-risk areas that also offer high potential rewards.

Semantic techniques: Because they tend to rely on standard semantic techniques, many of the proposed techniques can be run in the cloud, for example, in Amazon’s Neptune-centric KG ecosystem and supported by other Amazon Web Services for NLP and ML/DL. Cloud infrastructures give newsrooms a way to explore advanced computation- and storage-intensive KG-based solutions without investing heavily upfront in new infrastructure.

Other techniques: The demonstrated ability of KG-based approaches to work alongside a wide variety of other computing techniques and tools suggests that newsrooms that want to exploit semantic KGs should build on what they already have in place, using KG-based techniques to augment existing services and capabilities. For example, KGs are well suited to integrate diverse information sources through exchange standards such as RDF and SPARQL and ontologies expressed in RDFS and OWL. One possibility is therefore to introduce them in newsrooms as part of ML and DL initiatives that need input data from multiple and diverse sources, whether internal or external. Semantic analysis of natural language texts, audio, images, and video is rapidly becoming available as increasingly powerful commodity services. KGs in newsrooms could be positioned to enrich and exploit the outputs of such services, acting as a hub that can represent and integrate the results of ML- and DL-driven analysis tools and prepare the data for journalists and others.

News domain: For newsrooms that want to exploit KGs, the most mature domains are business and finance. For example, ViewerPro, an industrial tool for ontology-based semantic analysis and annotation of news texts, has been applied to gain effective access to relevant finance news. The proposed tools and techniques are often transferable across domains and purposes. Good candidates for industrial uptake are domains that are characterised by data streams that are reliable and high-quality, but insufficiently structured for currently available tools, e.g., for robot journalism [118]. Using KG-techniques to expand the reach and capabilities of existing journalistic robots may be a path to reap quick benefits from KGs on top of existing infrastructures.

Language: Given the focus on English in the research on semantic KG for the news and on NLP in general, international news is a natural starting point for newsrooms in non-English speaking countries that want to explore KG-based solutions. For newsrooms in English and other major-language countries, KG-powered cross-lingual and language-agnostic services can be used to simplify searching, accessing, and analysing minor-language resources, offering a low-effort/high-reward path to introducing semantic KGs.

4.3 Implications for Research

Based on our analysis of main papers, this section proposes paths for further research.

Technical result types: More industrial-grade prototypes and platforms are needed in response to the call for industrial testing. Much of the current research, such as the exploration of deep learning and other AI areas for news purposes, is technology-driven and needs to be balanced by investigations of the needs of journalists, newsrooms, news users, and other stakeholders.

Empirical result types: To better understand industrial needs, challenges, opportunities, and experiences, empirical studies are called for, using the full battery of research approaches, including case- and action-research, interview- and survey-based research, and ethnographic studies of newsrooms. Research on semantic knowledge graphs for the news might benefit from the growing and complementary body of literature on augmented, computational, and digital journalism (e.g., References [9297129134]), which focuses on the needs of newsrooms and journalists, but goes less into detail about the facilitating technologies, whether semantic or not. Indeed, the research on semantic KGs for the news hardly mentions the literature on augmented/digital/data journalism which, vice versa, does not go into the specifics of KGs.

Most papers that propose new techniques or tools offer at least some empirical evaluation of their own proposals. Experimental evaluations using gold-standard datasets and information-retrieval measures are becoming increasingly common, but there is no convergence yet towards particular gold-standard datasets and measures, which makes it hard to compare proposals and assess overall progress. This is an important methodological challenge for further research. We also find no papers that focus on evaluating tools or techniques proposed by others. Also, the papers that develop pipelines and prototypes are seldom explicit about the design research method they have followed.

Intended users: We found no papers discussing semantic knowledge graphs and related techniques for citizen journalism, for example, investigating social semantic journalism as outlined in Reference [105]. Local journalism [122133] is also not a current focus, and we found few papers that explicitly mention newsrooms or consider the social and organisational sides of news production and journalism. There is also no mentioning of robot journalism in the main papers.

Tasks: More research is needed in areas that are critical for representing news content on a deeper level, beyond semantic annotation with named entities, concepts, and topics. Central evolving areas are event detection, relation extraction, and KG updating, in particular identification and semantic analysis of dark entities and relations.

There is little research on the quality of data behind semantic KGs for news. Aspects of semantic data quality, such as privacy, provenance, ownership, and terms of use, need more attention. Few research proposals target or undertake multimedia analysis natively (i.e., without going through text) and specifically for news.

Input data: The research on social media tends to focus on short texts, which are hard to analyse, because they provide less context and use abbreviations, neologisms, and hashtags [131]. More context can be provided by integrating newer techniques that also analyse the audio, image, and video content in social messages. Some research approaches harvest citizen-provided data from social media, but there are no investigations of how to use semantic techniques and tools participatively for citizen journalism [105]. There is little research on KGs for news that exploits data from sensors and from the IoT in general [84], and there is little use of open web APIs outside a few domains (such as business/finance). We have already mentioned the ensuing possibility of combining semantic KGs with robot-journalism tools and techniques. GDELT is another untapped resource, although data quality and ownership is an issue. Research is needed on how its data quality can be corroborated and improved. Also, the low-level events in GDELT data streams need to be aggregated into news-level events.

News life cycle: Relatively little research targets detecting emerging news events, monitoring breaking news situations, and following developing stories. Event detection and tracking as well as detecting emerging entities and relations are important research challenges.

Semantic techniques: Most of the research uses existing news corpora or harvests news articles on-demand. There is less focus on building and curating journalistic knowledge graphs over time. Due to the high volume, velocity, and variety of news-related information, semantic news KGs are a potential driver and test bed for real-time and big-data semantic KGs. More research is therefore needed on combining KGs with state-of-the-art techniques for real-time processing and big data. Yet, none of the main papers have primary focus on the design of semantic data architectures/infrastructures for newsrooms, for example, using big-data infrastructures, data lakes, web-service orchestrations, and so on. The most big-data-ready research proposal is NewsReader, through its connection with the big-data ready KnowledgeStore repository. The News Hunter platform developed in the News Angler project [46] is also built on top of a big-data ready infrastructure [100]. In addition to supporting processing of big data in real time, these architectures and infrastructures must be forward-engineered to accommodate the increasing availability of high-quality, high-performance commodity cloud-services for NLP, ML, and DL that can be exploited by news organisations.

Other techniques: On the research side, few approaches to semantic KGs for news exploit recent advances in image understanding and speech recognition. There is a potential for cross-modal solutions that increase precision and recall by combining analyses of text, audio, images and, eventually, video. These solutions need to be integrated with semantic KGs, and their application should focus on areas where KGs bring additional benefits, such as infusing world and common-sense knowledge into existing analyses. Also, few approaches so far exploit big-data and real-time computing. Although some proposals express real-time ambitions, they are seldom evaluated on real-volume and -velocity data streams and, when they are (e.g., RDFLiveNews [21] and SPEED [24]), they do not approach web-scale performance. Although the proposed research pipelines may not be optimised for speed, performance-evaluation results suggest that more efficient algorithms are needed, for example, running on staged and parallel architectures. High-performance technologies for massively distributed news knowledge graphs are also called for, for example, exploiting big-graph databases such as Pregel and Giraph.

News domain: Whereas practical applications of KGs may be driven by economical (for economy/finance) and popular (e.g., for sports) interests, there is ample opportunity on the research side for adapting and tuning existing approaches to new and unexplored domains that have high societal value. One largely unexplored domain is corruption and political nepotism, along the lines suggested in Reference [128]. Misinformation is another area of great importance, and in the domain of crises and social unrest, the GDELT data streams may offer opportunities.

Language: Research is needed to make semantic KGs for news available for smaller languages. There is so far little uptake of cross-language models like multi-lingual BERT and little research on exploiting dedicated language models for smaller languages for news purposes.

4.4 Research Questions

We are now ready to answer the four research questions we posed in the Introduction.

RQ1: Which research problems and approaches are most common, and what are the central results? Our discussion in Section 4 and Table 6 answers this question for each of the main themes in our framework. The review shows that research on semantic knowledge graphs for news is highly diverse and in constant flux as the enabling technologies evolve. A frequent type of paper is one that develops new tools and techniques for representing news semantically to disseminate news content more effectively. In response to the increasing societal importance of information quality and misinformation, there is currently a rapidly growing interest in fake-news detection and fact checking. The tools and techniques are typically developed as pipelines or prototypes and evaluated using experiments, examples or use cases. The experimental methods used are maturing.

RQ2: Which research problems and approaches have received less attention, and what types of contributions are rarer? Our discussion in Section 4, and in 4.3, in particular, answers this question by identifying many under-researched areas. The review shows that there are very few industrial case studies. In our literature searches, we have found few surveys and reviews. There is also little research on issues such as privacy, ownership, terms of use, and provenance, although a few papers mention the latter. Only a few papers focus on evaluating their results in real-time and big-data settings and, when they do, the results are often in need of improvement. Other green-field areas include: exploiting location data and data from the Internet of Things, supporting social and citizen journalism, using semantic knowledge graphs to identify new newsworthy events as in Reuters Tracer, and using semantic knowledge graphs to construct narratives and generate news content.

Although the results suggest that semantic knowledge graphs can indeed support better organisation, management, retrieval, and dissemination of news content, there is still a potential for much larger uptake in industry. Empirical studies are needed to explain why. One possible explanation is that there is a mismatch between what the current tools and algorithms offer and what the industry needs. Another possible explanation is that the solutions themselves are immature, for example, that existing analysis techniques are not sufficiently precise or that the often crowd-sourced reference and training data used are perceived as less trustworthy.

RQ3: How is the research evolving? Our analysis in Section 3.13 answers this question by showing that the research broadly follows the development of the supporting technologies used. We identify four eras in the evolution of KGs for news, characterised by (1) applying early Semantic-Web ideas to the news domain, (2) exploiting the Linked Open Data (LOD) cloud for news purposes, (3) semantic knowledge graphs and machine learning and, most recently, (4) deep-learning approaches based on semantic knowledge graphs.

RQ4: Which are the most frequently cited papers and projects, and which papers and projects are citing one another? Our analyses in Sections 3.11 and 3.12 answer this question. The most cited papers are the ones about DKN [73] and KIM [37]. Many recent papers that use deep-learning techniques for fake-news detection or recommendation are already much cited, e.g., References [5269]. Among the central projects, main papers related to the Hermes [4], NewsReader [72], and NEWS [15] projects have been most cited. Another much-referenced group of papers centres around what we have called the “MediaLoep” collaboration. The citation analysis reported in the online Addendum (Figure 10) shows that the main paper from the Neptuno project [7] and the effort to make IPTC’s news architecture semantic [71] also are important.

4.5 Limitations

The most central limitation of our literature review is its scope. We only consider papers that use semantic knowledge graphs or related semantic techniques for news-related purposes, excluding papers that attempt to solve similar problems using other knowledge representation techniques or targeting other domains. There is also a growing body of research on representing texts in general as semantic knowledge graphs, proposing techniques and tools that could also be used to analyse news. There is another growing body of research on supporting news with knowledge graphs that are not semantically linked, i.e., with knowledge graphs whose nodes and edges do not link into the LOD cloud.Skip 5CONCLUSION Section

CONCLUSION

We have reported a systematic literature review of research on how semantic knowledge graphs can be used to facilitate all aspects of production, dissemination, and consumption of news. Starting with more than 6,000 papers, we identified 80 main papers that we analysed in depth according to an analysis framework that we kept refining as analysis progressed. As a result, we have been able answer research questions about past, current, and emerging research areas and trends, and Section 4.3 has offered many paths for further work. We hope the results of our study will be useful for practitioners and researchers who are interested specifically in semantic knowledge graphs for news or more generally in computational journalism or in semantic knowledge graphs.

CONFLICT OF INTEREST

The authors are themselves involved in the News Angler project reported in Reference [46].

3543508.appSupplementary material

Footnotes

  1. Hence, the article will use the term “semantic knowledge graph” or “semantic KG” in an inclusive way that also covers semantic technologies, computational ontology, Linked Open Data (LOD), and Semantic Web.Footnote
  2. Later superseded by the PROTON ontology.Footnote
  3. To weight the citation counts, the three most frequently cited papers (i.e., [375673]) are removed as outliers. Average citation counts are calculated for each year for the remaining main papers. A support-vector regression (SVR) model is trained using scikit-learn [119] with a radial-basis function (RBF) kernel, C = 1, 000, and γ = 0.001. Finally, the citation count for each paper is divided by the count predicted for a paper from that year.Footnote
  4. The online Addendum (Table 11) presents an extended top-15 list.Footnote
  5. We do not report weighted reference counts, because more recent papers are much more frequently cited in our dataset, giving unreasonably high relative weight to older papers even when they are referenced only once or a few times.Footnote
  6. The online Addendum (Table 12) again presents an extended top-15 list.Footnote
  7. Table 5 also introduces informal names such as “MediaLoep” and “Wuhan” for persistent collaborations that are not centred around a single named project.Footnote
  8. The timeline depicts three-year averages.Footnote

Skip Supplemental Material Section

Supplemental Material

Available for Download

pdf

3543508.app (1.4 MB)

Supplementary material

MAIN PAPERS

  1. [1]Ahmed Adeel and Saif Syed. 2017. DBpedia based ontological concepts driven information extraction from unstructured text. Int. J. Adv. Comput. Sci. Applic. 8, 9 (2017). DOI:Reference
  2. [2]Antonini Alessio, Pensa Ruggero G., Sapino Maria Luisa, Schifanella Claudio, Teraoni Prioletti Raffaele, and Vignaroli Luca. 2013. Tracking and analyzing TV content on the web through social and ontological knowledge. In Proceedings of the 11th European Conference on Interactive TV and Video — EuroITV’13. ACM Press, 13. DOI:Navigate to 
  3. [3]Berrizbeita Francisco and Vidal Maria-Esther. 2014. Traversing the linking open data cloud to create news from tweets. In On the Move to Meaningful Internet Systems: OTM 2014 Workshops, Meersman Robert, Panetto Hervé, Mishra Alok, Valencia-García Rafael, Soares António Lucas, Ciuciu Ioana, Ferri Fernando, Weichhart Georg, Moser Thomas, Bezzi Michele, and Chan Henry (Eds.), Vol. 8842. Springer, Berlin, 479–488. DOI:Reference 1Reference 2
  4. [4]Borsje Jethro, Levering Leonard, and Frasincar Flavius. 2008. Hermes: A semantic web-based news decision support system. In Proceedings of the ACM Symposium on Applied Computing — SAC’08. ACM Press, 2415. DOI:Navigate to 
  5. [5]Braşoveanu A. M. and Andonie R.. 2019. Semantic fake news detection: A machine learning perspective. In Proceedings of the International Work-Conference on Artificial Neural Networks. Springer, 656–667.Reference 1Reference 2Reference 3
  6. [6]Cantador Iván, Castells Pablo, and Bellogín Alejandro. 2011. An enhanced semantic layer for hybrid recommender systems: Application to news recommendation. Int. J. Semant. Web Inf. Syst. 7, 1 (2011), 44–78.Reference 1Reference 2
  7. [7]Castells Pablo, Perdrix Ferran, Pulido E., Mariano Rico, Benjamins R., Contreras Jesús, and Lorés J.. 2004. Neptuno: Semantic web technologies for a digital newspaper archive. In Proceedings of the European Semantic Web Symposium. Springer, Berlin, 445–458.Navigate to 
  8. [8]Davarpour Mohammad Hossein, Sohrabi Mohammad Karim, and Naderi Milad. 2019. Toward a semantic-based location tagging news feed system: Constructing a conceptual hierarchy on geographical hashtags. Comput. Electric. Eng. 78 (2019), 204–217. DOI:Reference
  9. [9]De Nies Tom, Coppens Sam, Van Deursen Davy, Mannens Erik, and Walle Rik Van de. 2012. Automatic discovery of high-level provenance using semantic similarity. In Provenance and Annotation of Data and Processes (Lecture Notes in Computer Science), Groth Paul and Frew James (Eds.). Springer, 97–110. Navigate to 
  10. [10]Debevere Pedro, Van Deursen Davy, Van Rijsselbergen Dieter, Mannens Erik, Matton Mike, De Sutter Robbie, and Walle Rik Van de. 2011. Enabling semantic search in a news production environment. In Semantic Multimedia (Lecture Notes in Computer Science), Declerck Thierry, Granitzer Michael, Grzegorzek Marcin, Romanelli Massimo, Rüger Stefan, and Sintek Michael (Eds.). Springer, 32–47. Navigate to 
  11. [11]Dowman Mike, Tablan Valentin, Cunningham Hamish, and Popov Borislav. 2005. Web-assisted annotation, semantic indexing and search of television and radio news. In Proceedings of the 14th International Conference on World Wide Web. ACM Press, 225. DOI:Reference
  12. [12]Duroyon Ludivine, Goasdoué François, Manolescu and Ioana. 2019. A linked data model for facts, statements and beliefs. In Proceedings of the World Wide Web Conference. ACM, 988–993. DOI:Reference 1Reference 2
  13. [13]Fallucchi Francesca, Di Stabile Rosario, Purificato Erasmo, Giuliano Romeo, and De Luca Ernesto William. 2021. Enriching videos with automatic place recognition in Google Maps. Multim. Tools Applic. 81 (2021), 23105–23121. DOI:Navigate to 
  14. [14]Fernàndez-Cañellas Dèlia, Espadaler Joan, Rodriguez David, Garolera Blai, Canet Gemma, Colom Aleix, Rimmek Joan Marco, Nieto Xavier Giro-i, Bou Elisenda, and Riveiro Juan Carlos. 2019. VLX-Stories: Building an online event knowledge base with emerging entity detection. In Proceedings of the International Semantic Web Conference (ISWC’19). Springer, 382–399.Navigate to 
  15. [15]Fernández Norberto, Blázquez José M., Fisteus Jesús A., Sánchez Luis, Sintek Michael, Bernardi Ansgar, Fuentes Manuel, Marrara Angelo, and Ben-Asher Zohar. 2006. NEWS: Bringing semantic web technologies into news agencies. In The Semantic Web — ISWC 2006, Hutchison David, Kanade Takeo, Kittler Josef, Kleinberg Jon M., Mattern Friedemann, Mitchell John C., Naor Moni, Nierstrasz Oscar, C. Pandu Rangan, Steffen Bernhard, Sudan Madhu, Terzopoulos Demetri, Tygar Dough, Vardi Moshe Y., Weikum Gerhard, Cruz Isabel, Decker Stefan, Allemang Dean, Preist Chris, Schwabe Daniel, Mika Peter, Uschold Mike, and Aroyo Lora M. (Eds.), Vol. 4273. Springer, Berlin, 778–791. DOI:Navigate to 
  16. [16]Fernández Norberto, Blázquez José M., Sánchez Luis, and Bernardi Ansgar. 2007. IdentityRank: Named entity disambiguation in the context of the NEWS project. In The Semantic Web: Research and Applications, Franconi Enrico, Kifer Michael, and May Wolfgang (Eds.), Vol. 4519. Springer, Berlin, 640–654. DOI:Reference 1Reference 2
  17. [17]Fernández Norberto, Fuentes Damaris, Sánchez Luis, and Fisteus Jesús A.. 2010. The NEWS ontology: Design and applications. Exp. Syst. Applic. 37, 12 (2010), 8694–8704. Navigate to 
  18. [18]Färber Michael, Rettinger Achim, and Harth Andreas. 2016. Towards monitoring of novel statements in the news. In The Semantic Web — Latest Advances and New Domains (Lecture Notes in Computer Science), Sack Harald, Blomqvist Eva, d’Aquin Mathieu, Ghidini Chiara, Ponzetto Simone Paolo, and Lange Christoph (Eds.). Springer, 285–299. Reference
  19. [19]Gao Jie, Xin Xin, Liu Junshuai, Wang Rui, Lu Jing, Li Biao, Fan Xin, and Guo Ping. 2018. Fine-grained deep knowledge-aware network for news recommendation with self-attention. In Proceedings of the IEEE/WIC/ACM International Conference on Web Intelligence (WI). IEEE, 81–88. DOI:Reference 1Reference 2Reference 3
  20. [20]García Roberto, Perdrix Ferran, Gil Rosa, and Oliva Marta. 2008. The semantic web as a newspaper media convergence facilitator. J. Web Semant. 6, 2 (Apr.2008), 151–161. DOI:Reference 1Reference 2Reference 3
  21. [21]Gerber Daniel, Hellmann Sebastian, Bühmann Lorenz, Soru Tommaso, Usbeck Ricardo, and Ngonga Ngomo Axel-Cyrille. 2013. Real-time RDF extraction from unstructured data streams. In Proceedings of the International Semantic Web Conference (ISWC’13). 135–150. DOI:Navigate to 
  22. [22]Golbeck J. and Halaschek-Wiener C.. 2009. Trust-based revision for expressive web syndication. J. Log. Computat. 19, 5 (Oct.2009), 771–790. DOI:Reference 1Reference 2Reference 3
  23. [23]Groza A. and Pop A.-D.. 2020. Fake news detector in the medical domain by reasoning with description logics. In Proceedings of the IEEE 16th International Conference on Intelligent Computer Communication and Processing (ICCP). 145–152. DOI:Navigate to 
  24. [24]Hogenboom Alexander, Hogenboom Frederik, Frasincar Flavius, Schouten Kim, and Meer Otto van der. 2013. Semantics-based information extraction for detecting economic events. Multim. Tools Applic. 64, 1 (May2013), 27–52. DOI:Navigate to 
  25. [25]Hogenboom Frederik, Vandic Damir, Frasincar Flavius, Verheij Arnout, and Kleijn Allard. 2014. A query language and ranking algorithm for news items in the Hermes news processing framework. Sci. Comput. Program. 94 (Nov.2014), 32–52. DOI:Reference 1Reference 2Reference 3
  26. [26]Hopfgartner Frank and Jose Joemon M.. 2010. Semantic user profiling techniques for personalised multimedia recommendation. Multim. Syst. 16, 4–5 (Aug.2010), 255–274. DOI:Reference 1Reference 2Reference 3
  27. [27]Hoxha Klesti, Baxhaku Artur, and Ninka Ilia. 2016. Bootstrapping an online news knowledge base. In Web Engineering, Bozzon Alessandro, Cudre-Maroux Philippe, and Pautasso Cesare (Eds.), Vol. 9671. Springer, 501–506. DOI:
  28. [28]Hsu I.-Ching. 2013. Personalized web feeds based on ontology technologies. Inf. Syst. Front. 15, 3 (July2013), 465–479. DOI:Reference 1Reference 2Reference 3
  29. [29]IJntema Wouter, Sangers Jordy, Hogenboom Frederik, and Frasincar Flavius. 2012. A lexico-semantic pattern language for learning ontology instances from text. J. Web Semant. 15 (2012), 37–50. DOI:Reference 1Reference 2
  30. [30]Java Akshay, Nirneburg Sergei, McShane Marjorie, Finin Timothy, English Jesse, and Joshi Anupam. 2007. Using a natural language understanding system to generate semantic web content. Int. J. Semant. Web Inf. Syst. 3, 4 (Oct.2007), 50–74. DOI:Reference 1Reference 2
  31. [31]Jing Yun, Zhiwei Xu, and Guanglai Gao. 2020. Context-driven image caption with global semantic relations of the named entities. IEEE Access 8 (2020), 143584–143594.Reference 1Reference 2Reference 3
  32. [32]Jongmans Maarten, Milea Viorel, and Frasincar Flavius. 2014. A semantic web approach for visualization-based news analytics. In Knowledge Management in Organizations, Uden Lorna, Fuenzaliza Oshee Darcy, Ting I-Hsien, and Liberona Dario (Eds.), Vol. 185. Springer, 195–204. DOI:Reference 1Reference 2
  33. [33]Joseph Kevin and Jiang Hui. 2019. Content based news recommendation via shortest entity distance over knowledge graphs. In Proceedings of the World Wide Web Conference. ACM, 690–699. DOI:Navigate to 
  34. [34]Kallipolitis Leonidas, Karpis Vassilis, and Karali Isambo. 2012. Semantic search in the world news domain using automatically extracted metadata files. Knowl.-based Syst. 27 (Mar.2012), 38–50. DOI:Reference 1Reference 2Reference 3
  35. [35]Kasper Walter, Steffen Jörg, and Zhang Yajing. 2008. News annotations for navigation by semantic similarity. In Proceedings of KI 2008: Advances in Artificial Intelligence, Dengel Andreas R., Berns Karsten, Breuel Thomas M., Bomarius Frank, and Roth-Berghofer Thomas R. (Eds.), Vol. 5243. Springer, Berlin, 233–240. DOI:Reference 1Reference 2
  36. [36]Kim Haklae. 2017. Building a K-Pop knowledge graph using an entertainment ontology. Knowl. Manag. Res. Pract. 15, 2 (May2017), 305–315. DOI:Navigate to 
  37. [37]Kiryakov Atanas, Popov Borislav, Terziev Ivan, Manov Dimitar, and Ognyanoff Damyan. 2004. Semantic annotation, indexing, and retrieval. J. Web Semant. 2, 1 (Dec.2004), 49–79. DOI:Navigate to 
  38. [38]Kuzey Erdal, Vreeken Jilles, and Weikum Gerhard. 2014. A fresh look on knowledge bases: Distilling named events from news. In Proceedings of the 23rd ACM International Conference on Conference on Information and Knowledge Management. ACM Press, 1689–1698. DOI:Reference 1Reference 2Reference 3
  39. [39]Lim Edward H. Y., Lee Raymond S. T., and Liu James N. K.. 2008. KnowledgeSeeker — An ontological agent-based system for retrieving and analyzing Chinese web articles. In Proceedings of the IEEE International Conference on Fuzzy Systems (IEEE World Congress on Computational Intelligence). IEEE, 1034–1041. DOI:Reference
  40. [40]Liu Danyang, Lian Jianxun, Liu Zheng, Wang Xiting, Sun Guangzhong, and Xie Xing. 2021. Reinforced anchor knowledge graph generation for news recommendation reasoning. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining. 1055–1065.Navigate to 
  41. [41]Liu Jue, Lu Zhuocheng, and Du Wei. 2019. Combining enterprise knowledge graph and news sentiment analysis for stock price prediction. In Proceedings of the 52nd Hawaii International Conference on System Sciences.Reference 1Reference 2Reference 3
  42. [42]Liu Jinshuo, Wang Chenyang, Li Chenxi, Li Ningxi, Deng Juan, and Pan Jeff Z.. 2021. DTN: Deep triple network for topic specific fake news detection. J. Web Semant. 70 (2021), 100646. DOI:Navigate to 
  43. [43]Lupiani-Ruiz Eduardo, García-Manotas Ignacio, Valencia-García Rafael, García-Sánchez Francisco, Castellanos-Nieves Dagoberto, Fernández-Breis Jesualdo Tomás, and Camón-Herrero Juan Bosco. 2011. Financial news semantic search engine. Exp. Syst. Applic. 38, 12 (Nov.2011), 15565–15572. DOI:Reference
  44. [44]Mannens Erik, Coppens Sam, De Pessemier Toon, Dacquin Hendrik, Van Deursen Davy, De Sutter Robbie, and Walle Rik Van de. 2013. Automatic news recommendations via aggregated profiling. Multim. Tools Applic. 63, 2 (Mar.2013), 407–425. DOI:Navigate to 
  45. [45]Mao Qianren, Li Xi, Peng Hao, Li Jianxin, He Dongxiao, Guo Shu, He Min, and Wang Lihong. 2021. Event prediction based on evolutionary event ontology knowledge. Fut. Gen. Comput. Syst. 115 (2021), 76–89.Navigate to 
  46. [46]Motta Enrico, Daga Enrico, Opdahl Andreas L., and Tessem Bjørnar. 2020. Analysis and design of computational news angles. IEEE Access 8 (2020), 120613–120626.Navigate to 
  47. [47]Mukherjee Saikat, Yang Guizhen, and Ramakrishnan I. V.. 2003. Automatic annotation of content-rich HTML documents: Structural and semantic analysis. In The Semantic Web — ISWC 2003, Goos Gerhard, Hartmanis Juris, Leeuwen Jan van, Fensel Dieter, Sycara Katia, and Mylopoulos John (Eds.), Vol. 2870. Springer, Berlin, 533–549. DOI:Reference
  48. [48]Müller-Budack Eric, Theiner Jonas, Diering Sebastian, Idahl Maximilian, and Ewerth Ralph. 2020. Multimodal analytics for real-world news using measures of cross-modal entity consistency. In Proceedings of the International Conference on Multimedia Retrieval. ACM, 16–25. DOI:Reference
  49. [49]Nguyen Quang-Minh and Cao Tuan-Dung. 2015. A novel approach for automatic extraction of semantic data about football transfer in sport news. Int. J. Pervas. Comput. Commun. 11, 2 (2015), 233–252. DOI:WOS:000212340300007.Navigate to 
  50. [50]Nguyen Quang-Minh, Nguyen Thanh-Tam, and Cao Tuan-Dung. 2016. Semantic-based recommendation for sport news aggregation system. In Research and Practical Issues of Enterprise Information Systems (Lecture Notes in Business Information Processing), Tjoa A Min, Xu Li Da, Raffai Maria, and Novak Niina Maarit (Eds.). Springer, 32–47. Reference 1Reference 2
  51. [51]Novalija Inna and Mladenić Dunja. 2013. Applying semantic technology to business news analysis. Appl. Artif. Intell. 27, 6 (July2013), 520–550. DOI:Navigate to 
  52. [52]Pan Jeff Z., Pavlova Siyana, Li Chenxi, Li Ningxi, Li Yangmei, and Liu Jinshuo