Enhancing Text Comprehension via Fusing Pre-trained Language Model with Knowledge Graph

Jing Qian, Gangmin Li, Katie Atkinson, Yong Yue*

*Corresponding author for this work

Research output: Chapter in Book or Report/Conference proceedingConference Proceedingpeer-review

Abstract

Pre-trained language models (PLMs) such as BERT and GPTs capture rich linguistic and syntactic knowledge from pre-training over large-scale text corpora, which can be further fine-tuned for specific downstream tasks. However, these models still have limitations as they rely on knowledge gained from plain text and ignore structured knowledge such as knowledge graphs (KGs). Recently, there has been a growing trend of explicitly integrating KGs into PLMs to improve their performance. For instance, K-BERT incorporates KG triples as domain-specific supplements into input sentences. Nevertheless, we have observed that such methods do not consider the semantic relevance between the introduced knowledge and the original input sentence, leading to the issue of knowledge impurities. To address this issue, we propose a semantic matching-based approach that enriches the input text with knowledge extracted from an external KG. The architecture of our model comprises three components: the knowledge retriever (KR), the knowledge injector (KI), and the knowledge aggregator (KA). The KR, built upon the sentence representation learning model (i.e. CoSENT), retrieves triples with high semantic relevance to the input sentence from an external KG to alleviate the issue of knowledge impurities. The KI then integrates the retrieved triples from the KR into the input text by converting the original sentence into a knowledge tree with multiple branches, the knowledge tree is transformed into an accessible sequence of text that can be fed into the KA. Finally, the KA takes the flattened knowledge tree and passes it through an embedding layer and a masked Transformer encoder. We conducted extensive evaluations on eight datasets covering five text comprehension tasks, and the experimental results demonstrate that our approach exhibits competitive advantages over popular knowledge-enhanced PLMs such as K-BERT and ERNIE.

Original languageEnglish
Title of host publicationACAI 2023 - Conference Program
Subtitle of host publication2023 6th International Conference on Algorithms, Computing and Artificial Intelligence
PublisherAssociation for Computing Machinery
Pages353-360
Number of pages8
ISBN (Electronic)9798400709203
DOIs
Publication statusPublished - 22 Dec 2023
Externally publishedYes
Event6th International Conference on Algorithms, Computing and Artificial Intelligence, ACAI 2023 - Sanya, China
Duration: 22 Dec 202324 Dec 2023

Publication series

NameACM International Conference Proceeding Series

Conference

Conference6th International Conference on Algorithms, Computing and Artificial Intelligence, ACAI 2023
Country/TerritoryChina
CitySanya
Period22/12/2324/12/23

Keywords

  • knowledge graphs
  • natural language understanding
  • sentence representation learning

Fingerprint

Dive into the research topics of 'Enhancing Text Comprehension via Fusing Pre-trained Language Model with Knowledge Graph'. Together they form a unique fingerprint.

Cite this