Injecting Commonsense Knowledge into Prompt Learning for Zero-Shot Text Classification

Jing Qian, Qi Chen, Yong Yue*, Katie Atkinson, Gangmin Li

*Corresponding author for this work

Research output: Chapter in Book or Report/Conference proceedingConference Proceedingpeer-review

Abstract

The combination of pre-training and fine-tuning has become a default solution to Natural Language Processing (NLP) tasks. The emergence of prompt learning breaks such routine, especially in the scenarios of low data resources. Insufficient labelled data or even unseen classes are frequent problems in text classification, equipping Pre-trained Language Models (PLMs) with task-specific prompts helps get rid of the dilemma. However, general PLMs are barely provided with commonsense knowledge. In this work, we propose a KG-driven verbalizer that leverages commonsense Knowledge Graph (KG) to map label words with predefined classes. Specifically, we transform the mapping relationships into semantic relevance in the commonsense-injected embedding space. For zero-shot text classification task, experimental results exhibit the effectiveness of our KG-driven verbalizer on a Twitter dataset for natural disasters (i.e. HumAID) compared with other baselines.

Original languageEnglish
Title of host publicationICMLC 2023 - Proceedings of the 2023 15th International Conference on Machine Learning and Computing
PublisherAssociation for Computing Machinery
Pages427-432
Number of pages6
ISBN (Electronic)9781450398411
DOIs
Publication statusPublished - 17 Feb 2023
Event15th International Conference on Machine Learning and Computing, ICMLC 2023 - Hybrid, Zhuhai, China
Duration: 17 Feb 202320 Feb 2023

Publication series

NameACM International Conference Proceeding Series

Conference

Conference15th International Conference on Machine Learning and Computing, ICMLC 2023
Country/TerritoryChina
CityHybrid, Zhuhai
Period17/02/2320/02/23

Keywords

  • knowledge graph
  • prompt learning
  • zero-shot text classification

Fingerprint

Dive into the research topics of 'Injecting Commonsense Knowledge into Prompt Learning for Zero-Shot Text Classification'. Together they form a unique fingerprint.

Cite this