Zero-shot text classification with knowledge resources under label-fully-unseen setting

Yuqi Wang, Wei Wang*, Qi Chen, Kaizhu Huang, Anh Nguyen, Suparna De

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Classification techniques are at the heart of many real-world applications, e.g. sentiment analysis, recommender systems and automatic text annotation, to process and analyse large-scale textual data in multiple fields. However, the effectiveness of natural language processing models can only be confirmed when a large amount of up-to-date training data is available. An unprecedented amount of data is continuously created, and new topics are introduced, making it less likely or even infeasible to collect labelled samples covering all topics for training models. We attempt to study the extreme case: there is no labelled data for model training, and the model, without being adapted to any specific dataset, will be directly applied to the testing samples. We propose a transformer-based framework to encode sentences in a contextualised way and leverage the existing knowledge resources, i.e. ConceptNet and WordNet, to integrate both descriptive and structural knowledge for better performance. To enhance the robustness of the model, we design an adversarial example generator based on relations from external knowledge bases. The framework is evaluated on both general and specific domain text classification datasets. Results show that the proposed framework can outperform the existing competitive state-of-the-art baselines, delivering new benchmark results.

Original languageEnglish
Article number128580
JournalNeurocomputing
Volume610
DOIs
Publication statusPublished - 28 Dec 2024

Keywords

  • Knowledge graph embedding
  • Multi-class classification
  • Natural language processing
  • Textual analysis
  • Zero-shot learning

Fingerprint

Dive into the research topics of 'Zero-shot text classification with knowledge resources under label-fully-unseen setting'. Together they form a unique fingerprint.

Cite this