TY - JOUR
T1 - Zero-shot text classification with knowledge resources under label-fully-unseen setting
AU - Wang, Yuqi
AU - Wang, Wei
AU - Chen, Qi
AU - Huang, Kaizhu
AU - Nguyen, Anh
AU - De, Suparna
N1 - Publisher Copyright:
© 2024 The Authors
PY - 2024/12/28
Y1 - 2024/12/28
N2 - Classification techniques are at the heart of many real-world applications, e.g. sentiment analysis, recommender systems and automatic text annotation, to process and analyse large-scale textual data in multiple fields. However, the effectiveness of natural language processing models can only be confirmed when a large amount of up-to-date training data is available. An unprecedented amount of data is continuously created, and new topics are introduced, making it less likely or even infeasible to collect labelled samples covering all topics for training models. We attempt to study the extreme case: there is no labelled data for model training, and the model, without being adapted to any specific dataset, will be directly applied to the testing samples. We propose a transformer-based framework to encode sentences in a contextualised way and leverage the existing knowledge resources, i.e. ConceptNet and WordNet, to integrate both descriptive and structural knowledge for better performance. To enhance the robustness of the model, we design an adversarial example generator based on relations from external knowledge bases. The framework is evaluated on both general and specific domain text classification datasets. Results show that the proposed framework can outperform the existing competitive state-of-the-art baselines, delivering new benchmark results.
AB - Classification techniques are at the heart of many real-world applications, e.g. sentiment analysis, recommender systems and automatic text annotation, to process and analyse large-scale textual data in multiple fields. However, the effectiveness of natural language processing models can only be confirmed when a large amount of up-to-date training data is available. An unprecedented amount of data is continuously created, and new topics are introduced, making it less likely or even infeasible to collect labelled samples covering all topics for training models. We attempt to study the extreme case: there is no labelled data for model training, and the model, without being adapted to any specific dataset, will be directly applied to the testing samples. We propose a transformer-based framework to encode sentences in a contextualised way and leverage the existing knowledge resources, i.e. ConceptNet and WordNet, to integrate both descriptive and structural knowledge for better performance. To enhance the robustness of the model, we design an adversarial example generator based on relations from external knowledge bases. The framework is evaluated on both general and specific domain text classification datasets. Results show that the proposed framework can outperform the existing competitive state-of-the-art baselines, delivering new benchmark results.
KW - Knowledge graph embedding
KW - Multi-class classification
KW - Natural language processing
KW - Textual analysis
KW - Zero-shot learning
UR - http://www.scopus.com/inward/record.url?scp=85203873970&partnerID=8YFLogxK
U2 - 10.1016/j.neucom.2024.128580
DO - 10.1016/j.neucom.2024.128580
M3 - Article
AN - SCOPUS:85203873970
SN - 0925-2312
VL - 610
JO - Neurocomputing
JF - Neurocomputing
M1 - 128580
ER -