TY - GEN
T1 - Generalised Zero-shot Learning for Entailment-based Text Classification with External Knowledge
AU - Wang, Yuqi
AU - Wang, Wei
AU - Chen, Qi
AU - Huang, Kaizhu
AU - Nguyen, Anh
AU - De, Suparna
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - Text classification techniques have been substantially important to many smart computing applications, e.g. topic extraction and event detection. However, classification is always challenging when only insufficient amount of labelled data for model training is available. To mitigate this issue, zero-shot learning (ZSL) has been introduced for models to recognise new classes that have not been observed during the training stage. We propose an entailment-based zero-shot text classification model, named as S-BERT-CAM, to better capture the relationship between the premise and hypothesis in the BERT embedding space. Two widely used textual datasets are utilised to conduct the experiments. We fine-tune our model using 50% of the labels for each dataset and evaluate it on the label space containing all labels (including both seen and unseen labels). The experimental results demonstrate that our model is more robust to the generalised ZSL and significantly improves the overall performance against baselines.
AB - Text classification techniques have been substantially important to many smart computing applications, e.g. topic extraction and event detection. However, classification is always challenging when only insufficient amount of labelled data for model training is available. To mitigate this issue, zero-shot learning (ZSL) has been introduced for models to recognise new classes that have not been observed during the training stage. We propose an entailment-based zero-shot text classification model, named as S-BERT-CAM, to better capture the relationship between the premise and hypothesis in the BERT embedding space. Two widely used textual datasets are utilised to conduct the experiments. We fine-tune our model using 50% of the labels for each dataset and evaluate it on the label space containing all labels (including both seen and unseen labels). The experimental results demonstrate that our model is more robust to the generalised ZSL and significantly improves the overall performance against baselines.
KW - BERT
KW - Zero-shot learning
KW - deep learning
KW - natural language processing
UR - http://www.scopus.com/inward/record.url?scp=85136157664&partnerID=8YFLogxK
U2 - 10.1109/SMARTCOMP55677.2022.00018
DO - 10.1109/SMARTCOMP55677.2022.00018
M3 - Conference Proceeding
AN - SCOPUS:85136157664
T3 - Proceedings - 2022 IEEE International Conference on Smart Computing, SMARTCOMP 2022
SP - 19
EP - 25
BT - Proceedings - 2022 IEEE International Conference on Smart Computing, SMARTCOMP 2022
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 8th IEEE International Conference on Smart Computing, SMARTCOMP 2022
Y2 - 20 June 2022 through 24 June 2022
ER -