TY - GEN
T1 - Empathizing Before Generation
T2 - 7th Chinese Conference on Pattern Recognition and Computer Vision, PRCV 2024
AU - Zhu, Jiahao
AU - Jiang, Zijian
AU - Zhou, Boyu
AU - Su, Jionglong
AU - Zhang, Jiaming
AU - Li, Zhihao
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2025.
PY - 2025
Y1 - 2025
N2 - Large Language Models (LLMs) have found extensive use across different applications due to its diverse capabilities and proficiency in executing instructions. In the case of chatbots, they are frequently required to show empathy when used in the context of emotional support. However, to date their performance is still not satisfactory due to the lack of deep understanding of user related issues. Hence, we introduce the Empathizing Before Generation (EBG), a two-step learning framework that allows LLMs to analyze the chain of thought (COT) prior to generating a response. This model also enables the inference of the 24 emotions conveyed in the user’s text as well as facilitates the generation of empathetic, high-quality and appropriate responses. We create a COT version of the dataset for sentiment inference by utilizing a publicly accessible sentiment dialogue. This dataset is then used as support for the training of two layers of EBG. Experiments indicate that models integrated with the EBG outperform other models in demonstrating empathy, with 98.2% and 92.8% accuracy in emotional attributes and labels respectively. Additionally, there is a notable enhancement in the model’s capacity to comprehend COT instructions, infer emotions, and generate answers that are more satisfactory than other models.
AB - Large Language Models (LLMs) have found extensive use across different applications due to its diverse capabilities and proficiency in executing instructions. In the case of chatbots, they are frequently required to show empathy when used in the context of emotional support. However, to date their performance is still not satisfactory due to the lack of deep understanding of user related issues. Hence, we introduce the Empathizing Before Generation (EBG), a two-step learning framework that allows LLMs to analyze the chain of thought (COT) prior to generating a response. This model also enables the inference of the 24 emotions conveyed in the user’s text as well as facilitates the generation of empathetic, high-quality and appropriate responses. We create a COT version of the dataset for sentiment inference by utilizing a publicly accessible sentiment dialogue. This dataset is then used as support for the training of two layers of EBG. Experiments indicate that models integrated with the EBG outperform other models in demonstrating empathy, with 98.2% and 92.8% accuracy in emotional attributes and labels respectively. Additionally, there is a notable enhancement in the model’s capacity to comprehend COT instructions, infer emotions, and generate answers that are more satisfactory than other models.
KW - Chain of thought
KW - Emotional support
KW - Large language model
UR - http://www.scopus.com/inward/record.url?scp=85209539740&partnerID=8YFLogxK
U2 - 10.1007/978-981-97-8490-5_35
DO - 10.1007/978-981-97-8490-5_35
M3 - Conference Proceeding
AN - SCOPUS:85209539740
SN - 9789819784899
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 490
EP - 503
BT - Pattern Recognition and Computer Vision - 7th Chinese Conference, PRCV 2024, Proceedings
A2 - Lin, Zhouchen
A2 - Zha, Hongbin
A2 - Cheng, Ming-Ming
A2 - He, Ran
A2 - Liu, Cheng-Lin
A2 - Ubul, Kurban
A2 - Silamu, Wushouer
A2 - Zhou, Jie
PB - Springer Science and Business Media Deutschland GmbH
Y2 - 18 October 2024 through 20 October 2024
ER -