TY - JOUR
T1 - Two-stage model re-optimization and application in face recognition
AU - Qian, Jianyu
AU - Mu, Shiyi
AU - Lu, Hengjie
AU - Xu, Shugong
N1 - Publisher Copyright:
© 2025
PY - 2025/10/28
Y1 - 2025/10/28
N2 - Face recognition technology has achieved mature performance in various fields and has been widely applied on edge devices with limited computational resources. However, in practical application scenarios, the deployment platforms are often predetermined, and the final deployed models are the result of trade-offs between the computational and storage capabilities of the platforms. These models are difficult to adjust once determined. Therefore, how to improve the performance of existing face recognition models without affecting the entire deployment process remains a topic of practical significance and research value. Additionally, with the rapid development of deep learning, there exists a vast array of high-performance open-source pre-trained models on the Internet. How to utilize these models efficiently is also a subject worthy of investigation. To address the aforementioned issues, this paper introduces a novel training framework called two-stage model re-optimization (TSMR). TSMR enhances face recognition performance by leveraging knowledge distillation, model re-parameterization and adversarial learning techniques. In particular, TSMR improves performance without increasing inference latency, modifying network architecture, or introducing additional computational cost. It is a versatile and effective framework that can be applied to any CNN-based recognition network. The experimental results show the effectiveness of TSMR in improving the performance of lightweight face recognition models. Across multiple datasets, including CALFW, CPLFW, YTF and AgeDB-30, TSMR achieves an average accuracy improvement ranging from 1% to 2%. Notably, when TSMR is applied with FDFaceNet as the baseline, it achieves an impressive accuracy of 95.58% on the LFW dataset, surpassing the state-of-the-art performance in lightweight face recognition networks.
AB - Face recognition technology has achieved mature performance in various fields and has been widely applied on edge devices with limited computational resources. However, in practical application scenarios, the deployment platforms are often predetermined, and the final deployed models are the result of trade-offs between the computational and storage capabilities of the platforms. These models are difficult to adjust once determined. Therefore, how to improve the performance of existing face recognition models without affecting the entire deployment process remains a topic of practical significance and research value. Additionally, with the rapid development of deep learning, there exists a vast array of high-performance open-source pre-trained models on the Internet. How to utilize these models efficiently is also a subject worthy of investigation. To address the aforementioned issues, this paper introduces a novel training framework called two-stage model re-optimization (TSMR). TSMR enhances face recognition performance by leveraging knowledge distillation, model re-parameterization and adversarial learning techniques. In particular, TSMR improves performance without increasing inference latency, modifying network architecture, or introducing additional computational cost. It is a versatile and effective framework that can be applied to any CNN-based recognition network. The experimental results show the effectiveness of TSMR in improving the performance of lightweight face recognition models. Across multiple datasets, including CALFW, CPLFW, YTF and AgeDB-30, TSMR achieves an average accuracy improvement ranging from 1% to 2%. Notably, when TSMR is applied with FDFaceNet as the baseline, it achieves an impressive accuracy of 95.58% on the LFW dataset, surpassing the state-of-the-art performance in lightweight face recognition networks.
KW - Face recognition
KW - Knowledge distillation
KW - Lightweight
KW - Re-parameterization
UR - https://www.scopus.com/pages/publications/105010845405
UR - http://scholar.xjtlu.edu.cn/en/publications/multi-dataset-fusion-for-multi-task-learning-on-face-attribute-re/
UR - https://scholar.xjtlu.edu.cn/en/publications/%E5%9F%BA%E4%BA%8E%E5%BF%AB%E9%80%9F%E4%B8%8B%E9%87%87%E6%A0%B7%E7%9A%84%E8%BD%BB%E9%87%8F%E5%8C%96%E7%BD%91%E7%BB%9C%E8%AE%BE%E8%AE%A1%E6%96%B9%E6%B3%95%E5%8F%8A%E4%BA%BA%E8%84%B8%E8%AF%86%E5%88%AB%E5%BA%94%E7%94%A8/
U2 - 10.1016/j.neucom.2025.130805
DO - 10.1016/j.neucom.2025.130805
M3 - Article
AN - SCOPUS:105010845405
SN - 0925-2312
VL - 651
JO - Neurocomputing
JF - Neurocomputing
M1 - 130805
ER -