TY - JOUR
T1 - Instance-Specific Model Perturbation Improves Generalized Zero-Shot Learning
AU - Yang, Guanyu
AU - Huang, Kaizhu
AU - Zhang, Rui
AU - Yang, Xi
N1 - Publisher Copyright:
© 2024 Massachusetts Institute of Technology.
PY - 2024/4/23
Y1 - 2024/4/23
N2 - Zero-shot learning (ZSL) refers to the design of predictive functions on new classes (unseen classes) of data that have never been seen during training. In a more practical scenario, generalized zero-shot learning (GZSL) requires predicting both seen and unseen classes accurately. In the absence of target samples, many GZSL models may overfit training data and are inclined to predict individuals as categories that have been seen in training. To alleviate this problem, we develop a parameter-wise adversarial training process that promotes robust recognition of seen classes while designing during the test a novel model perturbation mech-anism to ensure sufficient sensitivity to unseen classes. Concretely, adversarial perturbation is conducted on the model to obtain instance-specific parameters so that predictions can be biased to unseen classes in the test. Meanwhile, the robust training encourages the model robustness, leading to nearly unaffected prediction for seen classes. Moreover, perturbations in the parameter space, computed from multiple individuals simultaneously, can be used to avoid the effect of perturbations that are too extreme and ruin the predictions. Comparison results on four bench-mark ZSL data sets show the effective improvement that the proposed framework made on zero-shot methods with learned metrics.
AB - Zero-shot learning (ZSL) refers to the design of predictive functions on new classes (unseen classes) of data that have never been seen during training. In a more practical scenario, generalized zero-shot learning (GZSL) requires predicting both seen and unseen classes accurately. In the absence of target samples, many GZSL models may overfit training data and are inclined to predict individuals as categories that have been seen in training. To alleviate this problem, we develop a parameter-wise adversarial training process that promotes robust recognition of seen classes while designing during the test a novel model perturbation mech-anism to ensure sufficient sensitivity to unseen classes. Concretely, adversarial perturbation is conducted on the model to obtain instance-specific parameters so that predictions can be biased to unseen classes in the test. Meanwhile, the robust training encourages the model robustness, leading to nearly unaffected prediction for seen classes. Moreover, perturbations in the parameter space, computed from multiple individuals simultaneously, can be used to avoid the effect of perturbations that are too extreme and ruin the predictions. Comparison results on four bench-mark ZSL data sets show the effective improvement that the proposed framework made on zero-shot methods with learned metrics.
UR - http://www.scopus.com/inward/record.url?scp=85191615698&partnerID=8YFLogxK
U2 - 10.1162/neco_a_01639
DO - 10.1162/neco_a_01639
M3 - Article
C2 - 38457762
AN - SCOPUS:85191615698
SN - 0899-7667
VL - 36
SP - 936
EP - 962
JO - Neural Computation
JF - Neural Computation
IS - 5
ER -