TY - GEN
T1 - Improving deep neural network performance with kernelized min-max objective
AU - Yao, Kai
AU - Huang, Kaizhu
AU - Zhang, Rui
AU - Hussain, Amir
N1 - Publisher Copyright:
© 2018, Springer Nature Switzerland AG.
PY - 2018
Y1 - 2018
N2 - In this paper, we present a novel training strategy using kernelized Min-Max objective to enable improved object recognition performance on deep neural networks (DNN), e.g., convolutional neural networks (CNN). Without changing the other part of the original model, the kernelized Min-Max objective works by combining the kernel trick with the Min-Max objective and being embedded into a high layer of the networks in the training phase. The proposed kernelized objective explicitly enforces the learned object feature maps to maintain in a kernel space the least compactness for each category manifold and the biggest margin among different category manifolds. With very few additional computation costs, the proposed strategy can be widely used in different DNN models. Extensive experiments with shallow convolutional neural network model, deep convolutional neural network model, and deep residual neural network model on two benchmark datasets show that the proposed approach outperforms those competitive models.
AB - In this paper, we present a novel training strategy using kernelized Min-Max objective to enable improved object recognition performance on deep neural networks (DNN), e.g., convolutional neural networks (CNN). Without changing the other part of the original model, the kernelized Min-Max objective works by combining the kernel trick with the Min-Max objective and being embedded into a high layer of the networks in the training phase. The proposed kernelized objective explicitly enforces the learned object feature maps to maintain in a kernel space the least compactness for each category manifold and the biggest margin among different category manifolds. With very few additional computation costs, the proposed strategy can be widely used in different DNN models. Extensive experiments with shallow convolutional neural network model, deep convolutional neural network model, and deep residual neural network model on two benchmark datasets show that the proposed approach outperforms those competitive models.
UR - http://www.scopus.com/inward/record.url?scp=85059066513&partnerID=8YFLogxK
U2 - 10.1007/978-3-030-04167-0_17
DO - 10.1007/978-3-030-04167-0_17
M3 - Conference Proceeding
AN - SCOPUS:85059066513
SN - 9783030041663
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 182
EP - 191
BT - Neural Information Processing - 25th International Conference, ICONIP 2018, Proceedings
A2 - Cheng, Long
A2 - Leung, Andrew Chi Sing
A2 - Ozawa, Seiichi
PB - Springer Verlag
T2 - 25th International Conference on Neural Information Processing, ICONIP 2018
Y2 - 13 December 2018 through 16 December 2018
ER -