TY - JOUR
T1 - Learning from Few Samples with Memory Network
AU - Zhang, Shufei
AU - Huang, Kaizhu
AU - Zhang, Rui
AU - Hussain, Amir
N1 - Publisher Copyright:
© 2017, Springer Science+Business Media, LLC.
PY - 2018/2/1
Y1 - 2018/2/1
N2 - Neural networks (NN) have achieved great successes in pattern recognition and machine learning. However, the success of a NN usually relies on the provision of a sufficiently large number of data samples as training data. When fed with a limited data set, a NN’s performance may be degraded significantly. In this paper, a novel NN structure is proposed called a memory network. It is inspired by the cognitive mechanism of human beings, which can learn effectively, even from limited data. Taking advantage of the memory from previous samples, the new model achieves a remarkable improvement in performance when trained using limited data. The memory network is demonstrated here using the multi-layer perceptron (MLP) as a base model. However, it would be straightforward to extend the idea to other neural networks, e.g., convolutional neural networks (CNN). In this paper, the memory network structure is detailed, the training algorithm is presented, and a series of experiments are conducted to validate the proposed framework. Experimental results show that the proposed model outperforms traditional MLP-based models as well as other competitive algorithms in response to two real benchmark data sets.
AB - Neural networks (NN) have achieved great successes in pattern recognition and machine learning. However, the success of a NN usually relies on the provision of a sufficiently large number of data samples as training data. When fed with a limited data set, a NN’s performance may be degraded significantly. In this paper, a novel NN structure is proposed called a memory network. It is inspired by the cognitive mechanism of human beings, which can learn effectively, even from limited data. Taking advantage of the memory from previous samples, the new model achieves a remarkable improvement in performance when trained using limited data. The memory network is demonstrated here using the multi-layer perceptron (MLP) as a base model. However, it would be straightforward to extend the idea to other neural networks, e.g., convolutional neural networks (CNN). In this paper, the memory network structure is detailed, the training algorithm is presented, and a series of experiments are conducted to validate the proposed framework. Experimental results show that the proposed model outperforms traditional MLP-based models as well as other competitive algorithms in response to two real benchmark data sets.
KW - Memory
KW - Multi-layer perceptron
KW - Neural network
KW - Prior knowledge
KW - Recognition
UR - http://www.scopus.com/inward/record.url?scp=85032029937&partnerID=8YFLogxK
U2 - 10.1007/s12559-017-9507-z
DO - 10.1007/s12559-017-9507-z
M3 - Article
AN - SCOPUS:85032029937
SN - 1866-9956
VL - 10
SP - 15
EP - 22
JO - Cognitive Computation
JF - Cognitive Computation
IS - 1
ER -