Abstract
Over last fifty years, Neural Networks (NN) have been important and active models in machine learning and pattern recognition. Among different types of NNs, Back Propagation (BP) NN is one popular model, widely exploited in various applications. Recently, NNs attract even more attention in the community because a deep learning structure (if appropriately adopted) could significantly improve the learning performance. In this paper, based on a probabilistic assumption over the output neurons, we propose a hybrid strategy that manages to combine one typical deep NN, i.e., Convolutional NN (CNN) with the popular BP. We present the justification and describe the detailed learning formulations. A series of experiments validate that the hybrid approach could largely improve the accuracy for both CNN and BP on two large-scale benchmark data sets, i.e., MNIST and USPS. In particular, the proposed hybrid method significantly reduced the error rates of CNN and BP respectively by 11.72% and 28.89% on MNIST.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Bishop, C.M.: Neural Networks for Pattern Recognition. Oxford University Press, London (1995)
Ciresan, D.C., Meier, U., Masci, J., Gambardella, L.M., Schmidhuber, J.: High-performance neural networks for visual object classification. In: Proceedings of Internet Joint Conference on Artificial Intelligence, IJCAI (2011)
Hinton, G.E., Osindero, S., Teh, Y.: A fast learning algorithm for deep belief nets. Neural Computation 18, 1527–1554 (2006)
Hinton, G.E., Salakhutdinov, R.R.: Reducing the dimensionality of data with neural networks. Science 33(5786), 504–507 (2006)
Kohonen, T., Honkela, T.: Kohonen Network. Scholarpedia (2007)
Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: Advances in Neural Information Processing Systems (NIPS) 25 (2012)
LeCun, Y., Bottou, L., Bengio, Y., Hakner, P.: Gradient-based learning applied to document recognition. Proceedings of the IEEE 86(11), 2278–2324 (1998)
Storkey, A.J., Valabregue, R.: The basins of attraction of a new hopfield learning rule. Neural Networks 12(6), 869–876 (1999)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer International Publishing Switzerland
About this paper
Cite this paper
Zhang, R., Zhang, S., Huang, K. (2014). A Novel Hybrid Approach for Combining Deep and Traditional Neural Networks. In: Loo, C.K., Yap, K.S., Wong, K.W., Beng Jin, A.T., Huang, K. (eds) Neural Information Processing. ICONIP 2014. Lecture Notes in Computer Science, vol 8836. Springer, Cham. https://doi.org/10.1007/978-3-319-12643-2_43
Download citation
DOI: https://doi.org/10.1007/978-3-319-12643-2_43
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-12642-5
Online ISBN: 978-3-319-12643-2
eBook Packages: Computer ScienceComputer Science (R0)