Transfer Learning Using Rotated Image Data to Improve Deep Neural Network Performance
In this work we explore the idea that, in the presence of a small training set of images, it could be beneficial to use that set itself to obtain a transformed training set (by performing a random rotation on each sample), train a source network using the transformed data, then retrain the source network using the original data. Applying this transfer learning technique to three different types of character data, we achieve average relative improvements between 6 % and 16 % in the classification test error. Furthermore, we show that it is possible to achieve relative improvements between 8 % and 42 % in cases where the amount of original training samples is very limited (30 samples per class), by introducing not just one rotation but several random rotations per sample.
KeywordsTransfer learning Deep learning Stacked auto-encoders
Unable to display preview. Download preview PDF.
- Ciresan, D., Meier, U., Schmidhuber, J.: Transfer learning for Latin and Chinese characters with deep neural networks. In: International Joint Conference on Neural Networks (IJCNN), pp. 1–6 (2012)Google Scholar
- Deng, L., Yu, D.: Deep learning for signal and information processing. Microsoft Research Monograph (2013)Google Scholar
- Glorot, X., Bordes, A., Bengio, Y.: Domain adaptation for large-scale sentiment classification: a deep learning approach. In: International Conference on Machine Learning (ICML), pp. 513–520 (2011)Google Scholar
- Larochelle, H., Erhan, D., Courville, A., Bergstra, J., Bengio, Y.: An empirical evaluation of deep architectures on problems with many factors of variation. In: International Conference on Machine Learning (ICML), pp. 473–480 (2007)Google Scholar
- Simard, P., Steinkraus, D., Platt, J.C.: Best practices for convolutional neural networks applied to visual document analysis. In: International Conference on Document Analysis and Recognition (ICDAR), vol. 3, pp. 958–962 (2003)Google Scholar