Abstract
Auto-encoder as the representative method for data dimensionality reduction and feature extraction, plays a very important role on machine learning. However, the data in the actual research work or industrial production are not always normalized data, at this time, it will lead high reconstruction error and slow convergence speed. This study proposed an improved auto-encoder and a denoising auto-encoder based on 2-level prioritized experience replay, which can improve accuracy and reduce loss, while processing a dimensionality reduction or feature extraction problem on high dimension skewed data. In order to evaluate the effectiveness of the proposed method, three models of high dimension simulation dataset which on different skewed degrees are generated. The results of evaluation experiments show that the proposed method can get lower reconstruction error than conventional method for high dimension skewed simulation data.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Bishop, C.M.: Pattern Recognition and Machine Learning. Springer, Heidelberg (2006)
LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521(7553), 436 (2015)
Van Der Maaten, L., Postma, E., Van den Herik, J.: Dimensionality reduction: a comparative. J. Mach. Learn. Res. 10(66–71), 13 (2009)
Guyon, I., Gunn, S., Nikravesh, M., Zadeh, L.A.: Feature Extraction: Foundations and Applications, vol. 207. Springer, Heidelberg (2008)
Jolliffe, I.: Principal Component Analysis. Springer, Heidelberg (2011)
Hinton, G.E., Salakhutdinov, R.R.: Reducing the dimensionality of data with neural networks. Science 313(5786), 504–507 (2006)
Zurada, J.M.: Introduction to Artificial Neural Systems, vol. 8. West Publishing Company, St. Paul (1992)
Vincent, P., Larochelle, H., Bengio, Y., Manzagol, P.-A.: Extracting and composing robust features with denoising autoencoders. In: Proceedings of the 25th International Conference on Machine Learning, pp. 1096–1103. ACM (2008)
Zhang, B., Ma, K., Ji, Z.: The optimization of the hull form with the minimum wave making resistance based on rankine source method. J. Hydrodyn. 21(2), 277–284 (2009)
Surendran, S., Venkata Ramana Reddy, J.: Numerical simulation of ship stability for dynamic environment. Ocean Eng. 30(10), 1305–1317 (2003)
Machiraju, R., Yagel, R.: Reconstruction error characterization and control: a sampling theory approach. IEEE Trans. Visual Comput. Graph. 2(4), 364–378 (1996)
Barlow, H.B.: Unsupervised learning. Neural Comput. 1(3), 295–311 (1989)
Kingma, D.P., Welling, M.: Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114 (2013)
Higgins, I., Matthey, L., Pal, A., Burgess, C., Glorot, X., Botvinick, M., Mohamed, S., Lerchner, A.: \({\upbeta } \)-VAE: larning basic visual concepts with a constrained variational framework. ICLR 2(5), 6 (2017)
Schaul, T., Quan, J., Antonoglou, I., Silver, D.: Prioritized experience replay. arXiv preprint arXiv:1511.05952 (2015)
Szegedy, M.: The DLT priority sampling is essentially optimal. In: Proceedings of the Thirty-Eighth Annual ACM Symposium on Theory of Computing, pp. 150–158. ACM (2006)
Chollet, F., et al.: Keras: the python deep learning library. Astrophysics Source Code Library (2018)
Rasmussen, C.E.: Gaussian processes in machine learning. In: Summer School on Machine Learning, pp. 63–71. Springer (2003)
Kohavi, R., et al.: A study of cross-validation and bootstrap for accuracy estimation and model selection. In: IJCAI, Montreal, Canada, vol. 14, pp. 1137–1145 (1995)
Schmidhuber, J.: Deep learning in neural networks: an overview. Neural Netw. 61, 85–117 (2015)
Acknowledgment
Thanks to my professor Tomoki Hamagami, for helping me a lot in this research. Thanks to the China Scholarship Council (CSC) in 2016, pursue my study in Japan as a Ph.D. student under File no. 201608140085.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Li, X., Hamagami, T. (2020). An Improved Auto-encoder Based on 2-Level Prioritized Experience Replay for High Dimension Skewed Data. In: Sato, H., Iwanaga, S., Ishii, A. (eds) Proceedings of the 23rd Asia Pacific Symposium on Intelligent and Evolutionary Systems. IES 2019. Proceedings in Adaptation, Learning and Optimization, vol 12. Springer, Cham. https://doi.org/10.1007/978-3-030-37442-6_9
Download citation
DOI: https://doi.org/10.1007/978-3-030-37442-6_9
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-37441-9
Online ISBN: 978-3-030-37442-6
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)