Abstract
The present paper aims to propose a new type of learning method for interpreting relations between inputs and outputs in multi-layered neural networks. The method is composed of information augmentation, reduction and compression component. In the information augmentation component, information in inputs is forced to increase for the subsequent learning to choose appropriate information among many options. In the information reduction component, information is reduced by selectively choosing strong and active connection weights. Finally, in the information compression component, information contained in multi-layered neural networks is compressed by multiplying all connection weights in all layers for summarizing the main characteristics of connection weights. The method was applied to the improvement of an EC (electric commerce) web site for better profitability. The method could clarify relations between inputs and outputs and its interpretation was more natural than that by the conventional logistic regression analysis. The results suggest that multi-layered neural networks can be used to improve generalization and in addition to interpret final results, which is more important in many applications fields.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Guyon, I., Elisseeff, A.: An introduction to variable and feature selection. J. Mach. Learn. Res. 3, 1157–1182 (2003)
Rakotomamonjy, A.: Variable selection using SVM-based criteria. J. Mach. Learn. Res. 3, 1357–1370 (2003)
Perkins, S., Lacker, K., Theiler, J.: Grafting: fast, incremental feature selection by gradient descent in function space. J. Mach. Learn. Res. 3, 1333–1356 (2003)
Wang, J., Perez, L.: The Effectiveness of Data Augmentation in Image Classification Using Deep Learning. Technical report (2017)
Asperti, A., Mastronardo, C.: The effectiveness of data augmentation for detection of gastrointestinal diseases from endoscopical images. arXiv:1712.03689 (2017)
Xu, Y., Jia, R.., Mou, L., Li, G., Chen, Y., Lu, Y., Jin, Z.: Improved relation classification by deep recurrent neural networks with data augmentation. arXiv:1601.03651 (2016)
Marchesi, M.: Megapixel Size Image Creation Using Generative Adversarial Networks (2017)
Kingma, D.P., Welling, M.: Auto-Encoding Variational Bayes (2013)
Allen, D.M.: The relationship between variable selection and data agumentation and a method for prediction. Technometrics 16(1), 125–127 (1974)
Moody, J., Hanson, S., Krogh, A., Hertz, J.A.: A simple weight decay can improve generalization. Adv. Neural Inf. Process. Syst. 4, 950–957 (1995)
Hinton, G.E.: A practical guide to training restricted boltzmann machines. In: Neural Networks: Tricks of the Trade. Springer, Berlin, pp. 599–619 (2012)
Nishiuchi, H.: Statistical Analysis for Billion People (in Japanese). Nikkei BP marketing, Tokyo, Japan (2014)
Breiman, L.: Bagging predictors. Mach. Learn. 24(2), 123–140 (1996)
Breiman, L.: Random forests. Mach. Learn. 45(1), 5–32 (2001)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Kamimura, R. (2020). Information Augmentation, Reduction and Compression for Interpreting Multi-layered Neural Networks. In: Arai, K., Bhatia, R. (eds) Advances in Information and Communication. FICC 2019. Lecture Notes in Networks and Systems, vol 70. Springer, Cham. https://doi.org/10.1007/978-3-030-12385-7_17
Download citation
DOI: https://doi.org/10.1007/978-3-030-12385-7_17
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-12384-0
Online ISBN: 978-3-030-12385-7
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)