Abstract
A generated solution for an artificial neural network (ANN) may result in complex computations of neural networks, deployment, and usage of trained networks due to its inappropriate architecture. Therefore, modeling the hidden layer architecture of artificial neural networks remains as a research challenge. This paper presents a solution to achieve the hidden layer architecture of artificial neural networks which is inspired by some facts of neuroplasticity. The proposed method has two phases. First, it determines the number of hidden layers for the best architecture and then removes unnecessary hidden neurons from the network to enhance the performance. Experimental results in several benchmark problems show that the modified network shows better generalization than the original network.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Liu, Y., Starzyk, J.A., Zhu, Z.: Optimizing number of hidden neurons in neural networks. EeC 1(1), 6 (2007)
Amini, J.: Optimum learning rate in back-propagation neural network for classification of satellite images (IRS-1D). Sci. Iran. 15(6), 558–567 (2008)
Castellano, G., Fanelli, A.M., Pelillo, M.: An iterative pruning algorithm for feedforward neural networks. IEEE Trans. Neural Netw. 8(3), 519–531 (1997)
Lippmann, R.: An introduction to computing with neural nets. IEEE ASSP Mag. 4(2), 4–22 (1987)
LeCun, Y.: Generalization and network design strategies. In: Connectionism in Perspective, vol. 19, pp. 143–155. Elsevier, Amsterdam (1989)
Herculano-Houzel, S.: The remarkable, yet not extraordinary, human brain as a scaled-up primate brain and its associated cost. Proc. Natl. Acad. Sci. 109(Suppl. 1), 10661–10668 (2012)
Sousa, A.M., Meyer, K.A., Santpere, G., Gulden, F.O., Sestan, N.: Evolution of the human nervous system function, structure, and development. Cell 170(2), 226–247 (2017)
Wagarachchi, N.M., Karunananda, A.S.: Optimization of multi-layer artificial neural networks using delta values of hidden layers. In: IEEE Symposium on Computational Intelligence, Cognitive Algorithms, Mind, and Brain, pp. 80–86 (2013)
Vollmer, L.: Change Your Mind: Neuroplasticity & Buddhist Transformation. All Theses and Dissertations (ETDs), 928 (2010). http://openscholarship.wustl.edu/etd/928
Pascual-Leone, A., Amedi, A., Fregni, F., Merabet, L.B.: The plastic human brain cortex. Annu. Rev. Neurosci. 28, 377–401 (2005)
Eggermont, J.J.: The Correlative Brain: Theory and Experiment in Neural Interaction. Springer, New York (1990)
Chrol-Cannon, J., Jin, Y.: Computational modeling of neural plasticity for self-organization of neural networks. BioSystems 125, 43–54 (2014)
Wells, R.B.: Cortical neurons and circuits: a tutorial introduction. Unpublished Paper (2005) www Mrc Uidaho Edu
Mountcastle, V.B.: The columnar organization of the neocortex. Brain 120(4), 701–722 (1997)
Thomson, A.M.: Neocortical layer 6, a review. Front. Neuroanat. 4, 13 (2010)
Shaffer, J.: Neuroplasticity and positive psychology in clinical practice: a review for combined benefits. Psychology 3(12), 1110 (2012)
Children with Autism Have Extra Synapses in Brain. Columbia University Medical Center. http://newsroom.cumc.columbia.edu/blog/2014/08/21/children-autism-extra-synapses-brain/. Accessed: 10 Aug 2017
LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521(7553), 436–444 (2015)
Hinton, G.E., Osindero, S., Teh, Y.-W.: A fast learning algorithm for deep belief nets. Neural Comput. 18(7), 1527–1554 (2006)
Wagarachchi, M., Karunananda, A.: Optimization of artificial neural network architecture using neuroplasticity. Int. J. Artif. Intell. 15(1), 112–125 (2017)
Wagarachchi, N.M., Karunananda, A.S.: Novel technique for optimizing the hidden layer architecture in artificial neural networks. Am. Int. J. Res. Sci. Technol. Eng. Math. 4(1), 1–6 (2013)
Wagarachchi, N.M., Karunananda, A.S.: Towards a theoretical basis for modelling of hidden layer architecture in artificial neural networks. In: 2nd International Conference on Advances in Computing, Electronics and Communication, pp. 47–52 (2014)
Prechelt, L.: PROBEN 1: a set of benchmarks and benchmarking rules for neural network training algorithms. Univ., Fak. für Informatik (1994)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Wagarachchi, M., Karunananda, A. (2019). Modeling of Hidden Layer Architecture in Multilayer Artificial Neural Networks. In: Hemanth, J., Silva, T., Karunananda, A. (eds) Artificial Intelligence. SLAAI-ICAI 2018. Communications in Computer and Information Science, vol 890. Springer, Singapore. https://doi.org/10.1007/978-981-13-9129-3_5
Download citation
DOI: https://doi.org/10.1007/978-981-13-9129-3_5
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-13-9128-6
Online ISBN: 978-981-13-9129-3
eBook Packages: Computer ScienceComputer Science (R0)