Abstract
A very simple and efficient method for artificial neural networks training is proposed. Extensive simulation has established that it works very well as an ANN training law. It is faster than BackPropagation, it is suitable for on-chip learning and it can be implemented in parallel computers very easily.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Jacobs, R.A. (1988). Increased rates of Convergence through Learning Rate Adaptation. Neural Networks, vol.1. pp 295–307
Wasserman P.D. (1988). Experiments in Translating Chinese Characters Using Back-Propagation. Proc.of 33 IEEE Comp. Soc.IntConf., IEEE Comp. Society Press, pp 399–402.
Tollenaere T. (1990). Super SAB: Fast Adaptive Backpropagation with Good Scaling Properties. Neural Networks, vol.3, pp.561–573.
Samad T. (1991). Back Propagation With Expected Source Values. Neural Networks, vol.4, pp.615–618.
Beveridge G.S.G. and Schechter R.S. (1970). Optimization: Theory and Practice. McGraw-Hill, ISE.
Robert Hecht-Nielsen The munificence of high dimensionality. Artificial Neural Networks, 2 (I. Aleksander & J. Taylor Eds) pp. 1017–1030 North Holland, 1992.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 1993 Springer-Verlag London Limited
About this paper
Cite this paper
Petridis, V., Paraschidis, K. (1993). A Simple Training Law Suitable for On-Chip Learning.. In: Gielen, S., Kappen, B. (eds) ICANN ’93. ICANN 1993. Springer, London. https://doi.org/10.1007/978-1-4471-2063-6_315
Download citation
DOI: https://doi.org/10.1007/978-1-4471-2063-6_315
Published:
Publisher Name: Springer, London
Print ISBN: 978-3-540-19839-0
Online ISBN: 978-1-4471-2063-6
eBook Packages: Springer Book Archive