Log in

An Enhanced Training Algorithm for Multilayer Neural Networks Based on Reference Output of Hidden Layer

  • Original Article
  • Published:
Neural Computing & Applications Aims and scope Submit manuscript

Abstract

In this paper, the authors propose a new training algorithm which does not only rely upon the training samples, but also depends upon the output of the hidden layer. We adjust both the connecting weights and outputs of the hidden layer based on Least Square Backpropagation (LSB) algorithm. A set of ‘required’ outputs of the hidden layer is added to the input sets through a feedback path to accelerate the convergence speed. The numerical simulation results have demonstrated that the algorithm is better than conventional BP, Quasi-Newton BFGS (an alternative to the conjugate gradient methods for fast optimisation) and LSB algorithms in terms of convergence speed and training error. The proposed method does not suffer from the drawback of the LSB algorithm, for which the training error cannot be further reduced after three iterations.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
EUR 32.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or Ebook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

Author information

Authors and Affiliations

Authors

Rights and permissions

Reprints and permissions

About this article

Cite this article

Li, Y., Rad, A. & Peng, W. An Enhanced Training Algorithm for Multilayer Neural Networks Based on Reference Output of Hidden Layer. Neural Comput & Applic 8, 218–225 (1999). https://doi.org/10.1007/s005210050024

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s005210050024

Navigation