A novel fast backpropagation learning algorithm using parallel tangent and heuristic line search

  • Authors:
  • Payman Moallem;S. Amirhassan Monadjemi;Behzad Mirzaeian;Mohsen Ashourian

  • Affiliations:
  • Department of Electronic Engineering, Faculty of Engineering, University of Isfahan, Isfahan, Iran;Department of Computer Engineering, Faculty of Engineering, University of Isfahan, Isfahan, Iran;Department of Electronic Engineering, Faculty of Engineering, University of Isfahan, Isfahan, Iran;Department of Information Technology Engineering, Faculty of Engineering, University of Isfahan, Isfahan, Iran

  • Venue:
  • ICCOMP'06 Proceedings of the 10th WSEAS international conference on Computers
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

In gradient based learning algorithms, the momentum has usually an improving effect in convergence rate and decreasing the zigzagging phenomena. However it sometimes causes the convergence rate to decrease. The Parallel Tangent (ParTan) gradient is used as deflecting method to improve the convergence. From the implementation point of view, it is as simple as the momentum. In fact this method is one of the more practical implementation of conjugate gradient. ParTan tries to overcome the inefficiency of zigzagging of conventional backpropagation by deflecting the gradient through acceleration phase. In this paper, we use two learning rate, η for gradient search direction and µ for accelerating direction through parallel tangent. Moreover, an improved version of dynamic self adaptation of η and η is used to improve parallel tangent gradient learning method. In dynamic self adaptation, each learning rate is adapted locally to the cost function landscape and the previous learning rate. Finally we test the proposed algorithm on various MLP neural networks including a XOR 2×2×1, Encoder 16×4×16 and finally Parity 4×4×1. We compare the results with those of the dynamic self adaptation of gradient learning rate and momentum (DSη-α) and parallel tangent with dynamic self adaptation (PTDSηµ). Experimental results showed that the average number of epoch is decreased to around 66% and 50% for DSη-α and PTDSη-µ respectively. Moreover our proposed algorithm shows a good power for avoiding from local minimum.