By Topic

An adaptive step size for backpropagation using linear lower bounding functions

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$31 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

2 Author(s)
Mao Yu ; Dept. of Electr. & Comput. Eng., California Univ., Davis, CA, USA ; Chang, Tsu-Shuan

An adaptive step size is presented for the backpropagation algorithm in feedforward neural nets using linear lower bounding functions. Basically, a linear lower bounding function (LLBF) for a given function over an interval is a linear function that lies below the given function and matches the original function value at one end point. To search for an adaptive step size, an LLBF for the error function, which is expressed in terms of the step size, is derived. Since the error in a neural net can never be smaller than zero, it is plausible not to take a step larger than the step size when the associated LLBF reaches zero. In the paper, an adaptive learning algorithm based on the above idea is given. Numerical examples are used to illustrate its feasibility and to compare it with some previous results

Published in:

Signal Processing, IEEE Transactions on  (Volume:43 ,  Issue: 5 )