By Topic

A class of gradient-adaptive step size algorithms for complex-valued nonlinear neural adaptive filters

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$31 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

2 Author(s)
Su Lee Goh ; Imperial Coll., London, UK ; Mandic, D.P.

A class of variable step-size algorithms for complex-valued nonlinear neural adaptive finite impulse response (FIR) filters realised as a dynamical perceptron is proposed. The adaptive step-size is updated using gradient descent to give variable step-size complex-valued nonlinear gradient descent (VSCNGD) algorithms. These algorithms are shown to be capable of tracking signals with rich and unknown dynamics, and exhibit faster convergence and smaller steady state error than the standard algorithms. Further, the analysis of stability and computational complexity is provided. Simulations in the prediction setting support the approach.

Published in:

Acoustics, Speech, and Signal Processing, 2005. Proceedings. (ICASSP '05). IEEE International Conference on  (Volume:5 )

Date of Conference:

18-23 March 2005