By Topic

Computationally Efficient Sparse Bayesian Learning via Belief Propagation

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$31 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

2 Author(s)
Xing Tan ; Dept. of Electr. & Comput. Eng., Univ. of Florida, Gainesville, FL, USA ; Jian Li

We present a belief propagation (BP)-based sparse Bayesian learning (SBL) algorithm, referred to as the BP-SBL, to recover sparse transform coefficients in large scale compressed sensing problems. BP-SBL is based on a widely used hierarchical Bayesian model, which is turned into a factor graph so that BP can be applied to achieve computational efficiency. We prove that the messages in BP are Gaussian probability density functions and therefore, we only need to update their means and variances when we update the messages. The computational complexity of BP-SBL is proportional to the number of transform coefficients, allowing the algorithms to deal with large scale compressed sensing problems efficiently. Numerical examples are provided to demonstrate the effectiveness of BP-SBL.

Published in:

Signal Processing, IEEE Transactions on  (Volume:58 ,  Issue: 4 )