By Topic

Real-Time Perception-Based Clipping of Audio Signals Using Convex Optimization

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$31 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

5 Author(s)
Defraene, B. ; Department of Electrical Engineering, ESAT-SCD (SISTA), KU Leuven, Leuven, Belgium ; van Waterschoot, T. ; Ferreau, H.J. ; Diehl, M.
more authors

Clipping is an essential signal processing operation in many real-time audio applications, yet the use of existing clipping techniques generally has a detrimental effect on the perceived audio signal quality. In this paper, we present a novel multidisciplinary approach to clipping which aims to explicitly minimize the perceptible clipping-induced distortion by embedding a convex optimization criterion and a psychoacoustic model into a frame-based algorithm. The core of this perception-based clipping algorithm consists in solving a convex optimization problem for each time frame in a fast and reliable way. To this end, three different structure-exploiting optimization methods are derived in the common mathematical framework of convex optimization, and corresponding theoretical complexity bounds are provided. From comparative audio quality evaluation experiments, it is concluded that the perception-based clipping algorithm results in significantly higher objective audio quality scores than existing clipping techniques. Moreover, the algorithm is shown to be capable to adhere to real-time deadlines without making a sacrifice in terms of audio quality.

Published in:

Audio, Speech, and Language Processing, IEEE Transactions on  (Volume:20 ,  Issue: 10 )