By Topic

Coding of variable dimension speech spectral vectors using weighted nonsquare transform vector quantization

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$33 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

4 Author(s)
Chunyan Li ; Dept. of Electr. & Comput. Eng., California Univ., Santa Barbara, CA, USA ; Lupini, P. ; Shlomot, E. ; Cuperman, V.

This paper addresses the problem associated with variable-dimension vector quantization, and presents a new quantization technique that combines a variable-size nonsquare transform (NST) with a fixed-dimension vector quantizer. We show that all linear dimension conversion methods can be treated as special cases of a general approach for linear dimension conversion formulated as NST. By incorporating the speech perceptual properties, we introduce a technique called weighted nonsquare transform vector quantization (WNSTVQ) for the quantization of speech spectral vectors. We show that the total perceptual weighted distortion can be separated into the weighted modeling distortion, which is solely determined by the choice of transforms in WNSTVQ and the weighted quantizer distortion. We discuss the factors that influence the performance of the WNSTVQ system and provide a complexity analysis for two WNSTVQ implementations. Finally, experimental results are presented to show that the WNSTVQ system has the ability to trade performance for computational complexity and memory storage by selecting suitable transforms and/or the length of fixed-dimension vectors

Published in:

Speech and Audio Processing, IEEE Transactions on  (Volume:9 ,  Issue: 6 )