By Topic

Circuits and Systems for Video Technology, IEEE Transactions on

Issue 12 • Date Dec. 2007

Filter Results

Displaying Results 1 - 18 of 18
  • Table of contents

    Page(s): C1
    Save to Project icon | Request Permissions | PDF file iconPDF (70 KB)  
    Freely Available from IEEE
  • IEEE Transactions on Circuits and Systems for Video Technology publication information

    Page(s): C2
    Save to Project icon | Request Permissions | PDF file iconPDF (38 KB)  
    Freely Available from IEEE
  • Optimized Cross-Layer Design for Scalable Video Transmission Over the IEEE 802.11e Networks

    Page(s): 1665 - 1678
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (901 KB) |  | HTML iconHTML  

    A cross-layer design for optimizing 3-D wavelet scalable video transmission over the IEEE 802.11e networks is proposed. A thorough study on the behavior of the IEEE 802.11e protocol is conducted. Based on our findings, all timescales rate control is developed featuring a unique property of soft capacity support for multimedia delivery. The design consists of a macro timescale and a micro timescale rate control schemes residing at the application layer and the network sublayer respectively. The macro rate control uses bandwidth estimation to achieve optimal bit allocation with minimum distortion. The micro rate control employs an adaptive mapping of packets from video classifications to appropriate network priorities which preemptively drops less important video packets to maximize the transmission protection to the important video packets. The performance is investigated by simulations highlighting advantages of our cross-layer design. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Fast Matching Pursuit Video Coding by Combining Dictionary Approximation and Atom Extraction

    Page(s): 1679 - 1689
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (1817 KB) |  | HTML iconHTML  

    In this paper, we propose a systematic approach that approximates a target dictionary to reduce the complexity of a matching pursuit encoder. We combine calculation of the inner products and maximum atom extraction of a matching pursuit video coding scheme based on eigendictionary approximation and tree-based vector quantization. The approach makes the codec design and optimization cleaner and more systematic than previous dictionary approximation methods. We vary the quality of approximation to demonstrate the tradeoff between computational complexity and coding efficiency. The experiment results show that our codec achieves speed-up factors of up to 100 with a performance loss of less than 0.1 dB. We use double-stimulus impairment scale scores to evaluate the perceptual quality of our approach for different levels of complexity. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • A Real-Time Technique for Spatio–Temporal Video Noise Estimation

    Page(s): 1690 - 1699
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (2424 KB) |  | HTML iconHTML  

    This paper proposes a spatio-temporal technique for estimating the noise variance in noisy video signals, where the noise is assumed to be additive white Gaussian noise. The proposed technique utilizes domain-wise (spatial, temporal, and spatio-temporal) video information independently for improved reliability. It divides the video signal into cubes and measures their homogeneity using Laplacian of Gaussian based operators. Then, the variances of homogeneous cubes are selected to estimate the noise variance. A least median of squares robust estimator is used to reject outliers and produce domain-wise noise variance estimates which are adaptively integrated to obtain the final frame-wise estimate. The proposed technique estimates the noise variance reliably in video sequences with both low and high video activities (e.g., fast motion or high spatial structure) and it produces a maximum estimation error of 1.7-dB peak signal-to-noise ratio. The proposed method is fast when compared to referenced methods. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Quality Enhancement of Frame Rate Up-Converted Video by Adaptive Frame Skip and Reliable Motion Extraction

    Page(s): 1700 - 1713
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (2139 KB) |  | HTML iconHTML  

    Frame rate up-conversion is a postprocessing tool to convert the frame rate from a lower number to a higher one. It is a useful technique for a lot of practical applications, such as display format conversion, low bit rate video coding and slow motion playback. Unlike traditional approaches, such as frame repetition or linear frame interpolation, motion-compensated frame interpolation (MCFI) technique is more efficient since it takes block motion into account. In this paper, by considering the deficiencies of previous works, new criteria and coding schemes for improving motion derivation and interpolation processes are proposed. Next, for video coding applications, adaptive frame skip is executed at the encoder side to maximize the power of MCFI so that the quality of interpolated frames is guaranteed. Experimental results show that our proposal effectively enhances the overall quality of the frame rate up-converted video sequence, both subjectively and objectively. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • An Improved Motion-Compensated 3-D LLMMSE Filter With Spatio–Temporal Adaptive Filtering Support

    Page(s): 1714 - 1727
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (11999 KB) |  | HTML iconHTML  

    In this paper, we propose a motion-compensated spatio-temporal Locally adaptive linear minimum mean squared-error (LLMMSE) filter for noisy image sequences with both temporal and spatially adaptive filtering support. Motion compensation and an adaptive temporal filtering support (TFS) structure guarantee the uniformity of the TFS. An intelligent pixel aggregation algorithm is proposed to include homogeneous neighboring pixels and exclude the outlier pixels, resulting in uniform spatial filtering support. By using the proposed spatio-temporal LLMMSE filter with uniform spatio-temporal support, we can reduce noise efficiently without introducing visually disturbing blurring artifacts. Furthermore, we employ an adaptive weighted local mean and variance estimation algorithm to improve the accuracy of estimation. The weights provide an implicit mechanism for deemphasizing the contribution of the outlier pixels which are wrongly kept within the support to avoid blurring. The performance of the proposed filter is quantitatively evaluated and compared with state-of-the-art methods. The results demonstrate that the proposed filter can achieve superior filtering performance, both subjectively and objectively. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Video Shot Characterization Using Principles of Perceptual Prominence and Perceptual Grouping in Spatio–Temporal Domain

    Page(s): 1728 - 1741
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (1066 KB) |  | HTML iconHTML  

    We present a novel approach for applying perceptual grouping principles to the spatio-temporal domain of video. Our perceptual grouping scheme, applied on blobs, makes use of a specified spatio-temporal coherence model. The grouping scheme identifies the blob cliques or perceptual clusters in the scene. We propose a computational model for analyzing a video shot based on a novel principle of perceptual prominence. The principle of perceptual prominence captures the key aspects of mise-en-scene required for characterizing a video scene. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Unsupervized Video Segmentation With Low Depth of Field

    Page(s): 1742 - 1751
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (2554 KB) |  | HTML iconHTML  

    In this paper, a novel segmentation algorithm based on matting model is proposed to extract the focused objects in low depth-of-field (DoF) video images. The proposed algorithm is fully automatic and can be used to partition the video image into focused objects and defocused background. This method consists of three stages. The first stage is to generate a saliency map of the input image by the reblurring model. In the second stage, bilateral and morphological filtering are employed to smooth and accentuate the salient regions. Then a trimap with three regions is calculated by an adaptive thresholding method. The third stage involves the proposed adaptive error control matting scheme to extract the boundaries of the focused objects accurately. Experimental evaluation on test sequences shows that the proposed method is capable of segmenting the focused region effectively and accurately. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Error Resilient Decoding of JPEG2000

    Page(s): 1752 - 1757
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (477 KB) |  | HTML iconHTML  

    In this letter, we investigate the error resilience properties of JPEG2000. We identify the dependencies among coding passes of a codeblock codestream, and determine the sections of the codestream that can be salvaged in the presence of errors. In our analysis, we consider the effects of mode variations provided by the standard. The proposed methods are derived using the existing dependency structure of coding passes and do not require a substantial increase in computational complexity of the decoder. Experimental results indicate that the proposed methods can improve the error resilience performance substantially. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Statistical Background Subtraction Using Spatial Cues

    Page(s): 1758 - 1763
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (695 KB) |  | HTML iconHTML  

    Most statistical background subtraction techniques are based on the analysis of temporal color/intensity distribution. However, learning statistics on a series of time frames can be problematic, especially when no frame absent of moving objects is available or when the available memory is not sufficient to store the series of frames needed for learning. In this letter, we propose a spatial variation to the traditional temporal framework. The proposed framework allows statistical motion detection with methods trained on one background frame instead of a series of frames as is usually the case. Our framework includes two spatial background subtraction approaches suitable for different applications. The first approach is meant for scenes having a nonstatic background due to noise, camera jitter or animation in the scene (e.g.,waving trees, fluttering leaves). This approach models each pixel with two PDFs: one unimodal PDF and one multimodal PDF, both trained on one background frame. In this way, the method can handle backgrounds with static and nonstatic areas. The second spatial approach is designed to use as little processing time and memory as possible. Based on the assumption that neighboring pixels often share similar temporal distribution, this second approach models the background with one global mixture of Gaussians. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Polynomial Weighted Median Image Sequence Prediction

    Page(s): 1764 - 1770
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (1682 KB) |  | HTML iconHTML  

    Image sequence prediction is widely used in image compression and transmission schemes such as differential pulse code modulation. In traditional predictive coding, linear predictors are usually adopted for simplicity. The nonlinear Volterra predictor can be employed as an alternative to linear predictors to compensate for the nonstationary and non-Gaussian nature of image sequences. Although the Volterra predictor avoids the smoothing effects introduced by linear predictors, it generally amplifies noise contamination present in the images. In this letter, we propose a nonlinear polynomial weighted median (PWM) predictor for image sequence. The proposed PWM predictor is more robust to noise, while still retaining the information of higher order statistics of pixel values. Experimental results illustrate that the PWM predictor yields good results in both high and low motion video. It is especially suitable for high motion sequence in noisy case. The proposed scheme can be incorporated in new predictive coding systems. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • In this issue - Technically

    Page(s): 1771
    Save to Project icon | Request Permissions | PDF file iconPDF (229 KB)  
    Freely Available from IEEE
  • 2008 International Conference on Multimedia & Expo (ICME 2008)

    Page(s): 1772
    Save to Project icon | Request Permissions | PDF file iconPDF (561 KB)  
    Freely Available from IEEE
  • Special issue on event analysis in videos

    Page(s): 1773
    Save to Project icon | Request Permissions | PDF file iconPDF (119 KB)  
    Freely Available from IEEE
  • 2007 Index IEEE Transactions on Circuits and Systems for Video Technology Vol. 17

    Page(s): 1774 - 1792
    Save to Project icon | Request Permissions | PDF file iconPDF (221 KB)  
    Freely Available from IEEE
  • IEEE Circuits and Systems Society Information

    Page(s): C3
    Save to Project icon | Request Permissions | PDF file iconPDF (30 KB)  
    Freely Available from IEEE
  • IEEE Transactions on Circuits and Systems for Video Technology Information for authors

    Page(s): C4
    Save to Project icon | Request Permissions | PDF file iconPDF (31 KB)  
    Freely Available from IEEE

Aims & Scope

The emphasis is focused on, but not limited to:
1. Video A/D and D/ A
2. Video Compression Techniques and Signal Processing
3. Multi-Dimensional Filters and Transforms
4. High Speed Real-Tune Circuits
5. Multi-Processors Systems—Hardware and Software
6. VLSI Architecture and Implementation for Video Technology 

 

Full Aims & Scope

Meet Our Editors

Editor-in-Chief
Dan Schonfeld
Multimedia Communications Laboratory
ECE Dept. (M/C 154)
University of Illinois at Chicago (UIC)
Chicago, IL 60607-7053
tcsvt-eic@tcad.polito.it

Managing Editor
Jaqueline Zelkowitz
tcsvt@tcad.polito.it