Scheduled System Maintenance on May 29th, 2015:
IEEE Xplore will be upgraded between 11:00 AM and 10:00 PM EDT. During this time there may be intermittent impact on performance. We apologize for any inconvenience.
By Topic

Circuits and Systems for Video Technology, IEEE Transactions on

Issue 3 • Date March 2005

Filter Results

Displaying Results 1 - 19 of 19
  • Table of contents

    Publication Year: 2005 , Page(s): c1
    Save to Project icon | Request Permissions | PDF file iconPDF (73 KB)  
    Freely Available from IEEE
  • IEEE Transactions on Circuits and Systems for Video Technology publication information

    Publication Year: 2005 , Page(s): c2
    Save to Project icon | Request Permissions | PDF file iconPDF (36 KB)  
    Freely Available from IEEE
  • Multiple description video coding for scalable and robust transmission over IP

    Publication Year: 2005 , Page(s): 321 - 334
    Cited by:  Papers (52)  |  Patents (1)
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (776 KB) |  | HTML iconHTML  

    In this paper, we address the problem of video transmission over unreliable networks, such as the Internet, where packet losses occur. The most recent literature indicates multiple description (MD) as a promising coding approach to handle this issue. Moreover, it has been shown also how important the use of motion compensation prediction is in an MD-coding scheme. This paper proposes two architectures for multiple description video coding, both of them are based on the motion compensation prediction loop. The common characteristic of the two architectures is the use of a polyphase down-sampling technique to create the MDs and to introduce cross-redundancy among the descriptions. The first scheme, that we call drift-compensation multiple description video coder (DC-MDVC) appears very robust when used in an error-prone environment, but it can provide only two descriptions. The second architecture, called independent flow multiple description video coder (IF-MDVC), generates multiple sets of data before the motion compensation loop; in this case, there are no severe limitations in the selection of the number of descriptions used by the coder. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Efficient rate control for JPEG2000 image coding

    Publication Year: 2005 , Page(s): 335 - 344
    Cited by:  Papers (24)  |  Patents (1)
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (616 KB) |  | HTML iconHTML  

    JPEG2000 is the new image coding standard which can provide superior rate-distortion performance over the old JPEG standard. However, the conventional post-compression rate-distortion (PCRD) optimization scheme in JPEG2000 is not efficient. It requires entropy encoding all available data even though a large portion of them will not be included in the final output. In this paper, three rate control methods are proposed to efficiently reduce both the computational complexity and memory usage over the conventional PCRD method. The first method, called successive bit-plane rate allocation (SBRA), allocates the bit rate by using the currently available rate-distortion information only. The second method, called priority scanning rate allocation (PSRA), allocates bits according to certain prioritized ordering. The third method uses PSRA to achieve optimal truncation as PCRD without encoding of all the image details and is called priority scanning with optimal truncation (PSOT). Simulation results suggest that the three proposed methods provide different tradeoff among visual quality, computational complexity, coding delay and working memory size. SBRA is memoryless and causal and requires the least computational complexity, lowest coding delay and achieves good visual quality. PSRA achieves higher PSNR than SBRA at the expense of larger working memory size and longer coding delay. PSOT gives the best PSNR but requires even more computation, delay and memory. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Image-dependent shape coding and representation

    Publication Year: 2005 , Page(s): 345 - 354
    Cited by:  Papers (5)
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (1376 KB) |  | HTML iconHTML  

    We present a new shape-coding algorithm to support object-based representation, which differs from previous algorithms in that it encodes shape as dependent meta data for image description. Therefore, both the shape-coding and decoding processes of this algorithm are designed to be dependent on the underlying image in which the object (described by the shape) is contained. This way, the correlation between image and shape is effectively removed and the shape-coding efficiency is improved on average by three times over the state-of-the-art algorithms. To facilitate comparison, a generalized "contour-generating" framework is introduced to formulate the shape-coding problem. From this framework we derive both the proposed algorithm and a number of state-of-the-art algorithms, and show that the rate-distortion (RD) criterion can be studied in a uniform way under this framework. Specifically, a dynamic-programming-based algorithm is designed to find the RD optimal coding result for the proposed algorithm. As an extension, we also discuss the complexity and scalability issues related to the application design of the proposed algorithm. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Compressing the illumination-adjustable images with principal component analysis

    Publication Year: 2005 , Page(s): 355 - 364
    Cited by:  Papers (8)  |  Patents (1)
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (1632 KB) |  | HTML iconHTML  

    The ability to change illumination is a crucial factor in image-based modeling and rendering. Image-based relighting offers such capability. However, the tradeoff is the enormous increase of storage requirement. In this paper, we propose a compression scheme that effectively reduces the data volume while maintaining the real-time relighting capability. The proposed method is based on principal component analysis (PCA). A block-wise PCA is used to practically process the huge input data. The output of PCA is a set of eigenimages and the corresponding relighting coefficients. By dropping those low-energy eigenimages, the data size is drastically reduced. To further compress the data, eigenimages left are compressed using transform coding and quantization while the relighting coefficients are compressed using uniform quantization. We also suggest the suitable target bit rate for each phase of the compression method in order to preserve the visual quality. Finally, we propose a real-time engine that relights images from the compressed data. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Foveated shot detection for video segmentation

    Publication Year: 2005 , Page(s): 365 - 377
    Cited by:  Papers (29)
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (2152 KB) |  | HTML iconHTML  

    We view scenes in the real world by moving our eyes three to four times each second and integrating information across subsequent fixations (foveation points). By taking advantage of this fact, in this paper we propose an original approach to partitioning of a video into shots based on a foveated representation of the video. More precisely, the shot-change detection method is related to the computation, at each time instant, of a consistency measure of the fixation sequences generated by an ideal observer looking at the video. The proposed scheme aims at detecting both abrupt and gradual transitions between shots using a single technique, rather than a set of dedicated methods. Results on videos of various content types are reported and validate the proposed approach. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Analysis, fast algorithm, and VLSI architecture design for H.264/AVC intra frame coder

    Publication Year: 2005 , Page(s): 378 - 401
    Cited by:  Papers (149)  |  Patents (10)
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (2840 KB) |  | HTML iconHTML  

    Intra prediction with rate-distortion constrained mode decision is the most important technology in H.264/AVC intra frame coder, which is competitive with the latest image coding standard JPEG2000, in terms of both coding performance and computational complexity. The predictor generation engine for intra prediction and the transform engine for mode decision are critical because the operations require a lot of memory access and occupy 80% of the computation time of the entire intra compression process. A low cost general purpose processor cannot process these operations in real time. In this paper, we proposed two solutions for platform-based design of H.264/AVC intra frame coder. One solution is a software implementation targeted at low-end applications. Context-based decimation of unlikely candidates, subsampling of matching operations, bit-width truncation to reduce the computations, and interleaved full-search/partial-search strategy to stop the error propagation and to maintain the image quality, are proposed and combined as our fast algorithm. Experimental results show that our method can reduce 60% of the computation used for intra prediction and mode decision while keeping the peak signal-to-noise ratio degradation less than 0.3 dB. The other solution is a hardware accelerator targeted at high-end applications. After comprehensive analysis of instructions and exploration of parallelism, we proposed our system architecture with four-parallel intra prediction and mode decision to enhance the processing capability. Hadamard-based mode decision is modified as discrete cosine transform-based version to reduce 40% of memory access. Two-stage macroblock pipelining is also proposed to double the processing speed and hardware utilization. The other features of our design are reconfigurable predictor generator supporting all of the 13 intra prediction modes, parallel multitransform and inverse transform engine, and CAVLC bitstream engine. A prototype chip is fabricated with TSMC 0.25-μm CMOS 1P5M technology. Simulation results show that our implementation can process 16 mega-pixels (4096×4096) within 1 s, or namely 720×480 4:2:0 30 Hz video in real time, at the operating frequency of 54 MHz. The transistor count is 429 K, and the core - size is only 1.855×1.885 mm2. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • A proxy-assisted adaptation framework for object video multicasting

    Publication Year: 2005 , Page(s): 402 - 411
    Cited by:  Papers (3)
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (536 KB) |  | HTML iconHTML  

    Video multicast is a challenging problem due to the heterogeneous and best-effort nature of the Internet. In this paper, we present a novel video multicast framework that exploits the potential of object scalability offered by MPEG-4. Specifically, we introduce the concept of object transmission proxy (OTP), which filters incoming streams using object-based bandwidth adaptation to meet dynamic network conditions. Multiple OPTs can form an overlay network that interconnects diverse multicast islands with semi-uniform demands within each single island. We concur with the wisdom that an application best knows the utility of its data. Hence, the bandwidth-adaptation algorithm for the OTPs adaptively allocates bandwidth among video objects according to their respective utilities and then performs application-level filtering based on an effective stream classification and packetization scheme. Extensive simulation results demonstrate that our framework has substantial performance improvement over conventional bandwidth-adaptation schemes. It is particularly suitable for object-based video multicasting where the objects are of different importance. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Unequal loss protection for H.263 compressed video

    Publication Year: 2005 , Page(s): 412 - 419
    Cited by:  Papers (14)
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (296 KB) |  | HTML iconHTML  

    We study the application of unequal loss protection (ULP) algorithms to motion-compensated video over lossy packet networks. In particular, we focus on streaming video applications over the Internet. The original ULP framework applies unequal amounts of forward error correction to embedded data to provide graceful degradation of quality in the presence of increasing packet loss. In this letter, we apply the ULP framework to baseline H.263, a video compression standard that targets low bit rates, by investigating reorderings of the bitstream to make it embedded. The reordering process allows a receiver to display quality video, even at the loss rates encountered in wireless transmissions and the current Internet. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • A JPEG 2000 error resilience method using uneven block-sized information included markers

    Publication Year: 2005 , Page(s): 420 - 424
    Cited by:  Papers (5)
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (456 KB) |  | HTML iconHTML  

    This paper proposes an uneven block-sized information included marker (UBIIM) in the use of JPEG 2000 to achieve both compression efficiency and error resilience capability. Included in the UBIIM approach is the marker adopted in JPEG 2000 devised to incorporate also error detection and correction capabilities. The UBIIM consists of VS Index and TR Index, where VS Index is used to indicate a value support block, one of the neighboring blocks of the erroneous block of which the pixel values will be most similar as the erroneous block. TR Index is used to indicate a texture reference block, the block in the corresponding position of the erroneous block in the same subband of one-higher level of which the pixel value distribution is similar as the erroneous block. When errors occur, a new block is formed to replace the erroneous block by value support block and texture reference block, to achieve error concealment. The sizes of the blocks in UBIIM are determined based on the importance factor of various levels of subbands, that is, different levels have different block sizes. With such approach, the possible overhead due to the addition of markers between every two blocks could be dramatically reduced, while error resilience capability significantly improved. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Trinocular stereo sequence coding based on MPEG-2

    Publication Year: 2005 , Page(s): 425 - 429
    Cited by:  Papers (2)
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (904 KB) |  | HTML iconHTML  

    This paper presents a trinocular stereo sequence-coding algorithm based on moving picture experts group (MPEG)-2. Trinocular stereo images are a set of three images; center, left, and right view images are taken at the same time. To reduce the redundancy of a trinocular stereo sequence, we use disparity information, as well as motion information. Three reference images are employed: two reference (previous and future) images for the forward and backward motion vectors and the center view image for the disparity vector (DV). We investigate three temporal prediction structures used for trinocular stereo sequence coding. New macroblock types and variable length code tables that include DV are designed. Experimental results with real test sequences show that the proposed algorithm gives a higher peak signal-to-noise ratio (PSNR) than MPEG-2 and MPEG-2 multiview profile. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • A deblocking method for BDCT compressed images based on adaptive projections

    Publication Year: 2005 , Page(s): 430 - 435
    Cited by:  Papers (25)
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (864 KB) |  | HTML iconHTML  

    A deblocking method based on projection onto convex sets (POCS) is proposed to reduce blocking artifacts in compressed images coded by the block discrete cosine transform. The method differs from existing POCS-based methods in three aspects. Firstly, the adjustment of a pixel's intensity is determined by local properties of the pixel. Secondly, three locally adaptive constraint sets are introduced to improve deblocking results. Thirdly, the human visual system modeling and relaxed projections are incorporated to make pixels adjust appropriately. The method is tested on typical images with excellent results. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • A pipelined architecture for real-time correction of barrel distortion in wide-angle camera images

    Publication Year: 2005 , Page(s): 436 - 444
    Cited by:  Papers (22)  |  Patents (17)
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (824 KB) |  | HTML iconHTML  

    An efficient pipelined architecture for the real-time correction of barrel distortion in wide-angle camera images is presented in this paper. The distortion correction model is based on least-squares estimation to correct the nonlinear distortion in images. The model parameters include the expanded/corrected image size, the back-mapping coefficients, distortion center, and corrected center. The coordinate rotation digital computer (CORDIC) based hardware design is suitable for an input image size of 1028×1028 pixels and is pipelined to operate at a clock frequency of 40 MHz. The VLSI system will facilitate the use of a dedicated hardware that could be mounted along with the camera unit. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • A memory-efficient realization of cyclic convolution and its application to discrete cosine transform

    Publication Year: 2005 , Page(s): 445 - 453
    Cited by:  Papers (20)  |  Patents (1)
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (752 KB) |  | HTML iconHTML  

    This paper presents a memory-efficient approach to realize the cyclic convolution and its application to the discrete cosine transform (DCT). We adopt the way of distributed arithmetic (DA) computation, exploit the symmetry property of DCT coefficients to merge the elements in the matrix of DCT kernel, separate the kernel to be two perfect cyclic forms, and partition the content of ROM into groups to facilitate an efficient realization of a one-dimensional (1-D) N-point DCT kernel using (N-1)/2 adders or subtractors, one small ROM module, a barrel shifter, and ((N-1)/2)+1 accumulators. The proposed memory-efficient design technique is characterized by rearranging the content of the ROM using the conventional DA approach into several groups such that all the elements in a group can be accessed simultaneously in accumulating all the DCT outputs for increasing the ROM utilization. Considering an example using 16-bit coefficients, the proposed design can save more than 57% of the delay-area product, as compare with the existing DA-based designs in the case of the 1-D seven-point DCT. Finally, a 1-D DCT chip was implemented to illustrate the efficiency associated with the proposed approach. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Comments on "Winscale: an image-scaling algorithm using an area pixel Model"

    Publication Year: 2005 , Page(s): 454 - 455
    Cited by:  Papers (4)  |  Patents (1)
    Save to Project icon | Request Permissions | Click to expandQuick Abstract | PDF file iconPDF (112 KB) |  | HTML iconHTML  

    In the paper by Kim et al. (2003), the authors propose a new image scaling method called winscale. The presented method can be used for scaling up and down. However, scaling down utilizing the winscale concept gives exactly the same results as the well-known bilinear interpolation. Furthermore, compared to bilinear, scaling up with the proposed winscale "overlap stamping" method has very similar calculations. The basic winscale upscaling differs from the bilinear method. View full abstract»

    Full text access may be available. Click article title to sign in or learn about subscription options.
  • Call for participation for 2005 IEEE International Symposium on Circuits and Systems (ISCAS 2005)

    Publication Year: 2005 , Page(s): 456
    Save to Project icon | Request Permissions | PDF file iconPDF (523 KB) |  | HTML iconHTML  
    Freely Available from IEEE
  • IEEE Circuits and Systems Society Information

    Publication Year: 2005 , Page(s): c3
    Save to Project icon | Request Permissions | PDF file iconPDF (35 KB)  
    Freely Available from IEEE
  • IEEE Transactions on Circuits and Systems for Video Technology Information for authors

    Publication Year: 2005 , Page(s): c4
    Save to Project icon | Request Permissions | PDF file iconPDF (30 KB) |  | HTML iconHTML  
    Freely Available from IEEE

Aims & Scope

The emphasis is focused on, but not limited to:
1. Video A/D and D/ A
2. Video Compression Techniques and Signal Processing
3. Multi-Dimensional Filters and Transforms
4. High Speed Real-Tune Circuits
5. Multi-Processors Systems—Hardware and Software
6. VLSI Architecture and Implementation for Video Technology 

 

Full Aims & Scope

Meet Our Editors

Editor-in-Chief
Dan Schonfeld
Multimedia Communications Laboratory
ECE Dept. (M/C 154)
University of Illinois at Chicago (UIC)
Chicago, IL 60607-7053
tcsvt-eic@tcad.polito.it

Managing Editor
Jaqueline Zelkowitz
tcsvt@tcad.polito.it